You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by ct...@apache.org on 2015/01/09 03:44:05 UTC

[01/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Repository: accumulo
Updated Branches:
  refs/heads/1.5 07422efc5 -> 901d60ef1
  refs/heads/1.6 3bf46661e -> 9ca1ff02e
  refs/heads/master 44e2b2cdf -> 392d9d6a4


http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/trace/src/main/java/org/apache/accumulo/trace/instrument/CountSampler.java
----------------------------------------------------------------------
diff --git a/trace/src/main/java/org/apache/accumulo/trace/instrument/CountSampler.java b/trace/src/main/java/org/apache/accumulo/trace/instrument/CountSampler.java
index 56a3989..f6d256a 100644
--- a/trace/src/main/java/org/apache/accumulo/trace/instrument/CountSampler.java
+++ b/trace/src/main/java/org/apache/accumulo/trace/instrument/CountSampler.java
@@ -24,7 +24,7 @@ public class CountSampler extends org.htrace.impl.CountSampler implements Sample
   public CountSampler(long frequency) {
     super(frequency);
   }
-  
+
   @Override
   public boolean next() {
     return super.next(null);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/trace/src/main/java/org/apache/accumulo/trace/instrument/Sampler.java
----------------------------------------------------------------------
diff --git a/trace/src/main/java/org/apache/accumulo/trace/instrument/Sampler.java b/trace/src/main/java/org/apache/accumulo/trace/instrument/Sampler.java
index 94e5e67..9563f8f 100644
--- a/trace/src/main/java/org/apache/accumulo/trace/instrument/Sampler.java
+++ b/trace/src/main/java/org/apache/accumulo/trace/instrument/Sampler.java
@@ -21,7 +21,7 @@ package org.apache.accumulo.trace.instrument;
  */
 @Deprecated
 public interface Sampler extends org.htrace.Sampler<Object> {
-  
+
   boolean next();
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/trace/src/main/java/org/apache/accumulo/trace/instrument/Span.java
----------------------------------------------------------------------
diff --git a/trace/src/main/java/org/apache/accumulo/trace/instrument/Span.java b/trace/src/main/java/org/apache/accumulo/trace/instrument/Span.java
index 7c0e8bc..dbd8882 100644
--- a/trace/src/main/java/org/apache/accumulo/trace/instrument/Span.java
+++ b/trace/src/main/java/org/apache/accumulo/trace/instrument/Span.java
@@ -75,7 +75,7 @@ public class Span extends org.apache.accumulo.core.trace.Span implements Cloudtr
   }
 
   @Override
-  public Map<String, String> getData() {
+  public Map<String,String> getData() {
     Map<byte[],byte[]> data = span.getKVAnnotations();
     HashMap<String,String> stringData = new HashMap<>();
     for (Entry<byte[],byte[]> d : data.entrySet()) {


[46/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ConditionalWriterImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ConditionalWriterImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ConditionalWriterImpl.java
index 8440c1c..f0d9e0c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ConditionalWriterImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ConditionalWriterImpl.java
@@ -91,9 +91,9 @@ import org.slf4j.LoggerFactory;
 import com.google.common.net.HostAndPort;
 
 class ConditionalWriterImpl implements ConditionalWriter {
-  
+
   private static ThreadPoolExecutor cleanupThreadPool = new ThreadPoolExecutor(1, 1, 10, TimeUnit.SECONDS, new LinkedBlockingQueue<Runnable>());
-  
+
   static {
     cleanupThreadPool.allowCoreThreadTimeOut(true);
   }
@@ -101,7 +101,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
   private static final Logger log = LoggerFactory.getLogger(ConditionalWriterImpl.class);
 
   private static final int MAX_SLEEP = 30000;
-  
+
   private Authorizations auths;
   private VisibilityEvaluator ve;
   @SuppressWarnings("unchecked")
@@ -111,44 +111,44 @@ class ConditionalWriterImpl implements ConditionalWriter {
   private String tableId;
   private long timeout;
   private final Durability durability;
-  
+
   private static class ServerQueue {
     BlockingQueue<TabletServerMutations<QCMutation>> queue = new LinkedBlockingQueue<TabletServerMutations<QCMutation>>();
     boolean taskQueued = false;
   }
-  
+
   private Map<String,ServerQueue> serverQueues;
   private DelayQueue<QCMutation> failedMutations = new DelayQueue<QCMutation>();
   private ScheduledThreadPoolExecutor threadPool;
-  
+
   private class RQIterator implements Iterator<Result> {
-    
+
     private BlockingQueue<Result> rq;
     private int count;
-    
+
     public RQIterator(BlockingQueue<Result> resultQueue, int count) {
       this.rq = resultQueue;
       this.count = count;
     }
-    
+
     @Override
     public boolean hasNext() {
       return count > 0;
     }
-    
+
     @Override
     public Result next() {
       if (count <= 0)
         throw new NoSuchElementException();
-      
+
       try {
         Result result = rq.poll(1, TimeUnit.SECONDS);
         while (result == null) {
-          
+
           if (threadPool.isShutdown()) {
             throw new NoSuchElementException("ConditionalWriter closed");
           }
-          
+
           result = rq.poll(1, TimeUnit.SECONDS);
         }
         count--;
@@ -157,32 +157,32 @@ class ConditionalWriterImpl implements ConditionalWriter {
         throw new RuntimeException(e);
       }
     }
-    
+
     @Override
     public void remove() {
       throw new UnsupportedOperationException();
     }
-    
+
   }
-  
+
   private static class QCMutation extends ConditionalMutation implements Delayed {
     private BlockingQueue<Result> resultQueue;
     private long resetTime;
     private long delay = 50;
     private long entryTime;
-    
+
     QCMutation(ConditionalMutation cm, BlockingQueue<Result> resultQueue, long entryTime) {
       super(cm);
       this.resultQueue = resultQueue;
       this.entryTime = entryTime;
     }
-    
+
     @Override
     public int compareTo(Delayed o) {
       QCMutation oqcm = (QCMutation) o;
       return Long.valueOf(resetTime).compareTo(Long.valueOf(oqcm.resetTime));
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof QCMutation)
@@ -194,45 +194,45 @@ class ConditionalWriterImpl implements ConditionalWriter {
     public long getDelay(TimeUnit unit) {
       return unit.convert(delay - (System.currentTimeMillis() - resetTime), TimeUnit.MILLISECONDS);
     }
-    
+
     void resetDelay() {
       delay = Math.min(delay * 2, MAX_SLEEP);
       resetTime = System.currentTimeMillis();
     }
-    
+
     void queueResult(Result result) {
       resultQueue.add(result);
     }
   }
-  
+
   private ServerQueue getServerQueue(String location) {
     ServerQueue serverQueue;
     synchronized (serverQueues) {
       serverQueue = serverQueues.get(location);
       if (serverQueue == null) {
-        
+
         serverQueue = new ServerQueue();
         serverQueues.put(location, serverQueue);
       }
     }
     return serverQueue;
   }
-  
+
   private class CleanupTask implements Runnable {
     private List<SessionID> sessions;
-    
+
     CleanupTask(List<SessionID> activeSessions) {
       this.sessions = activeSessions;
     }
-    
+
     @Override
     public void run() {
       TabletClientService.Iface client = null;
-      
+
       for (SessionID sid : sessions) {
         if (!sid.isActive())
           continue;
-        
+
         TInfo tinfo = Tracer.traceInfo();
         try {
           client = getClient(sid.location);
@@ -240,7 +240,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
         } catch (Exception e) {} finally {
           ThriftUtil.returnClient((TServiceClient) client);
         }
-        
+
       }
     }
   }
@@ -248,11 +248,11 @@ class ConditionalWriterImpl implements ConditionalWriter {
   private void queueRetry(List<QCMutation> mutations, HostAndPort server) {
 
     if (timeout < Long.MAX_VALUE) {
-      
+
       long time = System.currentTimeMillis();
-      
+
       ArrayList<QCMutation> mutations2 = new ArrayList<ConditionalWriterImpl.QCMutation>(mutations.size());
-      
+
       for (QCMutation qcm : mutations) {
         qcm.resetDelay();
         if (time + qcm.getDelay(TimeUnit.MILLISECONDS) > qcm.entryTime + timeout) {
@@ -267,52 +267,52 @@ class ConditionalWriterImpl implements ConditionalWriter {
           mutations2.add(qcm);
         }
       }
-      
+
       if (mutations2.size() > 0)
         failedMutations.addAll(mutations2);
-      
+
     } else {
       for (QCMutation qcm : mutations)
         qcm.resetDelay();
       failedMutations.addAll(mutations);
     }
   }
-  
+
   private void queue(List<QCMutation> mutations) {
     List<QCMutation> failures = new ArrayList<QCMutation>();
     Map<String,TabletServerMutations<QCMutation>> binnedMutations = new HashMap<String,TabletLocator.TabletServerMutations<QCMutation>>();
-    
+
     try {
       locator.binMutations(context, mutations, binnedMutations, failures);
-      
+
       if (failures.size() == mutations.size())
         if (!Tables.exists(context.getInstance(), tableId))
           throw new TableDeletedException(tableId);
         else if (Tables.getTableState(context.getInstance(), tableId) == TableState.OFFLINE)
           throw new TableOfflineException(context.getInstance(), tableId);
-      
+
     } catch (Exception e) {
       for (QCMutation qcm : mutations)
         qcm.queueResult(new Result(e, qcm, null));
-      
+
       // do not want to queue anything that was put in before binMutations() failed
       failures.clear();
       binnedMutations.clear();
     }
-    
+
     if (failures.size() > 0)
       queueRetry(failures, null);
-    
+
     for (Entry<String,TabletServerMutations<QCMutation>> entry : binnedMutations.entrySet()) {
       queue(entry.getKey(), entry.getValue());
     }
-    
+
   }
-  
+
   private void queue(String location, TabletServerMutations<QCMutation> mutations) {
-    
+
     ServerQueue serverQueue = getServerQueue(location);
-    
+
     synchronized (serverQueue) {
       serverQueue.queue.add(mutations);
       // never execute more than one task per server
@@ -321,9 +321,9 @@ class ConditionalWriterImpl implements ConditionalWriter {
         serverQueue.taskQueued = true;
       }
     }
-    
+
   }
-  
+
   private void reschedule(SendTask task) {
     ServerQueue serverQueue = getServerQueue(task.location);
     // just finished processing work for this server, could reschedule if it has more work or immediately process the work
@@ -331,31 +331,31 @@ class ConditionalWriterImpl implements ConditionalWriter {
     // more data that need to be processed... also it will give the current server time to build
     // up more data... the thinking is that rescheduling instead or processing immediately will result
     // in bigger batches and less RPC overhead
-    
+
     synchronized (serverQueue) {
       if (serverQueue.queue.size() > 0)
         threadPool.execute(new LoggingRunnable(log, Trace.wrap(task)));
       else
         serverQueue.taskQueued = false;
     }
-    
+
   }
-  
+
   private TabletServerMutations<QCMutation> dequeue(String location) {
     BlockingQueue<TabletServerMutations<QCMutation>> queue = getServerQueue(location).queue;
-    
+
     ArrayList<TabletServerMutations<QCMutation>> mutations = new ArrayList<TabletLocator.TabletServerMutations<QCMutation>>();
     queue.drainTo(mutations);
-    
+
     if (mutations.size() == 0)
       return null;
-    
+
     if (mutations.size() == 1) {
       return mutations.get(0);
     } else {
       // merge multiple request to a single tablet server
       TabletServerMutations<QCMutation> tsm = mutations.get(0);
-      
+
       for (int i = 1; i < mutations.size(); i++) {
         for (Entry<KeyExtent,List<QCMutation>> entry : mutations.get(i).getMutations().entrySet()) {
           List<QCMutation> list = tsm.getMutations().get(entry.getKey());
@@ -363,15 +363,15 @@ class ConditionalWriterImpl implements ConditionalWriter {
             list = new ArrayList<QCMutation>();
             tsm.getMutations().put(entry.getKey(), list);
           }
-          
+
           list.addAll(entry.getValue());
         }
       }
-      
+
       return tsm;
     }
   }
-  
+
   ConditionalWriterImpl(ClientContext context, String tableId, ConditionalWriterConfig config) {
     this.context = context;
     this.auths = config.getAuthorizations();
@@ -382,9 +382,9 @@ class ConditionalWriterImpl implements ConditionalWriter {
     this.tableId = tableId;
     this.timeout = config.getTimeout(TimeUnit.MILLISECONDS);
     this.durability = config.getDurability();
-    
+
     Runnable failureHandler = new Runnable() {
-      
+
       @Override
       public void run() {
         List<QCMutation> mutations = new ArrayList<QCMutation>();
@@ -393,27 +393,27 @@ class ConditionalWriterImpl implements ConditionalWriter {
           queue(mutations);
       }
     };
-    
+
     failureHandler = new LoggingRunnable(log, failureHandler);
-    
+
     threadPool.scheduleAtFixedRate(failureHandler, 250, 250, TimeUnit.MILLISECONDS);
   }
-  
+
   @Override
   public Iterator<Result> write(Iterator<ConditionalMutation> mutations) {
-    
+
     BlockingQueue<Result> resultQueue = new LinkedBlockingQueue<Result>();
-    
+
     List<QCMutation> mutationList = new ArrayList<QCMutation>();
-    
+
     int count = 0;
-    
+
     long entryTime = System.currentTimeMillis();
-    
+
     mloop: while (mutations.hasNext()) {
       ConditionalMutation mut = mutations.next();
       count++;
-      
+
       if (mut.getConditions().size() == 0)
         throw new IllegalArgumentException("ConditionalMutation had no conditions " + new String(mut.getRow()));
 
@@ -423,26 +423,26 @@ class ConditionalWriterImpl implements ConditionalWriter {
           continue mloop;
         }
       }
-      
+
       // copy the mutations so that even if caller changes it, it will not matter
       mutationList.add(new QCMutation(mut, resultQueue, entryTime));
     }
-    
+
     queue(mutationList);
-    
+
     return new RQIterator(resultQueue, count);
-    
+
   }
-  
+
   private class SendTask implements Runnable {
-    
+
     String location;
-    
+
     public SendTask(String location) {
       this.location = location;
-      
+
     }
-    
+
     @Override
     public void run() {
       try {
@@ -454,18 +454,18 @@ class ConditionalWriterImpl implements ConditionalWriter {
       }
     }
   }
-  
+
   private static class CMK {
-    
+
     QCMutation cm;
     KeyExtent ke;
-    
+
     public CMK(KeyExtent ke, QCMutation cm) {
       this.ke = ke;
       this.cm = cm;
     }
   }
-  
+
   private static class SessionID {
     HostAndPort location;
     String lockId;
@@ -473,7 +473,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
     boolean reserved;
     long lastAccessTime;
     long ttl;
-    
+
     boolean isActive() {
       return System.currentTimeMillis() - lastAccessTime < ttl * .95;
     }
@@ -488,7 +488,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
       if (sid != null) {
         if (sid.reserved)
           throw new IllegalStateException();
-        
+
         if (!sid.isActive()) {
           cachedSessionIDs.remove(location);
         } else {
@@ -497,10 +497,10 @@ class ConditionalWriterImpl implements ConditionalWriter {
         }
       }
     }
-    
+
     TConditionalSession tcs = client.startConditionalUpdate(tinfo, context.rpcCreds(), ByteBufferUtil.toByteBuffers(auths.getAuthorizations()), tableId,
         DurabilityImpl.toThrift(durability));
-    
+
     synchronized (cachedSessionIDs) {
       SessionID sid = new SessionID();
       sid.reserved = true;
@@ -510,17 +510,17 @@ class ConditionalWriterImpl implements ConditionalWriter {
       sid.location = location;
       if (cachedSessionIDs.put(location, sid) != null)
         throw new IllegalStateException();
-      
+
       return sid;
     }
-    
+
   }
 
   private void invalidateSessionID(HostAndPort location) {
     synchronized (cachedSessionIDs) {
       cachedSessionIDs.remove(location);
     }
-    
+
   }
 
   private void unreserveSessionID(HostAndPort location) {
@@ -534,7 +534,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
       }
     }
   }
-  
+
   List<SessionID> getActiveSessions() {
     ArrayList<SessionID> activeSessions = new ArrayList<SessionID>();
     for (SessionID sid : cachedSessionIDs.values())
@@ -554,23 +554,23 @@ class ConditionalWriterImpl implements ConditionalWriter {
 
   private void sendToServer(HostAndPort location, TabletServerMutations<QCMutation> mutations) {
     TabletClientService.Iface client = null;
-    
+
     TInfo tinfo = Tracer.traceInfo();
-    
+
     Map<Long,CMK> cmidToCm = new HashMap<Long,CMK>();
     MutableLong cmid = new MutableLong(0);
-    
+
     SessionID sessionId = null;
-    
-    try {        
+
+    try {
       Map<TKeyExtent,List<TConditionalMutation>> tmutations = new HashMap<TKeyExtent,List<TConditionalMutation>>();
-      
+
       CompressedIterators compressedIters = new CompressedIterators();
       convertMutations(mutations, cmidToCm, cmid, tmutations, compressedIters);
-      
-      //getClient() call must come after converMutations in case it throws a TException
+
+      // getClient() call must come after converMutations in case it throws a TException
       client = getClient(location);
-      
+
       List<TCMResult> tresults = null;
       while (tresults == null) {
         try {
@@ -581,11 +581,11 @@ class ConditionalWriterImpl implements ConditionalWriter {
           invalidateSessionID(location);
         }
       }
-      
+
       HashSet<KeyExtent> extentsToInvalidate = new HashSet<KeyExtent>();
-      
+
       ArrayList<QCMutation> ignored = new ArrayList<QCMutation>();
-      
+
       for (TCMResult tcmResult : tresults) {
         if (tcmResult.status == TCMStatus.IGNORED) {
           CMK cmk = cmidToCm.get(tcmResult.cmid);
@@ -596,13 +596,13 @@ class ConditionalWriterImpl implements ConditionalWriter {
           qcm.queueResult(new Result(fromThrift(tcmResult.status), qcm, location.toString()));
         }
       }
-      
+
       for (KeyExtent ke : extentsToInvalidate) {
         locator.invalidateCache(ke);
       }
-      
+
       queueRetry(ignored, location);
-      
+
     } catch (ThriftSecurityException tse) {
       AccumuloSecurityException ase = new AccumuloSecurityException(context.getCredentials().getPrincipal(), tse.getCode(), Tables.getPrintableTableInfoFromId(
           context.getInstance(), tableId), tse);
@@ -618,7 +618,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
     } catch (Exception e) {
       queueException(location, cmidToCm, e);
     } finally {
-      if(sessionId != null)
+      if (sessionId != null)
         unreserveSessionID(location);
       ThriftUtil.returnClient((TServiceClient) client);
     }
@@ -649,22 +649,22 @@ class ConditionalWriterImpl implements ConditionalWriter {
       }
     }
   }
-  
+
   /*
    * The purpose of this code is to ensure that a conditional mutation will not execute when its status is unknown. This allows a user to read the row when the
    * status is unknown and not have to worry about the tserver applying the mutation after the scan.
-   * 
+   *
    * If a conditional mutation is taking a long time to process, then this method will wait for it to finish... unless this exceeds timeout.
    */
   private void invalidateSession(SessionID sessionId, HostAndPort location) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
 
     long sleepTime = 50;
-    
+
     long startTime = System.currentTimeMillis();
-    
+
     Instance instance = context.getInstance();
     LockID lid = new LockID(ZooUtil.getRoot(instance) + Constants.ZTSERVERS, sessionId.lockId);
-    
+
     ZooCacheFactory zcf = new ZooCacheFactory();
     while (true) {
       if (!ZooLock.isLockHeld(zcf.getZooCache(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut()), lid)) {
@@ -673,33 +673,33 @@ class ConditionalWriterImpl implements ConditionalWriter {
         locator.invalidateCache(context.getInstance(), location.toString());
         return;
       }
-      
+
       try {
         // if the mutation is currently processing, this method will block until its done or times out
         invalidateSession(sessionId.sessionID, location);
-        
+
         return;
       } catch (TApplicationException tae) {
         throw new AccumuloServerException(location.toString(), tae);
       } catch (TException e) {
         locator.invalidateCache(context.getInstance(), location.toString());
       }
-      
+
       if ((System.currentTimeMillis() - startTime) + sleepTime > timeout)
         throw new TimedOutException(Collections.singleton(location.toString()));
 
       UtilWaitThread.sleep(sleepTime);
       sleepTime = Math.min(2 * sleepTime, MAX_SLEEP);
-      
+
     }
-    
+
   }
 
   private void invalidateSession(long sessionId, HostAndPort location) throws TException {
     TabletClientService.Iface client = null;
-    
+
     TInfo tinfo = Tracer.traceInfo();
-    
+
     try {
       client = getClient(location);
       client.invalidateConditionalUpdate(tinfo, sessionId);
@@ -707,7 +707,7 @@ class ConditionalWriterImpl implements ConditionalWriter {
       ThriftUtil.returnClient((TServiceClient) client);
     }
   }
-  
+
   private Status fromThrift(TCMStatus status) {
     switch (status) {
       case ACCEPTED:
@@ -720,63 +720,63 @@ class ConditionalWriterImpl implements ConditionalWriter {
         throw new IllegalArgumentException(status.toString());
     }
   }
-  
+
   private void convertMutations(TabletServerMutations<QCMutation> mutations, Map<Long,CMK> cmidToCm, MutableLong cmid,
       Map<TKeyExtent,List<TConditionalMutation>> tmutations, CompressedIterators compressedIters) {
-    
+
     for (Entry<KeyExtent,List<QCMutation>> entry : mutations.getMutations().entrySet()) {
       TKeyExtent tke = entry.getKey().toThrift();
       ArrayList<TConditionalMutation> tcondMutaions = new ArrayList<TConditionalMutation>();
-      
+
       List<QCMutation> condMutations = entry.getValue();
-      
+
       for (QCMutation cm : condMutations) {
         TMutation tm = cm.toThrift();
-        
+
         List<TCondition> conditions = convertConditions(cm, compressedIters);
-        
+
         cmidToCm.put(cmid.longValue(), new CMK(entry.getKey(), cm));
         TConditionalMutation tcm = new TConditionalMutation(conditions, tm, cmid.longValue());
         cmid.increment();
         tcondMutaions.add(tcm);
       }
-      
+
       tmutations.put(tke, tcondMutaions);
     }
   }
-  
+
   private List<TCondition> convertConditions(ConditionalMutation cm, CompressedIterators compressedIters) {
     List<TCondition> conditions = new ArrayList<TCondition>(cm.getConditions().size());
-    
+
     for (Condition cond : cm.getConditions()) {
       long ts = 0;
       boolean hasTs = false;
-      
+
       if (cond.getTimestamp() != null) {
         ts = cond.getTimestamp();
         hasTs = true;
       }
-      
+
       ByteBuffer iters = compressedIters.compress(cond.getIterators());
-      
+
       TCondition tc = new TCondition(ByteBufferUtil.toByteBuffers(cond.getFamily()), ByteBufferUtil.toByteBuffers(cond.getQualifier()),
           ByteBufferUtil.toByteBuffers(cond.getVisibility()), ts, hasTs, ByteBufferUtil.toByteBuffers(cond.getValue()), iters);
-      
+
       conditions.add(tc);
     }
-    
+
     return conditions;
   }
-  
+
   private boolean isVisible(ByteSequence cv) {
     Text testVis = new Text(cv.toArray());
     if (testVis.getLength() == 0)
       return true;
-    
+
     Boolean b = cache.get(testVis);
     if (b != null)
       return b;
-    
+
     try {
       Boolean bb = ve.evaluate(new ColumnVisibility(testVis));
       cache.put(new Text(testVis), bb);
@@ -787,16 +787,16 @@ class ConditionalWriterImpl implements ConditionalWriter {
       return false;
     }
   }
-  
+
   @Override
   public Result write(ConditionalMutation mutation) {
     return write(Collections.singleton(mutation).iterator()).next();
   }
-  
+
   @Override
   public void close() {
     threadPool.shutdownNow();
     cleanupThreadPool.execute(new CleanupTask(getActiveSessions()));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/DurabilityImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/DurabilityImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/DurabilityImpl.java
index b2a0a98..bc82629 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/DurabilityImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/DurabilityImpl.java
@@ -64,5 +64,5 @@ public class DurabilityImpl {
     }
     return durability;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/InstanceOperationsImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/InstanceOperationsImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/InstanceOperationsImpl.java
index b95d8b2..7b3ee12 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/InstanceOperationsImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/InstanceOperationsImpl.java
@@ -219,6 +219,6 @@ public class InstanceOperationsImpl implements InstanceOperations {
       // should never happen
       throw new RuntimeException("Unexpected exception thrown", ex);
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/IsolationException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/IsolationException.java b/core/src/main/java/org/apache/accumulo/core/client/impl/IsolationException.java
index f09e1cb..c544ca3 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/IsolationException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/IsolationException.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.core.client.impl;
 
 public class IsolationException extends RuntimeException {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/MasterClient.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/MasterClient.java b/core/src/main/java/org/apache/accumulo/core/client/impl/MasterClient.java
index a9ad8a1..fcbf9f9 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/MasterClient.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/MasterClient.java
@@ -86,8 +86,8 @@ public class MasterClient {
     }
   }
 
-  public static <T> T execute(ClientContext context, ClientExecReturn<T,MasterClientService.Client> exec) throws AccumuloException,
-      AccumuloSecurityException, TableNotFoundException {
+  public static <T> T execute(ClientContext context, ClientExecReturn<T,MasterClientService.Client> exec) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException {
     MasterClientService.Client client = null;
     while (true) {
       try {
@@ -118,8 +118,8 @@ public class MasterClient {
     }
   }
 
-  public static void executeGeneric(ClientContext context, ClientExec<MasterClientService.Client> exec) throws AccumuloException,
-      AccumuloSecurityException, TableNotFoundException {
+  public static void executeGeneric(ClientContext context, ClientExec<MasterClientService.Client> exec) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException {
     MasterClientService.Client client = null;
     while (true) {
       try {
@@ -151,13 +151,13 @@ public class MasterClient {
     }
   }
 
-  public static void executeTable(ClientContext context, ClientExec<MasterClientService.Client> exec) throws AccumuloException,
-      AccumuloSecurityException, TableNotFoundException {
+  public static void executeTable(ClientContext context, ClientExec<MasterClientService.Client> exec) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException {
     executeGeneric(context, exec);
   }
 
-  public static void executeNamespace(ClientContext context, ClientExec<MasterClientService.Client> exec) throws AccumuloException,
-      AccumuloSecurityException, NamespaceNotFoundException {
+  public static void executeNamespace(ClientContext context, ClientExec<MasterClientService.Client> exec) throws AccumuloException, AccumuloSecurityException,
+      NamespaceNotFoundException {
     try {
       executeGeneric(context, exec);
     } catch (TableNotFoundException e) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/MultiTableBatchWriterImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/MultiTableBatchWriterImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/MultiTableBatchWriterImpl.java
index ace8701..82aa714 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/MultiTableBatchWriterImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/MultiTableBatchWriterImpl.java
@@ -155,7 +155,7 @@ public class MultiTableBatchWriterImpl implements MultiTableBatchWriter {
 
   /**
    * Returns the table ID for the given table name.
-   * 
+   *
    * @param tableName
    *          The name of the table which to find the ID for
    * @return The table ID, or null if the table name doesn't exist
@@ -203,7 +203,7 @@ public class MultiTableBatchWriterImpl implements MultiTableBatchWriter {
 
       // cacheResetCount could change after this point in time, but I think thats ok because just want to ensure this methods sees changes
       // made before it was called.
-      
+
       long internalResetCount = cacheLastState.get();
 
       if (cacheResetCount > internalResetCount) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/OfflineScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/OfflineScanner.java b/core/src/main/java/org/apache/accumulo/core/client/impl/OfflineScanner.java
index 2552682..d679ffa 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/OfflineScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/OfflineScanner.java
@@ -17,6 +17,7 @@
 package org.apache.accumulo.core.client.impl;
 
 import static com.google.common.base.Preconditions.checkArgument;
+
 import java.io.IOException;
 import java.util.ArrayList;
 import java.util.Arrays;
@@ -71,35 +72,35 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.io.Text;
 
 class OfflineIterator implements Iterator<Entry<Key,Value>> {
-  
+
   static class OfflineIteratorEnvironment implements IteratorEnvironment {
     @Override
     public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public AccumuloConfiguration getConfig() {
       return AccumuloConfiguration.getDefaultConfiguration();
     }
-    
+
     @Override
     public IteratorScope getIteratorScope() {
       return IteratorScope.scan;
     }
-    
+
     @Override
     public boolean isFullMajorCompaction() {
       return false;
     }
-    
+
     private ArrayList<SortedKeyValueIterator<Key,Value>> topLevelIterators = new ArrayList<SortedKeyValueIterator<Key,Value>>();
-    
+
     @Override
     public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
       topLevelIterators.add(iter);
     }
-    
+
     SortedKeyValueIterator<Key,Value> getTopLevelIterator(SortedKeyValueIterator<Key,Value> iter) {
       if (topLevelIterators.isEmpty())
         return iter;
@@ -108,7 +109,7 @@ class OfflineIterator implements Iterator<Entry<Key,Value>> {
       return new MultiIterator(allIters, false);
     }
   }
-  
+
   private SortedKeyValueIterator<Key,Value> iter;
   private Range range;
   private KeyExtent currentExtent;
@@ -124,83 +125,83 @@ class OfflineIterator implements Iterator<Entry<Key,Value>> {
     this.options = new ScannerOptions(options);
     this.instance = instance;
     this.range = range;
-    
+
     if (this.options.fetchedColumns.size() > 0) {
       this.range = range.bound(this.options.fetchedColumns.first(), this.options.fetchedColumns.last());
     }
-    
+
     this.tableId = table.toString();
     this.authorizations = authorizations;
     this.readers = new ArrayList<SortedKeyValueIterator<Key,Value>>();
-    
+
     try {
       conn = instance.getConnector(credentials.getPrincipal(), credentials.getToken());
       config = new ConfigurationCopy(conn.instanceOperations().getSiteConfiguration());
       nextTablet();
-      
+
       while (iter != null && !iter.hasTop())
         nextTablet();
-      
+
     } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public boolean hasNext() {
     return iter != null && iter.hasTop();
   }
-  
+
   @Override
   public Entry<Key,Value> next() {
     try {
       byte[] v = iter.getTopValue().get();
       // copy just like tablet server does, do this before calling next
       KeyValue ret = new KeyValue(new Key(iter.getTopKey()), Arrays.copyOf(v, v.length));
-      
+
       iter.next();
-      
+
       while (iter != null && !iter.hasTop())
         nextTablet();
-      
+
       return ret;
     } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   private void nextTablet() throws TableNotFoundException, AccumuloException, IOException {
-    
+
     Range nextRange = null;
-    
+
     if (currentExtent == null) {
       Text startRow;
-      
+
       if (range.getStartKey() != null)
         startRow = range.getStartKey().getRow();
       else
         startRow = new Text();
-      
+
       nextRange = new Range(new KeyExtent(new Text(tableId), startRow, null).getMetadataEntry(), true, null, false);
     } else {
-      
+
       if (currentExtent.getEndRow() == null) {
         iter = null;
         return;
       }
-      
+
       if (range.afterEndKey(new Key(currentExtent.getEndRow()).followingKey(PartialKey.ROW))) {
         iter = null;
         return;
       }
-      
+
       nextRange = new Range(currentExtent.getMetadataEntry(), false, null, false);
     }
-    
+
     List<String> relFiles = new ArrayList<String>();
-    
+
     Pair<KeyExtent,String> eloc = getTabletFiles(nextRange, relFiles);
-    
+
     while (eloc.getSecond() != null) {
       if (Tables.getTableState(instance, tableId) != TableState.OFFLINE) {
         Tables.clearCache(instance);
@@ -208,18 +209,18 @@ class OfflineIterator implements Iterator<Entry<Key,Value>> {
           throw new AccumuloException("Table is online " + tableId + " cannot scan tablet in offline mode " + eloc.getFirst());
         }
       }
-      
+
       UtilWaitThread.sleep(250);
-      
+
       eloc = getTabletFiles(nextRange, relFiles);
     }
-    
+
     KeyExtent extent = eloc.getFirst();
-    
+
     if (!extent.getTableId().toString().equals(tableId)) {
       throw new AccumuloException(" did not find tablets for table " + tableId + " " + extent);
     }
-    
+
     if (currentExtent != null && !extent.isPreviousExtent(currentExtent))
       throw new AccumuloException(" " + currentExtent + " is not previous extent " + extent);
 
@@ -241,108 +242,108 @@ class OfflineIterator implements Iterator<Entry<Key,Value>> {
         }
       }
     }
-    
+
     iter = createIterator(extent, absFiles);
     iter.seek(range, LocalityGroupUtil.families(options.fetchedColumns), options.fetchedColumns.size() == 0 ? false : true);
     currentExtent = extent;
-    
+
   }
-  
+
   private Pair<KeyExtent,String> getTabletFiles(Range nextRange, List<String> relFiles) throws TableNotFoundException {
     Scanner scanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.setBatchSize(100);
     scanner.setRange(nextRange);
-    
+
     RowIterator rowIter = new RowIterator(scanner);
     Iterator<Entry<Key,Value>> row = rowIter.next();
-    
+
     KeyExtent extent = null;
     String location = null;
-    
+
     while (row.hasNext()) {
       Entry<Key,Value> entry = row.next();
       Key key = entry.getKey();
-      
+
       if (key.getColumnFamily().equals(DataFileColumnFamily.NAME)) {
         relFiles.add(key.getColumnQualifier().toString());
       }
-      
+
       if (key.getColumnFamily().equals(TabletsSection.CurrentLocationColumnFamily.NAME)
           || key.getColumnFamily().equals(TabletsSection.FutureLocationColumnFamily.NAME)) {
         location = entry.getValue().toString();
       }
-      
+
       if (TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.hasColumns(key)) {
         extent = new KeyExtent(key.getRow(), entry.getValue());
       }
-      
+
     }
     return new Pair<KeyExtent,String>(extent, location);
   }
-  
+
   private SortedKeyValueIterator<Key,Value> createIterator(KeyExtent extent, List<String> absFiles) throws TableNotFoundException, AccumuloException,
       IOException {
-    
+
     // TODO share code w/ tablet - ACCUMULO-1303
     AccumuloConfiguration acuTableConf = AccumuloConfiguration.getTableConfiguration(conn, tableId);
-    
+
     Configuration conf = CachedConfiguration.getInstance();
 
     for (SortedKeyValueIterator<Key,Value> reader : readers) {
       ((FileSKVIterator) reader).close();
     }
-    
+
     readers.clear();
-    
+
     // TODO need to close files - ACCUMULO-1303
     for (String file : absFiles) {
       FileSystem fs = VolumeConfiguration.getVolume(file, conf, config).getFileSystem();
       FileSKVIterator reader = FileOperations.getInstance().openReader(file, false, fs, conf, acuTableConf, null, null);
       readers.add(reader);
     }
-    
+
     MultiIterator multiIter = new MultiIterator(readers, extent);
-    
+
     OfflineIteratorEnvironment iterEnv = new OfflineIteratorEnvironment();
-    
+
     DeletingIterator delIter = new DeletingIterator(multiIter, false);
-    
+
     ColumnFamilySkippingIterator cfsi = new ColumnFamilySkippingIterator(delIter);
-    
+
     ColumnQualifierFilter colFilter = new ColumnQualifierFilter(cfsi, new HashSet<Column>(options.fetchedColumns));
-    
+
     byte[] defaultSecurityLabel;
-    
+
     ColumnVisibility cv = new ColumnVisibility(acuTableConf.get(Property.TABLE_DEFAULT_SCANTIME_VISIBILITY));
     defaultSecurityLabel = cv.getExpression();
-    
+
     VisibilityFilter visFilter = new VisibilityFilter(colFilter, authorizations, defaultSecurityLabel);
-    
+
     return iterEnv.getTopLevelIterator(IteratorUtil.loadIterators(IteratorScope.scan, visFilter, extent, acuTableConf, options.serverSideIteratorList,
         options.serverSideIteratorOptions, iterEnv, false));
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
 }
 
 /**
- * 
+ *
  */
 public class OfflineScanner extends ScannerOptions implements Scanner {
-  
+
   private int batchSize;
   private int timeOut;
   private Range range;
-  
+
   private Instance instance;
   private Credentials credentials;
   private Authorizations authorizations;
   private Text tableId;
-  
+
   public OfflineScanner(Instance instance, Credentials credentials, String tableId, Authorizations authorizations) {
     checkArgument(instance != null, "instance is null");
     checkArgument(credentials != null, "credentials is null");
@@ -352,55 +353,55 @@ public class OfflineScanner extends ScannerOptions implements Scanner {
     this.credentials = credentials;
     this.tableId = new Text(tableId);
     this.range = new Range((Key) null, (Key) null);
-    
+
     this.authorizations = authorizations;
-    
+
     this.batchSize = Constants.SCAN_BATCH_SIZE;
     this.timeOut = Integer.MAX_VALUE;
   }
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {
     this.timeOut = timeOut;
   }
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
     return timeOut;
   }
-  
+
   @Override
   public void setRange(Range range) {
     this.range = range;
   }
-  
+
   @Override
   public Range getRange() {
     return range;
   }
-  
+
   @Override
   public void setBatchSize(int size) {
     this.batchSize = size;
   }
-  
+
   @Override
   public int getBatchSize() {
     return batchSize;
   }
-  
+
   @Override
   public void enableIsolation() {
-    
+
   }
-  
+
   @Override
   public void disableIsolation() {
-    
+
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     return new OfflineIterator(this, instance, credentials, authorizations, tableId, range);
@@ -415,5 +416,5 @@ public class OfflineScanner extends ScannerOptions implements Scanner {
   public void setReadaheadThreshold(long batches) {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ReplicationClient.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ReplicationClient.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ReplicationClient.java
index 95b71ee..a449389 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ReplicationClient.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ReplicationClient.java
@@ -115,7 +115,7 @@ public class ReplicationClient {
 
   /**
    * Attempt a single time to create a ReplicationServicer client to the given host
-   * 
+   *
    * @param context
    *          The client session for the peer replicant
    * @param server
@@ -152,8 +152,8 @@ public class ReplicationClient {
     }
   }
 
-  public static <T> T executeCoordinatorWithReturn(ClientContext context, ClientExecReturn<T,ReplicationCoordinator.Client> exec)
-      throws AccumuloException, AccumuloSecurityException {
+  public static <T> T executeCoordinatorWithReturn(ClientContext context, ClientExecReturn<T,ReplicationCoordinator.Client> exec) throws AccumuloException,
+      AccumuloSecurityException {
     ReplicationCoordinator.Client client = null;
     for (int i = 0; i < 10; i++) {
       try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/RootTabletLocator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/RootTabletLocator.java b/core/src/main/java/org/apache/accumulo/core/client/impl/RootTabletLocator.java
index 2a18765..49d5c9e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/RootTabletLocator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/RootTabletLocator.java
@@ -41,10 +41,10 @@ import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 
 public class RootTabletLocator extends TabletLocator {
-  
+
   private final TabletServerLockChecker lockChecker;
   private final ZooCacheFactory zcf;
-  
+
   RootTabletLocator(TabletServerLockChecker lockChecker) {
     this(lockChecker, new ZooCacheFactory());
   }
@@ -53,7 +53,7 @@ public class RootTabletLocator extends TabletLocator {
     this.lockChecker = lockChecker;
     this.zcf = zcf;
   }
-  
+
   @Override
   public <T extends Mutation> void binMutations(ClientContext context, List<T> mutations, Map<String,TabletServerMutations<T>> binnedMutations, List<T> failures)
       throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
@@ -68,11 +68,11 @@ public class RootTabletLocator extends TabletLocator {
       failures.addAll(mutations);
     }
   }
-  
+
   @Override
   public List<Range> binRanges(ClientContext context, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
-    
+
     TabletLocation rootTabletLocation = getRootTabletLocation(context);
     if (rootTabletLocation != null) {
       for (Range range : ranges) {
@@ -82,38 +82,38 @@ public class RootTabletLocator extends TabletLocator {
     }
     return ranges;
   }
-  
+
   @Override
   public void invalidateCache(KeyExtent failedExtent) {}
-  
+
   @Override
   public void invalidateCache(Collection<KeyExtent> keySet) {}
-  
+
   @Override
   public void invalidateCache(Instance instance, String server) {
     ZooCache zooCache = zcf.getZooCache(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
     String root = ZooUtil.getRoot(instance) + Constants.ZTSERVERS;
     zooCache.clear(root + "/" + server);
   }
-  
+
   @Override
   public void invalidateCache() {}
-  
+
   protected TabletLocation getRootTabletLocation(ClientContext context) {
     Instance instance = context.getInstance();
     String zRootLocPath = ZooUtil.getRoot(instance) + RootTable.ZROOT_TABLET_LOCATION;
     ZooCache zooCache = zcf.getZooCache(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
-    
+
     OpTimer opTimer = new OpTimer(Logger.getLogger(this.getClass()), Level.TRACE).start("Looking up root tablet location in zookeeper.");
     byte[] loc = zooCache.get(zRootLocPath);
     opTimer.stop("Found root tablet at " + (loc == null ? null : new String(loc)) + " in %DURATION%");
-    
+
     if (loc == null) {
       return null;
     }
-    
+
     String[] tokens = new String(loc).split("\\|");
-    
+
     if (lockChecker.isLockHeld(tokens[0], tokens[1]))
       return new TabletLocation(RootTable.EXTENT, tokens[0], tokens[1]);
     else
@@ -129,8 +129,8 @@ public class RootTabletLocator extends TabletLocator {
       UtilWaitThread.sleep(500);
       location = getRootTabletLocation(context);
     }
-    
+
     return location;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerImpl.java
index 666a8af..3f73f04 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerImpl.java
@@ -32,29 +32,29 @@ import org.apache.hadoop.io.Text;
 
 /**
  * provides scanner functionality
- * 
+ *
  * "Clients can iterate over multiple column families, and there are several mechanisms for limiting the rows, columns, and timestamps traversed by a scan. For
  * example, we could restrict [a] scan ... to only produce anchors whose columns match [a] regular expression ..., or to only produce anchors whose timestamps
  * fall within ten days of the current time."
- * 
+ *
  */
 public class ScannerImpl extends ScannerOptions implements Scanner {
-  
+
   // keep a list of columns over which to scan
   // keep track of the last thing read
   // hopefully, we can track all the state in the scanner on the client
   // and just query for the next highest row from the tablet server
-  
+
   private final ClientContext context;
   private Authorizations authorizations;
   private Text table;
-  
+
   private int size;
-  
+
   private Range range;
   private boolean isolated = false;
   private long readaheadThreshold = Constants.SCANNER_DEFAULT_READAHEAD_THRESHOLD;
-  
+
   public ScannerImpl(ClientContext context, String table, Authorizations authorizations) {
     checkArgument(context != null, "context is null");
     checkArgument(table != null, "table is null");
@@ -63,21 +63,21 @@ public class ScannerImpl extends ScannerOptions implements Scanner {
     this.table = new Text(table);
     this.range = new Range((Key) null, (Key) null);
     this.authorizations = authorizations;
-    
+
     this.size = Constants.SCAN_BATCH_SIZE;
   }
-  
+
   @Override
   public synchronized void setRange(Range range) {
     checkArgument(range != null, "range is null");
     this.range = range;
   }
-  
+
   @Override
   public synchronized Range getRange() {
     return range;
   }
-  
+
   @Override
   public synchronized void setBatchSize(int size) {
     if (size > 0)
@@ -85,12 +85,12 @@ public class ScannerImpl extends ScannerOptions implements Scanner {
     else
       throw new IllegalArgumentException("size must be greater than zero");
   }
-  
+
   @Override
   public synchronized int getBatchSize() {
     return size;
   }
-  
+
   /**
    * Returns an iterator over an accumulo table. This iterator uses the options that are currently set on the scanner for its lifetime. So setting options on a
    * Scanner object will have no effect on existing iterators.
@@ -99,17 +99,17 @@ public class ScannerImpl extends ScannerOptions implements Scanner {
   public synchronized Iterator<Entry<Key,Value>> iterator() {
     return new ScannerIterator(context, table, authorizations, range, size, getTimeOut(), this, isolated, readaheadThreshold);
   }
-  
+
   @Override
   public synchronized void enableIsolation() {
     this.isolated = true;
   }
-  
+
   @Override
   public synchronized void disableIsolation() {
     this.isolated = false;
   }
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {
@@ -118,7 +118,7 @@ public class ScannerImpl extends ScannerOptions implements Scanner {
     else
       setTimeout(timeOut, TimeUnit.SECONDS);
   }
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
@@ -127,16 +127,16 @@ public class ScannerImpl extends ScannerOptions implements Scanner {
       return Integer.MAX_VALUE;
     return (int) timeout;
   }
-  
+
   @Override
   public synchronized void setReadaheadThreshold(long batches) {
     if (0 > batches) {
       throw new IllegalArgumentException("Number of batches before read-ahead must be non-negative");
     }
-    
+
     readaheadThreshold = batches;
   }
-  
+
   @Override
   public synchronized long getReadaheadThreshold() {
     return readaheadThreshold;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerIterator.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerIterator.java
index 1e0ac99..276e1d6 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerIterator.java
@@ -44,49 +44,49 @@ import org.apache.hadoop.io.Text;
 import org.apache.log4j.Logger;
 
 public class ScannerIterator implements Iterator<Entry<Key,Value>> {
-  
+
   private static final Logger log = Logger.getLogger(ScannerIterator.class);
-  
+
   // scanner options
   private Text tableId;
   private int timeOut;
-  
+
   // scanner state
   private Iterator<KeyValue> iter;
   private ScanState scanState;
-  
+
   private ScannerOptions options;
-  
+
   private ArrayBlockingQueue<Object> synchQ;
-  
+
   private boolean finished = false;
-  
+
   private boolean readaheadInProgress = false;
   private long batchCount = 0;
   private long readaheadThreshold;
-  
+
   private static final List<KeyValue> EMPTY_LIST = Collections.emptyList();
-  
+
   private static ThreadPoolExecutor readaheadPool = new ThreadPoolExecutor(0, Integer.MAX_VALUE, 3l, TimeUnit.SECONDS, new SynchronousQueue<Runnable>(),
       new NamingThreadFactory("Accumulo scanner read ahead thread"));
-  
+
   private class Reader implements Runnable {
-    
+
     @Override
     public void run() {
-      
+
       try {
         while (true) {
           List<KeyValue> currentBatch = ThriftScanner.scan(scanState.context, scanState, timeOut);
-          
+
           if (currentBatch == null) {
             synchQ.add(EMPTY_LIST);
             return;
           }
-          
+
           if (currentBatch.size() == 0)
             continue;
-          
+
           synchQ.add(currentBatch);
           return;
         }
@@ -116,63 +116,62 @@ public class ScannerIterator implements Iterator<Entry<Key,Value>> {
         synchQ.add(e);
       }
     }
-    
+
   }
-  
-  ScannerIterator(ClientContext context, Text table, Authorizations authorizations, Range range, int size, int timeOut, ScannerOptions options,
-      boolean isolated) {
+
+  ScannerIterator(ClientContext context, Text table, Authorizations authorizations, Range range, int size, int timeOut, ScannerOptions options, boolean isolated) {
     this(context, table, authorizations, range, size, timeOut, options, isolated, Constants.SCANNER_DEFAULT_READAHEAD_THRESHOLD);
   }
-  
+
   ScannerIterator(ClientContext context, Text table, Authorizations authorizations, Range range, int size, int timeOut, ScannerOptions options,
       boolean isolated, long readaheadThreshold) {
     this.tableId = new Text(table);
     this.timeOut = timeOut;
     this.readaheadThreshold = readaheadThreshold;
-    
+
     this.options = new ScannerOptions(options);
-    
+
     synchQ = new ArrayBlockingQueue<Object>(1);
-    
+
     if (this.options.fetchedColumns.size() > 0) {
       range = range.bound(this.options.fetchedColumns.first(), this.options.fetchedColumns.last());
     }
-    
+
     scanState = new ScanState(context, tableId, authorizations, new Range(range), options.fetchedColumns, size, options.serverSideIteratorList,
         options.serverSideIteratorOptions, isolated, readaheadThreshold);
-    
+
     // If we want to start readahead immediately, don't wait for hasNext to be called
     if (0l == readaheadThreshold) {
       initiateReadAhead();
     }
     iter = null;
   }
-  
+
   private void initiateReadAhead() {
     readaheadInProgress = true;
     readaheadPool.execute(new Reader());
   }
-  
+
   @Override
   @SuppressWarnings("unchecked")
   public boolean hasNext() {
     if (finished)
       return false;
-    
+
     if (iter != null && iter.hasNext()) {
       return true;
     }
-    
+
     // this is done in order to find see if there is another batch to get
-    
+
     try {
       if (!readaheadInProgress) {
         // no read ahead run, fetch the next batch right now
         new Reader().run();
       }
-      
+
       Object obj = synchQ.take();
-      
+
       if (obj instanceof Exception) {
         finished = true;
         if (obj instanceof RuntimeException)
@@ -180,9 +179,9 @@ public class ScannerIterator implements Iterator<Entry<Key,Value>> {
         else
           throw new RuntimeException((Exception) obj);
       }
-      
+
       List<KeyValue> currentBatch = (List<KeyValue>) obj;
-      
+
       if (currentBatch.size() == 0) {
         currentBatch = null;
         finished = true;
@@ -190,31 +189,31 @@ public class ScannerIterator implements Iterator<Entry<Key,Value>> {
       }
       iter = currentBatch.iterator();
       batchCount++;
-      
+
       if (batchCount > readaheadThreshold) {
         // start a thread to read the next batch
         initiateReadAhead();
       }
-      
+
     } catch (InterruptedException e1) {
       throw new RuntimeException(e1);
     }
-    
+
     return true;
   }
-  
+
   @Override
   public Entry<Key,Value> next() {
     if (hasNext())
       return iter.next();
     throw new NoSuchElementException();
   }
-  
+
   // just here to satisfy the interface
   // could make this actually delete things from the database
   @Override
   public void remove() {
     throw new UnsupportedOperationException("remove is not supported in Scanner");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerOptions.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerOptions.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerOptions.java
index 9726266..d6c50c7 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerOptions.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ScannerOptions.java
@@ -17,6 +17,7 @@
 package org.apache.accumulo.core.client.impl;
 
 import static com.google.common.base.Preconditions.checkArgument;
+
 import java.util.ArrayList;
 import java.util.Collections;
 import java.util.HashMap;
@@ -39,70 +40,70 @@ import org.apache.accumulo.core.util.TextUtil;
 import org.apache.hadoop.io.Text;
 
 public class ScannerOptions implements ScannerBase {
-  
+
   protected List<IterInfo> serverSideIteratorList = Collections.emptyList();
   protected Map<String,Map<String,String>> serverSideIteratorOptions = Collections.emptyMap();
-  
+
   protected SortedSet<Column> fetchedColumns = new TreeSet<Column>();
-  
+
   protected long timeOut = Long.MAX_VALUE;
-  
+
   private String regexIterName = null;
-  
+
   protected ScannerOptions() {}
-  
+
   public ScannerOptions(ScannerOptions so) {
     setOptions(this, so);
   }
-  
+
   /**
    * Adds server-side scan iterators.
-   * 
+   *
    */
   @Override
   public synchronized void addScanIterator(IteratorSetting si) {
     checkArgument(si != null, "si is null");
     if (serverSideIteratorList.size() == 0)
       serverSideIteratorList = new ArrayList<IterInfo>();
-    
+
     for (IterInfo ii : serverSideIteratorList) {
       if (ii.iterName.equals(si.getName()))
         throw new IllegalArgumentException("Iterator name is already in use " + si.getName());
       if (ii.getPriority() == si.getPriority())
         throw new IllegalArgumentException("Iterator priority is already in use " + si.getPriority());
     }
-    
+
     serverSideIteratorList.add(new IterInfo(si.getPriority(), si.getIteratorClass(), si.getName()));
-    
+
     if (serverSideIteratorOptions.size() == 0)
       serverSideIteratorOptions = new HashMap<String,Map<String,String>>();
-    
+
     Map<String,String> opts = serverSideIteratorOptions.get(si.getName());
-    
+
     if (opts == null) {
       opts = new HashMap<String,String>();
       serverSideIteratorOptions.put(si.getName(), opts);
     }
     opts.putAll(si.getOptions());
   }
-  
+
   @Override
   public synchronized void removeScanIterator(String iteratorName) {
     checkArgument(iteratorName != null, "iteratorName is null");
     // if no iterators are set, we don't have it, so it is already removed
     if (serverSideIteratorList.size() == 0)
       return;
-    
+
     for (IterInfo ii : serverSideIteratorList) {
       if (ii.iterName.equals(iteratorName)) {
         serverSideIteratorList.remove(ii);
         break;
       }
     }
-    
+
     serverSideIteratorOptions.remove(iteratorName);
   }
-  
+
   /**
    * Override any existing options on the given named iterator
    */
@@ -113,29 +114,29 @@ public class ScannerOptions implements ScannerBase {
     checkArgument(value != null, "value is null");
     if (serverSideIteratorOptions.size() == 0)
       serverSideIteratorOptions = new HashMap<String,Map<String,String>>();
-    
+
     Map<String,String> opts = serverSideIteratorOptions.get(iteratorName);
-    
+
     if (opts == null) {
       opts = new HashMap<String,String>();
       serverSideIteratorOptions.put(iteratorName, opts);
     }
     opts.put(key, value);
   }
-  
+
   /**
    * Limit a scan to the specified column family. This can limit which locality groups are read on the server side.
-   * 
+   *
    * To fetch multiple column families call this function multiple times.
    */
-  
+
   @Override
   public synchronized void fetchColumnFamily(Text col) {
     checkArgument(col != null, "col is null");
     Column c = new Column(TextUtil.getBytes(col), null, null);
     fetchedColumns.add(c);
   }
-  
+
   @Override
   public synchronized void fetchColumn(Text colFam, Text colQual) {
     checkArgument(colFam != null, "colFam is null");
@@ -143,21 +144,21 @@ public class ScannerOptions implements ScannerBase {
     Column c = new Column(TextUtil.getBytes(colFam), TextUtil.getBytes(colQual), null);
     fetchedColumns.add(c);
   }
-  
+
   public synchronized void fetchColumn(Column column) {
     checkArgument(column != null, "column is null");
     fetchedColumns.add(column);
   }
-  
+
   @Override
   public synchronized void clearColumns() {
     fetchedColumns.clear();
   }
-  
+
   public synchronized SortedSet<Column> getFetchedColumns() {
     return fetchedColumns;
   }
-  
+
   /**
    * Clears scan iterators prior to returning a scanner to the pool.
    */
@@ -167,14 +168,14 @@ public class ScannerOptions implements ScannerBase {
     serverSideIteratorOptions = Collections.emptyMap();
     regexIterName = null;
   }
-  
+
   protected static void setOptions(ScannerOptions dst, ScannerOptions src) {
     synchronized (dst) {
       synchronized (src) {
         dst.regexIterName = src.regexIterName;
         dst.fetchedColumns = new TreeSet<Column>(src.fetchedColumns);
         dst.serverSideIteratorList = new ArrayList<IterInfo>(src.serverSideIteratorList);
-        
+
         dst.serverSideIteratorOptions = new HashMap<String,Map<String,String>>();
         Set<Entry<String,Map<String,String>>> es = src.serverSideIteratorOptions.entrySet();
         for (Entry<String,Map<String,String>> entry : es)
@@ -182,29 +183,29 @@ public class ScannerOptions implements ScannerBase {
       }
     }
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void setTimeout(long timeout, TimeUnit timeUnit) {
     if (timeOut < 0) {
       throw new IllegalArgumentException("TimeOut must be positive : " + timeOut);
     }
-    
+
     if (timeout == 0)
       this.timeOut = Long.MAX_VALUE;
     else
       this.timeOut = timeUnit.toMillis(timeout);
   }
-  
+
   @Override
   public long getTimeout(TimeUnit timeunit) {
     return timeunit.convert(timeOut, TimeUnit.MILLISECONDS);
   }
-  
+
   @Override
   public void close() {
     // Nothing needs to be closed

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ServerClient.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ServerClient.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ServerClient.java
index 84124ca..bbf2d3f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ServerClient.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ServerClient.java
@@ -42,7 +42,7 @@ import org.apache.thrift.transport.TTransportException;
 
 public class ServerClient {
   private static final Logger log = Logger.getLogger(ServerClient.class);
-  
+
   public static <T> T execute(ClientContext context, ClientExecReturn<T,ClientService.Client> exec) throws AccumuloException, AccumuloSecurityException {
     try {
       return executeRaw(context, exec);
@@ -54,7 +54,7 @@ public class ServerClient {
       throw new AccumuloException(e);
     }
   }
-  
+
   public static void execute(ClientContext context, ClientExec<ClientService.Client> exec) throws AccumuloException, AccumuloSecurityException {
     try {
       executeRaw(context, exec);
@@ -66,7 +66,7 @@ public class ServerClient {
       throw new AccumuloException(e);
     }
   }
-  
+
   public static <T> T executeRaw(ClientContext context, ClientExecReturn<T,ClientService.Client> exec) throws Exception {
     while (true) {
       ClientService.Client client = null;
@@ -85,7 +85,7 @@ public class ServerClient {
       }
     }
   }
-  
+
   public static void executeRaw(ClientContext context, ClientExec<ClientService.Client> exec) throws Exception {
     while (true) {
       ClientService.Client client = null;
@@ -105,23 +105,23 @@ public class ServerClient {
       }
     }
   }
-  
+
   static volatile boolean warnedAboutTServersBeingDown = false;
 
   public static Pair<String,ClientService.Client> getConnection(ClientContext context) throws TTransportException {
     return getConnection(context, true);
   }
-  
+
   public static Pair<String,ClientService.Client> getConnection(ClientContext context, boolean preferCachedConnections) throws TTransportException {
     return getConnection(context, preferCachedConnections, context.getClientTimeoutInMillis());
   }
-  
+
   public static Pair<String,ClientService.Client> getConnection(ClientContext context, boolean preferCachedConnections, long rpcTimeout)
       throws TTransportException {
     checkArgument(context != null, "context is null");
     // create list of servers
     ArrayList<ThriftTransportKey> servers = new ArrayList<ThriftTransportKey>();
-    
+
     // add tservers
     Instance instance = context.getInstance();
     ZooCache zc = new ZooCacheFactory().getZooCache(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
@@ -134,7 +134,7 @@ public class ServerClient {
           servers.add(new ThriftTransportKey(new ServerServices(strData).getAddress(Service.TSERV_CLIENT), rpcTimeout, context));
       }
     }
-    
+
     boolean opened = false;
     try {
       Pair<String,TTransport> pair = ThriftTransportPool.getInstance().getAnyTransport(servers, preferCachedConnections);
@@ -155,7 +155,7 @@ public class ServerClient {
       }
     }
   }
-  
+
   public static void close(ClientService.Client client) {
     if (client != null && client.getInputProtocol() != null && client.getInputProtocol().getTransport() != null) {
       ThriftTransportPool.getInstance().returnTransport(client.getInputProtocol().getTransport());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsHelper.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsHelper.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsHelper.java
index cea934d..4068dee 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsHelper.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsHelper.java
@@ -17,6 +17,7 @@
 package org.apache.accumulo.core.client.impl;
 
 import static com.google.common.base.Preconditions.checkArgument;
+
 import java.util.EnumSet;
 import java.util.HashMap;
 import java.util.Map;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsImpl.java
index dca40da..e32a9e1 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TableOperationsImpl.java
@@ -220,15 +220,14 @@ public class TableOperationsImpl extends TableOperationsHelper {
    * @param ntc
    *          specifies the new table's configuration. It determines whether the versioning iterator is enabled or disabled, logical or real-time based time
    *          recording for entries in the table
-   * 
+   *
    */
   @Override
   public void create(String tableName, NewTableConfiguration ntc) throws AccumuloException, AccumuloSecurityException, TableExistsException {
     checkArgument(tableName != null, "tableName is null");
     checkArgument(ntc != null, "ntc is null");
 
-    List<ByteBuffer> args = Arrays.asList(ByteBuffer.wrap(tableName.getBytes(UTF_8)),
-        ByteBuffer.wrap(ntc.getTimeType().name().getBytes(UTF_8)));
+    List<ByteBuffer> args = Arrays.asList(ByteBuffer.wrap(tableName.getBytes(UTF_8)), ByteBuffer.wrap(ntc.getTimeType().name().getBytes(UTF_8)));
 
     Map<String,String> opts = ntc.getProperties();
 
@@ -816,9 +815,9 @@ public class TableOperationsImpl extends TableOperationsHelper {
     if (config.getFlush())
       _flush(tableId, start, end, true);
 
-    List<ByteBuffer> args = Arrays.asList(ByteBuffer.wrap(tableId.getBytes(UTF_8)), start == null ? EMPTY : TextUtil.getByteBuffer(start),
-        end == null ? EMPTY : TextUtil.getByteBuffer(end), ByteBuffer.wrap(IteratorUtil.encodeIteratorSettings(config.getIterators())),
-        ByteBuffer.wrap(CompactionStrategyConfigUtil.encode(config.getCompactionStrategy())));
+    List<ByteBuffer> args = Arrays.asList(ByteBuffer.wrap(tableId.getBytes(UTF_8)), start == null ? EMPTY : TextUtil.getByteBuffer(start), end == null ? EMPTY
+        : TextUtil.getByteBuffer(end), ByteBuffer.wrap(IteratorUtil.encodeIteratorSettings(config.getIterators())), ByteBuffer
+        .wrap(CompactionStrategyConfigUtil.encode(config.getCompactionStrategy())));
 
     Map<String,String> opts = new HashMap<String,String>();
     try {
@@ -1474,11 +1473,11 @@ public class TableOperationsImpl extends TableOperationsHelper {
         throw new AccumuloSecurityException(e.getUser(), e.getCode());
       } catch (TTransportException e) {
         // some sort of communication error occurred, retry
-	if (pair == null) {
+        if (pair == null) {
           log.debug("Disk usage request failed.  Pair is null.  Retrying request...", e);
-	} else {
+        } else {
           log.debug("Disk usage request failed " + pair.getFirst() + ", retrying ... ", e);
-	}
+        }
         UtilWaitThread.sleep(100);
       } catch (TException e) {
         // may be a TApplicationException which indicates error on the server side

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/Tables.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/Tables.java b/core/src/main/java/org/apache/accumulo/core/client/impl/Tables.java
index c2b8001..20b1639 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/Tables.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/Tables.java
@@ -43,7 +43,7 @@ public class Tables {
   private static final Logger log = Logger.getLogger(Tables.class);
 
   public static final String VALID_NAME_REGEX = "^(\\w+\\.)?(\\w+)$";
-  
+
   private static final SecurityPermission TABLES_PERMISSION = new SecurityPermission("tablesPermission");
   private static final AtomicLong cacheResetCount = new AtomicLong(0);
 
@@ -222,7 +222,7 @@ public class Tables {
 
   /**
    * Returns the namespace id for a given table ID.
-   * 
+   *
    * @param instance
    *          The Accumulo Instance
    * @param tableId

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocator.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocator.java
index 3cb0fde..782a599 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocator.java
@@ -39,66 +39,66 @@ import org.apache.accumulo.core.metadata.RootTable;
 import org.apache.hadoop.io.Text;
 
 public abstract class TabletLocator {
-  
+
   public abstract TabletLocation locateTablet(ClientContext context, Text row, boolean skipRow, boolean retry) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException;
-  
+
   public abstract <T extends Mutation> void binMutations(ClientContext context, List<T> mutations, Map<String,TabletServerMutations<T>> binnedMutations,
       List<T> failures) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
-  
+
   public abstract List<Range> binRanges(ClientContext context, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges)
       throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
-  
+
   public abstract void invalidateCache(KeyExtent failedExtent);
-  
+
   public abstract void invalidateCache(Collection<KeyExtent> keySet);
-  
+
   /**
    * Invalidate entire cache
    */
   public abstract void invalidateCache();
-  
+
   /**
    * Invalidate all metadata entries that point to server
    */
   public abstract void invalidateCache(Instance instance, String server);
-  
+
   private static class LocatorKey {
     String instanceId;
     Text tableName;
-    
+
     LocatorKey(String instanceId, Text table) {
       this.instanceId = instanceId;
       this.tableName = table;
     }
-    
+
     @Override
     public int hashCode() {
       return instanceId.hashCode() + tableName.hashCode();
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof LocatorKey)
         return equals((LocatorKey) o);
       return false;
     }
-    
+
     public boolean equals(LocatorKey lk) {
       return instanceId.equals(lk.instanceId) && tableName.equals(lk.tableName);
     }
-    
+
   }
-  
+
   private static HashMap<LocatorKey,TabletLocator> locators = new HashMap<LocatorKey,TabletLocator>();
-  
+
   public static synchronized TabletLocator getLocator(ClientContext context, Text tableId) {
     Instance instance = context.getInstance();
     LocatorKey key = new LocatorKey(instance.getInstanceID(), tableId);
     TabletLocator tl = locators.get(key);
     if (tl == null) {
       MetadataLocationObtainer mlo = new MetadataLocationObtainer();
-      
+
       if (tableId.toString().equals(RootTable.ID)) {
         tl = new RootTabletLocator(new ZookeeperLockChecker(instance));
       } else if (tableId.toString().equals(MetadataTable.ID)) {
@@ -108,32 +108,32 @@ public abstract class TabletLocator {
       }
       locators.put(key, tl);
     }
-    
+
     return tl;
   }
-  
+
   public static class TabletLocations {
-    
+
     private final List<TabletLocation> locations;
     private final List<KeyExtent> locationless;
-    
+
     public TabletLocations(List<TabletLocation> locations, List<KeyExtent> locationless) {
       this.locations = locations;
       this.locationless = locationless;
     }
-    
+
     public List<TabletLocation> getLocations() {
       return locations;
     }
-    
+
     public List<KeyExtent> getLocationless() {
       return locationless;
     }
   }
-  
+
   public static class TabletLocation implements Comparable<TabletLocation> {
     private static final WeakHashMap<String,WeakReference<String>> tabletLocs = new WeakHashMap<String,WeakReference<String>>();
-    
+
     private static String dedupeLocation(String tabletLoc) {
       synchronized (tabletLocs) {
         WeakReference<String> lref = tabletLocs.get(tabletLoc);
@@ -143,17 +143,17 @@ public abstract class TabletLocator {
             return loc;
           }
         }
-        
+
         tabletLoc = new String(tabletLoc);
         tabletLocs.put(tabletLoc, new WeakReference<String>(tabletLoc));
         return tabletLoc;
       }
     }
-    
+
     public final KeyExtent tablet_extent;
     public final String tablet_location;
     public final String tablet_session;
-    
+
     public TabletLocation(KeyExtent tablet_extent, String tablet_location, String session) {
       checkArgument(tablet_extent != null, "tablet_extent is null");
       checkArgument(tablet_location != null, "tablet_location is null");
@@ -162,7 +162,7 @@ public abstract class TabletLocator {
       this.tablet_location = dedupeLocation(tablet_location);
       this.tablet_session = dedupeLocation(session);
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof TabletLocation) {
@@ -171,17 +171,17 @@ public abstract class TabletLocator {
       }
       return false;
     }
-    
+
     @Override
     public int hashCode() {
       throw new UnsupportedOperationException("hashcode is not implemented for class " + this.getClass().toString());
     }
-    
+
     @Override
     public String toString() {
       return "(" + tablet_extent + "," + tablet_location + "," + tablet_session + ")";
     }
-    
+
     @Override
     public int compareTo(TabletLocation o) {
       int result = tablet_extent.compareTo(o.tablet_extent);
@@ -193,7 +193,7 @@ public abstract class TabletLocator {
       return result;
     }
   }
-  
+
   public static class TabletServerMutations<T extends Mutation> {
     private Map<KeyExtent,List<T>> mutations;
     private String tserverSession;
@@ -209,14 +209,14 @@ public abstract class TabletLocator {
         mutList = new ArrayList<T>();
         mutations.put(ke, mutList);
       }
-      
+
       mutList.add(m);
     }
-    
+
     public Map<KeyExtent,List<T>> getMutations() {
       return mutations;
     }
-    
+
     final String getSession() {
       return tserverSession;
     }


[47/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/MutationsRejectedException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/MutationsRejectedException.java b/core/src/main/java/org/apache/accumulo/core/client/MutationsRejectedException.java
index 37233f5..962ef25 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/MutationsRejectedException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/MutationsRejectedException.java
@@ -32,16 +32,16 @@ import org.apache.accumulo.core.data.KeyExtent;
 
 /**
  * Communicate the failed mutations of a BatchWriter back to the client.
- * 
+ *
  */
 public class MutationsRejectedException extends AccumuloException {
   private static final long serialVersionUID = 1L;
-  
+
   private List<ConstraintViolationSummary> cvsl;
   private Map<KeyExtent,Set<SecurityErrorCode>> af;
   private Collection<String> es;
   private int unknownErrors;
-  
+
   /**
    * @param cvsList
    *          list of constraint violations
@@ -51,7 +51,7 @@ public class MutationsRejectedException extends AccumuloException {
    *          server side errors
    * @param unknownErrors
    *          number of unknown errors
-   *          
+   *
    * @deprecated since 1.6.0, see {@link #MutationsRejectedException(Instance, List, HashMap, Collection, int, Throwable)}
    */
   @Deprecated
@@ -64,7 +64,7 @@ public class MutationsRejectedException extends AccumuloException {
     this.es = serverSideErrors;
     this.unknownErrors = unknownErrors;
   }
-  
+
   /**
    * @param cvsList
    *          list of constraint violations
@@ -84,30 +84,30 @@ public class MutationsRejectedException extends AccumuloException {
     this.es = serverSideErrors;
     this.unknownErrors = unknownErrors;
   }
-  
+
   private static String format(HashMap<KeyExtent,Set<SecurityErrorCode>> hashMap, Instance instance) {
     Map<String,Set<SecurityErrorCode>> result = new HashMap<String,Set<SecurityErrorCode>>();
-    
+
     for (Entry<KeyExtent,Set<SecurityErrorCode>> entry : hashMap.entrySet()) {
       String tableInfo = Tables.getPrintableTableInfoFromId(instance, entry.getKey().getTableId().toString());
-      
+
       if (!result.containsKey(tableInfo)) {
         result.put(tableInfo, new HashSet<SecurityErrorCode>());
       }
-      
+
       result.get(tableInfo).addAll(hashMap.get(entry.getKey()));
     }
-    
+
     return result.toString();
   }
-  
+
   /**
    * @return the internal list of constraint violations
    */
   public List<ConstraintViolationSummary> getConstraintViolationSummaries() {
     return cvsl;
   }
-  
+
   /**
    * @return the internal list of authorization failures
    * @deprecated since 1.5, see {@link #getAuthorizationFailuresMap()}
@@ -116,7 +116,7 @@ public class MutationsRejectedException extends AccumuloException {
   public List<KeyExtent> getAuthorizationFailures() {
     return new ArrayList<KeyExtent>(af.keySet());
   }
-  
+
   /**
    * @return the internal mapping of keyextent mappings to SecurityErrorCode
    * @since 1.5.0
@@ -124,18 +124,18 @@ public class MutationsRejectedException extends AccumuloException {
   public Map<KeyExtent,Set<SecurityErrorCode>> getAuthorizationFailuresMap() {
     return af;
   }
-  
+
   /**
-   * 
+   *
    * @return A list of servers that had internal errors when mutations were written
-   * 
+   *
    */
   public Collection<String> getErrorServers() {
     return es;
   }
-  
+
   /**
-   * 
+   *
    * @return a count of unknown exceptions that occurred during processing
    */
   public int getUnknownExceptions() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/NamespaceExistsException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/NamespaceExistsException.java b/core/src/main/java/org/apache/accumulo/core/client/NamespaceExistsException.java
index d2cb607..223ab3f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/NamespaceExistsException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/NamespaceExistsException.java
@@ -24,7 +24,7 @@ import org.apache.accumulo.core.client.impl.thrift.ThriftTableOperationException
 public class NamespaceExistsException extends Exception {
   /**
    * Exception to throw if an operation is attempted on a namespace that already exists.
-   * 
+   *
    */
   private static final long serialVersionUID = 1L;
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/NamespaceNotFoundException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/NamespaceNotFoundException.java b/core/src/main/java/org/apache/accumulo/core/client/NamespaceNotFoundException.java
index 031cf16..e330d69 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/NamespaceNotFoundException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/NamespaceNotFoundException.java
@@ -24,7 +24,7 @@ import org.apache.accumulo.core.client.impl.thrift.ThriftTableOperationException
 public class NamespaceNotFoundException extends Exception {
   /**
    * Exception to throw if an operation is attempted on a namespace that doesn't exist.
-   * 
+   *
    */
   private static final long serialVersionUID = 1L;
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/RowIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/RowIterator.java b/core/src/main/java/org/apache/accumulo/core/client/RowIterator.java
index f5e9547..190b0b2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/RowIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/RowIterator.java
@@ -27,13 +27,13 @@ import org.apache.hadoop.io.Text;
 
 /**
  * Group Key/Value pairs into Iterators over rows. Suggested usage:
- * 
+ *
  * <pre>
  * RowIterator rowIterator = new RowIterator(connector.createScanner(tableName, authorizations));
  * </pre>
  */
 public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
-  
+
   /**
    * Iterate over entries in a single row.
    */
@@ -42,7 +42,7 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
     private Text currentRow = null;
     private long count = 0;
     private boolean disabled = false;
-    
+
     /**
      * SingleRowIter must be passed a PeekingIterator so that it can peek at the next entry to see if it belongs in the current row or not.
      */
@@ -51,21 +51,21 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
       if (source.hasNext())
         currentRow = source.peek().getKey().getRow();
     }
-    
+
     @Override
     public boolean hasNext() {
       if (disabled)
         throw new IllegalStateException("SingleRowIter no longer valid");
       return currentRow != null;
     }
-    
+
     @Override
     public Entry<Key,Value> next() {
       if (disabled)
         throw new IllegalStateException("SingleRowIter no longer valid");
       return _next();
     }
-    
+
     private Entry<Key,Value> _next() {
       if (currentRow == null)
         throw new NoSuchElementException();
@@ -76,19 +76,19 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
       }
       return kv;
     }
-    
+
     @Override
     public void remove() {
       throw new UnsupportedOperationException();
     }
-    
+
     /**
      * Get a count of entries read from the row (only equals the number of entries in the row when the row has been read fully).
      */
     public long getCount() {
       return count;
     }
-    
+
     /**
      * Consume the rest of the row. Disables the iterator from future use.
      */
@@ -98,28 +98,28 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
         _next();
     }
   }
-  
+
   private final PeekingIterator<Entry<Key,Value>> iter;
   private long count = 0;
   private SingleRowIter lastIter = null;
-  
+
   /**
    * Create an iterator from an (ordered) sequence of KeyValue pairs.
    */
   public RowIterator(Iterator<Entry<Key,Value>> iterator) {
     this.iter = new PeekingIterator<Entry<Key,Value>>(iterator);
   }
-  
+
   /**
    * Create an iterator from an Iterable.
    */
   public RowIterator(Iterable<Entry<Key,Value>> iterable) {
     this(iterable.iterator());
   }
-  
+
   /**
    * Returns true if there is at least one more row to get.
-   * 
+   *
    * If the last row hasn't been fully read, this method will read through the end of the last row so it can determine if the underlying iterator has a next
    * row. The last row is disabled from future use.
    */
@@ -132,7 +132,7 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
     }
     return iter.hasNext();
   }
-  
+
   /**
    * Fetch the next row.
    */
@@ -142,7 +142,7 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
       throw new NoSuchElementException();
     return lastIter = new SingleRowIter(iter);
   }
-  
+
   /**
    * Unsupported.
    */
@@ -150,7 +150,7 @@ public class RowIterator implements Iterator<Iterator<Entry<Key,Value>>> {
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * Get a count of the total number of entries in all rows read so far.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/Scanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/Scanner.java b/core/src/main/java/org/apache/accumulo/core/client/Scanner.java
index 112179e..372ee42 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/Scanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/Scanner.java
@@ -20,83 +20,86 @@ import org.apache.accumulo.core.data.Range;
 
 /**
  * Scans a table over a given range.
- * 
+ *
  * "Clients can iterate over multiple column families, and there are several mechanisms for limiting the rows, columns, and timestamps traversed by a scan. For
  * example, we could restrict [a] scan ... to only produce anchors whose columns match [a] regular expression ..., or to only produce anchors whose timestamps
  * fall within ten days of the current time."
  */
 public interface Scanner extends ScannerBase {
-  
+
   /**
    * This setting determines how long a scanner will automatically retry when a failure occurs. By default a scanner will retry forever.
-   * 
+   *
    * @param timeOut
    *          in seconds
    * @deprecated Since 1.5. See {@link ScannerBase#setTimeout(long, java.util.concurrent.TimeUnit)}
    */
   @Deprecated
   void setTimeOut(int timeOut);
-  
+
   /**
    * Returns the setting for how long a scanner will automatically retry when a failure occurs.
-   * 
+   *
    * @return the timeout configured for this scanner
    * @deprecated Since 1.5. See {@link ScannerBase#getTimeout(java.util.concurrent.TimeUnit)}
    */
   @Deprecated
   int getTimeOut();
-  
+
   /**
    * Sets the range of keys to scan over.
-   * 
+   *
    * @param range
    *          key range to begin and end scan
    */
   void setRange(Range range);
-  
+
   /**
    * Returns the range of keys to scan over.
-   * 
+   *
    * @return the range configured for this scanner
    */
   Range getRange();
-  
+
   /**
    * Sets the number of Key/Value pairs that will be fetched at a time from a tablet server.
-   * 
+   *
    * @param size
    *          the number of Key/Value pairs to fetch per call to Accumulo
    */
   void setBatchSize(int size);
-  
+
   /**
    * Returns the batch size (number of Key/Value pairs) that will be fetched at a time from a tablet server.
-   * 
+   *
    * @return the batch size configured for this scanner
    */
   int getBatchSize();
-  
+
   /**
    * Enables row isolation. Writes that occur to a row after a scan of that row has begun will not be seen if this option is enabled.
    */
   void enableIsolation();
-  
+
   /**
    * Disables row isolation. Writes that occur to a row after a scan of that row has begun may be seen if this option is enabled.
    */
   void disableIsolation();
-  
+
   /**
    * The number of batches of Key/Value pairs returned before the {@link Scanner} will begin to prefetch the next batch
+   *
    * @return Number of batches before read-ahead begins
    * @since 1.6.0
    */
   long getReadaheadThreshold();
-  
+
   /**
    * Sets the number of batches of Key/Value pairs returned before the {@link Scanner} will begin to prefetch the next batch
-   * @param batches Non-negative number of batches
+   *
+   * @param batches
+   *          Non-negative number of batches
    * @since 1.6.0
    */
-  void setReadaheadThreshold(long batches); 
+  void setReadaheadThreshold(long batches);
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/ScannerBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/ScannerBase.java b/core/src/main/java/org/apache/accumulo/core/client/ScannerBase.java
index 335b63a..82a3299 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/ScannerBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/ScannerBase.java
@@ -26,13 +26,13 @@ import org.apache.hadoop.io.Text;
 
 /**
  * This class hosts configuration methods that are shared between different types of scanners.
- * 
+ *
  */
 public interface ScannerBase extends Iterable<Entry<Key,Value>> {
-  
+
   /**
    * Add a server-side scan iterator.
-   * 
+   *
    * @param cfg
    *          fully specified scan-time iterator, including all options for the iterator. Any changes to the iterator setting after this call are not propagated
    *          to the stored iterator.
@@ -40,19 +40,19 @@ public interface ScannerBase extends Iterable<Entry<Key,Value>> {
    *           if the setting conflicts with existing iterators
    */
   void addScanIterator(IteratorSetting cfg);
-  
+
   /**
    * Remove an iterator from the list of iterators.
-   * 
+   *
    * @param iteratorName
    *          nickname used for the iterator
    */
   void removeScanIterator(String iteratorName);
-  
+
   /**
    * Update the options for an iterator. Note that this does <b>not</b> change the iterator options during a scan, it just replaces the given option on a
    * configured iterator before a scan is started.
-   * 
+   *
    * @param iteratorName
    *          the name of the iterator to change
    * @param key
@@ -61,61 +61,61 @@ public interface ScannerBase extends Iterable<Entry<Key,Value>> {
    *          the new value for the named option
    */
   void updateScanIteratorOption(String iteratorName, String key, String value);
-  
+
   /**
    * Adds a column family to the list of columns that will be fetched by this scanner. By default when no columns have been added the scanner fetches all
    * columns.
-   * 
+   *
    * @param col
    *          the column family to be fetched
    */
   void fetchColumnFamily(Text col);
-  
+
   /**
    * Adds a column to the list of columns that will be fetched by this scanner. The column is identified by family and qualifier. By default when no columns
    * have been added the scanner fetches all columns.
-   * 
+   *
    * @param colFam
    *          the column family of the column to be fetched
    * @param colQual
    *          the column qualifier of the column to be fetched
    */
   void fetchColumn(Text colFam, Text colQual);
-  
+
   /**
    * Clears the columns to be fetched (useful for resetting the scanner for reuse). Once cleared, the scanner will fetch all columns.
    */
   void clearColumns();
-  
+
   /**
    * Clears scan iterators prior to returning a scanner to the pool.
    */
   void clearScanIterators();
-  
+
   /**
    * Returns an iterator over an accumulo table. This iterator uses the options that are currently set for its lifetime, so setting options will have no effect
    * on existing iterators.
-   * 
+   *
    * Keys returned by the iterator are not guaranteed to be in sorted order.
-   * 
+   *
    * @return an iterator over Key,Value pairs which meet the restrictions set on the scanner
    */
   Iterator<Entry<Key,Value>> iterator();
-  
+
   /**
    * This setting determines how long a scanner will automatically retry when a failure occurs. By default a scanner will retry forever.
-   * 
+   *
    * Setting to zero or Long.MAX_VALUE and TimeUnit.MILLISECONDS means to retry forever.
-   * 
+   *
    * @param timeUnit
    *          determines how timeout is interpreted
    * @since 1.5.0
    */
   void setTimeout(long timeOut, TimeUnit timeUnit);
-  
+
   /**
    * Returns the setting for how long a scanner will automatically retry when a failure occurs.
-   * 
+   *
    * @return the timeout configured for this scanner
    * @since 1.5.0
    */
@@ -123,6 +123,7 @@ public interface ScannerBase extends Iterable<Entry<Key,Value>> {
 
   /**
    * Closes any underlying connections on the scanner
+   *
    * @since 1.5.0
    */
   void close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/TableDeletedException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/TableDeletedException.java b/core/src/main/java/org/apache/accumulo/core/client/TableDeletedException.java
index bc0b105..c1b1d6f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/TableDeletedException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/TableDeletedException.java
@@ -18,21 +18,21 @@ package org.apache.accumulo.core.client;
 
 /**
  * This exception is thrown if a table is deleted after an operation starts.
- * 
+ *
  * For example if table A exist when a scan is started, but is deleted during the scan then this exception is thrown.
- * 
+ *
  */
 
 public class TableDeletedException extends RuntimeException {
-  
+
   private static final long serialVersionUID = 1L;
   private String tableId;
-  
+
   public TableDeletedException(String tableId) {
     super("Table ID " + tableId + " was deleted");
     this.tableId = tableId;
   }
-  
+
   public String getTableId() {
     return tableId;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/TableExistsException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/TableExistsException.java b/core/src/main/java/org/apache/accumulo/core/client/TableExistsException.java
index bec008d..72c633c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/TableExistsException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/TableExistsException.java
@@ -24,10 +24,10 @@ import org.apache.accumulo.core.client.impl.thrift.ThriftTableOperationException
 public class TableExistsException extends Exception {
   /**
    * Exception to throw if an operation is attempted on a table that already exists.
-   * 
+   *
    */
   private static final long serialVersionUID = 1L;
-  
+
   /**
    * @param tableId
    *          the internal id of the table that exists
@@ -40,7 +40,7 @@ public class TableExistsException extends Exception {
     super("Table" + (tableName != null && !tableName.isEmpty() ? " " + tableName : "") + (tableId != null && !tableId.isEmpty() ? " (Id=" + tableId + ")" : "")
         + " exists" + (description != null && !description.isEmpty() ? " (" + description + ")" : ""));
   }
-  
+
   /**
    * @param tableId
    *          the internal id of the table that exists
@@ -55,7 +55,7 @@ public class TableExistsException extends Exception {
     this(tableId, tableName, description);
     super.initCause(cause);
   }
-  
+
   /**
    * @param e
    *          constructs an exception from a thrift exception

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/TableNotFoundException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/TableNotFoundException.java b/core/src/main/java/org/apache/accumulo/core/client/TableNotFoundException.java
index 6d27336..1e53936 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/TableNotFoundException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/TableNotFoundException.java
@@ -25,7 +25,7 @@ import org.apache.accumulo.core.client.impl.thrift.ThriftTableOperationException
 public class TableNotFoundException extends Exception {
   /**
    * Exception to throw if an operation is attempted on a table that doesn't exist.
-   * 
+   *
    */
   private static final long serialVersionUID = 1L;
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/TableOfflineException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/TableOfflineException.java b/core/src/main/java/org/apache/accumulo/core/client/TableOfflineException.java
index 9edb904..ef63228 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/TableOfflineException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/TableOfflineException.java
@@ -19,9 +19,9 @@ package org.apache.accumulo.core.client;
 import org.apache.accumulo.core.client.impl.Tables;
 
 public class TableOfflineException extends RuntimeException {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   private static String getTableName(Instance instance, String tableId) {
     if (tableId == null)
       return " <unknown table> ";
@@ -32,7 +32,7 @@ public class TableOfflineException extends RuntimeException {
       return " <unknown table> (" + tableId + ")";
     }
   }
-  
+
   public TableOfflineException(Instance instance, String tableId) {
     super("Table " + getTableName(instance, tableId) + " is offline");
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/TimedOutException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/TimedOutException.java b/core/src/main/java/org/apache/accumulo/core/client/TimedOutException.java
index 4b142e9..5ec9c59 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/TimedOutException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/TimedOutException.java
@@ -24,23 +24,23 @@ import java.util.Set;
  * @since 1.5.0
  */
 public class TimedOutException extends RuntimeException {
-  
+
   private Set<String> timedoutServers;
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   private static String shorten(Set<String> set) {
     if (set.size() < 10) {
       return set.toString();
     }
-    
+
     return new ArrayList<String>(set).subList(0, 10).toString() + " ... " + (set.size() - 10) + " servers not shown";
   }
 
   public TimedOutException(Set<String> timedoutServers) {
     super("Servers timed out " + shorten(timedoutServers));
     this.timedoutServers = timedoutServers;
-    
+
   }
 
   public TimedOutException(String msg) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/ZooKeeperInstance.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/ZooKeeperInstance.java b/core/src/main/java/org/apache/accumulo/core/client/ZooKeeperInstance.java
index b12e189..7c8f2e2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/ZooKeeperInstance.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/ZooKeeperInstance.java
@@ -129,8 +129,8 @@ public class ZooKeeperInstance implements Instance {
 
   /**
    * @param config
-   *          Client configuration for specifying connection options.
-   *          See {@link ClientConfiguration} which extends Configuration with convenience methods specific to Accumulo.
+   *          Client configuration for specifying connection options. See {@link ClientConfiguration} which extends Configuration with convenience methods
+   *          specific to Accumulo.
    * @since 1.6.0
    */
   public ZooKeeperInstance(Configuration config) {
@@ -140,7 +140,7 @@ public class ZooKeeperInstance implements Instance {
   ZooKeeperInstance(Configuration config, ZooCacheFactory zcf) {
     checkArgument(config != null, "config is null");
     if (config instanceof ClientConfiguration) {
-      this.clientConf = (ClientConfiguration)config;
+      this.clientConf = (ClientConfiguration) config;
     } else {
       this.clientConf = new ClientConfiguration(config);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveCompaction.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveCompaction.java b/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveCompaction.java
index 41b9c67..b9de4c2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveCompaction.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveCompaction.java
@@ -22,9 +22,8 @@ import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.data.KeyExtent;
 
-
 /**
- * 
+ *
  * @since 1.5.0
  */
 public abstract class ActiveCompaction {
@@ -47,7 +46,7 @@ public abstract class ActiveCompaction {
      */
     FULL
   };
-  
+
   public static enum CompactionReason {
     /**
      * compaction initiated by user
@@ -70,48 +69,48 @@ public abstract class ActiveCompaction {
      */
     CLOSE
   };
-  
+
   /**
-   * 
+   *
    * @return name of the table the compaction is running against
    */
   public abstract String getTable() throws TableNotFoundException;
-  
+
   /**
    * @return tablet thats is compacting
    */
   public abstract KeyExtent getExtent();
-  
+
   /**
    * @return how long the compaction has been running in milliseconds
    */
   public abstract long getAge();
-  
+
   /**
    * @return the files the compaction is reading from
    */
   public abstract List<String> getInputFiles();
-  
+
   /**
    * @return file compactions is writing too
    */
   public abstract String getOutputFile();
-  
+
   /**
    * @return the type of compaction
    */
   public abstract CompactionType getType();
-  
+
   /**
    * @return the reason the compaction was started
    */
   public abstract CompactionReason getReason();
-  
+
   /**
    * @return the locality group that is compacting
    */
   public abstract String getLocalityGroup();
-  
+
   /**
    * @return the number of key/values read by the compaction
    */
@@ -121,7 +120,7 @@ public abstract class ActiveCompaction {
    * @return the number of key/values written by the compaction
    */
   public abstract long getEntriesWritten();
-  
+
   /**
    * @return the per compaction iterators configured
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveScan.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveScan.java b/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveScan.java
index fc9808f..cde5d27 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveScan.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/ActiveScan.java
@@ -25,7 +25,7 @@ import org.apache.accumulo.core.security.Authorizations;
 
 /**
  * A class that contains information about an ActiveScan
- * 
+ *
  */
 public abstract class ActiveScan {
 
@@ -33,62 +33,62 @@ public abstract class ActiveScan {
    * @return an id that uniquely identifies that scan on the server
    */
   public abstract long getScanid();
-  
+
   /**
    * @return the address of the client that initiated the scan
    */
   public abstract String getClient();
-  
+
   /**
    * @return the user that initiated the scan
    */
   public abstract String getUser();
-  
+
   /**
    * @return the table the scan is running against
    */
   public abstract String getTable();
-  
+
   /**
    * @return the age of the scan in milliseconds
    */
   public abstract long getAge();
-  
+
   /**
    * @return milliseconds since last time client read data from the scan
    */
   public abstract long getLastContactTime();
-  
+
   public abstract ScanType getType();
-  
+
   public abstract ScanState getState();
-  
+
   /**
    * @return tablet the scan is running against, if a batch scan may be one of many or null
    */
   public abstract KeyExtent getExtent();
-  
+
   /**
    * @return columns requested by the scan
    */
   public abstract List<Column> getColumns();
-  
+
   /**
    * @return server side iterators used by the scan
    */
   public abstract List<String> getSsiList();
-  
+
   /**
    * @return server side iterator options
    */
   public abstract Map<String,Map<String,String>> getSsio();
-  
+
   /**
    * @return the authorizations being used for this scan
    * @since 1.5.0
    */
   public abstract Authorizations getAuthorizations();
-  
+
   /**
    * @return the time this scan has been idle in the tablet server
    * @since 1.5.0

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionConfig.java b/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionConfig.java
index f59c70b..38e5efd 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionConfig.java
@@ -22,13 +22,14 @@ import java.util.Collections;
 import java.util.List;
 import java.util.Map;
 
-import com.google.common.base.Preconditions;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.hadoop.io.Text;
 
+import com.google.common.base.Preconditions;
+
 /**
  * This class exist to pass parameters to {@link TableOperations#compact(String, CompactionConfig)}
- * 
+ *
  * @since 1.7.0
  */
 
@@ -64,7 +65,7 @@ public class CompactionConfig {
   }
 
   /**
-   * 
+   *
    * @param end
    *          Last tablet to be compacted contains this row, null means the last tablet in table. The default is null.
    * @return this
@@ -110,7 +111,7 @@ public class CompactionConfig {
   }
 
   /**
-   * 
+   *
    * @return The previously set wait. The default is true.
    */
   public boolean getWait() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionStrategyConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionStrategyConfig.java b/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionStrategyConfig.java
index 14b275e..c23b511 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionStrategyConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/CompactionStrategyConfig.java
@@ -24,7 +24,7 @@ import java.util.Map;
 import com.google.common.base.Preconditions;
 
 /**
- * 
+ *
  * @since 1.7.0
  */
 
@@ -33,7 +33,7 @@ public class CompactionStrategyConfig {
   private Map<String,String> options = Collections.emptyMap();
 
   /**
-   * 
+   *
    * @param className
    *          The name of a class that implements org.apache.accumulo.tserver.compaction.CompactionStrategy. This class must be exist on tservers.
    */
@@ -64,7 +64,7 @@ public class CompactionStrategyConfig {
   }
 
   /**
-   * 
+   *
    * @return The previously set options. Returns an unmodifiable map. The default is an empty map.
    */
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/DiskUsage.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/DiskUsage.java b/core/src/main/java/org/apache/accumulo/core/client/admin/DiskUsage.java
index 5c549ff..b62843b 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/DiskUsage.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/DiskUsage.java
@@ -19,47 +19,47 @@ package org.apache.accumulo.core.client.admin;
 import java.util.SortedSet;
 
 public class DiskUsage {
-  
+
   protected final SortedSet<String> tables;
   protected Long usage;
-  
+
   public DiskUsage(SortedSet<String> tables, Long usage) {
     this.tables = tables;
     this.usage = usage;
   }
-  
+
   public SortedSet<String> getTables() {
     return tables;
   }
-  
+
   public Long getUsage() {
     return usage;
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (this == o)
       return true;
     if (!(o instanceof DiskUsage))
       return false;
-    
+
     DiskUsage diskUsage = (DiskUsage) o;
-    
+
     if (tables != null ? !tables.equals(diskUsage.tables) : diskUsage.tables != null)
       return false;
     if (usage != null ? !usage.equals(diskUsage.usage) : diskUsage.usage != null)
       return false;
-    
+
     return true;
   }
-  
+
   @Override
   public int hashCode() {
     int result = tables != null ? tables.hashCode() : 0;
     result = 31 * result + (usage != null ? usage.hashCode() : 0);
     return result;
   }
-  
+
   @Override
   public String toString() {
     return "DiskUsage{" + "tables=" + tables + ", usage=" + usage + '}';

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/FindMax.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/FindMax.java b/core/src/main/java/org/apache/accumulo/core/client/admin/FindMax.java
index 3bb56a8..00fe950 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/FindMax.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/FindMax.java
@@ -36,35 +36,35 @@ public class FindMax {
     for (int i = 0; i < num; i++)
       baos.write(0);
   }
-  
+
   private static Text findMidPoint(Text minBS, Text maxBS) {
     ByteArrayOutputStream startOS = new ByteArrayOutputStream();
     startOS.write(0); // add a leading zero so bigint does not think its negative
     startOS.write(minBS.getBytes(), 0, minBS.getLength());
-    
+
     ByteArrayOutputStream endOS = new ByteArrayOutputStream();
     endOS.write(0);// add a leading zero so bigint does not think its negative
     endOS.write(maxBS.getBytes(), 0, maxBS.getLength());
-    
+
     // make the numbers of the same magnitude
     if (startOS.size() < endOS.size())
       appendZeros(startOS, endOS.size() - startOS.size());
     else if (endOS.size() < startOS.size())
       appendZeros(endOS, startOS.size() - endOS.size());
-    
+
     BigInteger min = new BigInteger(startOS.toByteArray());
     BigInteger max = new BigInteger(endOS.toByteArray());
-    
+
     BigInteger mid = max.subtract(min).divide(BigInteger.valueOf(2)).add(min);
-    
+
     byte[] ba = mid.toByteArray();
-    
+
     Text ret = new Text();
-    
+
     if (ba.length == startOS.size()) {
       if (ba[0] != 0)
         throw new RuntimeException();
-      
+
       // big int added a zero so it would not be negative, drop it
       ret.set(ba, 1, ba.length - 1);
     } else {
@@ -73,27 +73,27 @@ public class FindMax {
       for (int i = ba.length; i < expLen; i++) {
         ret.append(new byte[] {0}, 0, 1);
       }
-      
+
       ret.append(ba, 0, ba.length);
     }
-    
+
     // remove trailing 0x0 bytes
     while (ret.getLength() > 0 && ret.getBytes()[ret.getLength() - 1] == 0 && ret.compareTo(minBS) > 0) {
       Text t = new Text();
       t.set(ret.getBytes(), 0, ret.getLength() - 1);
       ret = t;
     }
-    
+
     return ret;
   }
-  
+
   private static Text _findMax(Scanner scanner, Text start, boolean inclStart, Text end, boolean inclEnd) {
-    
+
     // System.out.printf("findMax(%s, %s, %s, %s)%n", Key.toPrintableString(start.getBytes(), 0, start.getLength(), 1000), inclStart,
     // Key.toPrintableString(end.getBytes(), 0, end.getLength(), 1000), inclEnd);
-    
+
     int cmp = start.compareTo(end);
-    
+
     if (cmp >= 0) {
       if (inclStart && inclEnd && cmp == 0) {
         scanner.setRange(new Range(start, true, end, true));
@@ -101,45 +101,45 @@ public class FindMax {
         if (iter.hasNext())
           return iter.next().getKey().getRow();
       }
-      
+
       return null;
     }
-    
+
     Text mid = findMidPoint(start, end);
     // System.out.println("mid = :"+Key.toPrintableString(mid.getBytes(), 0, mid.getLength(), 1000)+":");
-    
+
     scanner.setRange(new Range(mid, mid.equals(start) ? inclStart : true, end, inclEnd));
-    
+
     Iterator<Entry<Key,Value>> iter = scanner.iterator();
-    
+
     if (iter.hasNext()) {
       Key next = iter.next().getKey();
-      
+
       int count = 0;
       while (count < 10 && iter.hasNext()) {
         next = iter.next().getKey();
         count++;
       }
-      
+
       if (!iter.hasNext())
         return next.getRow();
-      
+
       Text ret = _findMax(scanner, next.followingKey(PartialKey.ROW).getRow(), true, end, inclEnd);
       if (ret == null)
         return next.getRow();
       else
         return ret;
     } else {
-      
+
       return _findMax(scanner, start, inclStart, mid, mid.equals(start) ? inclStart : false);
     }
   }
-  
+
   private static Text findInitialEnd(Scanner scanner) {
     Text end = new Text(new byte[] {(byte) 0xff, (byte) 0xff, (byte) 0xff, (byte) 0xff});
-    
+
     scanner.setRange(new Range(end, null));
-    
+
     while (scanner.iterator().hasNext()) {
       Text t = new Text();
       t.append(end.getBytes(), 0, end.getLength());
@@ -147,24 +147,24 @@ public class FindMax {
       end = t;
       scanner.setRange(new Range(end, null));
     }
-    
+
     return end;
   }
-  
+
   public static Text findMax(Scanner scanner, Text start, boolean is, Text end, boolean ie) throws TableNotFoundException {
-    
+
     scanner.setBatchSize(12);
     IteratorSetting cfg = new IteratorSetting(Integer.MAX_VALUE, SortedKeyIterator.class);
     scanner.addScanIterator(cfg);
-    
+
     if (start == null) {
       start = new Text();
       is = true;
     }
-    
+
     if (end == null)
       end = findInitialEnd(scanner);
-    
+
     return _findMax(scanner, start, is, end, ie);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
index 04e1d0d..4bb4747 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
@@ -23,14 +23,14 @@ import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 
 /**
- * 
+ *
  */
 public interface InstanceOperations {
-  
+
   /**
    * Sets an system property in zookeeper. Tablet servers will pull this setting and override the equivalent setting in accumulo-site.xml. Changes can be seen
    * using {@link #getSystemConfiguration()}
-   * 
+   *
    * @param property
    *          the name of a per-table property
    * @param value
@@ -41,10 +41,10 @@ public interface InstanceOperations {
    *           if the user does not have permission
    */
   void setProperty(final String property, final String value) throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
    * Removes a system property from zookeeper. Changes can be seen using {@link #getSystemConfiguration()}
-   * 
+   *
    * @param property
    *          the name of a per-table property
    * @throws AccumuloException
@@ -53,72 +53,73 @@ public interface InstanceOperations {
    *           if the user does not have permission
    */
   void removeProperty(final String property) throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
-   * 
+   *
    * @return A map of system properties set in zookeeper. If a property is not set in zookeeper, then it will return the value set in accumulo-site.xml on some
    *         server. If nothing is set in an accumulo-site.xml file it will return the default value for each property.
    */
 
   Map<String,String> getSystemConfiguration() throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
-   * 
+   *
    * @return A map of system properties set in accumulo-site.xml on some server. If nothing is set in an accumulo-site.xml file it will return the default value
    *         for each property.
    */
 
   Map<String,String> getSiteConfiguration() throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
    * List the currently active tablet servers participating in the accumulo instance
-   * 
+   *
    * @return A list of currently active tablet servers.
    */
-  
+
   List<String> getTabletServers();
-  
+
   /**
    * List the active scans on tablet server.
-   * 
+   *
    * @param tserver
    *          The tablet server address should be of the form <ip address>:<port>
    * @return A list of active scans on tablet server.
    */
-  
+
   List<ActiveScan> getActiveScans(String tserver) throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
    * List the active compaction running on a tablet server
-   * 
+   *
    * @param tserver
    *          The tablet server address should be of the form <ip address>:<port>
    * @return the list of active compactions
    * @since 1.5.0
    */
-  
+
   List<ActiveCompaction> getActiveCompactions(String tserver) throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
    * Throws an exception if a tablet server can not be contacted.
-   * 
+   *
    * @param tserver
    *          The tablet server address should be of the form <ip address>:<port>
    * @since 1.5.0
    */
   void ping(String tserver) throws AccumuloException;
-  
+
   /**
    * Test to see if the instance can load the given class as the given type. This check does not consider per table classpaths, see
    * {@link TableOperations#testClassLoad(String, String, String)}
-   * 
+   *
    * @return true if the instance can load the given class as the given type, false otherwise
    */
   boolean testClassLoad(final String className, final String asTypeName) throws AccumuloException, AccumuloSecurityException;
-  
+
   /**
-   *  Waits for the tablet balancer to run and return no migrations.
-   *  @since 1.7.0
+   * Waits for the tablet balancer to run and return no migrations.
+   *
+   * @since 1.7.0
    */
   void waitForBalance() throws AccumuloException;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/NamespaceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/NamespaceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/NamespaceOperations.java
index c6dff92..b2e7198 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/NamespaceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/NamespaceOperations.java
@@ -31,17 +31,17 @@ import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 
 /**
  * Provides an API for administering namespaces
- * 
+ *
  * All tables exist in a namespace. The default namespace has no name, and is used if an explicit namespace is not specified. Fully qualified table names look
  * like "namespaceName.tableName". Tables in the default namespace are fully qualified simply as "tableName".
- * 
+ *
  * @since 1.6.0
  */
 public interface NamespaceOperations {
 
   /**
    * Returns the name of the system reserved namespace
-   * 
+   *
    * @return the name of the system namespace
    * @since 1.6.0
    */
@@ -49,7 +49,7 @@ public interface NamespaceOperations {
 
   /**
    * Returns the name of the default namespace
-   * 
+   *
    * @return the name of the default namespace
    * @since 1.6.0
    */
@@ -57,7 +57,7 @@ public interface NamespaceOperations {
 
   /**
    * Retrieve a list of namespaces in Accumulo.
-   * 
+   *
    * @return List of namespaces in accumulo
    * @throws AccumuloException
    *           if a general error occurs
@@ -69,7 +69,7 @@ public interface NamespaceOperations {
 
   /**
    * A method to check if a namespace exists in Accumulo.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @return true if the namespace exists
@@ -83,7 +83,7 @@ public interface NamespaceOperations {
 
   /**
    * Create an empty namespace with no initial configuration. Valid names for a namespace contain letters, numbers, and the underscore character.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @throws AccumuloException
@@ -98,7 +98,7 @@ public interface NamespaceOperations {
 
   /**
    * Delete an empty namespace
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @throws AccumuloException
@@ -115,7 +115,7 @@ public interface NamespaceOperations {
 
   /**
    * Rename a namespace
-   * 
+   *
    * @param oldNamespaceName
    *          the old namespace name
    * @param newNamespaceName
@@ -135,7 +135,7 @@ public interface NamespaceOperations {
 
   /**
    * Sets a property on a namespace which applies to all tables in the namespace. Note that it may take a few seconds to propagate the change everywhere.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param property
@@ -154,7 +154,7 @@ public interface NamespaceOperations {
 
   /**
    * Removes a property from a namespace. Note that it may take a few seconds to propagate the change everywhere.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param property
@@ -171,7 +171,7 @@ public interface NamespaceOperations {
 
   /**
    * Gets properties of a namespace, which are inherited by tables in this namespace. Note that recently changed properties may not be available immediately.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @return all properties visible by this namespace (system and per-table properties). Note that recently changed properties may not be visible immediately.
@@ -187,7 +187,7 @@ public interface NamespaceOperations {
 
   /**
    * Get a mapping of namespace name to internal namespace id.
-   * 
+   *
    * @return the map from namespace name to internal namespace id
    * @throws AccumuloException
    *           if a general error occurs
@@ -199,7 +199,7 @@ public interface NamespaceOperations {
 
   /**
    * Add an iterator to a namespace on all scopes.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param setting
@@ -216,7 +216,7 @@ public interface NamespaceOperations {
 
   /**
    * Add an iterator to a namespace on the given scopes.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param setting
@@ -236,7 +236,7 @@ public interface NamespaceOperations {
 
   /**
    * Remove an iterator from a namespace by name.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param name
@@ -256,7 +256,7 @@ public interface NamespaceOperations {
 
   /**
    * Get the settings for an iterator.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param name
@@ -277,7 +277,7 @@ public interface NamespaceOperations {
 
   /**
    * Get a list of iterators for this namespace.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @return a set of iterator names
@@ -294,7 +294,7 @@ public interface NamespaceOperations {
   /**
    * Check whether a given iterator configuration conflicts with existing configuration; in particular, determine if the name or priority are already in use for
    * the specified scopes. If so, an IllegalArgumentException is thrown, wrapped in an AccumuloException.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param setting
@@ -314,7 +314,7 @@ public interface NamespaceOperations {
 
   /**
    * Add a new constraint to a namespace.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param constraintClassName
@@ -332,7 +332,7 @@ public interface NamespaceOperations {
 
   /**
    * Remove a constraint from a namespace.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param id
@@ -349,7 +349,7 @@ public interface NamespaceOperations {
 
   /**
    * List constraints on a namespace with their assigned numbers.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @return a map from constraint class name to assigned number
@@ -365,7 +365,7 @@ public interface NamespaceOperations {
 
   /**
    * Test to see if the instance can load the given class as the given type. This check uses the table classpath property if it is set.
-   * 
+   *
    * @param namespace
    *          the name of the namespace
    * @param className

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/ReplicationOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/ReplicationOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/ReplicationOperations.java
index 4e24552..699d9d5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/ReplicationOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/ReplicationOperations.java
@@ -27,52 +27,63 @@ import org.apache.accumulo.core.client.replication.ReplicaSystem;
 
 /**
  * Supports replication configuration
+ *
  * @since 1.7.0
  */
 public interface ReplicationOperations {
 
   /**
    * Defines a cluster with the given name using the given {@link ReplicaSystem} implementation.
-   * @param name Name of the cluster, used for configuring replication on tables
-   * @param system Type of system to be replicated to
+   *
+   * @param name
+   *          Name of the cluster, used for configuring replication on tables
+   * @param system
+   *          Type of system to be replicated to
    */
   public void addPeer(String name, Class<? extends ReplicaSystem> system) throws AccumuloException, AccumuloSecurityException, PeerExistsException;
 
   /**
    * Defines a cluster with the given name and the given name system.
-   * @param name Unique name for the cluster
-   * @param replicaType {@link ReplicaSystem} class name to use to replicate the data
+   *
+   * @param name
+   *          Unique name for the cluster
+   * @param replicaType
+   *          {@link ReplicaSystem} class name to use to replicate the data
    */
   public void addPeer(String name, String replicaType) throws AccumuloException, AccumuloSecurityException, PeerExistsException;
 
   /**
    * Removes a cluster with the given name.
-   * @param name Name of the cluster to remove
+   *
+   * @param name
+   *          Name of the cluster to remove
    */
   public void removePeer(String name) throws AccumuloException, AccumuloSecurityException, PeerNotFoundException;
 
   /**
-   * Waits for a table to be fully replicated, given the state of files pending replication for the provided table
-   * at the point in time which this method is invoked.
-   * @param tableName The table to wait for
+   * Waits for a table to be fully replicated, given the state of files pending replication for the provided table at the point in time which this method is
+   * invoked.
+   *
+   * @param tableName
+   *          The table to wait for
    */
   public void drain(String tableName) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
 
   /**
-   * Given the provided set of files that are pending replication for a table, wait for those
-   * files to be fully replicated to all configured peers. This allows for the accurate calculation
-   * when a table, at a given point in time, has been fully replicated.
-   * @param tableName The table to wait for
+   * Given the provided set of files that are pending replication for a table, wait for those files to be fully replicated to all configured peers. This allows
+   * for the accurate calculation when a table, at a given point in time, has been fully replicated.
+   *
+   * @param tableName
+   *          The table to wait for
    */
   public void drain(String tableName, Set<String> files) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
 
   /**
-   * Gets all of the referenced files for a table from the metadata table. The result of this method
-   * is intended to be directly supplied to {@link #drain(String, Set)}. This helps determine when all
-   * data from a given point in time has been fully replicated.
+   * Gets all of the referenced files for a table from the metadata table. The result of this method is intended to be directly supplied to
+   * {@link #drain(String, Set)}. This helps determine when all data from a given point in time has been fully replicated.
    * <p>
-   * This also allows callers to get the {@link Set} of files for a table at some time, and later provide that
-   * {@link Set} to {@link #drain(String,Set)} to wait for all of those files to be replicated.
+   * This also allows callers to get the {@link Set} of files for a table at some time, and later provide that {@link Set} to {@link #drain(String,Set)} to wait
+   * for all of those files to be replicated.
    */
   public Set<String> referencedFiles(String tableName) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/SecurityOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/SecurityOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/SecurityOperations.java
index ea1fde1..efeafc0 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/SecurityOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/SecurityOperations.java
@@ -34,7 +34,7 @@ public interface SecurityOperations {
 
   /**
    * Create a user
-   * 
+   *
    * @param user
    *          the name of the user to create
    * @param password
@@ -52,7 +52,7 @@ public interface SecurityOperations {
 
   /**
    * Create a user
-   * 
+   *
    * @param principal
    *          the name of the user to create
    * @param password
@@ -67,7 +67,7 @@ public interface SecurityOperations {
 
   /**
    * Delete a user
-   * 
+   *
    * @param user
    *          the user name to delete
    * @throws AccumuloException
@@ -81,7 +81,7 @@ public interface SecurityOperations {
 
   /**
    * Delete a user
-   * 
+   *
    * @param principal
    *          the user name to delete
    * @throws AccumuloException
@@ -94,7 +94,7 @@ public interface SecurityOperations {
 
   /**
    * Verify a username/password combination is valid
-   * 
+   *
    * @param user
    *          the name of the user to authenticate
    * @param password
@@ -111,7 +111,7 @@ public interface SecurityOperations {
 
   /**
    * Verify a username/password combination is valid
-   * 
+   *
    * @param principal
    *          the name of the user to authenticate
    * @param token
@@ -127,7 +127,7 @@ public interface SecurityOperations {
 
   /**
    * Set the user's password
-   * 
+   *
    * @param user
    *          the name of the user to modify
    * @param password
@@ -144,7 +144,7 @@ public interface SecurityOperations {
 
   /**
    * Set the user's password
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param token
@@ -159,7 +159,7 @@ public interface SecurityOperations {
 
   /**
    * Set the user's record-level authorizations
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param authorizations
@@ -173,7 +173,7 @@ public interface SecurityOperations {
 
   /**
    * Retrieves the user's authorizations for scanning
-   * 
+   *
    * @param principal
    *          the name of the user to query
    * @return the set of authorizations the user has available for scanning
@@ -186,7 +186,7 @@ public interface SecurityOperations {
 
   /**
    * Verify the user has a particular system permission
-   * 
+   *
    * @param principal
    *          the name of the user to query
    * @param perm
@@ -201,7 +201,7 @@ public interface SecurityOperations {
 
   /**
    * Verify the user has a particular table permission
-   * 
+   *
    * @param principal
    *          the name of the user to query
    * @param table
@@ -218,7 +218,7 @@ public interface SecurityOperations {
 
   /**
    * Verify the user has a particular namespace permission
-   * 
+   *
    * @param principal
    *          the name of the user to query
    * @param namespace
@@ -235,7 +235,7 @@ public interface SecurityOperations {
 
   /**
    * Grant a user a system permission
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param permission
@@ -249,7 +249,7 @@ public interface SecurityOperations {
 
   /**
    * Grant a user a specific permission for a specific table
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param table
@@ -265,7 +265,7 @@ public interface SecurityOperations {
 
   /**
    * Grant a user a specific permission for a specific namespace
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param namespace
@@ -281,7 +281,7 @@ public interface SecurityOperations {
 
   /**
    * Revoke a system permission from a user
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param permission
@@ -295,7 +295,7 @@ public interface SecurityOperations {
 
   /**
    * Revoke a table permission for a specific user on a specific table
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param table
@@ -311,7 +311,7 @@ public interface SecurityOperations {
 
   /**
    * Revoke a namespace permission for a specific user on a specific namespace
-   * 
+   *
    * @param principal
    *          the name of the user to modify
    * @param namespace
@@ -327,7 +327,7 @@ public interface SecurityOperations {
 
   /**
    * Return a list of users in accumulo
-   * 
+   *
    * @return a set of user names
    * @throws AccumuloException
    *           if a general error occurs
@@ -340,7 +340,7 @@ public interface SecurityOperations {
 
   /**
    * Return a list of users in accumulo
-   * 
+   *
    * @return a set of user names
    * @throws AccumuloException
    *           if a general error occurs

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/TableOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/TableOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/TableOperations.java
index 63ac7b4..5c1260c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/TableOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/TableOperations.java
@@ -38,21 +38,21 @@ import org.apache.hadoop.io.Text;
 
 /**
  * Provides a class for administering tables
- * 
+ *
  */
 
 public interface TableOperations {
 
   /**
    * Retrieve a list of tables in Accumulo.
-   * 
+   *
    * @return List of tables in accumulo
    */
   SortedSet<String> list();
 
   /**
    * A method to check if a table exists in Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @return true if the table exists
@@ -61,7 +61,7 @@ public interface TableOperations {
 
   /**
    * Create a table with no special configuration
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @throws AccumuloException
@@ -126,7 +126,7 @@ public interface TableOperations {
 
   /**
    * Imports a table exported via exportTable and copied via hadoop distcp.
-   * 
+   *
    * @param tableName
    *          Name of a table to create and import into.
    * @param importDir
@@ -138,10 +138,10 @@ public interface TableOperations {
   /**
    * Exports a table. The tables data is not exported, just table metadata and a list of files to distcp. The table being exported must be offline and stay
    * offline for the duration of distcp. To avoid losing access to a table it can be cloned and the clone taken offline for export.
-   * 
+   *
    * <p>
    * See docs/examples/README.export
-   * 
+   *
    * @param tableName
    *          Name of the table to export.
    * @param exportDir
@@ -153,7 +153,8 @@ public interface TableOperations {
   /**
    * Ensures that tablets are split along a set of keys.
    * <p>
-   * Note that while the documentation for Text specifies that its bytestream should be UTF-8, the encoding is not enforced by operations that work with byte arrays.
+   * Note that while the documentation for Text specifies that its bytestream should be UTF-8, the encoding is not enforced by operations that work with byte
+   * arrays.
    * <p>
    * For example, you can create 256 evenly-sliced splits via the following code sample even though the given byte sequences are not valid UTF-8.
    *
@@ -232,7 +233,7 @@ public interface TableOperations {
 
   /**
    * Finds the max row within a given range. To find the max row in a table, pass null for start and end row.
-   * 
+   *
    * @param auths
    *          find the max row that can seen with these auths
    * @param startRow
@@ -243,7 +244,7 @@ public interface TableOperations {
    *          row to stop looking at, null means Infinity
    * @param endInclusive
    *          determines if the end row is included
-   * 
+   *
    * @return The max row in the range, or null if there is no visible data in the range.
    */
   Text getMaxRow(String tableName, Authorizations auths, Text startRow, boolean startInclusive, Text endRow, boolean endInclusive)
@@ -251,7 +252,7 @@ public interface TableOperations {
 
   /**
    * Merge tablets between (start, end]
-   * 
+   *
    * @param tableName
    *          the table to merge
    * @param start
@@ -263,7 +264,7 @@ public interface TableOperations {
 
   /**
    * Delete rows between (start, end]
-   * 
+   *
    * @param tableName
    *          the table to merge
    * @param start
@@ -275,7 +276,7 @@ public interface TableOperations {
 
   /**
    * Starts a full major compaction of the tablets in the range (start, end]. The compaction is preformed even for tablets that have only one file.
-   * 
+   *
    * @param tableName
    *          the table to compact
    * @param start
@@ -287,12 +288,11 @@ public interface TableOperations {
    * @param wait
    *          when true, the call will not return until compactions are finished
    */
-  void compact(String tableName, Text start, Text end, boolean flush, boolean wait) throws AccumuloSecurityException, TableNotFoundException,
-      AccumuloException;
+  void compact(String tableName, Text start, Text end, boolean flush, boolean wait) throws AccumuloSecurityException, TableNotFoundException, AccumuloException;
 
   /**
    * Starts a full major compaction of the tablets in the range (start, end]. The compaction is preformed even for tablets that have only one file.
-   * 
+   *
    * @param tableName
    *          the table to compact
    * @param start
@@ -313,16 +313,16 @@ public interface TableOperations {
   /**
    * Starts a full major compaction of the tablets in the range (start, end]. If the config does not specify a compaction strategy, then all files in a tablet
    * are compacted. The compaction is performed even for tablets that have only one file.
-   * 
+   *
    * <p>
    * Only one compact call at a time can pass iterators and/or a compaction strategy. If two threads call compaction with iterators and/or a copmaction
    * strategy, then one will fail.
-   * 
+   *
    * @param tableName
    *          the table to compact
    * @param config
    *          the configuration to use
-   * 
+   *
    * @since 1.7.0
    */
   void compact(String tableName, CompactionConfig config) throws AccumuloSecurityException, TableNotFoundException, AccumuloException;
@@ -331,7 +331,7 @@ public interface TableOperations {
    * Cancels a user initiated major compaction of a table initiated with {@link #compact(String, Text, Text, boolean, boolean)} or
    * {@link #compact(String, Text, Text, List, boolean, boolean)}. Compactions of tablets that are currently running may finish, but new compactions of tablets
    * will not start.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @throws AccumuloException
@@ -346,7 +346,7 @@ public interface TableOperations {
 
   /**
    * Delete a table
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @throws AccumuloException
@@ -362,9 +362,9 @@ public interface TableOperations {
    * Clone a table from an existing table. The cloned table will have the same data as the source table it was created from. After cloning, the two tables can
    * mutate independently. Initially the cloned table should not use any extra space, however as the source table and cloned table major compact extra space
    * will be used by the clone.
-   * 
+   *
    * Initially the cloned table is only readable and writable by the user who created it.
-   * 
+   *
    * @param srcTableName
    *          the table to clone
    * @param newTableName
@@ -382,7 +382,7 @@ public interface TableOperations {
 
   /**
    * Rename a table
-   * 
+   *
    * @param oldTableName
    *          the old table name
    * @param newTableName
@@ -396,19 +396,18 @@ public interface TableOperations {
    * @throws TableExistsException
    *           if the new table name already exists
    */
-  void rename(String oldTableName, String newTableName) throws AccumuloSecurityException, TableNotFoundException, AccumuloException,
-      TableExistsException;
+  void rename(String oldTableName, String newTableName) throws AccumuloSecurityException, TableNotFoundException, AccumuloException, TableExistsException;
 
   /**
    * Initiate a flush of a table's data that is in memory
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @throws AccumuloException
    *           if a general error occurs
    * @throws AccumuloSecurityException
    *           if the user does not have permission
-   * 
+   *
    * @deprecated As of release 1.4, replaced by {@link #flush(String, Text, Text, boolean)}
    */
   @Deprecated
@@ -416,7 +415,7 @@ public interface TableOperations {
 
   /**
    * Flush a table's data that is currently in memory.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param wait
@@ -431,7 +430,7 @@ public interface TableOperations {
 
   /**
    * Sets a property on a table. Note that it may take a short period of time (a second) to propagate the change everywhere.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param property
@@ -447,7 +446,7 @@ public interface TableOperations {
 
   /**
    * Removes a property from a table. Note that it may take a short period of time (a second) to propagate the change everywhere.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param property
@@ -461,7 +460,7 @@ public interface TableOperations {
 
   /**
    * Gets properties of a table. Note that recently changed properties may not be available immediately.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @return all properties visible by this table (system and per-table properties). Note that recently changed properties may not be visible immediately.
@@ -472,7 +471,7 @@ public interface TableOperations {
 
   /**
    * Sets a table's locality groups. A table's locality groups can be changed at any time.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param groups
@@ -487,9 +486,9 @@ public interface TableOperations {
   void setLocalityGroups(String tableName, Map<String,Set<Text>> groups) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
 
   /**
-   * 
+   *
    * Gets the locality groups currently set for a table.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @return mapping of locality group names to column families in the locality group
@@ -515,12 +514,11 @@ public interface TableOperations {
    * @throws TableNotFoundException
    *           if the table does not exist
    */
-  Set<Range> splitRangeByTablets(String tableName, Range range, int maxSplits) throws AccumuloException, AccumuloSecurityException,
-      TableNotFoundException;
+  Set<Range> splitRangeByTablets(String tableName, Range range, int maxSplits) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
 
   /**
    * Bulk import all the files in a directory into a table.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param dir
@@ -537,14 +535,14 @@ public interface TableOperations {
    *           when the user does not have the proper permissions
    * @throws TableNotFoundException
    *           when the table no longer exists
-   * 
+   *
    */
   void importDirectory(String tableName, String dir, String failureDir, boolean setTime) throws TableNotFoundException, IOException, AccumuloException,
       AccumuloSecurityException;
 
   /**
    * Initiates taking a table offline, but does not wait for action to complete
-   * 
+   *
    * @param tableName
    *          the table to take offline
    * @throws AccumuloException
@@ -555,7 +553,7 @@ public interface TableOperations {
   void offline(String tableName) throws AccumuloSecurityException, AccumuloException, TableNotFoundException;
 
   /**
-   * 
+   *
    * @param tableName
    *          the table to take offline
    * @param wait
@@ -570,7 +568,7 @@ public interface TableOperations {
 
   /**
    * Initiates bringing a table online, but does not wait for action to complete
-   * 
+   *
    * @param tableName
    *          the table to take online
    * @throws AccumuloException
@@ -581,7 +579,7 @@ public interface TableOperations {
   void online(String tableName) throws AccumuloSecurityException, AccumuloException, TableNotFoundException;
 
   /**
-   * 
+   *
    * @param tableName
    *          the table to take online
    * @param wait
@@ -596,7 +594,7 @@ public interface TableOperations {
 
   /**
    * Clears the tablet locator cache for a specified table
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @throws TableNotFoundException
@@ -606,14 +604,14 @@ public interface TableOperations {
 
   /**
    * Get a mapping of table name to internal table id.
-   * 
+   *
    * @return the map from table name to internal table id
    */
   Map<String,String> tableIdMap();
 
   /**
    * Add an iterator to a table on all scopes.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param setting
@@ -629,7 +627,7 @@ public interface TableOperations {
 
   /**
    * Add an iterator to a table on the given scopes.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param setting
@@ -646,7 +644,7 @@ public interface TableOperations {
 
   /**
    * Remove an iterator from a table by name.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param name
@@ -658,12 +656,11 @@ public interface TableOperations {
    * @throws TableNotFoundException
    *           throw if the table no longer exists
    */
-  void removeIterator(String tableName, String name, EnumSet<IteratorScope> scopes) throws AccumuloSecurityException, AccumuloException,
-      TableNotFoundException;
+  void removeIterator(String tableName, String name, EnumSet<IteratorScope> scopes) throws AccumuloSecurityException, AccumuloException, TableNotFoundException;
 
   /**
    * Get the settings for an iterator.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param name
@@ -681,7 +678,7 @@ public interface TableOperations {
 
   /**
    * Get a list of iterators for this table.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @return a set of iterator names
@@ -691,7 +688,7 @@ public interface TableOperations {
   /**
    * Check whether a given iterator configuration conflicts with existing configuration; in particular, determine if the name or priority are already in use for
    * the specified scopes.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param setting
@@ -701,7 +698,7 @@ public interface TableOperations {
 
   /**
    * Add a new constraint to a table.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param constraintClassName
@@ -717,7 +714,7 @@ public interface TableOperations {
 
   /**
    * Remove a constraint from a table.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @param number
@@ -730,7 +727,7 @@ public interface TableOperations {
 
   /**
    * List constraints on a table with their assigned numbers.
-   * 
+   *
    * @param tableName
    *          the name of the table
    * @return a map from constraint class name to assigned number
@@ -742,7 +739,7 @@ public interface TableOperations {
 
   /**
    * Gets the number of bytes being used in the files for a set of tables
-   * 
+   *
    * @param tables
    *          a set of tables
    * @return a list of disk usage objects containing linked table names and sizes
@@ -752,9 +749,9 @@ public interface TableOperations {
 
   /**
    * Test to see if the instance can load the given class as the given type. This check uses the table classpath if it is set.
-   * 
+   *
    * @return true if the instance can load the given class as the given type, false otherwise
-   * 
+   *
    * @since 1.5.0
    */
   boolean testClassLoad(String tableName, final String className, final String asTypeName) throws AccumuloException, AccumuloSecurityException,

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/admin/TimeType.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/TimeType.java b/core/src/main/java/org/apache/accumulo/core/client/admin/TimeType.java
index 3f5de94..426d384 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/TimeType.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/TimeType.java
@@ -24,7 +24,7 @@ public enum TimeType {
    * Used to guarantee ordering of data sequentially as inserted
    */
   LOGICAL,
-  
+
   /**
    * This is the default. Tries to ensure that inserted data is stored with the timestamp closest to the machine's time to the nearest millisecond, without
    * going backwards to guarantee insertion sequence. Note that using this time type can cause time to "skip" forward if a machine has a time that is too far

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/AccumuloServerException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/AccumuloServerException.java b/core/src/main/java/org/apache/accumulo/core/client/impl/AccumuloServerException.java
index a440aaf..c689c91 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/AccumuloServerException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/AccumuloServerException.java
@@ -21,24 +21,24 @@ import org.apache.thrift.TApplicationException;
 
 /**
  * This class is intended to encapsulate errors that occurred on the server side.
- * 
+ *
  */
 
 public class AccumuloServerException extends AccumuloException {
   private static final long serialVersionUID = 1L;
   private String server;
-  
+
   public AccumuloServerException(final String server, final TApplicationException tae) {
     super("Error on server " + server, tae);
     this.setServer(server);
   }
-  
+
   private void setServer(final String server) {
     this.server = server;
   }
-  
+
   public String getServer() {
     return server;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveCompactionImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveCompactionImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveCompactionImpl.java
index a413195..46259d1 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveCompactionImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveCompactionImpl.java
@@ -27,7 +27,6 @@ import org.apache.accumulo.core.client.admin.ActiveCompaction;
 import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.accumulo.core.data.thrift.IterInfo;
 
-
 /**
  *
  * @since 1.6.0

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveScanImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveScanImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveScanImpl.java
index 0f0e64c..5f953e2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveScanImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ActiveScanImpl.java
@@ -23,8 +23,8 @@ import java.util.Map;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.admin.ActiveScan;
-import org.apache.accumulo.core.client.admin.ScanType;
 import org.apache.accumulo.core.client.admin.ScanState;
+import org.apache.accumulo.core.client.admin.ScanType;
 import org.apache.accumulo.core.data.Column;
 import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.accumulo.core.data.thrift.IterInfo;
@@ -33,6 +33,7 @@ import org.apache.accumulo.core.security.Authorizations;
 
 /**
  * A class that contains information about an ActiveScan
+ *
  * @since 1.6.0
  */
 public class ActiveScanImpl extends ActiveScan {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/BatchWriterImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/BatchWriterImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/BatchWriterImpl.java
index 08f1475..c173333 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/BatchWriterImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/BatchWriterImpl.java
@@ -24,10 +24,10 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.data.Mutation;
 
 public class BatchWriterImpl implements BatchWriter {
-  
+
   private final String table;
   private final TabletServerBatchWriter bw;
-  
+
   public BatchWriterImpl(ClientContext context, String table, BatchWriterConfig config) {
     checkArgument(context != null, "context is null");
     checkArgument(table != null, "table is null");
@@ -36,27 +36,27 @@ public class BatchWriterImpl implements BatchWriter {
     this.table = table;
     this.bw = new TabletServerBatchWriter(context, config);
   }
-  
+
   @Override
   public void addMutation(Mutation m) throws MutationsRejectedException {
     checkArgument(m != null, "m is null");
     bw.addMutation(table, m);
   }
-  
+
   @Override
   public void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException {
     checkArgument(iterable != null, "iterable is null");
     bw.addMutation(table, iterable.iterator());
   }
-  
+
   @Override
   public void close() throws MutationsRejectedException {
     bw.close();
   }
-  
+
   @Override
   public void flush() throws MutationsRejectedException {
     bw.flush();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/CompressedIterators.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/CompressedIterators.java b/core/src/main/java/org/apache/accumulo/core/client/impl/CompressedIterators.java
index 549322e..3fcce90 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/CompressedIterators.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/CompressedIterators.java
@@ -33,7 +33,7 @@ public class CompressedIterators {
   private Map<String,Integer> symbolMap;
   private List<String> symbolTable;
   private Map<ByteSequence,IterConfig> cache;
-  
+
   public static class IterConfig {
     public List<IterInfo> ssiList = new ArrayList<IterInfo>();
     public Map<String,Map<String,String>> ssio = new HashMap<String,Map<String,String>>();
@@ -43,7 +43,7 @@ public class CompressedIterators {
     symbolMap = new HashMap<String,Integer>();
     symbolTable = new ArrayList<String>();
   }
-  
+
   public CompressedIterators(List<String> symbols) {
     this.symbolTable = symbols;
     this.cache = new HashMap<ByteSequence,IterConfig>();
@@ -56,36 +56,36 @@ public class CompressedIterators {
       symbolTable.add(symbol);
       symbolMap.put(symbol, id);
     }
-    
+
     return id;
   }
-  
+
   public ByteBuffer compress(IteratorSetting[] iterators) {
-    
+
     UnsynchronizedBuffer.Writer out = new UnsynchronizedBuffer.Writer(iterators.length * 8);
-    
+
     out.writeVInt(iterators.length);
 
     for (IteratorSetting is : iterators) {
       out.writeVInt(getSymbolID(is.getName()));
       out.writeVInt(getSymbolID(is.getIteratorClass()));
       out.writeVInt(is.getPriority());
-      
+
       Map<String,String> opts = is.getOptions();
       out.writeVInt(opts.size());
-      
+
       for (Entry<String,String> entry : opts.entrySet()) {
         out.writeVInt(getSymbolID(entry.getKey()));
         out.writeVInt(getSymbolID(entry.getValue()));
       }
     }
-    
+
     return out.toByteBuffer();
-    
+
   }
-  
+
   public IterConfig decompress(ByteBuffer iterators) {
-    
+
     ByteSequence iterKey = new ArrayByteSequence(iterators);
     IterConfig config = cache.get(iterKey);
     if (config != null) {
@@ -97,27 +97,27 @@ public class CompressedIterators {
     UnsynchronizedBuffer.Reader in = new UnsynchronizedBuffer.Reader(iterators);
 
     int num = in.readVInt();
-    
+
     for (int i = 0; i < num; i++) {
       String name = symbolTable.get(in.readVInt());
       String iterClass = symbolTable.get(in.readVInt());
       int prio = in.readVInt();
-      
+
       config.ssiList.add(new IterInfo(prio, iterClass, name));
-      
+
       int numOpts = in.readVInt();
-      
+
       HashMap<String,String> opts = new HashMap<String,String>();
-      
+
       for (int j = 0; j < numOpts; j++) {
         String key = symbolTable.get(in.readVInt());
         String val = symbolTable.get(in.readVInt());
-        
+
         opts.put(key, val);
       }
-      
+
       config.ssio.put(name, opts);
-      
+
     }
 
     cache.put(iterKey, config);


[33/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/StatusUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/StatusUtil.java b/core/src/main/java/org/apache/accumulo/core/replication/StatusUtil.java
index 482fc62..a7cd3f5 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/StatusUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/StatusUtil.java
@@ -59,7 +59,7 @@ public class StatusUtil {
 
   /**
    * Creates a {@link Status} for newly-created data that must be replicated
-   * 
+   *
    * @param recordsIngested
    *          Offset of records which need to be replicated
    * @return A {@link Status} tracking data that must be replicated
@@ -94,7 +94,7 @@ public class StatusUtil {
 
   /**
    * Creates a @{link Status} for a file which has new data and data which has been replicated
-   * 
+   *
    * @param recordsReplicated
    *          Offset of records which have been replicated
    * @param recordsIngested
@@ -107,7 +107,7 @@ public class StatusUtil {
 
   /**
    * Same as {@link #replicatedAndIngested(long, long)} but uses the provided {@link Builder}
-   * 
+   *
    * @param builder
    *          An existing builder
    * @param recordsReplicated
@@ -165,7 +165,8 @@ public class StatusUtil {
   }
 
   /**
-   * @param v Value with serialized Status
+   * @param v
+   *          Value with serialized Status
    * @return A Status created from the Value
    */
   public static Status fromValue(Value v) throws InvalidProtocolBufferException {
@@ -174,7 +175,9 @@ public class StatusUtil {
 
   /**
    * Is the given Status fully replicated and is its file ready for deletion on the source
-   * @param status a Status protobuf
+   *
+   * @param status
+   *          a Status protobuf
    * @return True if the file this Status references can be deleted.
    */
   public static boolean isSafeForRemoval(Status status) {
@@ -183,7 +186,9 @@ public class StatusUtil {
 
   /**
    * Is the given Status fully replicated but potentially not yet safe for deletion
-   * @param status a Status protobuf
+   *
+   * @param status
+   *          a Status protobuf
    * @return True if the file this Status references is fully replicated so far
    */
   public static boolean isFullyReplicated(Status status) {
@@ -196,7 +201,9 @@ public class StatusUtil {
 
   /**
    * Given the {@link Status}, is there replication work to be done
-   * @param status Status for a file
+   *
+   * @param status
+   *          Status for a file
    * @return true if replication work is required
    */
   public static boolean isWorkRequired(Status status) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/proto/Replication.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/proto/Replication.java b/core/src/main/java/org/apache/accumulo/core/replication/proto/Replication.java
index 2bff020..07168a6 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/proto/Replication.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/proto/Replication.java
@@ -19,13 +19,13 @@
 
 package org.apache.accumulo.core.replication.proto;
 
-@SuppressWarnings("all") public final class Replication {
+@SuppressWarnings("all")
+public final class Replication {
   private Replication() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface StatusOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+
+  public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry registry) {}
+
+  public interface StatusOrBuilder extends com.google.protobuf.MessageOrBuilder {
 
     // optional int64 begin = 1 [default = 0];
     /**
@@ -36,6 +36,7 @@ package org.apache.accumulo.core.replication.proto;
      * </pre>
      */
     boolean hasBegin();
+
     /**
      * <code>optional int64 begin = 1 [default = 0];</code>
      *
@@ -54,6 +55,7 @@ package org.apache.accumulo.core.replication.proto;
      * </pre>
      */
     boolean hasEnd();
+
     /**
      * <code>optional int64 end = 2 [default = 0];</code>
      *
@@ -72,6 +74,7 @@ package org.apache.accumulo.core.replication.proto;
      * </pre>
      */
     boolean hasInfiniteEnd();
+
     /**
      * <code>optional bool infiniteEnd = 3 [default = false];</code>
      *
@@ -90,6 +93,7 @@ package org.apache.accumulo.core.replication.proto;
      * </pre>
      */
     boolean hasClosed();
+
     /**
      * <code>optional bool closed = 4 [default = false];</code>
      *
@@ -108,6 +112,7 @@ package org.apache.accumulo.core.replication.proto;
      * </pre>
      */
     boolean hasCreatedTime();
+
     /**
      * <code>optional int64 createdTime = 5 [default = 0];</code>
      *
@@ -117,20 +122,23 @@ package org.apache.accumulo.core.replication.proto;
      */
     long getCreatedTime();
   }
+
   /**
    * Protobuf type {@code Status}
    */
-  public static final class Status extends
-      com.google.protobuf.GeneratedMessage
-      implements StatusOrBuilder {
+  public static final class Status extends com.google.protobuf.GeneratedMessage implements StatusOrBuilder {
     // Use Status.newBuilder() to construct.
     private Status(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
       super(builder);
       this.unknownFields = builder.getUnknownFields();
     }
-    private Status(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+    private Status(boolean noInit) {
+      this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance();
+    }
 
     private static final Status defaultInstance;
+
     public static Status getDefaultInstance() {
       return defaultInstance;
     }
@@ -140,19 +148,17 @@ package org.apache.accumulo.core.replication.proto;
     }
 
     private final com.google.protobuf.UnknownFieldSet unknownFields;
+
     @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
+    public final com.google.protobuf.UnknownFieldSet getUnknownFields() {
       return this.unknownFields;
     }
-    private Status(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+
+    private Status(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
       initFields();
       int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder();
       try {
         boolean done = false;
         while (!done) {
@@ -162,8 +168,7 @@ package org.apache.accumulo.core.replication.proto;
               done = true;
               break;
             default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
+              if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) {
                 done = true;
               }
               break;
@@ -198,30 +203,24 @@ package org.apache.accumulo.core.replication.proto;
       } catch (com.google.protobuf.InvalidProtocolBufferException e) {
         throw e.setUnfinishedMessage(this);
       } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
+        throw new com.google.protobuf.InvalidProtocolBufferException(e.getMessage()).setUnfinishedMessage(this);
       } finally {
         this.unknownFields = unknownFields.build();
         makeExtensionsImmutable();
       }
     }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
+
+    public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() {
       return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.accumulo.core.replication.proto.Replication.Status.class, org.apache.accumulo.core.replication.proto.Replication.Status.Builder.class);
+    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() {
+      return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_fieldAccessorTable.ensureFieldAccessorsInitialized(
+          org.apache.accumulo.core.replication.proto.Replication.Status.class, org.apache.accumulo.core.replication.proto.Replication.Status.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<Status> PARSER =
-        new com.google.protobuf.AbstractParser<Status>() {
-      public Status parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+    public static com.google.protobuf.Parser<Status> PARSER = new com.google.protobuf.AbstractParser<Status>() {
+      public Status parsePartialFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws com.google.protobuf.InvalidProtocolBufferException {
         return new Status(input, extensionRegistry);
       }
@@ -236,6 +235,7 @@ package org.apache.accumulo.core.replication.proto;
     // optional int64 begin = 1 [default = 0];
     public static final int BEGIN_FIELD_NUMBER = 1;
     private long begin_;
+
     /**
      * <code>optional int64 begin = 1 [default = 0];</code>
      *
@@ -246,6 +246,7 @@ package org.apache.accumulo.core.replication.proto;
     public boolean hasBegin() {
       return ((bitField0_ & 0x00000001) == 0x00000001);
     }
+
     /**
      * <code>optional int64 begin = 1 [default = 0];</code>
      *
@@ -260,6 +261,7 @@ package org.apache.accumulo.core.replication.proto;
     // optional int64 end = 2 [default = 0];
     public static final int END_FIELD_NUMBER = 2;
     private long end_;
+
     /**
      * <code>optional int64 end = 2 [default = 0];</code>
      *
@@ -270,6 +272,7 @@ package org.apache.accumulo.core.replication.proto;
     public boolean hasEnd() {
       return ((bitField0_ & 0x00000002) == 0x00000002);
     }
+
     /**
      * <code>optional int64 end = 2 [default = 0];</code>
      *
@@ -284,6 +287,7 @@ package org.apache.accumulo.core.replication.proto;
     // optional bool infiniteEnd = 3 [default = false];
     public static final int INFINITEEND_FIELD_NUMBER = 3;
     private boolean infiniteEnd_;
+
     /**
      * <code>optional bool infiniteEnd = 3 [default = false];</code>
      *
@@ -294,6 +298,7 @@ package org.apache.accumulo.core.replication.proto;
     public boolean hasInfiniteEnd() {
       return ((bitField0_ & 0x00000004) == 0x00000004);
     }
+
     /**
      * <code>optional bool infiniteEnd = 3 [default = false];</code>
      *
@@ -308,6 +313,7 @@ package org.apache.accumulo.core.replication.proto;
     // optional bool closed = 4 [default = false];
     public static final int CLOSED_FIELD_NUMBER = 4;
     private boolean closed_;
+
     /**
      * <code>optional bool closed = 4 [default = false];</code>
      *
@@ -318,6 +324,7 @@ package org.apache.accumulo.core.replication.proto;
     public boolean hasClosed() {
       return ((bitField0_ & 0x00000008) == 0x00000008);
     }
+
     /**
      * <code>optional bool closed = 4 [default = false];</code>
      *
@@ -332,6 +339,7 @@ package org.apache.accumulo.core.replication.proto;
     // optional int64 createdTime = 5 [default = 0];
     public static final int CREATEDTIME_FIELD_NUMBER = 5;
     private long createdTime_;
+
     /**
      * <code>optional int64 createdTime = 5 [default = 0];</code>
      *
@@ -342,6 +350,7 @@ package org.apache.accumulo.core.replication.proto;
     public boolean hasCreatedTime() {
       return ((bitField0_ & 0x00000010) == 0x00000010);
     }
+
     /**
      * <code>optional int64 createdTime = 5 [default = 0];</code>
      *
@@ -360,17 +369,19 @@ package org.apache.accumulo.core.replication.proto;
       closed_ = false;
       createdTime_ = 0L;
     }
+
     private byte memoizedIsInitialized = -1;
+
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized != -1)
+        return isInitialized == 1;
 
       memoizedIsInitialized = 1;
       return true;
     }
 
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
+    public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException {
       getSerializedSize();
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         output.writeInt64(1, begin_);
@@ -391,30 +402,27 @@ package org.apache.accumulo.core.replication.proto;
     }
 
     private int memoizedSerializedSize = -1;
+
     public int getSerializedSize() {
       int size = memoizedSerializedSize;
-      if (size != -1) return size;
+      if (size != -1)
+        return size;
 
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(1, begin_);
+        size += com.google.protobuf.CodedOutputStream.computeInt64Size(1, begin_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(2, end_);
+        size += com.google.protobuf.CodedOutputStream.computeInt64Size(2, end_);
       }
       if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBoolSize(3, infiniteEnd_);
+        size += com.google.protobuf.CodedOutputStream.computeBoolSize(3, infiniteEnd_);
       }
       if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBoolSize(4, closed_);
+        size += com.google.protobuf.CodedOutputStream.computeBoolSize(4, closed_);
       }
       if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(5, createdTime_);
+        size += com.google.protobuf.CodedOutputStream.computeInt64Size(5, createdTime_);
       }
       size += getUnknownFields().getSerializedSize();
       memoizedSerializedSize = size;
@@ -422,94 +430,94 @@ package org.apache.accumulo.core.replication.proto;
     }
 
     private static final long serialVersionUID = 0L;
+
     @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
+    protected java.lang.Object writeReplace() throws java.io.ObjectStreamException {
       return super.writeReplace();
     }
 
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(
-        com.google.protobuf.ByteString data)
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(com.google.protobuf.ByteString data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data, extensionRegistry);
     }
+
     public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(byte[] data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data, extensionRegistry);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(java.io.InputStream input) throws java.io.IOException {
       return PARSER.parseFrom(input);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException {
       return PARSER.parseFrom(input, extensionRegistry);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException {
       return PARSER.parseDelimitedFrom(input);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseDelimitedFrom(java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException {
       return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(
-        com.google.protobuf.CodedInputStream input)
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
       return PARSER.parseFrom(input);
     }
-    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
+
+    public static org.apache.accumulo.core.replication.proto.Replication.Status parseFrom(com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException {
       return PARSER.parseFrom(input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return Builder.create();
+    }
+
+    public Builder newBuilderForType() {
+      return newBuilder();
+    }
+
     public static Builder newBuilder(org.apache.accumulo.core.replication.proto.Replication.Status prototype) {
       return newBuilder().mergeFrom(prototype);
     }
-    public Builder toBuilder() { return newBuilder(this); }
+
+    public Builder toBuilder() {
+      return newBuilder(this);
+    }
 
     @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+    protected Builder newBuilderForType(com.google.protobuf.GeneratedMessage.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
+
     /**
      * Protobuf type {@code Status}
      */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.accumulo.core.replication.proto.Replication.StatusOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
+    public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements
+        org.apache.accumulo.core.replication.proto.Replication.StatusOrBuilder {
+      public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() {
         return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.accumulo.core.replication.proto.Replication.Status.class, org.apache.accumulo.core.replication.proto.Replication.Status.Builder.class);
+      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() {
+        return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_fieldAccessorTable.ensureFieldAccessorsInitialized(
+            org.apache.accumulo.core.replication.proto.Replication.Status.class, org.apache.accumulo.core.replication.proto.Replication.Status.Builder.class);
       }
 
       // Construct using org.apache.accumulo.core.replication.proto.Replication.Status.newBuilder()
@@ -517,15 +525,15 @@ package org.apache.accumulo.core.replication.proto;
         maybeForceBuilderInitialization();
       }
 
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+      private Builder(com.google.protobuf.GeneratedMessage.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
+
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
+        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {}
       }
+
       private static Builder create() {
         return new Builder();
       }
@@ -549,8 +557,7 @@ package org.apache.accumulo.core.replication.proto;
         return create().mergeFrom(buildPartial());
       }
 
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
+      public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() {
         return org.apache.accumulo.core.replication.proto.Replication.internal_static_Status_descriptor;
       }
 
@@ -597,7 +604,7 @@ package org.apache.accumulo.core.replication.proto;
 
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.accumulo.core.replication.proto.Replication.Status) {
-          return mergeFrom((org.apache.accumulo.core.replication.proto.Replication.Status)other);
+          return mergeFrom((org.apache.accumulo.core.replication.proto.Replication.Status) other);
         } else {
           super.mergeFrom(other);
           return this;
@@ -605,7 +612,8 @@ package org.apache.accumulo.core.replication.proto;
       }
 
       public Builder mergeFrom(org.apache.accumulo.core.replication.proto.Replication.Status other) {
-        if (other == org.apache.accumulo.core.replication.proto.Replication.Status.getDefaultInstance()) return this;
+        if (other == org.apache.accumulo.core.replication.proto.Replication.Status.getDefaultInstance())
+          return this;
         if (other.hasBegin()) {
           setBegin(other.getBegin());
         }
@@ -629,9 +637,7 @@ package org.apache.accumulo.core.replication.proto;
         return true;
       }
 
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+      public Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
         org.apache.accumulo.core.replication.proto.Replication.Status parsedMessage = null;
         try {
@@ -646,10 +652,12 @@ package org.apache.accumulo.core.replication.proto;
         }
         return this;
       }
+
       private int bitField0_;
 
       // optional int64 begin = 1 [default = 0];
-      private long begin_ ;
+      private long begin_;
+
       /**
        * <code>optional int64 begin = 1 [default = 0];</code>
        *
@@ -660,6 +668,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean hasBegin() {
         return ((bitField0_ & 0x00000001) == 0x00000001);
       }
+
       /**
        * <code>optional int64 begin = 1 [default = 0];</code>
        *
@@ -670,6 +679,7 @@ package org.apache.accumulo.core.replication.proto;
       public long getBegin() {
         return begin_;
       }
+
       /**
        * <code>optional int64 begin = 1 [default = 0];</code>
        *
@@ -683,6 +693,7 @@ package org.apache.accumulo.core.replication.proto;
         onChanged();
         return this;
       }
+
       /**
        * <code>optional int64 begin = 1 [default = 0];</code>
        *
@@ -698,7 +709,8 @@ package org.apache.accumulo.core.replication.proto;
       }
 
       // optional int64 end = 2 [default = 0];
-      private long end_ ;
+      private long end_;
+
       /**
        * <code>optional int64 end = 2 [default = 0];</code>
        *
@@ -709,6 +721,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean hasEnd() {
         return ((bitField0_ & 0x00000002) == 0x00000002);
       }
+
       /**
        * <code>optional int64 end = 2 [default = 0];</code>
        *
@@ -719,6 +732,7 @@ package org.apache.accumulo.core.replication.proto;
       public long getEnd() {
         return end_;
       }
+
       /**
        * <code>optional int64 end = 2 [default = 0];</code>
        *
@@ -732,6 +746,7 @@ package org.apache.accumulo.core.replication.proto;
         onChanged();
         return this;
       }
+
       /**
        * <code>optional int64 end = 2 [default = 0];</code>
        *
@@ -747,7 +762,8 @@ package org.apache.accumulo.core.replication.proto;
       }
 
       // optional bool infiniteEnd = 3 [default = false];
-      private boolean infiniteEnd_ ;
+      private boolean infiniteEnd_;
+
       /**
        * <code>optional bool infiniteEnd = 3 [default = false];</code>
        *
@@ -758,6 +774,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean hasInfiniteEnd() {
         return ((bitField0_ & 0x00000004) == 0x00000004);
       }
+
       /**
        * <code>optional bool infiniteEnd = 3 [default = false];</code>
        *
@@ -768,6 +785,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean getInfiniteEnd() {
         return infiniteEnd_;
       }
+
       /**
        * <code>optional bool infiniteEnd = 3 [default = false];</code>
        *
@@ -781,6 +799,7 @@ package org.apache.accumulo.core.replication.proto;
         onChanged();
         return this;
       }
+
       /**
        * <code>optional bool infiniteEnd = 3 [default = false];</code>
        *
@@ -796,7 +815,8 @@ package org.apache.accumulo.core.replication.proto;
       }
 
       // optional bool closed = 4 [default = false];
-      private boolean closed_ ;
+      private boolean closed_;
+
       /**
        * <code>optional bool closed = 4 [default = false];</code>
        *
@@ -807,6 +827,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean hasClosed() {
         return ((bitField0_ & 0x00000008) == 0x00000008);
       }
+
       /**
        * <code>optional bool closed = 4 [default = false];</code>
        *
@@ -817,6 +838,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean getClosed() {
         return closed_;
       }
+
       /**
        * <code>optional bool closed = 4 [default = false];</code>
        *
@@ -830,6 +852,7 @@ package org.apache.accumulo.core.replication.proto;
         onChanged();
         return this;
       }
+
       /**
        * <code>optional bool closed = 4 [default = false];</code>
        *
@@ -845,7 +868,8 @@ package org.apache.accumulo.core.replication.proto;
       }
 
       // optional int64 createdTime = 5 [default = 0];
-      private long createdTime_ ;
+      private long createdTime_;
+
       /**
        * <code>optional int64 createdTime = 5 [default = 0];</code>
        *
@@ -856,6 +880,7 @@ package org.apache.accumulo.core.replication.proto;
       public boolean hasCreatedTime() {
         return ((bitField0_ & 0x00000010) == 0x00000010);
       }
+
       /**
        * <code>optional int64 createdTime = 5 [default = 0];</code>
        *
@@ -866,6 +891,7 @@ package org.apache.accumulo.core.replication.proto;
       public long getCreatedTime() {
         return createdTime_;
       }
+
       /**
        * <code>optional int64 createdTime = 5 [default = 0];</code>
        *
@@ -879,6 +905,7 @@ package org.apache.accumulo.core.replication.proto;
         onChanged();
         return this;
       }
+
       /**
        * <code>optional int64 createdTime = 5 [default = 0];</code>
        *
@@ -904,45 +931,30 @@ package org.apache.accumulo.core.replication.proto;
     // @@protoc_insertion_point(class_scope:Status)
   }
 
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_Status_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_Status_fieldAccessorTable;
+  private static com.google.protobuf.Descriptors.Descriptor internal_static_Status_descriptor;
+  private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_Status_fieldAccessorTable;
 
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
+  public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() {
     return descriptor;
   }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
+
+  private static com.google.protobuf.Descriptors.FileDescriptor descriptor;
   static {
-    java.lang.String[] descriptorData = {
-      "\n#src/main/protobuf/replication.proto\"u\n" +
-      "\006Status\022\020\n\005begin\030\001 \001(\003:\0010\022\016\n\003end\030\002 \001(\003:\001" +
-      "0\022\032\n\013infiniteEnd\030\003 \001(\010:\005false\022\025\n\006closed\030" +
-      "\004 \001(\010:\005false\022\026\n\013createdTime\030\005 \001(\003:\0010B.\n*" +
-      "org.apache.accumulo.core.replication.pro" +
-      "toH\001"
+    java.lang.String[] descriptorData = {"\n#src/main/protobuf/replication.proto\"u\n"
+        + "\006Status\022\020\n\005begin\030\001 \001(\003:\0010\022\016\n\003end\030\002 \001(\003:\001"
+        + "0\022\032\n\013infiniteEnd\030\003 \001(\010:\005false\022\025\n\006closed\030"
+        + "\004 \001(\010:\005false\022\026\n\013createdTime\030\005 \001(\003:\0010B.\n*" + "org.apache.accumulo.core.replication.pro" + "toH\001"};
+    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
+      public com.google.protobuf.ExtensionRegistry assignDescriptors(com.google.protobuf.Descriptors.FileDescriptor root) {
+        descriptor = root;
+        internal_static_Status_descriptor = getDescriptor().getMessageTypes().get(0);
+        internal_static_Status_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable(internal_static_Status_descriptor,
+            new java.lang.String[] {"Begin", "End", "InfiniteEnd", "Closed", "CreatedTime",});
+        return null;
+      }
     };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_Status_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_Status_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_Status_descriptor,
-              new java.lang.String[] { "Begin", "End", "InfiniteEnd", "Closed", "CreatedTime", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
+    com.google.protobuf.Descriptors.FileDescriptor.internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] {},
+        assigner);
   }
 
   // @@protoc_insertion_point(outer_class_scope)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/rpc/TBufferedSocket.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/rpc/TBufferedSocket.java b/core/src/main/java/org/apache/accumulo/core/rpc/TBufferedSocket.java
index 87b7c13..942c554 100644
--- a/core/src/main/java/org/apache/accumulo/core/rpc/TBufferedSocket.java
+++ b/core/src/main/java/org/apache/accumulo/core/rpc/TBufferedSocket.java
@@ -24,16 +24,16 @@ import org.apache.thrift.transport.TIOStreamTransport;
 import org.apache.thrift.transport.TSocket;
 
 public class TBufferedSocket extends TIOStreamTransport {
-  
+
   String client;
-  
+
   public TBufferedSocket(TSocket sock, int bufferSize) throws IOException {
     super(new BufferedInputStream(sock.getSocket().getInputStream(), bufferSize), new BufferedOutputStream(sock.getSocket().getOutputStream(), bufferSize));
     client = sock.getSocket().getInetAddress().getHostAddress() + ":" + sock.getSocket().getPort();
   }
-  
+
   public String getClientString() {
     return client;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/rpc/TTimeoutTransport.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/rpc/TTimeoutTransport.java b/core/src/main/java/org/apache/accumulo/core/rpc/TTimeoutTransport.java
index 6eace77..8c23555 100644
--- a/core/src/main/java/org/apache/accumulo/core/rpc/TTimeoutTransport.java
+++ b/core/src/main/java/org/apache/accumulo/core/rpc/TTimeoutTransport.java
@@ -34,16 +34,16 @@ import org.apache.thrift.transport.TTransport;
 import com.google.common.net.HostAndPort;
 
 public class TTimeoutTransport {
-  
+
   private static InputStream getInputStream(Socket socket, long timeout) {
     try {
       Method m = NetUtils.class.getMethod("getInputStream", Socket.class, Long.TYPE);
-      return (InputStream)m.invoke(null, socket, timeout);
+      return (InputStream) m.invoke(null, socket, timeout);
     } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   public static TTransport create(HostAndPort addr, long timeoutMillis) throws IOException {
     return create(new InetSocketAddress(addr.getHostText(), addr.getPort()), timeoutMillis);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/rpc/TraceProtocol.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/rpc/TraceProtocol.java b/core/src/main/java/org/apache/accumulo/core/rpc/TraceProtocol.java
index 74aad57..e7c6a44 100644
--- a/core/src/main/java/org/apache/accumulo/core/rpc/TraceProtocol.java
+++ b/core/src/main/java/org/apache/accumulo/core/rpc/TraceProtocol.java
@@ -44,4 +44,4 @@ public class TraceProtocol extends TCompactProtocol {
   public TraceProtocol(TTransport transport) {
     super(transport);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/AuthorizationContainer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/AuthorizationContainer.java b/core/src/main/java/org/apache/accumulo/core/security/AuthorizationContainer.java
index 029b622..01ad46d 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/AuthorizationContainer.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/AuthorizationContainer.java
@@ -25,7 +25,8 @@ public interface AuthorizationContainer {
   /**
    * Checks whether this object contains the given authorization.
    *
-   * @param auth authorization, as a string encoded in UTF-8
+   * @param auth
+   *          authorization, as a string encoded in UTF-8
    * @return true if authorization is in this collection
    */
   boolean contains(ByteSequence auth);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/Authorizations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/Authorizations.java b/core/src/main/java/org/apache/accumulo/core/security/Authorizations.java
index b6ce5b7..001fed2 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/Authorizations.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/Authorizations.java
@@ -54,7 +54,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * A special header string used when serializing instances of this class.
-   * 
+   *
    * @see #serialize()
    */
   public static final String HEADER = "!AUTH1:";
@@ -102,7 +102,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
   /**
    * Constructs an authorization object from a collection of string authorizations that have each already been encoded as UTF-8 bytes. Warning: This method does
    * not verify that each encoded string is valid UTF-8.
-   * 
+   *
    * @param authorizations
    *          collection of authorizations, as strings encoded in UTF-8
    * @throws IllegalArgumentException
@@ -119,7 +119,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
   /**
    * Constructs an authorization object from a list of string authorizations that have each already been encoded as UTF-8 bytes. Warning: This method does not
    * verify that each encoded string is valid UTF-8.
-   * 
+   *
    * @param authorizations
    *          list of authorizations, as strings encoded in UTF-8 and placed in buffers
    * @throws IllegalArgumentException
@@ -137,7 +137,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
   /**
    * Constructs an authorizations object from a serialized form. This is NOT a constructor for a set of authorizations of size one. Warning: This method does
    * not verify that the encoded serialized form is valid UTF-8.
-   * 
+   *
    * @param authorizations
    *          a serialized authorizations string produced by {@link #getAuthorizationsArray()} or {@link #serialize()}, converted to UTF-8 bytes
    * @throws IllegalArgumentException
@@ -167,14 +167,14 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Constructs an empty set of authorizations.
-   * 
+   *
    * @see #Authorizations(String...)
    */
   public Authorizations() {}
 
   /**
    * Constructs an authorizations object from a set of human-readable authorizations.
-   * 
+   *
    * @param authorizations
    *          array of authorizations
    * @throws IllegalArgumentException
@@ -197,7 +197,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Returns a serialized form of these authorizations.
-   * 
+   *
    * @return serialized form of these authorizations, as a string encoded in UTF-8
    * @see #serialize()
    */
@@ -207,7 +207,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Gets the authorizations in sorted order. The returned list is not modifiable.
-   * 
+   *
    * @return authorizations, each as a string encoded in UTF-8
    * @see #Authorizations(Collection)
    */
@@ -223,7 +223,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Gets the authorizations in sorted order. The returned list is not modifiable.
-   * 
+   *
    * @return authorizations, each as a string encoded in UTF-8 and within a buffer
    */
   public List<ByteBuffer> getAuthorizationsBB() {
@@ -251,7 +251,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Checks whether this object contains the given authorization.
-   * 
+   *
    * @param auth
    *          authorization, as a string encoded in UTF-8
    * @return true if authorization is in this collection
@@ -262,7 +262,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Checks whether this object contains the given authorization. Warning: This method does not verify that the encoded string is valid UTF-8.
-   * 
+   *
    * @param auth
    *          authorization, as a string encoded in UTF-8
    * @return true if authorization is in this collection
@@ -274,7 +274,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Checks whether this object contains the given authorization.
-   * 
+   *
    * @param auth
    *          authorization
    * @return true if authorization is in this collection
@@ -308,7 +308,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Gets the size of this collection of authorizations.
-   * 
+   *
    * @return collection size
    */
   public int size() {
@@ -317,7 +317,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Checks if this collection of authorizations is empty.
-   * 
+   *
    * @return true if this collection contains no authorizations
    */
   public boolean isEmpty() {
@@ -331,7 +331,7 @@ public class Authorizations implements Iterable<byte[]>, Serializable, Authoriza
 
   /**
    * Returns a serialized form of these authorizations. Convert the returned string to UTF-8 bytes to deserialize with {@link #Authorizations(byte[])}.
-   * 
+   *
    * @return serialized form of authorizations
    */
   public String serialize() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
index 3e930f5..e6d26de 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
@@ -38,20 +38,19 @@ import org.apache.hadoop.io.WritableComparator;
  * definition of an expression.
  *
  * <P>
- * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the
- * binary operators "&amp;" and "|" indicating that both operands are necessary, or the either
- * is necessary. The following are valid expressions for visibility:
- * 
+ * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&amp;" and "|" indicating that both operands are
+ * necessary, or the either is necessary. The following are valid expressions for visibility:
+ *
  * <pre>
  * A
  * A|B
  * (A|B)&amp;(C|D)
  * orange|(red&amp;yellow)
  * </pre>
- * 
+ *
  * <P>
  * The following are not valid expressions for visibility:
- * 
+ *
  * <pre>
  * A|B&amp;C
  * A=B
@@ -61,19 +60,21 @@ import org.apache.hadoop.io.WritableComparator;
  * )
  * dog|!cat
  * </pre>
- * 
+ *
  * <P>
- * In addition to the base set of visibilities, any character can be used in the expression if it is quoted. If the quoted term contains '&quot;' or '\', then escape
- * the character with '\'. The {@link #quote(String)} method can be used to properly quote and escape terms automatically. The following is an example of a quoted term:
+ * In addition to the base set of visibilities, any character can be used in the expression if it is quoted. If the quoted term contains '&quot;' or '\', then
+ * escape the character with '\'. The {@link #quote(String)} method can be used to properly quote and escape terms automatically. The following is an example of
+ * a quoted term:
+ *
  * <pre>
  * &quot;A#C&quot;<span />&amp;<span />B
  * </pre>
  */
 public class ColumnVisibility {
-  
+
   Node node = null;
   private byte[] expression;
-  
+
   /**
    * Accessor for the underlying byte string.
    *
@@ -82,7 +83,7 @@ public class ColumnVisibility {
   public byte[] getExpression() {
     return expression;
   }
-  
+
   /**
    * The node types in a parse tree for a visibility expression.
    */
@@ -94,7 +95,7 @@ public class ColumnVisibility {
    * All empty nodes are equal and represent the same value.
    */
   private static final Node EMPTY_NODE = new Node(NodeType.EMPTY, 0);
-  
+
   /**
    * A node in the parse tree for a visibility expression.
    */
@@ -107,76 +108,76 @@ public class ColumnVisibility {
     int start;
     int end;
     List<Node> children = EMPTY;
-    
+
     public Node(NodeType type, int start) {
       this.type = type;
       this.start = start;
       this.end = start + 1;
     }
-    
+
     public Node(int start, int end) {
       this.type = NodeType.TERM;
       this.start = start;
       this.end = end;
     }
-    
+
     public void add(Node child) {
       if (children == EMPTY)
         children = new ArrayList<Node>();
-      
+
       children.add(child);
     }
-    
+
     public NodeType getType() {
       return type;
     }
-    
+
     public List<Node> getChildren() {
       return children;
     }
-    
+
     public int getTermStart() {
       return start;
     }
-    
+
     public int getTermEnd() {
       return end;
     }
-    
+
     public ByteSequence getTerm(byte expression[]) {
       if (type != NodeType.TERM)
         throw new RuntimeException();
-      
+
       if (expression[start] == '"') {
         // its a quoted term
         int qStart = start + 1;
         int qEnd = end - 1;
-        
+
         return new ArrayByteSequence(expression, qStart, qEnd - qStart);
       }
       return new ArrayByteSequence(expression, start, end - start);
     }
   }
-  
+
   /**
-   * A node comparator. Nodes sort according to node type, terms sort
-   * lexicographically. AND and OR nodes sort by number of children, or if
-   * the same by corresponding children.
+   * A node comparator. Nodes sort according to node type, terms sort lexicographically. AND and OR nodes sort by number of children, or if the same by
+   * corresponding children.
    */
   public static class NodeComparator implements Comparator<Node>, Serializable {
-    
+
     private static final long serialVersionUID = 1L;
     byte[] text;
-    
+
     /**
      * Creates a new comparator.
      *
-     * @param text expression string, encoded in UTF-8
+     * @param text
+     *          expression string, encoded in UTF-8
      */
     public NodeComparator(byte[] text) {
       this.text = text;
     }
-    
+
     @Override
     public int compare(Node a, Node b) {
       int diff = a.type.ordinal() - b.type.ordinal();
@@ -201,14 +202,14 @@ public class ColumnVisibility {
       return 0;
     }
   }
-  
+
   /*
    * Convience method that delegates to normalize with a new NodeComparator constructed using the supplied expression.
    */
   public static Node normalize(Node root, byte[] expression) {
     return normalize(root, expression, new NodeComparator(expression));
   }
-  
+
   // @formatter:off
   /*
    * Walks an expression's AST in order to:
@@ -231,16 +232,16 @@ public class ColumnVisibility {
       rolledUp.addAll(root.children);
       root.children.clear();
       root.children.addAll(rolledUp);
-      
+
       // need to promote a child if it's an only child
       if (root.children.size() == 1) {
         return root.children.get(0);
       }
     }
-    
+
     return root;
   }
-  
+
   /*
    * Walks an expression's AST and appends a string representation to a supplied StringBuilder. This method adds parens where necessary.
    */
@@ -261,7 +262,7 @@ public class ColumnVisibility {
       }
     }
   }
-  
+
   /**
    * Generates a byte[] that represents a normalized, but logically equivalent, form of this evaluator's expression.
    *
@@ -273,13 +274,13 @@ public class ColumnVisibility {
     stringify(normRoot, expression, builder);
     return builder.toString().getBytes(UTF_8);
   }
-  
+
   private static class ColumnVisibilityParser {
     private int index = 0;
     private int parens = 0;
-    
+
     public ColumnVisibilityParser() {}
-    
+
     Node parse(byte[] expression) {
       if (expression.length > 0) {
         Node node = parse_(expression);
@@ -293,7 +294,7 @@ public class ColumnVisibility {
       }
       return null;
     }
-    
+
     Node processTerm(int start, int end, Node expr, byte[] expression) {
       if (start != end) {
         if (expr != null)
@@ -304,14 +305,14 @@ public class ColumnVisibility {
         throw new BadArgumentException("empty term", new String(expression, UTF_8), start);
       return expr;
     }
-    
+
     Node parse_(byte[] expression) {
       Node result = null;
       Node expr = null;
       int wholeTermStart = index;
       int subtermStart = index;
       boolean subtermComplete = false;
-      
+
       while (index < expression.length) {
         switch (expression[index++]) {
           case '&': {
@@ -369,7 +370,7 @@ public class ColumnVisibility {
           case '"': {
             if (subtermStart != index - 1)
               throw new BadArgumentException("expression needs & or |", new String(expression, UTF_8), index - 1);
-            
+
             while (index < expression.length && expression[index] != '"') {
               if (expression[index] == '\\') {
                 index++;
@@ -378,23 +379,23 @@ public class ColumnVisibility {
               }
               index++;
             }
-            
+
             if (index == expression.length)
               throw new BadArgumentException("unclosed quote", new String(expression, UTF_8), subtermStart);
-            
+
             if (subtermStart + 1 == index)
               throw new BadArgumentException("empty term", new String(expression, UTF_8), subtermStart);
- 
+
             index++;
-            
+
             subtermComplete = true;
-            
+
             break;
           }
           default: {
             if (subtermComplete)
               throw new BadArgumentException("expression needs & or |", new String(expression, UTF_8), index - 1);
-            
+
             byte c = expression[index - 1];
             if (!Authorizations.isValidAuthChar(c))
               throw new BadArgumentException("bad character (" + c + ")", new String(expression, UTF_8), index - 1);
@@ -413,7 +414,7 @@ public class ColumnVisibility {
       return result;
     }
   }
-  
+
   private void validate(byte[] expression) {
     if (expression != null && expression.length > 0) {
       ColumnVisibilityParser p = new ColumnVisibilityParser();
@@ -423,16 +424,16 @@ public class ColumnVisibility {
     }
     this.expression = expression;
   }
-  
+
   /**
    * Creates an empty visibility. Normally, elements with empty visibility can be seen by everyone. Though, one could change this behavior with filters.
-   * 
+   *
    * @see #ColumnVisibility(String)
    */
   public ColumnVisibility() {
     this(new byte[] {});
   }
-  
+
   /**
    * Creates a column visibility for a Mutation.
    *
@@ -442,32 +443,34 @@ public class ColumnVisibility {
   public ColumnVisibility(String expression) {
     this(expression.getBytes(UTF_8));
   }
-  
+
   /**
    * Creates a column visibility for a Mutation.
    *
-   * @param expression visibility expression
+   * @param expression
+   *          visibility expression
    * @see #ColumnVisibility(String)
    */
   public ColumnVisibility(Text expression) {
     this(TextUtil.getBytes(expression));
   }
-  
+
   /**
    * Creates a column visibility for a Mutation from a string already encoded in UTF-8 bytes.
    *
-   * @param expression visibility expression, encoded as UTF-8 bytes
+   * @param expression
+   *          visibility expression, encoded as UTF-8 bytes
    * @see #ColumnVisibility(String)
    */
   public ColumnVisibility(byte[] expression) {
     validate(expression);
   }
-  
+
   @Override
   public String toString() {
     return "[" + new String(expression, UTF_8) + "]";
   }
-  
+
   /**
    * See {@link #equals(ColumnVisibility)}
    */
@@ -477,22 +480,23 @@ public class ColumnVisibility {
       return equals((ColumnVisibility) obj);
     return false;
   }
-  
+
   /**
    * Compares two ColumnVisibilities for string equivalence, not as a meaningful comparison of terms and conditions.
    *
-   * @param otherLe other column visibility
+   * @param otherLe
+   *          other column visibility
    * @return true if this visibility equals the other via string comparison
    */
   public boolean equals(ColumnVisibility otherLe) {
     return Arrays.equals(expression, otherLe.expression);
   }
-  
+
   @Override
   public int hashCode() {
     return Arrays.hashCode(expression);
   }
-  
+
   /**
    * Gets the parse tree for this column visibility.
    *
@@ -501,7 +505,7 @@ public class ColumnVisibility {
   public Node getParseTree() {
     return node;
   }
-  
+
   /**
    * Properly quotes terms in a column visibility expression. If no quoting is needed, then nothing is done.
    *
@@ -516,33 +520,35 @@ public class ColumnVisibility {
    * ColumnVisibility cv = new ColumnVisibility(quote(&quot;A#C&quot;) + &quot;&amp;&quot; + quote(&quot;FOO&quot;));
    * </pre>
    *
-   * @param term term to quote
+   * @param term
+   *          term to quote
    * @return quoted term (unquoted if unnecessary)
    */
   public static String quote(String term) {
     return new String(quote(term.getBytes(UTF_8)), UTF_8);
   }
-  
+
   /**
    * Properly quotes terms in a column visibility expression. If no quoting is needed, then nothing is done.
    *
-   * @param term term to quote, encoded as UTF-8 bytes
+   * @param term
+   *          term to quote, encoded as UTF-8 bytes
    * @return quoted term (unquoted if unnecessary), encoded as UTF-8 bytes
    * @see #quote(String)
    */
   public static byte[] quote(byte[] term) {
     boolean needsQuote = false;
-    
+
     for (int i = 0; i < term.length; i++) {
       if (!Authorizations.isValidAuthChar(term[i])) {
         needsQuote = true;
         break;
       }
     }
-    
+
     if (!needsQuote)
       return term;
-    
+
     return VisibilityEvaluator.escape(term, true);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/Credentials.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/Credentials.java b/core/src/main/java/org/apache/accumulo/core/security/Credentials.java
index 525a958..ed7789a 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/Credentials.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/Credentials.java
@@ -19,6 +19,7 @@ package org.apache.accumulo.core.security;
 import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.nio.ByteBuffer;
+
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.Instance;
@@ -33,7 +34,7 @@ import org.apache.accumulo.core.util.Base64;
  * important, so that the authentication token carried in a {@link Connector} can be destroyed, invalidating future RPC operations from that {@link Connector}.
  * <p>
  * See ACCUMULO-1312
- * 
+ *
  * @since 1.6.0
  */
 public class Credentials {
@@ -43,7 +44,7 @@ public class Credentials {
 
   /**
    * Creates a new credentials object.
-   * 
+   *
    * @param principal
    *          unique identifier for the entity (e.g. a user or service) authorized for these credentials
    * @param token
@@ -56,7 +57,7 @@ public class Credentials {
 
   /**
    * Gets the principal.
-   * 
+   *
    * @return unique identifier for the entity (e.g. a user or service) authorized for these credentials
    */
   public String getPrincipal() {
@@ -65,7 +66,7 @@ public class Credentials {
 
   /**
    * Gets the authentication token.
-   * 
+   *
    * @return authentication token used to prove that the principal for these credentials has been properly verified
    */
   public AuthenticationToken getToken() {
@@ -75,7 +76,7 @@ public class Credentials {
   /**
    * Converts the current object to the relevant thrift type. The object returned from this contains a non-destroyable version of the
    * {@link AuthenticationToken}, so this should be used just before placing on the wire, and references to it should be tightly controlled.
-   * 
+   *
    * @param instance
    *          client instance
    * @return Thrift credentials
@@ -92,7 +93,9 @@ public class Credentials {
 
   /**
    * Converts a given thrift object to our internal Credentials representation.
-   * @param serialized a Thrift encoded set of credentials
+   *
+   * @param serialized
+   *          a Thrift encoded set of credentials
    * @return a new Credentials instance; destroy the token when you're done.
    */
   public static Credentials fromThrift(TCredentials serialized) {
@@ -102,7 +105,7 @@ public class Credentials {
   /**
    * Converts the current object to a serialized form. The object returned from this contains a non-destroyable version of the {@link AuthenticationToken}, so
    * references to it should be tightly controlled.
-   * 
+   *
    * @return serialized form of these credentials
    */
   public final String serialize() {
@@ -113,7 +116,7 @@ public class Credentials {
 
   /**
    * Converts the serialized form to an instance of {@link Credentials}. The original serialized form will not be affected.
-   * 
+   *
    * @param serializedForm
    *          serialized form of credentials
    * @return deserialized credentials

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/NamespacePermission.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/NamespacePermission.java b/core/src/main/java/org/apache/accumulo/core/security/NamespacePermission.java
index 44cee0c..638f630 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/NamespacePermission.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/NamespacePermission.java
@@ -30,8 +30,8 @@ public enum NamespacePermission {
   GRANT((byte) 3),
   ALTER_TABLE((byte) 4),
   CREATE_TABLE((byte) 5),
-  DROP_TABLE((byte) 6), 
-  BULK_IMPORT((byte) 7), 
+  DROP_TABLE((byte) 6),
+  BULK_IMPORT((byte) 7),
   DROP_NAMESPACE((byte) 8);
 
   final private byte permID;
@@ -48,7 +48,7 @@ public enum NamespacePermission {
 
   /**
    * Gets the byte ID of this permission.
-   * 
+   *
    * @return byte ID
    */
   public byte getId() {
@@ -57,7 +57,7 @@ public enum NamespacePermission {
 
   /**
    * Returns a list of printable permission values.
-   * 
+   *
    * @return list of namespace permission values, as "Namespace." + permission name
    */
   public static List<String> printableValues() {
@@ -73,7 +73,7 @@ public enum NamespacePermission {
 
   /**
    * Gets the permission matching the given byte ID.
-   * 
+   *
    * @param id
    *          byte ID
    * @return system permission

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/SystemPermission.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/SystemPermission.java b/core/src/main/java/org/apache/accumulo/core/security/SystemPermission.java
index d9109ec..b998179 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/SystemPermission.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/SystemPermission.java
@@ -25,8 +25,7 @@ import java.util.List;
  */
 public enum SystemPermission {
   /*
-   * One may add new permissions, but new permissions must use new numbers.
-   * Current numbers in use must not be changed.
+   * One may add new permissions, but new permissions must use new numbers. Current numbers in use must not be changed.
    */
   GRANT((byte) 0),
   CREATE_TABLE((byte) 1),
@@ -39,20 +38,20 @@ public enum SystemPermission {
   CREATE_NAMESPACE((byte) 8),
   DROP_NAMESPACE((byte) 9),
   ALTER_NAMESPACE((byte) 10);
-  
+
   private byte permID;
-  
+
   private static HashMap<Byte,SystemPermission> mapping;
   static {
     mapping = new HashMap<Byte,SystemPermission>(SystemPermission.values().length);
     for (SystemPermission perm : SystemPermission.values())
       mapping.put(perm.permID, perm);
   }
-  
+
   private SystemPermission(byte id) {
     this.permID = id;
   }
-  
+
   /**
    * Gets the byte ID of this permission.
    *
@@ -61,7 +60,7 @@ public enum SystemPermission {
   public byte getId() {
     return this.permID;
   }
-  
+
   /**
    * Returns a list of printable permission values.
    *
@@ -69,21 +68,23 @@ public enum SystemPermission {
    */
   public static List<String> printableValues() {
     SystemPermission[] a = SystemPermission.values();
-    
+
     List<String> list = new ArrayList<String>(a.length);
-    
+
     for (SystemPermission p : a)
       list.add("System." + p);
-    
+
     return list;
   }
-  
+
   /**
    * Gets the permission matching the given byte ID.
    *
-   * @param id byte ID
+   * @param id
+   *          byte ID
    * @return system permission
-   * @throws IndexOutOfBoundsException if the byte ID is invalid
+   * @throws IndexOutOfBoundsException
+   *           if the byte ID is invalid
    */
   public static SystemPermission getPermissionById(byte id) {
     if (mapping.containsKey(id))

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/TablePermission.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/TablePermission.java b/core/src/main/java/org/apache/accumulo/core/security/TablePermission.java
index 05895f5..b6d122f 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/TablePermission.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/TablePermission.java
@@ -24,8 +24,7 @@ import java.util.List;
  */
 public enum TablePermission {
   /*
-   * One may add new permissions, but new permissions must use new numbers.
-   * Current numbers in use must not be changed.
+   * One may add new permissions, but new permissions must use new numbers. Current numbers in use must not be changed.
    */
   // CREATE_LOCALITY_GROUP(0),
   // DROP_LOCALITY_GROUP(1),
@@ -35,19 +34,19 @@ public enum TablePermission {
   ALTER_TABLE((byte) 5),
   GRANT((byte) 6),
   DROP_TABLE((byte) 7);
-  
+
   final private byte permID;
-  
+
   final private static TablePermission mapping[] = new TablePermission[8];
   static {
     for (TablePermission perm : TablePermission.values())
       mapping[perm.permID] = perm;
   }
-  
+
   private TablePermission(byte id) {
     this.permID = id;
   }
-  
+
   /**
    * Gets the byte ID of this permission.
    *
@@ -56,7 +55,7 @@ public enum TablePermission {
   public byte getId() {
     return this.permID;
   }
-  
+
   /**
    * Returns a list of printable permission values.
    *
@@ -64,21 +63,23 @@ public enum TablePermission {
    */
   public static List<String> printableValues() {
     TablePermission[] a = TablePermission.values();
-    
+
     List<String> list = new ArrayList<String>(a.length);
-    
+
     for (TablePermission p : a)
       list.add("Table." + p);
-    
+
     return list;
   }
-  
+
   /**
    * Gets the permission matching the given byte ID.
    *
-   * @param id byte ID
+   * @param id
+   *          byte ID
    * @return table permission
-   * @throws IndexOutOfBoundsException if the byte ID is invalid
+   * @throws IndexOutOfBoundsException
+   *           if the byte ID is invalid
    */
   public static TablePermission getPermissionById(byte id) {
     TablePermission result = mapping[id];
@@ -86,5 +87,5 @@ public enum TablePermission {
       return result;
     throw new IndexOutOfBoundsException("No such permission");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/VisibilityConstraint.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/VisibilityConstraint.java b/core/src/main/java/org/apache/accumulo/core/security/VisibilityConstraint.java
index c4fe5ac..90cf951 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/VisibilityConstraint.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/VisibilityConstraint.java
@@ -28,8 +28,7 @@ import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.util.BadArgumentException;
 
 /**
- * A constraint that checks the visibility of columns against the actor's
- * authorizations. Violation codes:
+ * A constraint that checks the visibility of columns against the actor's authorizations. Violation codes:
  * <p>
  * <ul>
  * <li>1 = failure to parse visibility expression</li>
@@ -37,7 +36,7 @@ import org.apache.accumulo.core.util.BadArgumentException;
  * </ul>
  */
 public class VisibilityConstraint implements Constraint {
-  
+
   @Override
   public String getViolationDescription(short violationCode) {
     switch (violationCode) {
@@ -46,47 +45,47 @@ public class VisibilityConstraint implements Constraint {
       case 2:
         return "User does not have authorization on column visibility";
     }
-    
+
     return null;
   }
-  
+
   @Override
   public List<Short> check(Environment env, Mutation mutation) {
     List<ColumnUpdate> updates = mutation.getUpdates();
-    
+
     HashSet<String> ok = null;
     if (updates.size() > 1)
       ok = new HashSet<String>();
-    
+
     VisibilityEvaluator ve = null;
-    
+
     for (ColumnUpdate update : updates) {
-      
+
       byte[] cv = update.getColumnVisibility();
       if (cv.length > 0) {
         String key = null;
         if (ok != null && ok.contains(key = new String(cv, UTF_8)))
           continue;
-        
+
         try {
-          
+
           if (ve == null)
             ve = new VisibilityEvaluator(env);
-          
+
           if (!ve.evaluate(new ColumnVisibility(cv)))
             return Collections.singletonList(Short.valueOf((short) 2));
-          
+
         } catch (BadArgumentException bae) {
           return Collections.singletonList(new Short((short) 1));
         } catch (VisibilityParseException e) {
           return Collections.singletonList(new Short((short) 1));
         }
-        
+
         if (ok != null)
           ok.add(key);
       }
     }
-    
+
     return null;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/VisibilityEvaluator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/VisibilityEvaluator.java b/core/src/main/java/org/apache/accumulo/core/security/VisibilityEvaluator.java
index 6baa17c..8535731 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/VisibilityEvaluator.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/VisibilityEvaluator.java
@@ -26,39 +26,40 @@ import org.apache.accumulo.core.security.ColumnVisibility.Node;
  */
 public class VisibilityEvaluator {
   private AuthorizationContainer auths;
-  
+
   /**
-   * Creates a new {@link Authorizations} object with escaped forms of the
-   * authorizations in the given object.
+   * Creates a new {@link Authorizations} object with escaped forms of the authorizations in the given object.
    *
-   * @param auths original authorizations
+   * @param auths
+   *          original authorizations
    * @return authorizations object with escaped authorization strings
    * @see #escape(byte[], boolean)
    */
   static Authorizations escape(Authorizations auths) {
     ArrayList<byte[]> retAuths = new ArrayList<byte[]>(auths.getAuthorizations().size());
-    
+
     for (byte[] auth : auths.getAuthorizations())
       retAuths.add(escape(auth, false));
-    
+
     return new Authorizations(retAuths);
   }
-  
+
   /**
-   * Properly escapes an authorization string. The string can be quoted if
-   * desired.
+   * Properly escapes an authorization string. The string can be quoted if desired.
    *
-   * @param auth authorization string, as UTF-8 encoded bytes
-   * @param quote true to wrap escaped authorization in quotes
+   * @param auth
+   *          authorization string, as UTF-8 encoded bytes
+   * @param quote
+   *          true to wrap escaped authorization in quotes
    * @return escaped authorization string
    */
   public static byte[] escape(byte[] auth, boolean quote) {
     int escapeCount = 0;
-    
+
     for (int i = 0; i < auth.length; i++)
       if (auth[i] == '"' || auth[i] == '\\')
         escapeCount++;
-    
+
     if (escapeCount > 0 || quote) {
       byte[] escapedAuth = new byte[auth.length + escapeCount + (quote ? 2 : 0)];
       int index = quote ? 1 : 0;
@@ -67,52 +68,53 @@ public class VisibilityEvaluator {
           escapedAuth[index++] = '\\';
         escapedAuth[index++] = auth[i];
       }
-      
+
       if (quote) {
         escapedAuth[0] = '"';
         escapedAuth[escapedAuth.length - 1] = '"';
       }
-      
+
       auth = escapedAuth;
     }
     return auth;
   }
-  
+
   /**
-   * Creates a new evaluator for the authorizations found in the given
-   * environment.
+   * Creates a new evaluator for the authorizations found in the given environment.
    *
-   * @param env environment containing authorizations
+   * @param env
+   *          environment containing authorizations
    */
   VisibilityEvaluator(Environment env) {
     this.auths = env.getAuthorizationsContainer();
   }
-  
+
   /**
-   * Creates a new evaluator for the given collection of authorizations.
-   * Each authorization string is escaped before handling, and the original
-   * strings are unchanged.
+   * Creates a new evaluator for the given collection of authorizations. Each authorization string is escaped before handling, and the original strings are
+   * unchanged.
    *
-   * @param authorizations authorizations object
+   * @param authorizations
+   *          authorizations object
    */
   public VisibilityEvaluator(Authorizations authorizations) {
-      this.auths = escape((Authorizations) authorizations);
+    this.auths = escape((Authorizations) authorizations);
   }
-  
+
   /**
-   * Evaluates the given column visibility against the authorizations provided to this evaluator.
-   * A visibility passes evaluation if all authorizations in it are contained in those known to the evaluator, and
-   * all AND and OR subexpressions have at least two children.
+   * Evaluates the given column visibility against the authorizations provided to this evaluator. A visibility passes evaluation if all authorizations in it are
+   * contained in those known to the evaluator, and all AND and OR subexpressions have at least two children.
    *
-   * @param visibility column visibility to evaluate
+   * @param visibility
+   *          column visibility to evaluate
    * @return true if visibility passes evaluation
-   * @throws VisibilityParseException if an AND or OR subexpression has less than two children, or a subexpression is of an unknown type
+   * @throws VisibilityParseException
+   *           if an AND or OR subexpression has less than two children, or a subexpression is of an unknown type
    */
   public boolean evaluate(ColumnVisibility visibility) throws VisibilityParseException {
     // The VisibilityEvaluator computes a trie from the given Authorizations, that ColumnVisibility expressions can be evaluated against.
     return evaluate(visibility.getExpression(), visibility.getParseTree());
   }
-  
+
   private final boolean evaluate(final byte[] expression, final Node root) throws VisibilityParseException {
     if (expression.length == 0)
       return true;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/VisibilityParseException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/VisibilityParseException.java b/core/src/main/java/org/apache/accumulo/core/security/VisibilityParseException.java
index ca24884..d48840e 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/VisibilityParseException.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/VisibilityParseException.java
@@ -30,15 +30,18 @@ public class VisibilityParseException extends ParseException {
   /**
    * Creates a new exception.
    *
-   * @param reason reason string
-   * @param visibility visibility that could not be parsed
-   * @param errorOffset offset into visibility where parsing failed
+   * @param reason
+   *          reason string
+   * @param visibility
+   *          visibility that could not be parsed
+   * @param errorOffset
+   *          offset into visibility where parsing failed
    */
   public VisibilityParseException(String reason, byte[] visibility, int errorOffset) {
     super(reason, errorOffset);
     this.visibility = new String(visibility, UTF_8);
   }
-  
+
   @Override
   public String getMessage() {
     return super.getMessage() + " in string '" + visibility + "' at position " + super.getErrorOffset();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/BlockedOutputStream.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/BlockedOutputStream.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/BlockedOutputStream.java
index 3ce648e..c929cab 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/BlockedOutputStream.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/BlockedOutputStream.java
@@ -21,7 +21,7 @@ import java.io.IOException;
 import java.io.OutputStream;
 import java.nio.ByteBuffer;
 
-// Buffers all input in a growing buffer until flush() is called. Then entire buffer is written, with size information, and padding to force the underlying 
+// Buffers all input in a growing buffer until flush() is called. Then entire buffer is written, with size information, and padding to force the underlying
 // crypto output stream to also fully flush
 public class BlockedOutputStream extends OutputStream {
   int blockSize;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/CachingHDFSSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/CachingHDFSSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/CachingHDFSSecretKeyEncryptionStrategy.java
index e5ad13d..b70cf57 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/CachingHDFSSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/CachingHDFSSecretKeyEncryptionStrategy.java
@@ -39,11 +39,10 @@ import org.apache.log4j.Logger;
  * A {@link SecretKeyEncryptionStrategy} that gets its key from HDFS and caches it for IO.
  */
 public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncryptionStrategy {
-  
+
   private static final Logger log = Logger.getLogger(CachingHDFSSecretKeyEncryptionStrategy.class);
   private SecretKeyCache secretKeyCache = new SecretKeyCache();
-  
-  
+
   @Override
   public CryptoModuleParameters encryptSecretKey(CryptoModuleParameters context) {
     try {
@@ -55,7 +54,7 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
     }
     return context;
   }
-  
+
   @Override
   public CryptoModuleParameters decryptSecretKey(CryptoModuleParameters context) {
     try {
@@ -67,7 +66,7 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
     }
     return context;
   }
-  
+
   private void doKeyEncryptionOperation(int encryptionMode, CryptoModuleParameters params) throws IOException {
     Cipher cipher = DefaultCryptoModuleUtils.getCipher(params.getAllOptions().get(Property.CRYPTO_DEFAULT_KEY_STRATEGY_CIPHER_SUITE.getKey()));
 
@@ -76,8 +75,8 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
     } catch (InvalidKeyException e) {
       log.error(e);
       throw new RuntimeException(e);
-    }      
-    
+    }
+
     if (Cipher.UNWRAP_MODE == encryptionMode) {
       try {
         Key plaintextKey = cipher.unwrap(params.getEncryptedKey(), params.getAlgorithmName(), Cipher.SECRET_KEY);
@@ -102,43 +101,41 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
         log.error(e);
         throw new RuntimeException(e);
       }
-      
-    } 
+
+    }
   }
-  
+
   private static class SecretKeyCache {
 
     private boolean initialized = false;
     private byte[] keyEncryptionKey;
     private String pathToKeyName;
-    
-    
-    public SecretKeyCache() {
-    };
-    
+
+    public SecretKeyCache() {};
+
     public synchronized void ensureSecretKeyCacheInitialized(CryptoModuleParameters context) throws IOException {
-      
+
       if (initialized) {
         return;
       }
-      
+
       // First identify if the KEK already exists
       pathToKeyName = getFullPathToKey(context);
-      
+
       if (pathToKeyName == null || pathToKeyName.equals("")) {
         pathToKeyName = Property.CRYPTO_DEFAULT_KEY_STRATEGY_KEY_LOCATION.getDefaultValue();
       }
 
       // TODO ACCUMULO-2530 Ensure volumes a properly supported
-      Path pathToKey = new Path(pathToKeyName);  
-      FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());   
+      Path pathToKey = new Path(pathToKeyName);
+      FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
 
       DataInputStream in = null;
       try {
         if (!fs.exists(pathToKey)) {
           initializeKeyEncryptionKey(fs, pathToKey, context);
         }
-        
+
         in = fs.open(pathToKey);
 
         int keyEncryptionKeyLength = in.readInt();
@@ -146,14 +143,14 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
         in.read(keyEncryptionKey);
 
         initialized = true;
-        
+
       } catch (IOException e) {
         log.error("Could not initialize key encryption cache", e);
       } finally {
         IOUtils.closeQuietly(in);
       }
     }
-    
+
     private void initializeKeyEncryptionKey(FileSystem fs, Path pathToKey, CryptoModuleParameters params) throws IOException {
       DataOutputStream out = null;
       try {
@@ -169,30 +166,29 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
         out.flush();
       } finally {
         if (out != null) {
-          out.close();        
+          out.close();
         }
       }
-      
+
     }
 
     @SuppressWarnings("deprecation")
     private String getFullPathToKey(CryptoModuleParameters params) {
       String pathToKeyName = params.getAllOptions().get(Property.CRYPTO_DEFAULT_KEY_STRATEGY_KEY_LOCATION.getKey());
       String instanceDirectory = params.getAllOptions().get(Property.INSTANCE_DFS_DIR.getKey());
-      
-      
+
       if (pathToKeyName == null) {
         pathToKeyName = Property.CRYPTO_DEFAULT_KEY_STRATEGY_KEY_LOCATION.getDefaultValue();
       }
-      
+
       if (instanceDirectory == null) {
         instanceDirectory = Property.INSTANCE_DFS_DIR.getDefaultValue();
       }
-      
+
       if (!pathToKeyName.startsWith("/")) {
         pathToKeyName = "/" + pathToKeyName;
       }
-      
+
       String fullPath = instanceDirectory + pathToKeyName;
       return fullPath;
     }
@@ -205,6 +201,5 @@ public class CachingHDFSSecretKeyEncryptionStrategy implements SecretKeyEncrypti
       return pathToKeyName;
     }
   }
-  
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModule.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModule.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModule.java
index 535c141..44531dc 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModule.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModule.java
@@ -27,70 +27,71 @@ import javax.crypto.CipherOutputStream;
 /**
  * Classes that obey this interface may be used to provide encrypting and decrypting streams to the rest of Accumulo. Classes that obey this interface may be
  * configured as the crypto module by setting the property crypto.module.class in the accumulo-site.xml file.
- * 
- * 
+ *
+ *
  */
 public interface CryptoModule {
-  
+
   /**
-   * Takes a {@link CryptoModuleParameters} object containing an {@link OutputStream} to wrap within a {@link CipherOutputStream}. The various other parts of the
-   * {@link CryptoModuleParameters} object specify the details about the type of encryption to use. Callers should pay special attention to the
+   * Takes a {@link CryptoModuleParameters} object containing an {@link OutputStream} to wrap within a {@link CipherOutputStream}. The various other parts of
+   * the {@link CryptoModuleParameters} object specify the details about the type of encryption to use. Callers should pay special attention to the
    * {@link CryptoModuleParameters#getRecordParametersToStream()} and {@link CryptoModuleParameters#getCloseUnderylingStreamAfterCryptoStreamClose()} flags
    * within the {@link CryptoModuleParameters} object, as they control whether or not this method will write to the given {@link OutputStream} in
    * {@link CryptoModuleParameters#getPlaintextOutputStream()}.
-   * 
+   *
    * <p>
-   * 
+   *
    * This method returns a {@link CryptoModuleParameters} object. Implementers of this interface maintain a contract that the returned object is <i>the same</i>
    * as the one passed in, always. Return values are enclosed within that object, as some other calls will typically return more than one value.
-   * 
+   *
    * @param params
    *          the {@link CryptoModuleParameters} object that specifies how to set up the encrypted stream.
    * @return the same {@link CryptoModuleParameters} object with the {@link CryptoModuleParameters#getEncryptedOutputStream()} set to a stream that is not null.
-   *         That stream may be exactly the same stream as {@link CryptoModuleParameters#getPlaintextInputStream()} if the params object specifies no cryptography.
+   *         That stream may be exactly the same stream as {@link CryptoModuleParameters#getPlaintextInputStream()} if the params object specifies no
+   *         cryptography.
    */
   CryptoModuleParameters getEncryptingOutputStream(CryptoModuleParameters params) throws IOException;
-  
-  
-  
+
   /**
    * Takes a {@link CryptoModuleParameters} object containing an {@link InputStream} to wrap within a {@link CipherInputStream}. The various other parts of the
    * {@link CryptoModuleParameters} object specify the details about the type of encryption to use. Callers should pay special attention to the
    * {@link CryptoModuleParameters#getRecordParametersToStream()} and {@link CryptoModuleParameters#getCloseUnderylingStreamAfterCryptoStreamClose()} flags
    * within the {@link CryptoModuleParameters} object, as they control whether or not this method will read from the given {@link InputStream} in
    * {@link CryptoModuleParameters#getEncryptedInputStream()}.
-   * 
+   *
    * <p>
-   * 
+   *
    * This method returns a {@link CryptoModuleParameters} object. Implementers of this interface maintain a contract that the returned object is <i>the same</i>
    * as the one passed in, always. Return values are enclosed within that object, as some other calls will typically return more than one value.
-   * 
+   *
    * @param params
    *          the {@link CryptoModuleParameters} object that specifies how to set up the encrypted stream.
    * @return the same {@link CryptoModuleParameters} object with the {@link CryptoModuleParameters#getPlaintextInputStream()} set to a stream that is not null.
-   *         That stream may be exactly the same stream as {@link CryptoModuleParameters#getEncryptedInputStream()} if the params object specifies no cryptography.
+   *         That stream may be exactly the same stream as {@link CryptoModuleParameters#getEncryptedInputStream()} if the params object specifies no
+   *         cryptography.
    */
   CryptoModuleParameters getDecryptingInputStream(CryptoModuleParameters params) throws IOException;
 
-  
   /**
-   * Generates a random session key and sets it into the {@link CryptoModuleParameters#getPlaintextKey()} property.  Saves callers from 
-   * having to set up their own secure random provider.  Also will set the {@link CryptoModuleParameters#getSecureRandom()} property if it
-   * has not already been set by some other function.
-   * 
-   * @param params a {@link CryptoModuleParameters} object contained a correctly instantiated set of properties.
+   * Generates a random session key and sets it into the {@link CryptoModuleParameters#getPlaintextKey()} property. Saves callers from having to set up their
+   * own secure random provider. Also will set the {@link CryptoModuleParameters#getSecureRandom()} property if it has not already been set by some other
+   * function.
+   *
+   * @param params
+   *          a {@link CryptoModuleParameters} object contained a correctly instantiated set of properties.
    * @return the same {@link CryptoModuleParameters} object with the plaintext key set
    */
   CryptoModuleParameters generateNewRandomSessionKey(CryptoModuleParameters params);
-  
+
   /**
    * Generates a {@link Cipher} object based on the parameters in the given {@link CryptoModuleParameters} object and places it into the
-   * {@link CryptoModuleParameters#getCipher()} property. Callers may choose to use this method if they want to get the initialization
-   * vector from the cipher before proceeding to create wrapped streams.  
-   * 
-   * @param params a {@link CryptoModuleParameters} object contained a correctly instantiated set of properties.
+   * {@link CryptoModuleParameters#getCipher()} property. Callers may choose to use this method if they want to get the initialization vector from the cipher
+   * before proceeding to create wrapped streams.
+   *
+   * @param params
+   *          a {@link CryptoModuleParameters} object contained a correctly instantiated set of properties.
    * @return the same {@link CryptoModuleParameters} object with the cipher set.
    */
   CryptoModuleParameters initializeCipher(CryptoModuleParameters params);
- 
+
 }


[29/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/ConditionTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/ConditionTest.java b/core/src/test/java/org/apache/accumulo/core/data/ConditionTest.java
index c76275b..8de9668 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/ConditionTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/ConditionTest.java
@@ -195,34 +195,34 @@ public class ConditionTest {
     c2.setIterators(ITERATORS);
     assertTrue(c.equals(c2));
     assertTrue(c2.equals(c));
-    
+
     // set everything but vis, so its null
     Condition c4 = new Condition(FAMILY, QUALIFIER);
     c4.setValue(VALUE);
     c4.setTimestamp(1234L);
     c4.setIterators(ITERATORS);
-    
+
     assertFalse(c.equals(c4));
     assertFalse(c4.equals(c));
-    
+
     // set everything but timestamp, so its null
     Condition c5 = new Condition(FAMILY, QUALIFIER);
     c5.setVisibility(cvis);
     c5.setValue(VALUE);
     c5.setIterators(ITERATORS);
-    
+
     assertFalse(c.equals(c5));
     assertFalse(c5.equals(c));
-    
+
     // set everything but value
     Condition c6 = new Condition(FAMILY, QUALIFIER);
     c6.setVisibility(cvis);
     c6.setTimestamp(1234L);
     c6.setIterators(ITERATORS);
-    
+
     assertFalse(c.equals(c6));
     assertFalse(c6.equals(c));
-    
+
     // test w/ no optional fields set
     Condition c7 = new Condition(FAMILY, QUALIFIER);
     Condition c8 = new Condition(FAMILY, QUALIFIER);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/ConstraintViolationSummaryTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/ConstraintViolationSummaryTest.java b/core/src/test/java/org/apache/accumulo/core/data/ConstraintViolationSummaryTest.java
index ec2110e..00e7c56 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/ConstraintViolationSummaryTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/ConstraintViolationSummaryTest.java
@@ -16,7 +16,7 @@
  */
 package org.apache.accumulo.core.data;
 
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
 
 import org.junit.Test;
 
@@ -24,14 +24,11 @@ public class ConstraintViolationSummaryTest {
 
   @Test
   public void testToString() {
-    ConstraintViolationSummary cvs = new ConstraintViolationSummary(
-        "fooClass", (short) 1, "fooDescription", 100L);
-    assertEquals("ConstraintViolationSummary(constrainClass:fooClass, violationCode:1, violationDescription:fooDescription, numberOfViolatingMutations:100)", 
-        cvs.toString());
-    
-    cvs = new ConstraintViolationSummary(
-        null, (short) 2, null, 101L);
-    assertEquals("ConstraintViolationSummary(constrainClass:null, violationCode:2, violationDescription:null, numberOfViolatingMutations:101)", 
+    ConstraintViolationSummary cvs = new ConstraintViolationSummary("fooClass", (short) 1, "fooDescription", 100L);
+    assertEquals("ConstraintViolationSummary(constrainClass:fooClass, violationCode:1, violationDescription:fooDescription, numberOfViolatingMutations:100)",
         cvs.toString());
+
+    cvs = new ConstraintViolationSummary(null, (short) 2, null, 101L);
+    assertEquals("ConstraintViolationSummary(constrainClass:null, violationCode:2, violationDescription:null, numberOfViolatingMutations:101)", cvs.toString());
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/KeyExtentTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/KeyExtentTest.java b/core/src/test/java/org/apache/accumulo/core/data/KeyExtentTest.java
index b32dd26..749f14c 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/KeyExtentTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/KeyExtentTest.java
@@ -45,7 +45,7 @@ public class KeyExtentTest {
   KeyExtent nke(String t, String er, String per) {
     return new KeyExtent(new Text(t), er == null ? null : new Text(er), per == null ? null : new Text(per));
   }
-  
+
   KeyExtent ke;
   TreeSet<KeyExtent> set0;
 
@@ -57,153 +57,153 @@ public class KeyExtentTest {
   @Test
   public void testDecodingMetadataRow() {
     Text flattenedExtent = new Text("foo;bar");
-    
+
     ke = new KeyExtent(flattenedExtent, (Text) null);
-    
+
     assertEquals(new Text("bar"), ke.getEndRow());
     assertEquals(new Text("foo"), ke.getTableId());
     assertNull(ke.getPrevEndRow());
-    
+
     flattenedExtent = new Text("foo<");
-    
+
     ke = new KeyExtent(flattenedExtent, (Text) null);
-    
+
     assertNull(ke.getEndRow());
     assertEquals(new Text("foo"), ke.getTableId());
     assertNull(ke.getPrevEndRow());
-    
+
     flattenedExtent = new Text("foo;bar;");
-    
+
     ke = new KeyExtent(flattenedExtent, (Text) null);
-    
+
     assertEquals(new Text("bar;"), ke.getEndRow());
     assertEquals(new Text("foo"), ke.getTableId());
     assertNull(ke.getPrevEndRow());
-    
+
   }
-  
+
   @Test
   public void testFindContainingExtents() {
     assertNull(KeyExtent.findContainingExtent(nke("t", null, null), set0));
     assertNull(KeyExtent.findContainingExtent(nke("t", "1", "0"), set0));
     assertNull(KeyExtent.findContainingExtent(nke("t", "1", null), set0));
     assertNull(KeyExtent.findContainingExtent(nke("t", null, "0"), set0));
-    
+
     TreeSet<KeyExtent> set1 = new TreeSet<KeyExtent>();
-    
+
     set1.add(nke("t", null, null));
-    
+
     assertEquals(nke("t", null, null), KeyExtent.findContainingExtent(nke("t", null, null), set1));
     assertEquals(nke("t", null, null), KeyExtent.findContainingExtent(nke("t", "1", "0"), set1));
     assertEquals(nke("t", null, null), KeyExtent.findContainingExtent(nke("t", "1", null), set1));
     assertEquals(nke("t", null, null), KeyExtent.findContainingExtent(nke("t", null, "0"), set1));
-    
+
     TreeSet<KeyExtent> set2 = new TreeSet<KeyExtent>();
-    
+
     set2.add(nke("t", "g", null));
     set2.add(nke("t", null, "g"));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", null, null), set2));
     assertEquals(nke("t", "g", null), KeyExtent.findContainingExtent(nke("t", "c", "a"), set2));
     assertEquals(nke("t", "g", null), KeyExtent.findContainingExtent(nke("t", "c", null), set2));
-    
+
     assertEquals(nke("t", "g", null), KeyExtent.findContainingExtent(nke("t", "g", "a"), set2));
     assertEquals(nke("t", "g", null), KeyExtent.findContainingExtent(nke("t", "g", null), set2));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", "h", "a"), set2));
     assertNull(KeyExtent.findContainingExtent(nke("t", "h", null), set2));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", "z", "f"), set2));
     assertNull(KeyExtent.findContainingExtent(nke("t", null, "f"), set2));
-    
+
     assertEquals(nke("t", null, "g"), KeyExtent.findContainingExtent(nke("t", "z", "g"), set2));
     assertEquals(nke("t", null, "g"), KeyExtent.findContainingExtent(nke("t", null, "g"), set2));
-    
+
     assertEquals(nke("t", null, "g"), KeyExtent.findContainingExtent(nke("t", "z", "h"), set2));
     assertEquals(nke("t", null, "g"), KeyExtent.findContainingExtent(nke("t", null, "h"), set2));
-    
+
     TreeSet<KeyExtent> set3 = new TreeSet<KeyExtent>();
-    
+
     set3.add(nke("t", "g", null));
     set3.add(nke("t", "s", "g"));
     set3.add(nke("t", null, "s"));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", null, null), set3));
-    
+
     assertEquals(nke("t", "g", null), KeyExtent.findContainingExtent(nke("t", "g", null), set3));
     assertEquals(nke("t", "s", "g"), KeyExtent.findContainingExtent(nke("t", "s", "g"), set3));
     assertEquals(nke("t", null, "s"), KeyExtent.findContainingExtent(nke("t", null, "s"), set3));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", "t", "g"), set3));
     assertNull(KeyExtent.findContainingExtent(nke("t", "t", "f"), set3));
     assertNull(KeyExtent.findContainingExtent(nke("t", "s", "f"), set3));
-    
+
     assertEquals(nke("t", "s", "g"), KeyExtent.findContainingExtent(nke("t", "r", "h"), set3));
     assertEquals(nke("t", "s", "g"), KeyExtent.findContainingExtent(nke("t", "s", "h"), set3));
     assertEquals(nke("t", "s", "g"), KeyExtent.findContainingExtent(nke("t", "r", "g"), set3));
-    
+
     assertEquals(nke("t", null, "s"), KeyExtent.findContainingExtent(nke("t", null, "t"), set3));
     assertNull(KeyExtent.findContainingExtent(nke("t", null, "r"), set3));
-    
+
     assertEquals(nke("t", "g", null), KeyExtent.findContainingExtent(nke("t", "f", null), set3));
     assertNull(KeyExtent.findContainingExtent(nke("t", "h", null), set3));
-    
+
     TreeSet<KeyExtent> set4 = new TreeSet<KeyExtent>();
-    
+
     set4.add(nke("t1", "d", null));
     set4.add(nke("t1", "q", "d"));
     set4.add(nke("t1", null, "q"));
     set4.add(nke("t2", "g", null));
     set4.add(nke("t2", "s", "g"));
     set4.add(nke("t2", null, "s"));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", null, null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("z", null, null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("t11", null, null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("t1", null, null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("t2", null, null), set4));
-    
+
     assertNull(KeyExtent.findContainingExtent(nke("t", "g", null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("z", "g", null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("t11", "g", null), set4));
     assertNull(KeyExtent.findContainingExtent(nke("t1", "g", null), set4));
-    
+
     assertEquals(nke("t2", "g", null), KeyExtent.findContainingExtent(nke("t2", "g", null), set4));
     assertEquals(nke("t2", "s", "g"), KeyExtent.findContainingExtent(nke("t2", "s", "g"), set4));
     assertEquals(nke("t2", null, "s"), KeyExtent.findContainingExtent(nke("t2", null, "s"), set4));
-    
+
     assertEquals(nke("t1", "d", null), KeyExtent.findContainingExtent(nke("t1", "d", null), set4));
     assertEquals(nke("t1", "q", "d"), KeyExtent.findContainingExtent(nke("t1", "q", "d"), set4));
     assertEquals(nke("t1", null, "q"), KeyExtent.findContainingExtent(nke("t1", null, "q"), set4));
-    
+
   }
-  
+
   private static boolean overlaps(KeyExtent extent, SortedMap<KeyExtent,Object> extents) {
     return !KeyExtent.findOverlapping(extent, extents).isEmpty();
   }
-  
+
   @Test
   public void testOverlaps() {
     SortedMap<KeyExtent,Object> set0 = new TreeMap<KeyExtent,Object>();
     set0.put(nke("a", null, null), null);
-    
+
     // Nothing overlaps with the empty set
     assertFalse(overlaps(nke("t", null, null), null));
     assertFalse(overlaps(nke("t", null, null), set0));
-    
+
     SortedMap<KeyExtent,Object> set1 = new TreeMap<KeyExtent,Object>();
-    
+
     // Everything overlaps with the infinite range
     set1.put(nke("t", null, null), null);
     assertTrue(overlaps(nke("t", null, null), set1));
     assertTrue(overlaps(nke("t", "b", "a"), set1));
     assertTrue(overlaps(nke("t", null, "a"), set1));
-    
+
     set1.put(nke("t", "b", "a"), null);
     assertTrue(overlaps(nke("t", null, null), set1));
     assertTrue(overlaps(nke("t", "b", "a"), set1));
     assertTrue(overlaps(nke("t", null, "a"), set1));
-    
+
     // simple overlaps
     SortedMap<KeyExtent,Object> set2 = new TreeMap<KeyExtent,Object>();
     set2.put(nke("a", null, null), null);
@@ -214,7 +214,7 @@ public class KeyExtentTest {
     assertTrue(overlaps(nke("t", "z", "a"), set2));
     assertFalse(overlaps(nke("t", "j", "a"), set2));
     assertFalse(overlaps(nke("t", "z", "m"), set2));
-    
+
     // non-overlaps
     assertFalse(overlaps(nke("t", "b", "a"), set2));
     assertFalse(overlaps(nke("t", "z", "y"), set2));
@@ -222,7 +222,7 @@ public class KeyExtentTest {
     assertFalse(overlaps(nke("t", null, "y"), set2));
     assertFalse(overlaps(nke("t", "j", null), set2));
     assertFalse(overlaps(nke("t", null, "m"), set2));
-    
+
     // infinite overlaps
     SortedMap<KeyExtent,Object> set3 = new TreeMap<KeyExtent,Object>();
     set3.put(nke("t", "j", null), null);
@@ -232,10 +232,10 @@ public class KeyExtentTest {
     assertTrue(overlaps(nke("t", "z", "k"), set3));
     assertTrue(overlaps(nke("t", null, "k"), set3));
     assertTrue(overlaps(nke("t", null, null), set3));
-    
+
     // falls between
     assertFalse(overlaps(nke("t", "l", "k"), set3));
-    
+
     SortedMap<KeyExtent,Object> set4 = new TreeMap<KeyExtent,Object>();
     set4.put(nke("t", null, null), null);
     assertTrue(overlaps(nke("t", "k", "a"), set4));
@@ -244,7 +244,7 @@ public class KeyExtentTest {
     assertTrue(overlaps(nke("t", null, "k"), set4));
     assertTrue(overlaps(nke("t", null, null), set4));
     assertTrue(overlaps(nke("t", null, null), set4));
-    
+
     for (String er : new String[] {"z", "y", "r", null}) {
       for (String per : new String[] {"a", "b", "d", null}) {
         assertTrue(nke("t", "y", "b").overlaps(nke("t", er, per)));
@@ -253,7 +253,7 @@ public class KeyExtentTest {
         assertTrue(nke("t", null, null).overlaps(nke("t", er, per)));
       }
     }
-    
+
     assertFalse(nke("t", "y", "b").overlaps(nke("t", "z", "y")));
     assertFalse(nke("t", "y", "b").overlaps(nke("t", null, "y")));
     assertFalse(nke("t", "y", null).overlaps(nke("t", "z", "y")));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/KeyTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/KeyTest.java b/core/src/test/java/org/apache/accumulo/core/data/KeyTest.java
index cc737db..9cee691 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/KeyTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/KeyTest.java
@@ -39,87 +39,87 @@ public class KeyTest {
     Key k2 = new Key("r1".getBytes(), "cf".getBytes(), "cq".getBytes(), new byte[0], 0, false);
     Key k3 = new Key("r1".getBytes(), "cf".getBytes(), "cq".getBytes(), new byte[0], 0, true);
     Key k4 = new Key("r1".getBytes(), "cf".getBytes(), "cq".getBytes(), new byte[0], 0, true);
-    
+
     assertTrue(k1.equals(k2));
     assertTrue(k3.equals(k4));
     assertTrue(k1.compareTo(k3) > 0);
     assertTrue(k3.compareTo(k1) < 0);
   }
-  
+
   @Test
   public void testCopyData() {
     byte row[] = "r".getBytes();
     byte cf[] = "cf".getBytes();
     byte cq[] = "cq".getBytes();
     byte cv[] = "cv".getBytes();
-    
+
     Key k1 = new Key(row, cf, cq, cv, 5l, false, false);
     Key k2 = new Key(row, cf, cq, cv, 5l, false, true);
-    
+
     assertSame(row, k1.getRowBytes());
     assertSame(cf, k1.getColFamily());
     assertSame(cq, k1.getColQualifier());
     assertSame(cv, k1.getColVisibility());
-    
+
     assertSame(row, k1.getRowData().getBackingArray());
     assertSame(cf, k1.getColumnFamilyData().getBackingArray());
     assertSame(cq, k1.getColumnQualifierData().getBackingArray());
     assertSame(cv, k1.getColumnVisibilityData().getBackingArray());
-    
+
     assertNotSame(row, k2.getRowBytes());
     assertNotSame(cf, k2.getColFamily());
     assertNotSame(cq, k2.getColQualifier());
     assertNotSame(cv, k2.getColVisibility());
-    
+
     assertNotSame(row, k2.getRowData().getBackingArray());
     assertNotSame(cf, k2.getColumnFamilyData().getBackingArray());
     assertNotSame(cq, k2.getColumnQualifierData().getBackingArray());
     assertNotSame(cv, k2.getColumnVisibilityData().getBackingArray());
-    
+
     assertEquals(k1, k2);
-    
+
   }
-  
+
   @Test
   public void testString() {
     Key k1 = new Key("r1");
     Key k2 = new Key(new Text("r1"));
     assertEquals(k2, k1);
-    
+
     k1 = new Key("r1", "cf1");
     k2 = new Key(new Text("r1"), new Text("cf1"));
     assertEquals(k2, k1);
-    
+
     k1 = new Key("r1", "cf2", "cq2");
     k2 = new Key(new Text("r1"), new Text("cf2"), new Text("cq2"));
     assertEquals(k2, k1);
-    
+
     k1 = new Key("r1", "cf2", "cq2", "cv");
     k2 = new Key(new Text("r1"), new Text("cf2"), new Text("cq2"), new Text("cv"));
     assertEquals(k2, k1);
-    
+
     k1 = new Key("r1", "cf2", "cq2", "cv", 89);
     k2 = new Key(new Text("r1"), new Text("cf2"), new Text("cq2"), new Text("cv"), 89);
     assertEquals(k2, k1);
-    
+
     k1 = new Key("r1", "cf2", "cq2", 89);
     k2 = new Key(new Text("r1"), new Text("cf2"), new Text("cq2"), 89);
     assertEquals(k2, k1);
-    
+
   }
-  
+
   @Test
   public void testVisibilityFollowingKey() {
     Key k = new Key("r", "f", "q", "v");
     assertEquals(k.followingKey(PartialKey.ROW_COLFAM_COLQUAL_COLVIS).toString(), "r f:q [v%00;] " + Long.MAX_VALUE + " false");
   }
-  
+
   public void testVisibilityGetters() {
     Key k = new Key("r", "f", "q", "v1|(v2&v3)");
-    
+
     Text expression = k.getColumnVisibility();
     ColumnVisibility parsed = k.getColumnVisibilityParsed();
-    
+
     assertEquals(expression, new Text(parsed.getExpression()));
   }
 
@@ -131,7 +131,7 @@ public class KeyTest {
     assertEquals(k, k2);
   }
 
-  @Test(expected=IllegalArgumentException.class)
+  @Test(expected = IllegalArgumentException.class)
   public void testThrift_Invalid() {
     Key k = new Key("r1", "cf2", "cq2", "cv");
     TKey tk = k.toThrift();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/MutationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/MutationTest.java b/core/src/test/java/org/apache/accumulo/core/data/MutationTest.java
index 33b060e..8b50788 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/MutationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/MutationTest.java
@@ -16,6 +16,10 @@
  */
 package org.apache.accumulo.core.data;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
+
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
 import java.io.DataInputStream;
@@ -24,17 +28,13 @@ import java.io.IOException;
 import java.util.Arrays;
 import java.util.List;
 
-import org.junit.Test;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
-
 import org.apache.accumulo.core.data.thrift.TMutation;
 import org.apache.accumulo.core.security.ColumnVisibility;
 import org.apache.hadoop.io.Text;
+import org.junit.Test;
+
+public class MutationTest {
 
-public class MutationTest  {
-  
   private static String toHexString(byte[] ba) {
     StringBuilder str = new StringBuilder();
     for (int i = 0; i < ba.length; i++) {
@@ -43,102 +43,101 @@ public class MutationTest  {
     return str.toString();
   }
 
-  /* Test constructing a Mutation using a byte buffer. The byte array
-   * returned as the row is converted to a hexadecimal string for easy
-   * comparision.
+  /*
+   * Test constructing a Mutation using a byte buffer. The byte array returned as the row is converted to a hexadecimal string for easy comparision.
    */
   public void testByteConstructor() {
     Mutation m = new Mutation("0123456789".getBytes());
     assertEquals("30313233343536373839", toHexString(m.getRow()));
   }
-  
+
   public void testLimitedByteConstructor() {
     Mutation m = new Mutation("0123456789".getBytes(), 2, 5);
     assertEquals("3233343536", toHexString(m.getRow()));
   }
-  
+
   @Test
   public void test1() {
     Mutation m = new Mutation(new Text("r1"));
     m.put(new Text("cf1"), new Text("cq1"), new Value("v1".getBytes()));
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(1, updates.size());
-    
+
     ColumnUpdate cu = updates.get(0);
-    
+
     assertEquals("cf1", new String(cu.getColumnFamily()));
     assertEquals("cq1", new String(cu.getColumnQualifier()));
     assertEquals("", new String(cu.getColumnVisibility()));
     assertFalse(cu.hasTimestamp());
-    
+
   }
-  
+
   @Test
   public void test2() throws IOException {
     Mutation m = new Mutation(new Text("r1"));
     m.put(new Text("cf1"), new Text("cq1"), new Value("v1".getBytes()));
     m.put(new Text("cf2"), new Text("cq2"), 56, new Value("v2".getBytes()));
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(2, updates.size());
-    
+
     assertEquals("r1", new String(m.getRow()));
     ColumnUpdate cu = updates.get(0);
-    
+
     assertEquals("cf1", new String(cu.getColumnFamily()));
     assertEquals("cq1", new String(cu.getColumnQualifier()));
     assertEquals("", new String(cu.getColumnVisibility()));
     assertFalse(cu.hasTimestamp());
-    
+
     cu = updates.get(1);
-    
+
     assertEquals("cf2", new String(cu.getColumnFamily()));
     assertEquals("cq2", new String(cu.getColumnQualifier()));
     assertEquals("", new String(cu.getColumnVisibility()));
     assertTrue(cu.hasTimestamp());
     assertEquals(56, cu.getTimestamp());
-    
+
     m = cloneMutation(m);
-    
+
     assertEquals("r1", new String(m.getRow()));
     updates = m.getUpdates();
-    
+
     assertEquals(2, updates.size());
-    
+
     cu = updates.get(0);
-    
+
     assertEquals("cf1", new String(cu.getColumnFamily()));
     assertEquals("cq1", new String(cu.getColumnQualifier()));
     assertEquals("", new String(cu.getColumnVisibility()));
     assertFalse(cu.hasTimestamp());
-    
+
     cu = updates.get(1);
-    
+
     assertEquals("cf2", new String(cu.getColumnFamily()));
     assertEquals("cq2", new String(cu.getColumnQualifier()));
     assertEquals("", new String(cu.getColumnVisibility()));
     assertTrue(cu.hasTimestamp());
     assertEquals(56, cu.getTimestamp());
-    
+
   }
-  
+
   private Mutation cloneMutation(Mutation m) throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
     m.write(dos);
     dos.close();
-    
+
     ByteArrayInputStream bais = new ByteArrayInputStream(baos.toByteArray());
     DataInputStream dis = new DataInputStream(bais);
-    
+
     m = new Mutation();
     m.readFields(dis);
     return m;
   }
-  
+
   @Test
   public void test3() throws IOException {
     Mutation m = new Mutation(new Text("r1"));
@@ -147,11 +146,11 @@ public class MutationTest  {
       byte val[] = new byte[len];
       for (int j = 0; j < len; j++)
         val[j] = (byte) i;
-      
+
       m.put(new Text("cf" + i), new Text("cq" + i), new Value(val));
-      
+
     }
-    
+
     for (int r = 0; r < 3; r++) {
       assertEquals("r1", new String(m.getRow()));
       List<ColumnUpdate> updates = m.getUpdates();
@@ -162,149 +161,149 @@ public class MutationTest  {
         assertEquals("cq" + i, new String(cu.getColumnQualifier()));
         assertEquals("", new String(cu.getColumnVisibility()));
         assertFalse(cu.hasTimestamp());
-        
+
         byte[] val = cu.getValue();
         int len = Mutation.VALUE_SIZE_COPY_CUTOFF - 2 + i;
         assertEquals(len, val.length);
         for (int j = 0; j < len; j++)
           assertEquals(i, val[j]);
       }
-      
+
       m = cloneMutation(m);
     }
   }
-  
+
   private Text nt(String s) {
     return new Text(s);
   }
-  
+
   private Value nv(String s) {
     return new Value(s.getBytes());
   }
-  
+
   @Test
   public void testPuts() {
     Mutation m = new Mutation(new Text("r1"));
-    
+
     m.put(nt("cf1"), nt("cq1"), nv("v1"));
     m.put(nt("cf2"), nt("cq2"), new ColumnVisibility("cv2"), nv("v2"));
     m.put(nt("cf3"), nt("cq3"), 3l, nv("v3"));
     m.put(nt("cf4"), nt("cq4"), new ColumnVisibility("cv4"), 4l, nv("v4"));
-    
+
     m.putDelete(nt("cf5"), nt("cq5"));
     m.putDelete(nt("cf6"), nt("cq6"), new ColumnVisibility("cv6"));
     m.putDelete(nt("cf7"), nt("cq7"), 7l);
     m.putDelete(nt("cf8"), nt("cq8"), new ColumnVisibility("cv8"), 8l);
-    
+
     assertEquals(8, m.size());
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(8, m.size());
     assertEquals(8, updates.size());
-    
+
     verifyColumnUpdate(updates.get(0), "cf1", "cq1", "", 0l, false, false, "v1");
     verifyColumnUpdate(updates.get(1), "cf2", "cq2", "cv2", 0l, false, false, "v2");
     verifyColumnUpdate(updates.get(2), "cf3", "cq3", "", 3l, true, false, "v3");
     verifyColumnUpdate(updates.get(3), "cf4", "cq4", "cv4", 4l, true, false, "v4");
-    
+
     verifyColumnUpdate(updates.get(4), "cf5", "cq5", "", 0l, false, true, "");
     verifyColumnUpdate(updates.get(5), "cf6", "cq6", "cv6", 0l, false, true, "");
     verifyColumnUpdate(updates.get(6), "cf7", "cq7", "", 7l, true, true, "");
     verifyColumnUpdate(updates.get(7), "cf8", "cq8", "cv8", 8l, true, true, "");
-    
+
   }
-  
+
   @Test
   public void testPutsString() {
     Mutation m = new Mutation("r1");
-    
+
     m.put("cf1", "cq1", nv("v1"));
     m.put("cf2", "cq2", new ColumnVisibility("cv2"), nv("v2"));
     m.put("cf3", "cq3", 3l, nv("v3"));
     m.put("cf4", "cq4", new ColumnVisibility("cv4"), 4l, nv("v4"));
-    
+
     m.putDelete("cf5", "cq5");
     m.putDelete("cf6", "cq6", new ColumnVisibility("cv6"));
     m.putDelete("cf7", "cq7", 7l);
     m.putDelete("cf8", "cq8", new ColumnVisibility("cv8"), 8l);
-    
+
     assertEquals(8, m.size());
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(8, m.size());
     assertEquals(8, updates.size());
-    
+
     verifyColumnUpdate(updates.get(0), "cf1", "cq1", "", 0l, false, false, "v1");
     verifyColumnUpdate(updates.get(1), "cf2", "cq2", "cv2", 0l, false, false, "v2");
     verifyColumnUpdate(updates.get(2), "cf3", "cq3", "", 3l, true, false, "v3");
     verifyColumnUpdate(updates.get(3), "cf4", "cq4", "cv4", 4l, true, false, "v4");
-    
+
     verifyColumnUpdate(updates.get(4), "cf5", "cq5", "", 0l, false, true, "");
     verifyColumnUpdate(updates.get(5), "cf6", "cq6", "cv6", 0l, false, true, "");
     verifyColumnUpdate(updates.get(6), "cf7", "cq7", "", 7l, true, true, "");
     verifyColumnUpdate(updates.get(7), "cf8", "cq8", "cv8", 8l, true, true, "");
   }
-  
+
   @Test
   public void testPutsStringString() {
     Mutation m = new Mutation("r1");
-    
+
     m.put("cf1", "cq1", "v1");
     m.put("cf2", "cq2", new ColumnVisibility("cv2"), "v2");
     m.put("cf3", "cq3", 3l, "v3");
     m.put("cf4", "cq4", new ColumnVisibility("cv4"), 4l, "v4");
-    
+
     m.putDelete("cf5", "cq5");
     m.putDelete("cf6", "cq6", new ColumnVisibility("cv6"));
     m.putDelete("cf7", "cq7", 7l);
     m.putDelete("cf8", "cq8", new ColumnVisibility("cv8"), 8l);
-    
+
     assertEquals(8, m.size());
     assertEquals("r1", new String(m.getRow()));
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(8, m.size());
     assertEquals(8, updates.size());
-    
+
     verifyColumnUpdate(updates.get(0), "cf1", "cq1", "", 0l, false, false, "v1");
     verifyColumnUpdate(updates.get(1), "cf2", "cq2", "cv2", 0l, false, false, "v2");
     verifyColumnUpdate(updates.get(2), "cf3", "cq3", "", 3l, true, false, "v3");
     verifyColumnUpdate(updates.get(3), "cf4", "cq4", "cv4", 4l, true, false, "v4");
-    
+
     verifyColumnUpdate(updates.get(4), "cf5", "cq5", "", 0l, false, true, "");
     verifyColumnUpdate(updates.get(5), "cf6", "cq6", "cv6", 0l, false, true, "");
     verifyColumnUpdate(updates.get(6), "cf7", "cq7", "", 7l, true, true, "");
     verifyColumnUpdate(updates.get(7), "cf8", "cq8", "cv8", 8l, true, true, "");
   }
-  
+
   public void testByteArrays() {
     Mutation m = new Mutation("r1".getBytes());
-    
+
     m.put("cf1".getBytes(), "cq1".getBytes(), "v1".getBytes());
     m.put("cf2".getBytes(), "cq2".getBytes(), new ColumnVisibility("cv2"), "v2".getBytes());
     m.put("cf3".getBytes(), "cq3".getBytes(), 3l, "v3".getBytes());
     m.put("cf4".getBytes(), "cq4".getBytes(), new ColumnVisibility("cv4"), 4l, "v4".getBytes());
-    
+
     m.putDelete("cf5".getBytes(), "cq5".getBytes());
     m.putDelete("cf6".getBytes(), "cq6".getBytes(), new ColumnVisibility("cv6"));
     m.putDelete("cf7".getBytes(), "cq7".getBytes(), 7l);
     m.putDelete("cf8".getBytes(), "cq8".getBytes(), new ColumnVisibility("cv8"), 8l);
-    
+
     assertEquals(8, m.size());
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(8, m.size());
     assertEquals(8, updates.size());
-    
+
     verifyColumnUpdate(updates.get(0), "cf1", "cq1", "", 0l, false, false, "v1");
     verifyColumnUpdate(updates.get(1), "cf2", "cq2", "cv2", 0l, false, false, "v2");
     verifyColumnUpdate(updates.get(2), "cf3", "cq3", "", 3l, true, false, "v3");
     verifyColumnUpdate(updates.get(3), "cf4", "cq4", "cv4", 4l, true, false, "v4");
-    
+
     verifyColumnUpdate(updates.get(4), "cf5", "cq5", "", 0l, false, true, "");
     verifyColumnUpdate(updates.get(5), "cf6", "cq6", "cv6", 0l, false, true, "");
     verifyColumnUpdate(updates.get(6), "cf7", "cq7", "", 7l, true, true, "");
@@ -327,28 +326,28 @@ public class MutationTest  {
     m1.put("cf1.3", "cq1.3", new ColumnVisibility("E|F"), new String(val1_3));
     int size1 = m1.size();
     long nb1 = m1.numBytes();
-    
+
     Mutation m2 = new Mutation("row2");
     byte[] val2 = new byte[Mutation.VALUE_SIZE_COPY_CUTOFF + 2];
     Arrays.fill(val2, (byte) 2);
     m2.put("cf2", "cq2", new ColumnVisibility("G|H"), 1234, new String(val2));
     int size2 = m2.size();
     long nb2 = m2.numBytes();
-    
+
     ByteArrayOutputStream bos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(bos);
     m1.write(dos);
     m2.write(dos);
     dos.close();
-    
+
     // Now read the mutations back in from the byte array, making sure to
     // reuse the same mutation object, and make sure everything is correct.
     ByteArrayInputStream bis = new ByteArrayInputStream(bos.toByteArray());
     DataInputStream dis = new DataInputStream(bis);
-    
+
     Mutation m = new Mutation();
     m.readFields(dis);
-    
+
     assertEquals("row1", new String(m.getRow()));
     assertEquals(size1, m.size());
     assertEquals(nb1, m.numBytes());
@@ -356,20 +355,20 @@ public class MutationTest  {
     verifyColumnUpdate(m.getUpdates().get(0), "cf1.1", "cq1.1", "A|B", 0L, false, false, "val1.1");
     verifyColumnUpdate(m.getUpdates().get(1), "cf1.2", "cq1.2", "C|D", 0L, false, false, "val1.2");
     verifyColumnUpdate(m.getUpdates().get(2), "cf1.3", "cq1.3", "E|F", 0L, false, false, new String(val1_3));
-    
+
     // Reuse the same mutation object (which is what happens in the hadoop framework
     // when objects are read by an input format)
     m.readFields(dis);
-    
+
     assertEquals("row2", new String(m.getRow()));
     assertEquals(size2, m.size());
     assertEquals(nb2, m.numBytes());
     assertEquals(1, m.getUpdates().size());
     verifyColumnUpdate(m.getUpdates().get(0), "cf2", "cq2", "G|H", 1234L, true, false, new String(val2));
   }
-  
+
   private void verifyColumnUpdate(ColumnUpdate cu, String cf, String cq, String cv, long ts, boolean timeSet, boolean deleted, String val) {
-    
+
     assertEquals(cf, new String(cu.getColumnFamily()));
     assertEquals(cq, new String(cu.getColumnQualifier()));
     assertEquals(cv, new String(cu.getColumnVisibility()));
@@ -379,38 +378,38 @@ public class MutationTest  {
     assertEquals(deleted, cu.isDeleted());
     assertEquals(val, new String(cu.getValue()));
   }
-  
+
   @Test
   public void test4() throws Exception {
     Mutation m1 = new Mutation(new Text("r1"));
-    
+
     m1.put(nt("cf1"), nt("cq1"), nv("v1"));
     m1.put(nt("cf2"), nt("cq2"), new ColumnVisibility("cv2"), nv("v2"));
-    
+
     ByteArrayOutputStream bos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(bos);
     m1.write(dos);
     dos.close();
-    
+
     Mutation m2 = new Mutation(new Text("r2"));
-    
+
     m2.put(nt("cf3"), nt("cq3"), nv("v3"));
     m2.put(nt("cf4"), nt("cq4"), new ColumnVisibility("cv2"), nv("v4"));
-    
+
     ByteArrayInputStream bis = new ByteArrayInputStream(bos.toByteArray());
     DataInputStream dis = new DataInputStream(bis);
-    
+
     // used to be a bug where puts done before readFields would be seen
     // after readFields
     m2.readFields(dis);
-    
+
     assertEquals("r1", new String(m2.getRow()));
     assertEquals(2, m2.getUpdates().size());
     assertEquals(2, m2.size());
     verifyColumnUpdate(m2.getUpdates().get(0), "cf1", "cq1", "", 0l, false, false, "v1");
     verifyColumnUpdate(m2.getUpdates().get(1), "cf2", "cq2", "cv2", 0l, false, false, "v2");
   }
-  
+
   Mutation convert(OldMutation old) throws IOException {
     ByteArrayOutputStream bos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(bos);
@@ -423,7 +422,7 @@ public class MutationTest  {
     dis.close();
     return m;
   }
-  
+
   @Test
   public void testNewSerialization() throws Exception {
     // write an old mutation
@@ -435,12 +434,12 @@ public class MutationTest  {
     DataOutputStream dos = new DataOutputStream(bos);
     m2.write(dos);
     dos.close();
-    long oldSize = dos.size(); 
+    long oldSize = dos.size();
     ByteArrayInputStream bis = new ByteArrayInputStream(bos.toByteArray());
     DataInputStream dis = new DataInputStream(bis);
     m2.readFields(dis);
     dis.close();
-    
+
     // check it
     assertEquals("r1", new String(m2.getRow()));
     assertEquals(3, m2.getUpdates().size());
@@ -450,14 +449,14 @@ public class MutationTest  {
     verifyColumnUpdate(m2.getUpdates().get(2), "cf3", "cq3", "", 0l, false, true, "");
 
     Mutation m1 = convert(m2);
-    
+
     assertEquals("r1", new String(m1.getRow()));
     assertEquals(3, m1.getUpdates().size());
     assertEquals(3, m1.size());
     verifyColumnUpdate(m1.getUpdates().get(0), "cf1", "cq1", "", 0l, false, false, "v1");
     verifyColumnUpdate(m1.getUpdates().get(1), "cf2", "cq2", "cv2", 0l, false, false, "v2");
     verifyColumnUpdate(m1.getUpdates().get(2), "cf3", "cq3", "", 0l, false, true, "");
-    
+
     Text exampleRow = new Text(" 123456789 123456789 123456789 123456789 123456789");
     int exampleLen = exampleRow.getLength();
     m1 = new Mutation(exampleRow);
@@ -472,7 +471,7 @@ public class MutationTest  {
     assertEquals(10, newSize - exampleLen);
     assertEquals(68, oldSize - exampleLen);
     // I am converting to integer to avoid comparing floats which are inaccurate
-    assertEquals(14705, (int)(((newSize-exampleLen) * 100. / (oldSize - exampleLen)) * 1000));
+    assertEquals(14705, (int) (((newSize - exampleLen) * 100. / (oldSize - exampleLen)) * 1000));
     StringBuilder sb = new StringBuilder();
     byte[] ba = bos.toByteArray();
     for (int i = 0; i < bos.size(); i += 4) {
@@ -481,10 +480,11 @@ public class MutationTest  {
       }
       sb.append(" ");
     }
-    assertEquals("80322031 32333435 36373839 20313233 34353637 38392031 32333435 36373839 20313233 34353637 38392031 32333435 36373839 06000000 00000001 ", sb.toString());
+    assertEquals("80322031 32333435 36373839 20313233 34353637 38392031 32333435 36373839 20313233 34353637 38392031 32333435 36373839 06000000 00000001 ",
+        sb.toString());
 
   }
-  
+
   @Test
   public void testReserialize() throws Exception {
     // test reading in a new mutation from an old mutation and reserializing the new mutation... this was failing
@@ -498,20 +498,19 @@ public class MutationTest  {
     }
     om.put("cf2", "big", bigVal);
 
-    
     Mutation m1 = convert(om);
 
     ByteArrayOutputStream bos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(bos);
     m1.write(dos);
     dos.close();
-    
+
     Mutation m2 = new Mutation();
-    
+
     ByteArrayInputStream bis = new ByteArrayInputStream(bos.toByteArray());
     DataInputStream dis = new DataInputStream(bis);
     m2.readFields(dis);
-    
+
     assertEquals("r1", new String(m1.getRow()));
     assertEquals(4, m2.getUpdates().size());
     assertEquals(4, m2.size());
@@ -520,11 +519,11 @@ public class MutationTest  {
     verifyColumnUpdate(m2.getUpdates().get(2), "cf3", "cq3", "", 0l, false, true, "");
     verifyColumnUpdate(m2.getUpdates().get(3), "cf2", "big", "", 0l, false, false, bigVal.toString());
   }
-  
+
   @Test
   public void testEquals() {
     Mutation m1 = new Mutation("r1");
-    
+
     m1.put("cf1", "cq1", "v1");
     m1.put("cf1", "cq1", new ColumnVisibility("A&B"), "v2");
     m1.put("cf1", "cq1", 3, "v3");
@@ -532,10 +531,10 @@ public class MutationTest  {
     m1.putDelete("cf2", "cf3");
     m1.putDelete("cf2", "cf4", 3);
     m1.putDelete("cf2", "cf4", new ColumnVisibility("A&B&C"), 3);
-    
+
     // m2 has same data as m1
     Mutation m2 = new Mutation("r1");
-    
+
     m2.put("cf1", "cq1", "v1");
     m2.put("cf1", "cq1", new ColumnVisibility("A&B"), "v2");
     m2.put("cf1", "cq1", 3, "v3");
@@ -543,10 +542,10 @@ public class MutationTest  {
     m2.putDelete("cf2", "cf3");
     m2.putDelete("cf2", "cf4", 3);
     m2.putDelete("cf2", "cf4", new ColumnVisibility("A&B&C"), 3);
-    
+
     // m3 has different row than m1
     Mutation m3 = new Mutation("r2");
-    
+
     m3.put("cf1", "cq1", "v1");
     m3.put("cf1", "cq1", new ColumnVisibility("A&B"), "v2");
     m3.put("cf1", "cq1", 3, "v3");
@@ -557,7 +556,7 @@ public class MutationTest  {
 
     // m4 has a different column than m1
     Mutation m4 = new Mutation("r1");
-    
+
     m4.put("cf2", "cq1", "v1");
     m4.put("cf1", "cq1", new ColumnVisibility("A&B"), "v2");
     m4.put("cf1", "cq1", 3, "v3");
@@ -565,10 +564,10 @@ public class MutationTest  {
     m4.putDelete("cf2", "cf3");
     m4.putDelete("cf2", "cf4", 3);
     m4.putDelete("cf2", "cf4", new ColumnVisibility("A&B&C"), 3);
-    
+
     // m5 has a different value than m1
     Mutation m5 = new Mutation("r1");
-    
+
     m5.put("cf1", "cq1", "v1");
     m5.put("cf1", "cq1", new ColumnVisibility("A&B"), "v2");
     m5.put("cf1", "cq1", 3, "v4");
@@ -599,8 +598,8 @@ public class MutationTest  {
     Mutation m2 = new Mutation(tm1);
     assertEquals(m1, m2);
   }
-  
-  @Test(expected=IllegalArgumentException.class)
+
+  @Test(expected = IllegalArgumentException.class)
   public void testThrift_Invalid() {
     Mutation m1 = new Mutation("r1");
     m1.put("cf1", "cq1", "v1");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/OldMutation.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/OldMutation.java b/core/src/test/java/org/apache/accumulo/core/data/OldMutation.java
index 5f6f993..a40f4e0 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/OldMutation.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/OldMutation.java
@@ -35,46 +35,46 @@ import org.apache.hadoop.io.Writable;
  * Will read/write old mutations.
  */
 public class OldMutation implements Writable {
-  
+
   static final int VALUE_SIZE_COPY_CUTOFF = 1 << 15;
-  
+
   private byte[] row;
   private byte[] data;
   private int entries;
   private List<byte[]> values;
-  
+
   // created this little class instead of using ByteArrayOutput stream and DataOutputStream
   // because both are synchronized... lots of small syncs slow things down
   private static class ByteBuffer {
-    
+
     int offset;
     byte data[] = new byte[64];
-    
+
     private void reserve(int l) {
       if (offset + l > data.length) {
         int newSize = data.length * 2;
         while (newSize <= offset + l)
           newSize = newSize * 2;
-        
+
         byte[] newData = new byte[newSize];
         System.arraycopy(data, 0, newData, 0, offset);
         data = newData;
       }
-      
+
     }
-    
+
     void add(byte[] b) {
       reserve(b.length);
       System.arraycopy(b, 0, data, offset, b.length);
       offset += b.length;
     }
-    
+
     public void add(byte[] bytes, int off, int length) {
       reserve(length);
       System.arraycopy(bytes, off, data, offset, length);
       offset += length;
     }
-    
+
     void add(boolean b) {
       reserve(1);
       if (b)
@@ -82,7 +82,7 @@ public class OldMutation implements Writable {
       else
         data[offset++] = 0;
     }
-    
+
     void add(long v) {
       reserve(8);
       data[offset++] = (byte) (v >>> 56);
@@ -94,7 +94,7 @@ public class OldMutation implements Writable {
       data[offset++] = (byte) (v >>> 8);
       data[offset++] = (byte) (v >>> 0);
     }
-    
+
     void add(int i) {
       reserve(4);
       data[offset++] = (byte) (i >>> 24);
@@ -102,69 +102,69 @@ public class OldMutation implements Writable {
       data[offset++] = (byte) (i >>> 8);
       data[offset++] = (byte) (i >>> 0);
     }
-    
+
     public byte[] toArray() {
       byte ret[] = new byte[offset];
       System.arraycopy(data, 0, ret, 0, offset);
       return ret;
     }
-    
+
   }
-  
+
   private static class SimpleReader {
     int offset;
     byte data[];
-    
+
     SimpleReader(byte b[]) {
       this.data = b;
     }
-    
+
     int readInt() {
       return (data[offset++] << 24) + ((data[offset++] & 255) << 16) + ((data[offset++] & 255) << 8) + ((data[offset++] & 255) << 0);
-      
+
     }
-    
+
     long readLong() {
       return (((long) data[offset++] << 56) + ((long) (data[offset++] & 255) << 48) + ((long) (data[offset++] & 255) << 40)
           + ((long) (data[offset++] & 255) << 32) + ((long) (data[offset++] & 255) << 24) + ((data[offset++] & 255) << 16) + ((data[offset++] & 255) << 8) + ((data[offset++] & 255) << 0));
     }
-    
+
     void readBytes(byte b[]) {
       System.arraycopy(data, offset, b, 0, b.length);
       offset += b.length;
     }
-    
+
     boolean readBoolean() {
       return (data[offset++] == 1);
     }
-    
+
   }
-  
+
   private ByteBuffer buffer;
-  
+
   private List<ColumnUpdate> updates;
-  
+
   private static final byte[] EMPTY_BYTES = new byte[0];
-  
+
   private void serialize() {
     if (buffer != null) {
       data = buffer.toArray();
       buffer = null;
     }
   }
-  
+
   public OldMutation(Text row) {
     this.row = new byte[row.getLength()];
     System.arraycopy(row.getBytes(), 0, this.row, 0, row.getLength());
     buffer = new ByteBuffer();
   }
-  
+
   public OldMutation(CharSequence row) {
     this(new Text(row.toString()));
   }
-  
+
   public OldMutation() {}
-  
+
   public OldMutation(TMutation tmutation) {
     this.row = ByteBufferUtil.toBytes(tmutation.row);
     this.data = ByteBufferUtil.toBytes(tmutation.data);
@@ -178,45 +178,45 @@ public class OldMutation implements Writable {
       throw new IllegalArgumentException("null serialized data");
     }
   }
-  
+
   public byte[] getRow() {
     return row;
   }
-  
+
   private void put(byte b[]) {
     buffer.add(b.length);
     buffer.add(b);
   }
-  
+
   private void put(Text t) {
     buffer.add(t.getLength());
     buffer.add(t.getBytes(), 0, t.getLength());
   }
-  
+
   private void put(boolean b) {
     buffer.add(b);
   }
-  
+
   private void put(int i) {
     buffer.add(i);
   }
-  
+
   private void put(long l) {
     buffer.add(l);
   }
-  
+
   private void put(Text cf, Text cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
-    
+
     if (buffer == null)
       throw new IllegalStateException("Can not add to mutation after serializing it");
-    
+
     put(cf);
     put(cq);
     put(cv);
     put(hasts);
     put(ts);
     put(deleted);
-    
+
     if (val.length < VALUE_SIZE_COPY_CUTOFF) {
       put(val);
     } else {
@@ -227,129 +227,129 @@ public class OldMutation implements Writable {
       values.add(copy);
       put(-1 * values.size());
     }
-    
+
     entries++;
   }
-  
+
   private void put(CharSequence cf, CharSequence cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
     put(new Text(cf.toString()), new Text(cq.toString()), cv, hasts, ts, deleted, val);
   }
-  
+
   private void put(CharSequence cf, CharSequence cq, byte[] cv, boolean hasts, long ts, boolean deleted, CharSequence val) {
     put(cf, cq, cv, hasts, ts, deleted, TextUtil.getBytes(new Text(val.toString())));
   }
-  
+
   public void put(Text columnFamily, Text columnQualifier, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value.get());
   }
-  
+
   public void put(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value.get());
   }
-  
+
   public void put(Text columnFamily, Text columnQualifier, long timestamp, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value.get());
   }
-  
+
   public void put(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility, long timestamp, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value.get());
   }
-  
+
   public void putDelete(Text columnFamily, Text columnQualifier) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, true, EMPTY_BYTES);
   }
-  
+
   public void putDelete(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, true, EMPTY_BYTES);
   }
-  
+
   public void putDelete(Text columnFamily, Text columnQualifier, long timestamp) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   public void putDelete(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility, long timestamp) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value.get());
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value.get());
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, long timestamp, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value.get());
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, long timestamp, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value.get());
   }
-  
+
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, true, EMPTY_BYTES);
   }
-  
+
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, true, EMPTY_BYTES);
   }
-  
+
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier, long timestamp) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, long timestamp) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, CharSequence value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value);
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, CharSequence value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value);
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, long timestamp, CharSequence value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value);
   }
-  
+
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, long timestamp, CharSequence value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value);
   }
-  
+
   private byte[] readBytes(SimpleReader in) {
     int len = in.readInt();
     if (len == 0)
       return EMPTY_BYTES;
-    
+
     byte bytes[] = new byte[len];
     in.readBytes(bytes);
     return bytes;
   }
-  
+
   public List<ColumnUpdate> getUpdates() {
     serialize();
-    
+
     SimpleReader in = new SimpleReader(data);
-    
+
     if (updates == null) {
       if (entries == 1) {
         updates = Collections.singletonList(deserializeColumnUpdate(in));
       } else {
         ColumnUpdate[] tmpUpdates = new ColumnUpdate[entries];
-        
+
         for (int i = 0; i < entries; i++)
           tmpUpdates[i] = deserializeColumnUpdate(in);
-        
+
         updates = Arrays.asList(tmpUpdates);
       }
     }
-    
+
     return updates;
   }
-  
+
   private ColumnUpdate deserializeColumnUpdate(SimpleReader in) {
     byte[] cf = readBytes(in);
     byte[] cq = readBytes(in);
@@ -357,10 +357,10 @@ public class OldMutation implements Writable {
     boolean hasts = in.readBoolean();
     long ts = in.readLong();
     boolean deleted = in.readBoolean();
-    
+
     byte[] val;
     int valLen = in.readInt();
-    
+
     if (valLen < 0) {
       val = values.get((-1 * valLen) - 1);
     } else if (valLen == 0) {
@@ -369,44 +369,44 @@ public class OldMutation implements Writable {
       val = new byte[valLen];
       in.readBytes(val);
     }
-    
+
     return new ColumnUpdate(cf, cq, cv, hasts, ts, deleted, val);
   }
-  
+
   private int cachedValLens = -1;
-  
+
   long getValueLengths() {
     if (values == null)
       return 0;
-    
+
     if (cachedValLens == -1) {
       int tmpCVL = 0;
       for (byte[] val : values)
         tmpCVL += val.length;
-      
+
       cachedValLens = tmpCVL;
     }
-    
+
     return cachedValLens;
-    
+
   }
-  
+
   public long numBytes() {
     serialize();
     return row.length + data.length + getValueLengths();
   }
-  
+
   public long estimatedMemoryUsed() {
     return numBytes() + 230;
   }
-  
+
   /**
    * @return the number of column value pairs added to the mutation
    */
   public int size() {
     return entries;
   }
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     // Clear out cached column updates and value lengths so
@@ -415,7 +415,7 @@ public class OldMutation implements Writable {
     updates = null;
     cachedValLens = -1;
     buffer = null;
-    
+
     int len = in.readInt();
     row = new byte[len];
     in.readFully(row);
@@ -423,7 +423,7 @@ public class OldMutation implements Writable {
     data = new byte[len];
     in.readFully(data);
     entries = in.readInt();
-    
+
     boolean valuesPresent = in.readBoolean();
     if (!valuesPresent) {
       values = null;
@@ -438,7 +438,7 @@ public class OldMutation implements Writable {
       }
     }
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     serialize();
@@ -447,7 +447,7 @@ public class OldMutation implements Writable {
     out.writeInt(data.length);
     out.write(data);
     out.writeInt(entries);
-    
+
     if (values == null)
       out.writeBoolean(false);
     else {
@@ -459,21 +459,21 @@ public class OldMutation implements Writable {
         out.write(val);
       }
     }
-    
+
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof OldMutation)
       return equals((OldMutation) o);
     return false;
   }
-  
+
   @Override
   public int hashCode() {
     return toThrift().hashCode();
   }
-  
+
   public boolean equals(OldMutation m) {
     serialize();
     if (!Arrays.equals(row, m.getRow()))
@@ -490,10 +490,10 @@ public class OldMutation implements Writable {
     }
     return false;
   }
-  
+
   public TMutation toThrift() {
     serialize();
     return new TMutation(java.nio.ByteBuffer.wrap(row), java.nio.ByteBuffer.wrap(data), ByteBufferUtil.toByteBuffers(values), entries);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/ValueTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/ValueTest.java b/core/src/test/java/org/apache/accumulo/core/data/ValueTest.java
index 0aa7649..9a9e21d 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/ValueTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/ValueTest.java
@@ -18,6 +18,15 @@
 package org.apache.accumulo.core.data;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.junit.Assert.assertArrayEquals;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNotSame;
+import static org.junit.Assert.assertSame;
+import static org.junit.Assert.assertTrue;
 
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
@@ -29,9 +38,6 @@ import java.util.List;
 import org.junit.Before;
 import org.junit.Test;
 
-import static org.easymock.EasyMock.*;
-import static org.junit.Assert.*;
-
 public class ValueTest {
   private static final byte[] toBytes(String s) {
     return s.getBytes(UTF_8);
@@ -54,22 +60,22 @@ public class ValueTest {
     assertEquals(0, v.get().length);
   }
 
-  @Test(expected=NullPointerException.class)
+  @Test(expected = NullPointerException.class)
   public void testNullBytesConstructor() {
     new Value((byte[]) null);
   }
 
-  @Test(expected=NullPointerException.class)
+  @Test(expected = NullPointerException.class)
   public void testNullCopyConstructor() {
     new Value((Value) null);
   }
 
-  @Test(expected=NullPointerException.class)
+  @Test(expected = NullPointerException.class)
   public void testNullByteBufferConstructor() {
-    new Value((ByteBuffer)null);
+    new Value((ByteBuffer) null);
   }
 
-  @Test(expected=NullPointerException.class)
+  @Test(expected = NullPointerException.class)
   public void testNullSet() {
     Value v = new Value();
     v.set(null);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/FileOperationsTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/FileOperationsTest.java b/core/src/test/java/org/apache/accumulo/core/file/FileOperationsTest.java
index 30c667d..3fdeb8a 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/FileOperationsTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/FileOperationsTest.java
@@ -30,15 +30,15 @@ import org.apache.hadoop.fs.FileSystem;
 import org.junit.Test;
 
 public class FileOperationsTest {
-  
+
   public FileOperationsTest() {}
-  
+
   /**
    * Test for filenames with +1 dot
    */
   @Test
   public void handlesFilenamesWithMoreThanOneDot() throws IOException {
-    
+
     Boolean caughtException = false;
     FileSKVWriter writer = null;
     String filename = "target/test.file." + RFile.EXTENSION;
@@ -61,7 +61,7 @@ public class FileOperationsTest {
       }
       FileUtils.forceDelete(testFile);
     }
-    
+
     assertFalse("Should not throw with more than 1 dot in filename.", caughtException);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestCachedBlockQueue.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestCachedBlockQueue.java b/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestCachedBlockQueue.java
index 19f279b..940fcf8 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestCachedBlockQueue.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestCachedBlockQueue.java
@@ -24,9 +24,9 @@ import java.util.LinkedList;
 import junit.framework.TestCase;
 
 public class TestCachedBlockQueue extends TestCase {
-  
+
   public void testQueue() throws Exception {
-    
+
     CachedBlock cb1 = new CachedBlock(1000, "cb1", 1);
     CachedBlock cb2 = new CachedBlock(1500, "cb2", 2);
     CachedBlock cb3 = new CachedBlock(1000, "cb3", 3);
@@ -37,9 +37,9 @@ public class TestCachedBlockQueue extends TestCase {
     CachedBlock cb8 = new CachedBlock(1500, "cb8", 8);
     CachedBlock cb9 = new CachedBlock(1000, "cb9", 9);
     CachedBlock cb10 = new CachedBlock(1500, "cb10", 10);
-    
+
     CachedBlockQueue queue = new CachedBlockQueue(10000, 1000);
-    
+
     queue.add(cb1);
     queue.add(cb2);
     queue.add(cb3);
@@ -50,12 +50,12 @@ public class TestCachedBlockQueue extends TestCase {
     queue.add(cb8);
     queue.add(cb9);
     queue.add(cb10);
-    
+
     // We expect cb1 through cb8 to be in the queue
     long expectedSize = cb1.heapSize() + cb2.heapSize() + cb3.heapSize() + cb4.heapSize() + cb5.heapSize() + cb6.heapSize() + cb7.heapSize() + cb8.heapSize();
-    
+
     assertEquals(queue.heapSize(), expectedSize);
-    
+
     LinkedList<org.apache.accumulo.core.file.blockfile.cache.CachedBlock> blocks = queue.getList();
     assertEquals(blocks.poll().getName(), "cb1");
     assertEquals(blocks.poll().getName(), "cb2");
@@ -65,11 +65,11 @@ public class TestCachedBlockQueue extends TestCase {
     assertEquals(blocks.poll().getName(), "cb6");
     assertEquals(blocks.poll().getName(), "cb7");
     assertEquals(blocks.poll().getName(), "cb8");
-    
+
   }
-  
+
   public void testQueueSmallBlockEdgeCase() throws Exception {
-    
+
     CachedBlock cb1 = new CachedBlock(1000, "cb1", 1);
     CachedBlock cb2 = new CachedBlock(1500, "cb2", 2);
     CachedBlock cb3 = new CachedBlock(1000, "cb3", 3);
@@ -80,9 +80,9 @@ public class TestCachedBlockQueue extends TestCase {
     CachedBlock cb8 = new CachedBlock(1500, "cb8", 8);
     CachedBlock cb9 = new CachedBlock(1000, "cb9", 9);
     CachedBlock cb10 = new CachedBlock(1500, "cb10", 10);
-    
+
     CachedBlockQueue queue = new CachedBlockQueue(10000, 1000);
-    
+
     queue.add(cb1);
     queue.add(cb2);
     queue.add(cb3);
@@ -93,20 +93,20 @@ public class TestCachedBlockQueue extends TestCase {
     queue.add(cb8);
     queue.add(cb9);
     queue.add(cb10);
-    
+
     CachedBlock cb0 = new CachedBlock(10 + CachedBlock.PER_BLOCK_OVERHEAD, "cb0", 0);
     queue.add(cb0);
-    
+
     // This is older so we must include it, but it will not end up kicking
     // anything out because (heapSize - cb8.heapSize + cb0.heapSize < maxSize)
     // and we must always maintain heapSize >= maxSize once we achieve it.
-    
+
     // We expect cb0 through cb8 to be in the queue
     long expectedSize = cb1.heapSize() + cb2.heapSize() + cb3.heapSize() + cb4.heapSize() + cb5.heapSize() + cb6.heapSize() + cb7.heapSize() + cb8.heapSize()
         + cb0.heapSize();
-    
+
     assertEquals(queue.heapSize(), expectedSize);
-    
+
     LinkedList<org.apache.accumulo.core.file.blockfile.cache.CachedBlock> blocks = queue.getList();
     assertEquals(blocks.poll().getName(), "cb0");
     assertEquals(blocks.poll().getName(), "cb1");
@@ -117,9 +117,9 @@ public class TestCachedBlockQueue extends TestCase {
     assertEquals(blocks.poll().getName(), "cb6");
     assertEquals(blocks.poll().getName(), "cb7");
     assertEquals(blocks.poll().getName(), "cb8");
-    
+
   }
-  
+
   private static class CachedBlock extends org.apache.accumulo.core.file.blockfile.cache.CachedBlock {
     public CachedBlock(long heapSize, String name, long accessTime) {
       super(name, new byte[(int) (heapSize - CachedBlock.PER_BLOCK_OVERHEAD)], accessTime, false);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestLruBlockCache.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestLruBlockCache.java b/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestLruBlockCache.java
index 8d6a61a..88d61ac 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestLruBlockCache.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/blockfile/cache/TestLruBlockCache.java
@@ -27,26 +27,26 @@ import junit.framework.TestCase;
 /**
  * Tests the concurrent LruBlockCache.
  * <p>
- * 
+ *
  * Tests will ensure it grows and shrinks in size properly, evictions run when they're supposed to and do what they should, and that cached blocks are
  * accessible when expected to be.
  */
 public class TestLruBlockCache extends TestCase {
-  
+
   public void testBackgroundEvictionThread() throws Exception {
-    
+
     long maxSize = 100000;
     long blockSize = calculateBlockSizeDefault(maxSize, 9); // room for 9, will evict
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize);
-    
+
     Block[] blocks = generateFixedBlocks(10, blockSize, "block");
-    
+
     // Add all the blocks
     for (Block block : blocks) {
       cache.cacheBlock(block.blockName, block.buf);
     }
-    
+
     // Let the eviction run
     int n = 0;
     while (cache.getEvictionCount() == 0) {
@@ -56,85 +56,85 @@ public class TestLruBlockCache extends TestCase {
     // A single eviction run should have occurred
     assertEquals(cache.getEvictionCount(), 1);
   }
-  
+
   public void testCacheSimple() throws Exception {
-    
+
     long maxSize = 1000000;
     long blockSize = calculateBlockSizeDefault(maxSize, 101);
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize);
-    
+
     Block[] blocks = generateRandomBlocks(100, blockSize);
-    
+
     long expectedCacheSize = cache.heapSize();
-    
+
     // Confirm empty
     for (Block block : blocks) {
       assertTrue(cache.getBlock(block.blockName) == null);
     }
-    
+
     // Add blocks
     for (Block block : blocks) {
       cache.cacheBlock(block.blockName, block.buf);
       expectedCacheSize += block.heapSize();
     }
-    
+
     // Verify correctly calculated cache heap size
     assertEquals(expectedCacheSize, cache.heapSize());
-    
+
     // Check if all blocks are properly cached and retrieved
     for (Block block : blocks) {
       CacheEntry ce = cache.getBlock(block.blockName);
       assertTrue(ce != null);
       assertEquals(ce.getBuffer().length, block.buf.length);
     }
-    
+
     // Verify correctly calculated cache heap size
     assertEquals(expectedCacheSize, cache.heapSize());
-    
+
     // Check if all blocks are properly cached and retrieved
     for (Block block : blocks) {
       CacheEntry ce = cache.getBlock(block.blockName);
       assertTrue(ce != null);
       assertEquals(ce.getBuffer().length, block.buf.length);
     }
-    
+
     // Expect no evictions
     assertEquals(0, cache.getEvictionCount());
     // Thread t = new LruBlockCache.StatisticsThread(cache);
     // t.start();
     // t.join();
   }
-  
+
   public void testCacheEvictionSimple() throws Exception {
-    
+
     long maxSize = 100000;
     long blockSize = calculateBlockSizeDefault(maxSize, 10);
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize, false);
-    
+
     Block[] blocks = generateFixedBlocks(10, blockSize, "block");
-    
+
     long expectedCacheSize = cache.heapSize();
-    
+
     // Add all the blocks
     for (Block block : blocks) {
       cache.cacheBlock(block.blockName, block.buf);
       expectedCacheSize += block.heapSize();
     }
-    
+
     // A single eviction run should have occurred
     assertEquals(1, cache.getEvictionCount());
-    
+
     // Our expected size overruns acceptable limit
     assertTrue(expectedCacheSize > (maxSize * LruBlockCache.DEFAULT_ACCEPTABLE_FACTOR));
-    
+
     // But the cache did not grow beyond max
     assertTrue(cache.heapSize() < maxSize);
-    
+
     // And is still below the acceptable limit
     assertTrue(cache.heapSize() < (maxSize * LruBlockCache.DEFAULT_ACCEPTABLE_FACTOR));
-    
+
     // All blocks except block 0 and 1 should be in the cache
     assertTrue(cache.getBlock(blocks[0].blockName) == null);
     assertTrue(cache.getBlock(blocks[1].blockName) == null);
@@ -142,292 +142,292 @@ public class TestLruBlockCache extends TestCase {
       assertTrue(Arrays.equals(cache.getBlock(blocks[i].blockName).getBuffer(), blocks[i].buf));
     }
   }
-  
+
   public void testCacheEvictionTwoPriorities() throws Exception {
-    
+
     long maxSize = 100000;
     long blockSize = calculateBlockSizeDefault(maxSize, 10);
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize, false, (int) Math.ceil(1.2 * maxSize / blockSize), LruBlockCache.DEFAULT_LOAD_FACTOR,
         LruBlockCache.DEFAULT_CONCURRENCY_LEVEL, 0.98f, // min
         0.99f, // acceptable
         0.25f, // single
         0.50f, // multi
         0.25f);// memory
-    
+
     Block[] singleBlocks = generateFixedBlocks(5, 10000, "single");
     Block[] multiBlocks = generateFixedBlocks(5, 10000, "multi");
-    
+
     long expectedCacheSize = cache.heapSize();
-    
+
     // Add and get the multi blocks
     for (Block block : multiBlocks) {
       cache.cacheBlock(block.blockName, block.buf);
       expectedCacheSize += block.heapSize();
       assertTrue(Arrays.equals(cache.getBlock(block.blockName).getBuffer(), block.buf));
     }
-    
+
     // Add the single blocks (no get)
     for (Block block : singleBlocks) {
       cache.cacheBlock(block.blockName, block.buf);
       expectedCacheSize += block.heapSize();
     }
-    
+
     // A single eviction run should have occurred
     assertEquals(cache.getEvictionCount(), 1);
-    
+
     // We expect two entries evicted
     assertEquals(cache.getEvictedCount(), 2);
-    
+
     // Our expected size overruns acceptable limit
     assertTrue(expectedCacheSize > (maxSize * LruBlockCache.DEFAULT_ACCEPTABLE_FACTOR));
-    
+
     // But the cache did not grow beyond max
     assertTrue(cache.heapSize() <= maxSize);
-    
+
     // And is now below the acceptable limit
     assertTrue(cache.heapSize() <= (maxSize * LruBlockCache.DEFAULT_ACCEPTABLE_FACTOR));
-    
+
     // We expect fairness across the two priorities.
     // This test makes multi go barely over its limit, in-memory
     // empty, and the rest in single. Two single evictions and
     // one multi eviction expected.
     assertTrue(cache.getBlock(singleBlocks[0].blockName) == null);
     assertTrue(cache.getBlock(multiBlocks[0].blockName) == null);
-    
+
     // And all others to be cached
     for (int i = 1; i < 4; i++) {
       assertTrue(Arrays.equals(cache.getBlock(singleBlocks[i].blockName).getBuffer(), singleBlocks[i].buf));
       assertTrue(Arrays.equals(cache.getBlock(multiBlocks[i].blockName).getBuffer(), multiBlocks[i].buf));
     }
   }
-  
+
   public void testCacheEvictionThreePriorities() throws Exception {
-    
+
     long maxSize = 100000;
     long blockSize = calculateBlockSize(maxSize, 10);
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize, false, (int) Math.ceil(1.2 * maxSize / blockSize), LruBlockCache.DEFAULT_LOAD_FACTOR,
         LruBlockCache.DEFAULT_CONCURRENCY_LEVEL, 0.98f, // min
         0.99f, // acceptable
         0.33f, // single
         0.33f, // multi
         0.34f);// memory
-    
+
     Block[] singleBlocks = generateFixedBlocks(5, blockSize, "single");
     Block[] multiBlocks = generateFixedBlocks(5, blockSize, "multi");
     Block[] memoryBlocks = generateFixedBlocks(5, blockSize, "memory");
-    
+
     long expectedCacheSize = cache.heapSize();
-    
+
     // Add 3 blocks from each priority
     for (int i = 0; i < 3; i++) {
-      
+
       // Just add single blocks
       cache.cacheBlock(singleBlocks[i].blockName, singleBlocks[i].buf);
       expectedCacheSize += singleBlocks[i].heapSize();
-      
+
       // Add and get multi blocks
       cache.cacheBlock(multiBlocks[i].blockName, multiBlocks[i].buf);
       expectedCacheSize += multiBlocks[i].heapSize();
       cache.getBlock(multiBlocks[i].blockName);
-      
+
       // Add memory blocks as such
       cache.cacheBlock(memoryBlocks[i].blockName, memoryBlocks[i].buf, true);
       expectedCacheSize += memoryBlocks[i].heapSize();
-      
+
     }
-    
+
     // Do not expect any evictions yet
     assertEquals(0, cache.getEvictionCount());
-    
+
     // Verify cache size
     assertEquals(expectedCacheSize, cache.heapSize());
-    
+
     // Insert a single block, oldest single should be evicted
     cache.cacheBlock(singleBlocks[3].blockName, singleBlocks[3].buf);
-    
+
     // Single eviction, one thing evicted
     assertEquals(1, cache.getEvictionCount());
     assertEquals(1, cache.getEvictedCount());
-    
+
     // Verify oldest single block is the one evicted
     assertEquals(null, cache.getBlock(singleBlocks[0].blockName));
-    
+
     // Change the oldest remaining single block to a multi
     cache.getBlock(singleBlocks[1].blockName);
-    
+
     // Insert another single block
     cache.cacheBlock(singleBlocks[4].blockName, singleBlocks[4].buf);
-    
+
     // Two evictions, two evicted.
     assertEquals(2, cache.getEvictionCount());
     assertEquals(2, cache.getEvictedCount());
-    
+
     // Oldest multi block should be evicted now
     assertEquals(null, cache.getBlock(multiBlocks[0].blockName));
-    
+
     // Insert another memory block
     cache.cacheBlock(memoryBlocks[3].blockName, memoryBlocks[3].buf, true);
-    
+
     // Three evictions, three evicted.
     assertEquals(3, cache.getEvictionCount());
     assertEquals(3, cache.getEvictedCount());
-    
+
     // Oldest memory block should be evicted now
     assertEquals(null, cache.getBlock(memoryBlocks[0].blockName));
-    
+
     // Add a block that is twice as big (should force two evictions)
     Block[] bigBlocks = generateFixedBlocks(3, blockSize * 3, "big");
     cache.cacheBlock(bigBlocks[0].blockName, bigBlocks[0].buf);
-    
+
     // Four evictions, six evicted (inserted block 3X size, expect +3 evicted)
     assertEquals(4, cache.getEvictionCount());
     assertEquals(6, cache.getEvictedCount());
-    
+
     // Expect three remaining singles to be evicted
     assertEquals(null, cache.getBlock(singleBlocks[2].blockName));
     assertEquals(null, cache.getBlock(singleBlocks[3].blockName));
     assertEquals(null, cache.getBlock(singleBlocks[4].blockName));
-    
+
     // Make the big block a multi block
     cache.getBlock(bigBlocks[0].blockName);
-    
+
     // Cache another single big block
     cache.cacheBlock(bigBlocks[1].blockName, bigBlocks[1].buf);
-    
+
     // Five evictions, nine evicted (3 new)
     assertEquals(5, cache.getEvictionCount());
     assertEquals(9, cache.getEvictedCount());
-    
+
     // Expect three remaining multis to be evicted
     assertEquals(null, cache.getBlock(singleBlocks[1].blockName));
     assertEquals(null, cache.getBlock(multiBlocks[1].blockName));
     assertEquals(null, cache.getBlock(multiBlocks[2].blockName));
-    
+
     // Cache a big memory block
     cache.cacheBlock(bigBlocks[2].blockName, bigBlocks[2].buf, true);
-    
+
     // Six evictions, twelve evicted (3 new)
     assertEquals(6, cache.getEvictionCount());
     assertEquals(12, cache.getEvictedCount());
-    
+
     // Expect three remaining in-memory to be evicted
     assertEquals(null, cache.getBlock(memoryBlocks[1].blockName));
     assertEquals(null, cache.getBlock(memoryBlocks[2].blockName));
     assertEquals(null, cache.getBlock(memoryBlocks[3].blockName));
-    
+
   }
-  
+
   // test scan resistance
   public void testScanResistance() throws Exception {
-    
+
     long maxSize = 100000;
     long blockSize = calculateBlockSize(maxSize, 10);
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize, false, (int) Math.ceil(1.2 * maxSize / blockSize), LruBlockCache.DEFAULT_LOAD_FACTOR,
         LruBlockCache.DEFAULT_CONCURRENCY_LEVEL, 0.66f, // min
         0.99f, // acceptable
         0.33f, // single
         0.33f, // multi
         0.34f);// memory
-    
+
     Block[] singleBlocks = generateFixedBlocks(20, blockSize, "single");
     Block[] multiBlocks = generateFixedBlocks(5, blockSize, "multi");
-    
+
     // Add 5 multi blocks
     for (Block block : multiBlocks) {
       cache.cacheBlock(block.blockName, block.buf);
       cache.getBlock(block.blockName);
     }
-    
+
     // Add 5 single blocks
     for (int i = 0; i < 5; i++) {
       cache.cacheBlock(singleBlocks[i].blockName, singleBlocks[i].buf);
     }
-    
+
     // An eviction ran
     assertEquals(1, cache.getEvictionCount());
-    
+
     // To drop down to 2/3 capacity, we'll need to evict 4 blocks
     assertEquals(4, cache.getEvictedCount());
-    
+
     // Should have been taken off equally from single and multi
     assertEquals(null, cache.getBlock(singleBlocks[0].blockName));
     assertEquals(null, cache.getBlock(singleBlocks[1].blockName));
     assertEquals(null, cache.getBlock(multiBlocks[0].blockName));
     assertEquals(null, cache.getBlock(multiBlocks[1].blockName));
-    
+
     // Let's keep "scanning" by adding single blocks. From here on we only
     // expect evictions from the single bucket.
-    
+
     // Every time we reach 10 total blocks (every 4 inserts) we get 4 single
     // blocks evicted. Inserting 13 blocks should yield 3 more evictions and
     // 12 more evicted.
-    
+
     for (int i = 5; i < 18; i++) {
       cache.cacheBlock(singleBlocks[i].blockName, singleBlocks[i].buf);
     }
-    
+
     // 4 total evictions, 16 total evicted
     assertEquals(4, cache.getEvictionCount());
     assertEquals(16, cache.getEvictedCount());
-    
+
     // Should now have 7 total blocks
     assertEquals(7, cache.size());
-    
+
   }
-  
+
   // test setMaxSize
   public void testResizeBlockCache() throws Exception {
-    
+
     long maxSize = 300000;
     long blockSize = calculateBlockSize(maxSize, 31);
-    
+
     LruBlockCache cache = new LruBlockCache(maxSize, blockSize, false, (int) Math.ceil(1.2 * maxSize / blockSize), LruBlockCache.DEFAULT_LOAD_FACTOR,
         LruBlockCache.DEFAULT_CONCURRENCY_LEVEL, 0.98f, // min
         0.99f, // acceptable
         0.33f, // single
         0.33f, // multi
         0.34f);// memory
-    
+
     Block[] singleBlocks = generateFixedBlocks(10, blockSize, "single");
     Block[] multiBlocks = generateFixedBlocks(10, blockSize, "multi");
     Block[] memoryBlocks = generateFixedBlocks(10, blockSize, "memory");
-    
+
     // Add all blocks from all priorities
     for (int i = 0; i < 10; i++) {
-      
+
       // Just add single blocks
       cache.cacheBlock(singleBlocks[i].blockName, singleBlocks[i].buf);
-      
+
       // Add and get multi blocks
       cache.cacheBlock(multiBlocks[i].blockName, multiBlocks[i].buf);
       cache.getBlock(multiBlocks[i].blockName);
-      
+
       // Add memory blocks as such
       cache.cacheBlock(memoryBlocks[i].blockName, memoryBlocks[i].buf, true);
     }
-    
+
     // Do not expect any evictions yet
     assertEquals(0, cache.getEvictionCount());
-    
+
     // Resize to half capacity plus an extra block (otherwise we evict an extra)
     cache.setMaxSize((long) (maxSize * 0.5f));
-    
+
     // Should have run a single eviction
     assertEquals(1, cache.getEvictionCount());
-    
+
     // And we expect 1/2 of the blocks to be evicted
     assertEquals(15, cache.getEvictedCount());
-    
+
     // And the oldest 5 blocks from each category should be gone
     for (int i = 0; i < 5; i++) {
       assertEquals(null, cache.getBlock(singleBlocks[i].blockName));
       assertEquals(null, cache.getBlock(multiBlocks[i].blockName));
       assertEquals(null, cache.getBlock(memoryBlocks[i].blockName));
     }
-    
+
     // And the newest 5 blocks should still be accessible
     for (int i = 5; i < 10; i++) {
       assertTrue(Arrays.equals(singleBlocks[i].buf, cache.getBlock(singleBlocks[i].blockName).getBuffer()));
@@ -435,7 +435,7 @@ public class TestLruBlockCache extends TestCase {
       assertTrue(Arrays.equals(memoryBlocks[i].buf, cache.getBlock(memoryBlocks[i].blockName).getBuffer()));
     }
   }
-  
+
   private Block[] generateFixedBlocks(int numBlocks, int size, String pfx) {
     Block[] blocks = new Block[numBlocks];
     for (int i = 0; i < numBlocks; i++) {
@@ -443,11 +443,11 @@ public class TestLruBlockCache extends TestCase {
     }
     return blocks;
   }
-  
+
   private Block[] generateFixedBlocks(int numBlocks, long size, String pfx) {
     return generateFixedBlocks(numBlocks, (int) size, pfx);
   }
-  
+
   private Block[] generateRandomBlocks(int numBlocks, long maxSize) {
     Block[] blocks = new Block[numBlocks];
     Random r = new Random();
@@ -456,7 +456,7 @@ public class TestLruBlockCache extends TestCase {
     }
     return blocks;
   }
-  
+
   private long calculateBlockSize(long maxSize, int numBlocks) {
     long roughBlockSize = maxSize / numBlocks;
     int numEntries = (int) Math.ceil((1.2) * maxSize / roughBlockSize);
@@ -466,7 +466,7 @@ public class TestLruBlockCache extends TestCase {
     negateBlockSize += CachedBlock.PER_BLOCK_OVERHEAD;
     return ClassSize.align((long) Math.floor((roughBlockSize - negateBlockSize) * 0.99f));
   }
-  
+
   private long calculateBlockSizeDefault(long maxSize, int numBlocks) {
     long roughBlockSize = maxSize / numBlocks;
     int numEntries = (int) Math.ceil((1.2) * maxSize / roughBlockSize);
@@ -476,16 +476,16 @@ public class TestLruBlockCache extends TestCase {
     negateBlockSize += CachedBlock.PER_BLOCK_OVERHEAD;
     return ClassSize.align((long) Math.floor((roughBlockSize - negateBlockSize) * LruBlockCache.DEFAULT_ACCEPTABLE_FACTOR));
   }
-  
+
   private static class Block implements HeapSize {
     String blockName;
     byte buf[];
-    
+
     Block(String blockName, int size) {
       this.blockName = blockName;
       this.buf = new byte[size];
     }
-    
+
     public long heapSize() {
       return CachedBlock.PER_BLOCK_OVERHEAD + ClassSize.align(blockName.length()) + ClassSize.align(buf.length);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/rfile/BlockIndexTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/rfile/BlockIndexTest.java b/core/src/test/java/org/apache/accumulo/core/file/rfile/BlockIndexTest.java
index f293718..1b2b2a6 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/rfile/BlockIndexTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/rfile/BlockIndexTest.java
@@ -31,28 +31,28 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class BlockIndexTest {
-  
+
   private static class MyCacheEntry implements CacheEntry {
     Object idx;
     byte[] data;
-    
+
     MyCacheEntry(byte[] d) {
       this.data = d;
     }
-    
+
     @Override
     public void setIndex(Object idx) {
       this.idx = idx;
     }
-    
+
     @Override
     public Object getIndex() {
       return idx;
     }
-    
+
     @Override
     public byte[] getBuffer() {
       return data;
@@ -63,43 +63,42 @@ public class BlockIndexTest {
   public void test1() throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream out = new DataOutputStream(baos);
-    
+
     Key prevKey = null;
-    
+
     int num = 1000;
-    
+
     for (int i = 0; i < num; i++) {
       Key key = new Key(RFileTest.nf("", i), "cf1", "cq1");
       new RelativeKey(prevKey, key).write(out);
       new Value(new byte[0]).write(out);
       prevKey = key;
     }
-    
+
     out.close();
     final byte[] data = baos.toByteArray();
-    
+
     CacheEntry ce = new MyCacheEntry(data);
 
     ABlockReader cacheBlock = new CachableBlockFile.CachedBlockRead(ce, data);
     BlockIndex blockIndex = null;
-    
+
     for (int i = 0; i < 129; i++)
       blockIndex = BlockIndex.getIndex(cacheBlock, new IndexEntry(prevKey, num, 0, 0, 0));
-    
+
     BlockIndexEntry[] indexEntries = blockIndex.getIndexEntries();
-    
+
     for (int i = 0; i < indexEntries.length; i++) {
       int row = Integer.parseInt(indexEntries[i].getPrevKey().getRowData().toString());
-      
+
       BlockIndexEntry bie;
-      
 
       bie = blockIndex.seekBlock(new Key(RFileTest.nf("", row), "cf1", "cq1"), cacheBlock);
       if (i == 0)
         Assert.assertSame(null, bie);
       else
         Assert.assertSame(indexEntries[i - 1], bie);
-      
+
       Assert.assertSame(bie, blockIndex.seekBlock(new Key(RFileTest.nf("", row - 1), "cf1", "cq1"), cacheBlock));
 
       bie = blockIndex.seekBlock(new Key(RFileTest.nf("", row + 1), "cf1", "cq1"), cacheBlock);
@@ -108,7 +107,7 @@ public class BlockIndexTest {
       RelativeKey rk = new RelativeKey();
       rk.setPrevKey(bie.getPrevKey());
       rk.readFields(cacheBlock);
-      
+
       Assert.assertEquals(rk.getKey(), new Key(RFileTest.nf("", row + 1), "cf1", "cq1"));
 
     }
@@ -119,56 +118,56 @@ public class BlockIndexTest {
   public void testSame() throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream out = new DataOutputStream(baos);
-    
+
     Key prevKey = null;
-    
+
     int num = 1000;
-    
+
     for (int i = 0; i < num; i++) {
       Key key = new Key(RFileTest.nf("", 1), "cf1", "cq1");
       new RelativeKey(prevKey, key).write(out);
       new Value(new byte[0]).write(out);
       prevKey = key;
     }
-    
+
     for (int i = 0; i < num; i++) {
       Key key = new Key(RFileTest.nf("", 3), "cf1", "cq1");
       new RelativeKey(prevKey, key).write(out);
       new Value(new byte[0]).write(out);
       prevKey = key;
     }
-    
+
     for (int i = 0; i < num; i++) {
       Key key = new Key(RFileTest.nf("", 5), "cf1", "cq1");
       new RelativeKey(prevKey, key).write(out);
       new Value(new byte[0]).write(out);
       prevKey = key;
     }
-    
+
     out.close();
     final byte[] data = baos.toByteArray();
-    
+
     CacheEntry ce = new MyCacheEntry(data);
 
     ABlockReader cacheBlock = new CachableBlockFile.CachedBlockRead(ce, data);
     BlockIndex blockIndex = null;
-    
+
     for (int i = 0; i < 257; i++)
       blockIndex = BlockIndex.getIndex(cacheBlock, new IndexEntry(prevKey, num, 0, 0, 0));
-    
+
     Assert.assertSame(null, blockIndex.seekBlock(new Key(RFileTest.nf("", 0), "cf1", "cq1"), cacheBlock));
     Assert.assertSame(null, blockIndex.seekBlock(new Key(RFileTest.nf("", 1), "cf1", "cq1"), cacheBlock));
-    
+
     for (int i = 2; i < 6; i++) {
       Key seekKey = new Key(RFileTest.nf("", i), "cf1", "cq1");
       BlockIndexEntry bie = blockIndex.seekBlock(seekKey, cacheBlock);
-      
+
       Assert.assertTrue(bie.getPrevKey().compareTo(seekKey) < 0);
 
       RelativeKey rk = new RelativeKey();
       rk.setPrevKey(bie.getPrevKey());
       rk.readFields(cacheBlock);
-      
+
       Assert.assertTrue(rk.getKey().compareTo(seekKey) <= 0);
     }
     cacheBlock.close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/rfile/CreateCompatTestFile.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/rfile/CreateCompatTestFile.java b/core/src/test/java/org/apache/accumulo/core/file/rfile/CreateCompatTestFile.java
index 46c2f0e..3eadc06 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/rfile/CreateCompatTestFile.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/rfile/CreateCompatTestFile.java
@@ -30,54 +30,54 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
 public class CreateCompatTestFile {
-  
+
   public static Set<ByteSequence> ncfs(String... colFams) {
     HashSet<ByteSequence> cfs = new HashSet<ByteSequence>();
-    
+
     for (String cf : colFams) {
       cfs.add(new ArrayByteSequence(cf));
     }
-    
+
     return cfs;
   }
-  
+
   private static Key nk(String row, String cf, String cq, String cv, long ts) {
     return new Key(row.getBytes(), cf.getBytes(), cq.getBytes(), cv.getBytes(), ts);
   }
-  
+
   private static Value nv(String val) {
     return new Value(val.getBytes());
   }
-  
+
   private static String nf(String prefix, int i) {
     return String.format(prefix + "%06d", i);
   }
-  
+
   public static void main(String[] args) throws Exception {
     Configuration conf = new Configuration();
     FileSystem fs = FileSystem.get(conf);
     CachableBlockFile.Writer _cbw = new CachableBlockFile.Writer(fs, new Path(args[0]), "gz", conf, AccumuloConfiguration.getDefaultConfiguration());
     RFile.Writer writer = new RFile.Writer(_cbw, 1000);
-    
+
     writer.startNewLocalityGroup("lg1", ncfs(nf("cf_", 1), nf("cf_", 2)));
-    
+
     for (int i = 0; i < 1000; i++) {
       writer.append(nk(nf("r_", i), nf("cf_", 1), nf("cq_", 0), "", 1000 - i), nv(i + ""));
       writer.append(nk(nf("r_", i), nf("cf_", 2), nf("cq_", 0), "", 1000 - i), nv(i + ""));
     }
-    
+
     writer.startNewLocalityGroup("lg2", ncfs(nf("cf_", 3)));
-    
+
     for (int i = 0; i < 1000; i++) {
       writer.append(nk(nf("r_", i), nf("cf_", 3), nf("cq_", 0), "", 1000 - i), nv(i + ""));
     }
-    
+
     writer.startDefaultLocalityGroup();
-    
+
     for (int i = 0; i < 1000; i++) {
       writer.append(nk(nf("r_", i), nf("cf_", 4), nf("cq_", 0), "", 1000 - i), nv(i + ""));
     }
-    
+
     writer.close();
     _cbw.close();
   }


[18/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
index 103ba05..58fa9d3 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
@@ -59,34 +59,34 @@ import org.apache.log4j.Logger;
 import com.google.common.base.Optional;
 
 public class FileUtil {
-  
+
   public static class FileInfo {
     Key firstKey = new Key();
     Key lastKey = new Key();
-    
+
     public FileInfo(Key firstKey, Key lastKey) {
       this.firstKey = firstKey;
       this.lastKey = lastKey;
     }
-    
+
     public Text getFirstRow() {
       return firstKey.getRow();
     }
-    
+
     public Text getLastRow() {
       return lastKey.getRow();
     }
   }
-  
+
   private static final Logger log = Logger.getLogger(FileUtil.class);
-  
+
   private static Path createTmpDir(AccumuloConfiguration acuConf, VolumeManager fs) throws IOException {
-    String accumuloDir = fs.choose(Optional.<String>absent(), ServerConstants.getBaseUris());
+    String accumuloDir = fs.choose(Optional.<String> absent(), ServerConstants.getBaseUris());
 
     Path result = null;
     while (result == null) {
       result = new Path(accumuloDir + Path.SEPARATOR + "tmp/idxReduce_" + String.format("%09d", new Random().nextInt(Integer.MAX_VALUE)));
-      
+
       try {
         fs.getFileStatus(result);
         result = null;
@@ -94,9 +94,9 @@ public class FileUtil {
       } catch (FileNotFoundException fne) {
         // found an unused temp directory
       }
-      
+
       fs.mkdirs(result);
-      
+
       // try to reserve the tmp dir
       // In some versions of hadoop, two clients concurrently trying to create the same directory might both return true
       // Creating a file is not subject to this, so create a special file to make sure we solely will use this directory
@@ -105,36 +105,36 @@ public class FileUtil {
     }
     return result;
   }
-  
+
   public static Collection<String> reduceFiles(AccumuloConfiguration acuConf, Configuration conf, VolumeManager fs, Text prevEndRow, Text endRow,
       Collection<String> mapFiles, int maxFiles, Path tmpDir, int pass) throws IOException {
     ArrayList<String> paths = new ArrayList<String>(mapFiles);
-    
+
     if (paths.size() <= maxFiles)
       return paths;
-    
+
     String newDir = String.format("%s/pass_%04d", tmpDir, pass);
-    
+
     int start = 0;
-    
+
     ArrayList<String> outFiles = new ArrayList<String>();
-    
+
     int count = 0;
-    
+
     while (start < paths.size()) {
       int end = Math.min(maxFiles + start, paths.size());
       List<String> inFiles = paths.subList(start, end);
-      
+
       start = end;
-      
+
       String newMapFile = String.format("%s/%04d.%s", newDir, count++, RFile.EXTENSION);
-      
+
       outFiles.add(newMapFile);
       FileSystem ns = fs.getVolumeByPath(new Path(newMapFile)).getFileSystem();
       FileSKVWriter writer = new RFileOperations().openWriter(newMapFile.toString(), ns, ns.getConf(), acuConf);
       writer.startDefaultLocalityGroup();
       List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(inFiles.size());
-      
+
       FileSKVIterator reader = null;
       try {
         for (String s : inFiles) {
@@ -142,21 +142,21 @@ public class FileUtil {
           reader = FileOperations.getInstance().openIndex(s, ns, ns.getConf(), acuConf);
           iters.add(reader);
         }
-        
+
         MultiIterator mmfi = new MultiIterator(iters, true);
-        
+
         while (mmfi.hasTop()) {
           Key key = mmfi.getTopKey();
-          
+
           boolean gtPrevEndRow = prevEndRow == null || key.compareRow(prevEndRow) > 0;
           boolean lteEndRow = endRow == null || key.compareRow(endRow) <= 0;
-          
+
           if (gtPrevEndRow && lteEndRow)
             writer.append(key, new Value(new byte[0]));
-          
+
           if (!lteEndRow)
             break;
-          
+
           mmfi.next();
         }
       } finally {
@@ -166,7 +166,7 @@ public class FileUtil {
         } catch (IOException e) {
           log.error(e, e);
         }
-        
+
         for (SortedKeyValueIterator<Key,Value> r : iters)
           try {
             if (r != null)
@@ -175,16 +175,16 @@ public class FileUtil {
             // continue closing
             log.error(e, e);
           }
-        
+
         try {
-            writer.close();
+          writer.close();
         } catch (IOException e) {
           log.error(e, e);
           throw e;
         }
       }
     }
-    
+
     return reduceFiles(acuConf, conf, fs, prevEndRow, endRow, outFiles, maxFiles, tmpDir, pass + 1);
   }
 
@@ -192,114 +192,114 @@ public class FileUtil {
       double minSplit) throws IOException {
     return findMidPoint(fs, acuConf, prevEndRow, endRow, mapFiles, minSplit, true);
   }
-  
+
   public static double estimatePercentageLTE(VolumeManager fs, AccumuloConfiguration acuconf, Text prevEndRow, Text endRow, Collection<String> mapFiles,
       Text splitRow) throws IOException {
-    
+
     Configuration conf = CachedConfiguration.getInstance();
-    
+
     Path tmpDir = null;
-    
+
     int maxToOpen = acuconf.getCount(Property.TSERV_TABLET_SPLIT_FINDMIDPOINT_MAXOPEN);
     ArrayList<FileSKVIterator> readers = new ArrayList<FileSKVIterator>(mapFiles.size());
-    
+
     try {
       if (mapFiles.size() > maxToOpen) {
         tmpDir = createTmpDir(acuconf, fs);
-        
+
         log.debug("Too many indexes (" + mapFiles.size() + ") to open at once for " + endRow + " " + prevEndRow + ", reducing in tmpDir = " + tmpDir);
-        
+
         long t1 = System.currentTimeMillis();
         mapFiles = reduceFiles(acuconf, conf, fs, prevEndRow, endRow, mapFiles, maxToOpen, tmpDir, 0);
         long t2 = System.currentTimeMillis();
-        
+
         log.debug("Finished reducing indexes for " + endRow + " " + prevEndRow + " in " + String.format("%6.2f secs", (t2 - t1) / 1000.0));
       }
-      
+
       if (prevEndRow == null)
         prevEndRow = new Text();
-      
+
       long numKeys = 0;
-      
+
       numKeys = countIndexEntries(acuconf, prevEndRow, endRow, mapFiles, true, conf, fs, readers);
-      
+
       if (numKeys == 0) {
         // not enough info in the index to answer the question, so instead of going to
         // the data just punt and return .5
         return .5;
       }
-      
+
       List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(readers);
       MultiIterator mmfi = new MultiIterator(iters, true);
-      
+
       // skip the prevendrow
       while (mmfi.hasTop() && mmfi.getTopKey().compareRow(prevEndRow) <= 0) {
         mmfi.next();
       }
-      
+
       int numLte = 0;
-      
+
       while (mmfi.hasTop() && mmfi.getTopKey().compareRow(splitRow) <= 0) {
         numLte++;
         mmfi.next();
       }
-      
+
       if (numLte > numKeys) {
         // something went wrong
         throw new RuntimeException("numLte > numKeys " + numLte + " " + numKeys + " " + prevEndRow + " " + endRow + " " + splitRow + " " + mapFiles);
       }
-      
+
       // do not want to return 0% or 100%, so add 1 and 2 below
       return (numLte + 1) / (double) (numKeys + 2);
-      
+
     } finally {
       cleanupIndexOp(acuconf, tmpDir, fs, readers);
     }
   }
-  
+
   /**
-   * 
+   *
    * @param mapFiles
    *          - list MapFiles to find the mid point key
-   * 
+   *
    *          ISSUES : This method used the index files to find the mid point. If the map files have different index intervals this method will not return an
    *          accurate mid point. Also, it would be tricky to use this method in conjunction with an in memory map because the indexing interval is unknown.
    */
   public static SortedMap<Double,Key> findMidPoint(VolumeManager fs, AccumuloConfiguration acuConf, Text prevEndRow, Text endRow, Collection<String> mapFiles,
       double minSplit, boolean useIndex) throws IOException {
     Configuration conf = CachedConfiguration.getInstance();
-    
+
     Collection<String> origMapFiles = mapFiles;
-    
+
     Path tmpDir = null;
-    
+
     int maxToOpen = acuConf.getCount(Property.TSERV_TABLET_SPLIT_FINDMIDPOINT_MAXOPEN);
     ArrayList<FileSKVIterator> readers = new ArrayList<FileSKVIterator>(mapFiles.size());
-    
+
     try {
       if (mapFiles.size() > maxToOpen) {
         if (!useIndex)
           throw new IOException("Cannot find mid point using data files, too many " + mapFiles.size());
         tmpDir = createTmpDir(acuConf, fs);
-        
+
         log.debug("Too many indexes (" + mapFiles.size() + ") to open at once for " + endRow + " " + prevEndRow + ", reducing in tmpDir = " + tmpDir);
-        
+
         long t1 = System.currentTimeMillis();
         mapFiles = reduceFiles(acuConf, conf, fs, prevEndRow, endRow, mapFiles, maxToOpen, tmpDir, 0);
         long t2 = System.currentTimeMillis();
-        
+
         log.debug("Finished reducing indexes for " + endRow + " " + prevEndRow + " in " + String.format("%6.2f secs", (t2 - t1) / 1000.0));
       }
-      
+
       if (prevEndRow == null)
         prevEndRow = new Text();
-      
+
       long t1 = System.currentTimeMillis();
-      
+
       long numKeys = 0;
-      
+
       numKeys = countIndexEntries(acuConf, prevEndRow, endRow, mapFiles, tmpDir == null ? useIndex : false, conf, fs, readers);
-      
+
       if (numKeys == 0) {
         if (useIndex) {
           log.warn("Failed to find mid point using indexes, falling back to data files which is slower. No entries between " + prevEndRow + " and " + endRow
@@ -309,48 +309,48 @@ public class FileUtil {
         }
         throw new IOException("Failed to find mid point, no entries between " + prevEndRow + " and " + endRow + " for " + mapFiles);
       }
-      
+
       List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(readers);
       MultiIterator mmfi = new MultiIterator(iters, true);
-      
+
       // skip the prevendrow
       while (mmfi.hasTop() && mmfi.getTopKey().compareRow(prevEndRow) <= 0)
         mmfi.next();
-      
+
       // read half of the keys in the index
       TreeMap<Double,Key> ret = new TreeMap<Double,Key>();
       Key lastKey = null;
       long keysRead = 0;
-      
+
       Key keyBeforeMidPoint = null;
       long keyBeforeMidPointPosition = 0;
-      
+
       while (keysRead < numKeys / 2) {
         if (lastKey != null && !lastKey.equals(mmfi.getTopKey(), PartialKey.ROW) && (keysRead - 1) / (double) numKeys >= minSplit) {
           keyBeforeMidPoint = new Key(lastKey);
           keyBeforeMidPointPosition = keysRead - 1;
         }
-        
+
         if (lastKey == null)
           lastKey = new Key();
-        
+
         lastKey.set(mmfi.getTopKey());
-        
+
         keysRead++;
-        
+
         // consume minimum
         mmfi.next();
       }
-      
+
       if (keyBeforeMidPoint != null)
         ret.put(keyBeforeMidPointPosition / (double) numKeys, keyBeforeMidPoint);
-      
+
       long t2 = System.currentTimeMillis();
-      
+
       log.debug(String.format("Found midPoint from indexes in %6.2f secs.%n", ((t2 - t1) / 1000.0)));
-      
+
       ret.put(.5, mmfi.getTopKey());
-      
+
       // sanity check
       for (Key key : ret.values()) {
         boolean inRange = (key.compareRow(prevEndRow) > 0 && (endRow == null || key.compareRow(endRow) < 1));
@@ -358,7 +358,7 @@ public class FileUtil {
           throw new IOException("Found mid point is not in range " + key + " " + prevEndRow + " " + endRow + " " + mapFiles);
         }
       }
-      
+
       return ret;
     } finally {
       cleanupIndexOp(acuConf, tmpDir, fs, readers);
@@ -390,9 +390,9 @@ public class FileUtil {
 
   private static long countIndexEntries(AccumuloConfiguration acuConf, Text prevEndRow, Text endRow, Collection<String> mapFiles, boolean useIndex,
       Configuration conf, VolumeManager fs, ArrayList<FileSKVIterator> readers) throws IOException {
-    
+
     long numKeys = 0;
-    
+
     // count the total number of index entries
     for (String ref : mapFiles) {
       FileSKVIterator reader = null;
@@ -402,16 +402,16 @@ public class FileUtil {
         if (useIndex)
           reader = FileOperations.getInstance().openIndex(path.toString(), ns, ns.getConf(), acuConf);
         else
-          reader = FileOperations.getInstance().openReader(path.toString(), new Range(prevEndRow, false, null, true), LocalityGroupUtil.EMPTY_CF_SET, false, ns, ns.getConf(),
-              acuConf);
-        
+          reader = FileOperations.getInstance().openReader(path.toString(), new Range(prevEndRow, false, null, true), LocalityGroupUtil.EMPTY_CF_SET, false,
+              ns, ns.getConf(), acuConf);
+
         while (reader.hasTop()) {
           Key key = reader.getTopKey();
           if (endRow != null && key.compareRow(endRow) > 0)
             break;
           else if (prevEndRow == null || key.compareRow(prevEndRow) > 0)
             numKeys++;
-          
+
           reader.next();
         }
       } finally {
@@ -422,35 +422,35 @@ public class FileUtil {
           log.error(e, e);
         }
       }
-      
+
       if (useIndex)
         readers.add(FileOperations.getInstance().openIndex(path.toString(), ns, ns.getConf(), acuConf));
       else
-        readers.add(FileOperations.getInstance().openReader(path.toString(), new Range(prevEndRow, false, null, true), LocalityGroupUtil.EMPTY_CF_SET, false, ns, ns.getConf(),
-            acuConf));
-      
+        readers.add(FileOperations.getInstance().openReader(path.toString(), new Range(prevEndRow, false, null, true), LocalityGroupUtil.EMPTY_CF_SET, false,
+            ns, ns.getConf(), acuConf));
+
     }
     return numKeys;
   }
-  
+
   public static Map<FileRef,FileInfo> tryToGetFirstAndLastRows(VolumeManager fs, AccumuloConfiguration acuConf, Set<FileRef> mapfiles) {
-    
+
     HashMap<FileRef,FileInfo> mapFilesInfo = new HashMap<FileRef,FileInfo>();
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     for (FileRef mapfile : mapfiles) {
-      
+
       FileSKVIterator reader = null;
       FileSystem ns = fs.getVolumeByPath(mapfile.path()).getFileSystem();
       try {
         reader = FileOperations.getInstance().openReader(mapfile.toString(), false, ns, ns.getConf(), acuConf);
-        
+
         Key firstKey = reader.getFirstKey();
         if (firstKey != null) {
           mapFilesInfo.put(mapfile, new FileInfo(firstKey, reader.getLastKey()));
         }
-        
+
       } catch (IOException ioe) {
         log.warn("Failed to read map file to determine first and last key : " + mapfile, ioe);
       } finally {
@@ -462,31 +462,31 @@ public class FileUtil {
           }
         }
       }
-      
+
     }
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     log.debug(String.format("Found first and last keys for %d map files in %6.2f secs", mapfiles.size(), (t2 - t1) / 1000.0));
-    
+
     return mapFilesInfo;
   }
-  
+
   public static WritableComparable<Key> findLastKey(VolumeManager fs, AccumuloConfiguration acuConf, Collection<FileRef> mapFiles) throws IOException {
     Key lastKey = null;
-    
+
     for (FileRef ref : mapFiles) {
       Path path = ref.path();
       FileSystem ns = fs.getVolumeByPath(path).getFileSystem();
       FileSKVIterator reader = FileOperations.getInstance().openReader(path.toString(), true, ns, ns.getConf(), acuConf);
-      
+
       try {
         if (!reader.hasTop())
           // file is empty, so there is no last key
           continue;
-        
+
         Key key = reader.getLastKey();
-        
+
         if (lastKey == null || key.compareTo(lastKey) > 0)
           lastKey = key;
       } finally {
@@ -498,41 +498,41 @@ public class FileUtil {
         }
       }
     }
-    
+
     return lastKey;
-    
+
   }
-  
+
   private static class MLong {
     public MLong(long i) {
       l = i;
     }
-    
+
     long l;
   }
-  
+
   public static Map<KeyExtent,Long> estimateSizes(AccumuloConfiguration acuConf, Path mapFile, long fileSize, List<KeyExtent> extents, Configuration conf,
       VolumeManager fs) throws IOException {
-    
+
     long totalIndexEntries = 0;
     Map<KeyExtent,MLong> counts = new TreeMap<KeyExtent,MLong>();
     for (KeyExtent keyExtent : extents)
       counts.put(keyExtent, new MLong(0));
-    
+
     Text row = new Text();
     FileSystem ns = fs.getVolumeByPath(mapFile).getFileSystem();
     FileSKVIterator index = FileOperations.getInstance().openIndex(mapFile.toString(), ns, ns.getConf(), acuConf);
-    
+
     try {
       while (index.hasTop()) {
         Key key = index.getTopKey();
         totalIndexEntries++;
         key.getRow(row);
-        
+
         for (Entry<KeyExtent,MLong> entry : counts.entrySet())
           if (entry.getKey().contains(row))
             entry.getValue().l++;
-        
+
         index.next();
       }
     } finally {
@@ -544,7 +544,7 @@ public class FileUtil {
         log.error(e, e);
       }
     }
-    
+
     Map<KeyExtent,Long> results = new TreeMap<KeyExtent,Long>();
     for (KeyExtent keyExtent : extents) {
       double numEntries = counts.get(keyExtent).l;
@@ -555,7 +555,7 @@ public class FileUtil {
     }
     return results;
   }
-  
+
   public static Collection<String> toPathStrings(Collection<FileRef> refs) {
     ArrayList<String> ret = new ArrayList<String>();
     for (FileRef fileRef : refs) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/Halt.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/Halt.java b/server/base/src/main/java/org/apache/accumulo/server/util/Halt.java
index 82eb639..aa45bff 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/Halt.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/Halt.java
@@ -22,7 +22,7 @@ import org.apache.log4j.Logger;
 
 public class Halt {
   static private final Logger log = Logger.getLogger(Halt.class);
-  
+
   public static void halt(final String msg) {
     halt(0, new Runnable() {
       public void run() {
@@ -30,7 +30,7 @@ public class Halt {
       }
     });
   }
-  
+
   public static void halt(final String msg, int status) {
     halt(status, new Runnable() {
       public void run() {
@@ -38,7 +38,7 @@ public class Halt {
       }
     });
   }
-  
+
   public static void halt(final int status, Runnable runnable) {
     try {
       // give ourselves a little time to try and do something
@@ -48,7 +48,7 @@ public class Halt {
           Runtime.getRuntime().halt(status);
         }
       }.start();
-      
+
       if (runnable != null)
         runnable.run();
       Runtime.getRuntime().halt(status);
@@ -57,5 +57,5 @@ public class Halt {
       Runtime.getRuntime().halt(-1);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/ListInstances.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/ListInstances.java b/server/base/src/main/java/org/apache/accumulo/server/util/ListInstances.java
index 8550d0b..c8b8dff 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/ListInstances.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/ListInstances.java
@@ -38,39 +38,40 @@ import org.apache.log4j.Logger;
 import com.beust.jcommander.Parameter;
 
 public class ListInstances {
-  
+
   private static final Logger log = Logger.getLogger(ListInstances.class);
-  
+
   private static final int NAME_WIDTH = 20;
   private static final int UUID_WIDTH = 37;
   private static final int MASTER_WIDTH = 30;
-  
+
   private static final int ZOOKEEPER_TIMER_MILLIS = 30 * 1000;
-  
+
   static class Opts extends Help {
-    @Parameter(names="--print-errors", description="display errors while listing instances")
+    @Parameter(names = "--print-errors", description = "display errors while listing instances")
     boolean printErrors = false;
-    @Parameter(names="--print-all", description="print information for all instances, not just those with names")
+    @Parameter(names = "--print-all", description = "print information for all instances, not just those with names")
     boolean printAll = false;
-    @Parameter(names={"-z", "--zookeepers"}, description="the zookeepers to contact")
+    @Parameter(names = {"-z", "--zookeepers"}, description = "the zookeepers to contact")
     String keepers = null;
   }
+
   static Opts opts = new Opts();
   static int errors = 0;
-  
+
   public static void main(String[] args) {
     opts.parseArgs(ListInstances.class.getName(), args);
-    
+
     if (opts.keepers == null) {
       opts.keepers = SiteConfiguration.getInstance().get(Property.INSTANCE_ZK_HOST);
     }
-    
+
     String keepers = opts.keepers;
     boolean printAll = opts.printAll;
     boolean printErrors = opts.printErrors;
-    
+
     listInstances(keepers, printAll, printErrors);
-    
+
   }
 
   static synchronized void listInstances(String keepers, boolean printAll, boolean printErrors) {
@@ -81,17 +82,17 @@ public class ListInstances {
     ZooCache cache = new ZooCache(keepers, ZOOKEEPER_TIMER_MILLIS);
 
     TreeMap<String,UUID> instanceNames = getInstanceNames(rdr, printErrors);
-    
+
     System.out.println();
     printHeader();
-    
+
     for (Entry<String,UUID> entry : instanceNames.entrySet()) {
       printInstanceInfo(cache, entry.getKey(), entry.getValue(), printErrors);
     }
-    
+
     TreeSet<UUID> instancedIds = getInstanceIDs(rdr, printErrors);
     instancedIds.removeAll(instanceNames.values());
-    
+
     if (printAll) {
       for (UUID uuid : instancedIds) {
         printInstanceInfo(cache, null, uuid, printErrors);
@@ -102,57 +103,57 @@ public class ListInstances {
     } else {
       System.out.println();
     }
-    
+
     if (!printErrors && errors > 0) {
       System.err.println("WARN : There were " + errors + " errors, run with --print-errors to see more info");
     }
   }
-  
+
   private static class CharFiller implements Formattable {
-    
+
     char c;
-    
+
     CharFiller(char c) {
       this.c = c;
     }
-    
+
     @Override
     public void formatTo(Formatter formatter, int flags, int width, int precision) {
-      
+
       StringBuilder sb = new StringBuilder();
       for (int i = 0; i < width; i++)
         sb.append(c);
       formatter.format(sb.toString());
     }
-    
+
   }
-  
+
   private static void printHeader() {
     System.out.printf(" %-" + NAME_WIDTH + "s| %-" + UUID_WIDTH + "s| %-" + MASTER_WIDTH + "s%n", "Instance Name", "Instance ID", "Master");
     System.out.printf("%" + (NAME_WIDTH + 1) + "s+%" + (UUID_WIDTH + 1) + "s+%" + (MASTER_WIDTH + 1) + "s%n", new CharFiller('-'), new CharFiller('-'),
         new CharFiller('-'));
-    
+
   }
-  
+
   private static void printInstanceInfo(ZooCache cache, String instanceName, UUID iid, boolean printErrors) {
     String master = getMaster(cache, iid, printErrors);
     if (instanceName == null) {
       instanceName = "";
     }
-    
+
     if (master == null) {
       master = "";
     }
-    
+
     System.out.printf("%" + NAME_WIDTH + "s |%" + UUID_WIDTH + "s |%" + MASTER_WIDTH + "s%n", "\"" + instanceName + "\"", iid, master);
   }
-  
+
   private static String getMaster(ZooCache cache, UUID iid, boolean printErrors) {
-    
+
     if (iid == null) {
       return null;
     }
-    
+
     try {
       String masterLocPath = Constants.ZROOT + "/" + iid + Constants.ZMASTER_LOCK;
       byte[] master = ZooLock.getLockData(cache, masterLocPath, null);
@@ -165,22 +166,22 @@ public class ListInstances {
       return null;
     }
   }
-  
+
   private static TreeMap<String,UUID> getInstanceNames(ZooReader zk, boolean printErrors) {
-    
+
     String instancesPath = Constants.ZROOT + Constants.ZINSTANCES;
-    
+
     TreeMap<String,UUID> tm = new TreeMap<String,UUID>();
-    
+
     List<String> names;
-    
+
     try {
       names = zk.getChildren(instancesPath);
     } catch (Exception e) {
       handleException(e, printErrors);
       return tm;
     }
-    
+
     for (String name : names) {
       String instanceNamePath = Constants.ZROOT + Constants.ZINSTANCES + "/" + name;
       try {
@@ -191,16 +192,16 @@ public class ListInstances {
         tm.put(name, null);
       }
     }
-    
+
     return tm;
   }
-  
+
   private static TreeSet<UUID> getInstanceIDs(ZooReader zk, boolean printErrors) {
     TreeSet<UUID> ts = new TreeSet<UUID>();
-    
+
     try {
       List<String> children = zk.getChildren(Constants.ZROOT);
-      
+
       for (String iid : children) {
         if (iid.equals("instances"))
           continue;
@@ -213,15 +214,15 @@ public class ListInstances {
     } catch (Exception e) {
       handleException(e, printErrors);
     }
-    
+
     return ts;
   }
-  
+
   private static void handleException(Exception e, boolean printErrors) {
     if (printErrors) {
       log.error(e);
     }
-    
+
     errors++;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/ListVolumesUsed.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/ListVolumesUsed.java b/server/base/src/main/java/org/apache/accumulo/server/util/ListVolumesUsed.java
index b288aac..e90d1dd 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/ListVolumesUsed.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/ListVolumesUsed.java
@@ -37,7 +37,7 @@ import org.apache.accumulo.server.fs.VolumeManager.FileType;
 import org.apache.hadoop.fs.Path;
 
 /**
- * 
+ *
  */
 public class ListVolumesUsed {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/LocalityCheck.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/LocalityCheck.java b/server/base/src/main/java/org/apache/accumulo/server/util/LocalityCheck.java
index 35bb8f5..5d49fa7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/LocalityCheck.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/LocalityCheck.java
@@ -41,22 +41,22 @@ import org.apache.hadoop.fs.Path;
 import com.google.common.net.HostAndPort;
 
 public class LocalityCheck {
-  
+
   public int run(String[] args) throws Exception {
     ClientOpts opts = new ClientOpts();
     opts.parseArgs(LocalityCheck.class.getName(), args);
-    
+
     VolumeManager fs = VolumeManagerImpl.get();
     Connector connector = opts.getConnector();
     Scanner scanner = connector.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.fetchColumnFamily(TabletsSection.CurrentLocationColumnFamily.NAME);
     scanner.fetchColumnFamily(DataFileColumnFamily.NAME);
     scanner.setRange(MetadataSchema.TabletsSection.getRange());
-    
+
     Map<String,Long> totalBlocks = new HashMap<String,Long>();
     Map<String,Long> localBlocks = new HashMap<String,Long>();
     ArrayList<String> files = new ArrayList<String>();
-    
+
     for (Entry<Key,Value> entry : scanner) {
       Key key = entry.getKey();
       if (key.compareColumnFamily(TabletsSection.CurrentLocationColumnFamily.NAME) == 0) {
@@ -66,7 +66,7 @@ public class LocalityCheck {
         addBlocks(fs, host, files, totalBlocks, localBlocks);
         files.clear();
       } else if (key.compareColumnFamily(DataFileColumnFamily.NAME) == 0) {
-        
+
         files.add(fs.getFullPath(key).toString());
       }
     }
@@ -78,7 +78,7 @@ public class LocalityCheck {
     }
     return 0;
   }
-  
+
   private void addBlocks(VolumeManager fs, String host, ArrayList<String> files, Map<String,Long> totalBlocks, Map<String,Long> localBlocks) throws Exception {
     long allBlocks = 0;
     long matchingBlocks = 0;
@@ -105,7 +105,7 @@ public class LocalityCheck {
     totalBlocks.put(host, allBlocks + totalBlocks.get(host));
     localBlocks.put(host, matchingBlocks + localBlocks.get(host));
   }
-  
+
   public static void main(String[] args) throws Exception {
     LocalityCheck check = new LocalityCheck();
     System.exit(check.run(args));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/LoginProperties.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/LoginProperties.java b/server/base/src/main/java/org/apache/accumulo/server/util/LoginProperties.java
index 241eef3..be5a7c8 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/LoginProperties.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/LoginProperties.java
@@ -30,29 +30,29 @@ import org.apache.accumulo.server.security.handler.Authenticator;
 import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
 
 /**
- * 
+ *
  */
 public class LoginProperties {
-  
+
   public static void main(String[] args) throws Exception {
     AccumuloConfiguration config = new ServerConfigurationFactory(HdfsZooInstance.getInstance()).getConfiguration();
     Authenticator authenticator = AccumuloVFSClassLoader.getClassLoader().loadClass(config.get(Property.INSTANCE_SECURITY_AUTHENTICATOR))
         .asSubclass(Authenticator.class).newInstance();
-    
+
     List<Set<TokenProperty>> tokenProps = new ArrayList<Set<TokenProperty>>();
-    
+
     for (Class<? extends AuthenticationToken> tokenType : authenticator.getSupportedTokenTypes()) {
       tokenProps.add(tokenType.newInstance().getProperties());
     }
-    
+
     System.out.println("Supported token types for " + authenticator.getClass().getName() + " are : ");
     for (Class<? extends AuthenticationToken> tokenType : authenticator.getSupportedTokenTypes()) {
       System.out.println("\t" + tokenType.getName() + ", which accepts the following properties : ");
-      
+
       for (TokenProperty tokenProperty : tokenType.newInstance().getProperties()) {
         System.out.println("\t\t" + tokenProperty);
       }
-      
+
       System.out.println();
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/MasterMetadataUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/MasterMetadataUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/MasterMetadataUtil.java
index 88316c6..80a6734 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/MasterMetadataUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/MasterMetadataUtil.java
@@ -60,94 +60,94 @@ import org.apache.log4j.Logger;
 import org.apache.zookeeper.KeeperException;
 
 /**
- * 
+ *
  */
 public class MasterMetadataUtil {
-  
+
   private static final Logger log = Logger.getLogger(MasterMetadataUtil.class);
-  
-  public static void addNewTablet(ClientContext context, KeyExtent extent, String path, TServerInstance location,
-      Map<FileRef,DataFileValue> datafileSizes, Map<FileRef,Long> bulkLoadedFiles, String time, long lastFlushID, long lastCompactID, ZooLock zooLock) {
+
+  public static void addNewTablet(ClientContext context, KeyExtent extent, String path, TServerInstance location, Map<FileRef,DataFileValue> datafileSizes,
+      Map<FileRef,Long> bulkLoadedFiles, String time, long lastFlushID, long lastCompactID, ZooLock zooLock) {
     Mutation m = extent.getPrevRowUpdateMutation();
-    
+
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(m, new Value(path.getBytes(UTF_8)));
     TabletsSection.ServerColumnFamily.TIME_COLUMN.put(m, new Value(time.getBytes(UTF_8)));
     if (lastFlushID > 0)
       TabletsSection.ServerColumnFamily.FLUSH_COLUMN.put(m, new Value(("" + lastFlushID).getBytes()));
     if (lastCompactID > 0)
       TabletsSection.ServerColumnFamily.COMPACT_COLUMN.put(m, new Value(("" + lastCompactID).getBytes()));
-    
+
     if (location != null) {
       location.putLocation(m);
       location.clearFutureLocation(m);
     }
-    
+
     for (Entry<FileRef,DataFileValue> entry : datafileSizes.entrySet()) {
       m.put(DataFileColumnFamily.NAME, entry.getKey().meta(), new Value(entry.getValue().encode()));
     }
-    
+
     for (Entry<FileRef,Long> entry : bulkLoadedFiles.entrySet()) {
       byte[] tidBytes = Long.toString(entry.getValue()).getBytes();
       m.put(TabletsSection.BulkFileColumnFamily.NAME, entry.getKey().meta(), new Value(tidBytes));
     }
-    
+
     MetadataTableUtil.update(context, zooLock, m, extent);
   }
-  
+
   public static KeyExtent fixSplit(ClientContext context, Text metadataEntry, SortedMap<ColumnFQ,Value> columns, TServerInstance tserver, ZooLock lock)
       throws AccumuloException, IOException {
     log.info("Incomplete split " + metadataEntry + " attempting to fix");
-    
+
     Value oper = columns.get(TabletsSection.TabletColumnFamily.OLD_PREV_ROW_COLUMN);
-    
+
     if (columns.get(TabletsSection.TabletColumnFamily.SPLIT_RATIO_COLUMN) == null) {
       throw new IllegalArgumentException("Metadata entry does not have split ratio (" + metadataEntry + ")");
     }
-    
+
     double splitRatio = Double.parseDouble(new String(columns.get(TabletsSection.TabletColumnFamily.SPLIT_RATIO_COLUMN).get(), UTF_8));
-    
+
     Value prevEndRowIBW = columns.get(TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN);
-    
+
     if (prevEndRowIBW == null) {
       throw new IllegalArgumentException("Metadata entry does not have prev row (" + metadataEntry + ")");
     }
-    
+
     Value time = columns.get(TabletsSection.ServerColumnFamily.TIME_COLUMN);
-    
+
     if (time == null) {
       throw new IllegalArgumentException("Metadata entry does not have time (" + metadataEntry + ")");
     }
-    
+
     Value flushID = columns.get(TabletsSection.ServerColumnFamily.FLUSH_COLUMN);
     long initFlushID = -1;
     if (flushID != null)
       initFlushID = Long.parseLong(flushID.toString());
-    
+
     Value compactID = columns.get(TabletsSection.ServerColumnFamily.COMPACT_COLUMN);
     long initCompactID = -1;
     if (compactID != null)
       initCompactID = Long.parseLong(compactID.toString());
-    
+
     Text metadataPrevEndRow = KeyExtent.decodePrevEndRow(prevEndRowIBW);
-    
+
     Text table = (new KeyExtent(metadataEntry, (Text) null)).getTableId();
-    
+
     return fixSplit(context, table, metadataEntry, metadataPrevEndRow, oper, splitRatio, tserver, time.toString(), initFlushID, initCompactID, lock);
   }
-  
+
   private static KeyExtent fixSplit(ClientContext context, Text table, Text metadataEntry, Text metadataPrevEndRow, Value oper, double splitRatio,
       TServerInstance tserver, String time, long initFlushID, long initCompactID, ZooLock lock) throws AccumuloException, IOException {
     if (metadataPrevEndRow == null)
       // something is wrong, this should not happen... if a tablet is split, it will always have a
       // prev end row....
       throw new AccumuloException("Split tablet does not have prev end row, something is amiss, extent = " + metadataEntry);
-    
+
     // check to see if prev tablet exist in metadata tablet
     Key prevRowKey = new Key(new Text(KeyExtent.getMetadataEntry(table, metadataPrevEndRow)));
-    
+
     ScannerImpl scanner2 = new ScannerImpl(context, MetadataTable.ID, Authorizations.EMPTY);
     scanner2.setRange(new Range(prevRowKey, prevRowKey.followingKey(PartialKey.ROW)));
-    
+
     VolumeManager fs = VolumeManagerImpl.get();
     if (!scanner2.iterator().hasNext()) {
       log.info("Rolling back incomplete split " + metadataEntry + " " + metadataPrevEndRow);
@@ -155,34 +155,34 @@ public class MasterMetadataUtil {
       return new KeyExtent(metadataEntry, KeyExtent.decodePrevEndRow(oper));
     } else {
       log.info("Finishing incomplete split " + metadataEntry + " " + metadataPrevEndRow);
-      
+
       List<FileRef> highDatafilesToRemove = new ArrayList<FileRef>();
-      
+
       Scanner scanner3 = new ScannerImpl(context, MetadataTable.ID, Authorizations.EMPTY);
       Key rowKey = new Key(metadataEntry);
-      
+
       SortedMap<FileRef,DataFileValue> origDatafileSizes = new TreeMap<FileRef,DataFileValue>();
       SortedMap<FileRef,DataFileValue> highDatafileSizes = new TreeMap<FileRef,DataFileValue>();
       SortedMap<FileRef,DataFileValue> lowDatafileSizes = new TreeMap<FileRef,DataFileValue>();
       scanner3.fetchColumnFamily(DataFileColumnFamily.NAME);
       scanner3.setRange(new Range(rowKey, rowKey.followingKey(PartialKey.ROW)));
-      
+
       for (Entry<Key,Value> entry : scanner3) {
         if (entry.getKey().compareColumnFamily(DataFileColumnFamily.NAME) == 0) {
           origDatafileSizes.put(new FileRef(fs, entry.getKey()), new DataFileValue(entry.getValue().get()));
         }
       }
-      
+
       MetadataTableUtil.splitDatafiles(table, metadataPrevEndRow, splitRatio, new HashMap<FileRef,FileUtil.FileInfo>(), origDatafileSizes, lowDatafileSizes,
           highDatafileSizes, highDatafilesToRemove);
-      
+
       MetadataTableUtil.finishSplit(metadataEntry, highDatafileSizes, highDatafilesToRemove, context, lock);
-      
+
       return new KeyExtent(metadataEntry, KeyExtent.encodePrevEndRow(metadataPrevEndRow));
     }
-    
+
   }
-  
+
   private static TServerInstance getTServerInstance(String address, ZooLock zooLock) {
     while (true) {
       try {
@@ -195,51 +195,51 @@ public class MasterMetadataUtil {
       UtilWaitThread.sleep(1000);
     }
   }
-  
+
   public static void replaceDatafiles(ClientContext context, KeyExtent extent, Set<FileRef> datafilesToDelete, Set<FileRef> scanFiles, FileRef path,
       Long compactionId, DataFileValue size, String address, TServerInstance lastLocation, ZooLock zooLock) throws IOException {
     replaceDatafiles(context, extent, datafilesToDelete, scanFiles, path, compactionId, size, address, lastLocation, zooLock, true);
   }
-  
+
   public static void replaceDatafiles(ClientContext context, KeyExtent extent, Set<FileRef> datafilesToDelete, Set<FileRef> scanFiles, FileRef path,
       Long compactionId, DataFileValue size, String address, TServerInstance lastLocation, ZooLock zooLock, boolean insertDeleteFlags) throws IOException {
-    
+
     if (insertDeleteFlags) {
       // add delete flags for those paths before the data file reference is removed
       MetadataTableUtil.addDeleteEntries(extent, datafilesToDelete, context);
     }
-    
+
     // replace data file references to old mapfiles with the new mapfiles
     Mutation m = new Mutation(extent.getMetadataEntry());
-    
+
     for (FileRef pathToRemove : datafilesToDelete)
       m.putDelete(DataFileColumnFamily.NAME, pathToRemove.meta());
-    
+
     for (FileRef scanFile : scanFiles)
       m.put(ScanFileColumnFamily.NAME, scanFile.meta(), new Value(new byte[0]));
-    
+
     if (size.getNumEntries() > 0)
       m.put(DataFileColumnFamily.NAME, path.meta(), new Value(size.encode()));
-    
+
     if (compactionId != null)
       TabletsSection.ServerColumnFamily.COMPACT_COLUMN.put(m, new Value(("" + compactionId).getBytes()));
-    
+
     TServerInstance self = getTServerInstance(address, zooLock);
     self.putLastLocation(m);
-    
+
     // remove the old location
     if (lastLocation != null && !lastLocation.equals(self))
       lastLocation.clearLastLocation(m);
-    
+
     MetadataTableUtil.update(context, zooLock, m, extent);
   }
-  
+
   /**
    * new data file update function adds one data file to a tablet's list
-   * 
+   *
    * @param path
    *          should be relative to the table directory
-   * 
+   *
    */
   public static void updateTabletDataFile(ClientContext context, KeyExtent extent, FileRef path, FileRef mergeFile, DataFileValue dfv, String time,
       Set<FileRef> filesInUseByScans, String address, ZooLock zooLock, Set<String> unusedWalLogs, TServerInstance lastLocation, long flushId) {
@@ -292,6 +292,7 @@ public class MasterMetadataUtil {
 
   /**
    * Create an update that updates a tablet
+   *
    * @return A Mutation to update a tablet from the given information
    */
   protected static Mutation getUpdateForTabletDataFile(KeyExtent extent, FileRef path, FileRef mergeFile, DataFileValue dfv, String time,

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/MetadataTableUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/MetadataTableUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/MetadataTableUtil.java
index 524abb0..ed7626e 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/MetadataTableUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/MetadataTableUtil.java
@@ -182,8 +182,7 @@ public class MetadataTableUtil {
     }
   }
 
-  public static void updateTabletDataFile(long tid, KeyExtent extent, Map<FileRef,DataFileValue> estSizes, String time, ClientContext context,
-      ZooLock zooLock) {
+  public static void updateTabletDataFile(long tid, KeyExtent extent, Map<FileRef,DataFileValue> estSizes, String time, ClientContext context, ZooLock zooLock) {
     Mutation m = new Mutation(extent.getMetadataEntry());
     byte[] tidBytes = Long.toString(tid).getBytes(UTF_8);
 
@@ -506,8 +505,8 @@ public class MetadataTableUtil {
     }
   }
 
-  public static Pair<List<LogEntry>,SortedMap<FileRef,DataFileValue>> getFileAndLogEntries(ClientContext context, KeyExtent extent)
-      throws KeeperException, InterruptedException, IOException {
+  public static Pair<List<LogEntry>,SortedMap<FileRef,DataFileValue>> getFileAndLogEntries(ClientContext context, KeyExtent extent) throws KeeperException,
+      InterruptedException, IOException {
     ArrayList<LogEntry> result = new ArrayList<LogEntry>();
     TreeMap<FileRef,DataFileValue> sizes = new TreeMap<FileRef,DataFileValue>();
 
@@ -891,8 +890,8 @@ public class MetadataTableUtil {
       Key k = entry.getKey();
       Mutation m = new Mutation(k.getRow());
       m.putDelete(k.getColumnFamily(), k.getColumnQualifier());
-      String dir = volumeManager.choose(Optional.of(tableId), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR + tableId + Path.SEPARATOR
-          + new String(FastFormat.toZeroPaddedString(dirCount++, 8, 16, Constants.CLONE_PREFIX_BYTES));
+      String dir = volumeManager.choose(Optional.of(tableId), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR + tableId
+          + Path.SEPARATOR + new String(FastFormat.toZeroPaddedString(dirCount++, 8, 16, Constants.CLONE_PREFIX_BYTES));
       TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(m, new Value(dir.getBytes(UTF_8)));
 
       bw.addMutation(m);
@@ -983,8 +982,8 @@ public class MetadataTableUtil {
    * During an upgrade from 1.6 to 1.7, we need to add the replication table
    */
   public static void createReplicationTable(ClientContext context) throws IOException {
-    String dir = VolumeManagerImpl.get().choose(Optional.of(ReplicationTable.ID), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR + ReplicationTable.ID
-        + Constants.DEFAULT_TABLET_LOCATION;
+    String dir = VolumeManagerImpl.get().choose(Optional.of(ReplicationTable.ID), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR
+        + ReplicationTable.ID + Constants.DEFAULT_TABLET_LOCATION;
 
     Mutation m = new Mutation(new Text(KeyExtent.getMetadataEntry(new Text(ReplicationTable.ID), null)));
     m.put(DIRECTORY_COLUMN.getColumnFamily(), DIRECTORY_COLUMN.getColumnQualifier(), 0, new Value(dir.getBytes(UTF_8)));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/RandomWriter.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/RandomWriter.java b/server/base/src/main/java/org/apache/accumulo/server/util/RandomWriter.java
index 0666297..b3c198e 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/RandomWriter.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/RandomWriter.java
@@ -31,26 +31,26 @@ import org.apache.log4j.Logger;
 import com.beust.jcommander.Parameter;
 
 public class RandomWriter {
-  
+
   private static String table_name = "test_write_table";
   private static int num_columns_per_row = 1;
   private static int num_payload_bytes = 1024;
   private static final Logger log = Logger.getLogger(RandomWriter.class);
-  
+
   public static class RandomMutationGenerator implements Iterable<Mutation>, Iterator<Mutation> {
     private long max_mutations;
     private int mutations_so_far = 0;
     private Random r = new Random();
     private static final Logger log = Logger.getLogger(RandomMutationGenerator.class);
-    
+
     public RandomMutationGenerator(long num_mutations) {
       max_mutations = num_mutations;
     }
-    
+
     public boolean hasNext() {
       return mutations_so_far < max_mutations;
     }
-    
+
     public Mutation next() {
       Text row_value = new Text(Long.toString(((r.nextLong() & 0x7fffffffffffffffl) / 177) % 100000000000l));
       Mutation m = new Mutation(row_value);
@@ -66,27 +66,31 @@ public class RandomWriter {
       }
       return m;
     }
-    
+
     public void remove() {
       mutations_so_far++;
     }
-    
+
     @Override
     public Iterator<Mutation> iterator() {
       return this;
     }
   }
+
   static class Opts extends ClientOnDefaultTable {
-    @Parameter(names="--count", description="number of mutations to write", required=true)
+    @Parameter(names = "--count", description = "number of mutations to write", required = true)
     long count;
-    
-    Opts(String table) { super(table); }
+
+    Opts(String table) {
+      super(table);
+    }
   }
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts(table_name);
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(RandomWriter.class.getName(), args, bwOpts);
-    
+
     long start = System.currentTimeMillis();
     log.info("starting at " + start + " for user " + opts.principal);
     try {
@@ -100,9 +104,9 @@ public class RandomWriter {
       throw e;
     }
     long stop = System.currentTimeMillis();
-    
+
     log.info("stopping at " + stop);
     log.info("elapsed: " + (((double) stop - (double) start) / 1000.0));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/RandomizeVolumes.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/RandomizeVolumes.java b/server/base/src/main/java/org/apache/accumulo/server/util/RandomizeVolumes.java
index de360fe..77ffca3 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/RandomizeVolumes.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/RandomizeVolumes.java
@@ -112,7 +112,8 @@ public class RandomizeVolumes {
       Key key = entry.getKey();
       Mutation m = new Mutation(key.getRow());
 
-      final String newLocation = vm.choose(Optional.of(tableId), ServerConstants.getBaseUris()) + Path.SEPARATOR + ServerConstants.TABLE_DIR + Path.SEPARATOR + tableId + Path.SEPARATOR + directory;
+      final String newLocation = vm.choose(Optional.of(tableId), ServerConstants.getBaseUris()) + Path.SEPARATOR + ServerConstants.TABLE_DIR + Path.SEPARATOR
+          + tableId + Path.SEPARATOR + directory;
       m.put(key.getColumnFamily(), key.getColumnQualifier(), new Value(newLocation.getBytes(UTF_8)));
       if (log.isTraceEnabled()) {
         log.trace("Replacing " + oldLocation + " with " + newLocation);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/RemoveEntriesForMissingFiles.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/RemoveEntriesForMissingFiles.java b/server/base/src/main/java/org/apache/accumulo/server/util/RemoveEntriesForMissingFiles.java
index d5b586c..b5cf510 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/RemoveEntriesForMissingFiles.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/RemoveEntriesForMissingFiles.java
@@ -57,7 +57,7 @@ import com.beust.jcommander.Parameter;
 
 /**
  * Remove file entries for map files that don't exist.
- * 
+ *
  */
 public class RemoveEntriesForMissingFiles {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/RestoreZookeeper.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/RestoreZookeeper.java b/server/base/src/main/java/org/apache/accumulo/server/util/RestoreZookeeper.java
index b08bf90..e5a2add 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/RestoreZookeeper.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/RestoreZookeeper.java
@@ -40,17 +40,17 @@ import org.xml.sax.helpers.DefaultHandler;
 import com.beust.jcommander.Parameter;
 
 public class RestoreZookeeper {
-  
+
   private static class Restore extends DefaultHandler {
     IZooReaderWriter zk = null;
     Stack<String> cwd = new Stack<String>();
     boolean overwrite = false;
-    
+
     Restore(IZooReaderWriter zk, boolean overwrite) {
       this.zk = zk;
       this.overwrite = overwrite;
     }
-    
+
     @Override
     public void startElement(String uri, String localName, String name, Attributes attributes) throws SAXException {
       if ("node".equals(name)) {
@@ -73,12 +73,12 @@ public class RestoreZookeeper {
         create(root, "", UTF_8.name());
       }
     }
-    
+
     @Override
     public void endElement(String uri, String localName, String name) throws SAXException {
       cwd.pop();
     }
-    
+
     // assume UTF-8 if not "base64"
     private void create(String path, String value, String encoding) {
       byte[] data = value.getBytes(UTF_8);
@@ -97,7 +97,7 @@ public class RestoreZookeeper {
       }
     }
   }
-  
+
   static class Opts extends Help {
     @Parameter(names = {"-z", "--keepers"})
     String keepers = "localhost:2181";
@@ -106,17 +106,17 @@ public class RestoreZookeeper {
     @Parameter(names = "--file")
     String file;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Logger.getRootLogger().setLevel(Level.WARN);
     Opts opts = new Opts();
     opts.parseArgs(RestoreZookeeper.class.getName(), args);
-    
+
     InputStream in = System.in;
     if (opts.file != null) {
       in = new FileInputStream(opts.file);
     }
-    
+
     SAXParserFactory factory = SAXParserFactory.newInstance();
     SAXParser parser = factory.newSAXParser();
     parser.parse(in, new Restore(ZooReaderWriter.getInstance(), opts.overwrite));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/SendLogToChainsaw.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/SendLogToChainsaw.java b/server/base/src/main/java/org/apache/accumulo/server/util/SendLogToChainsaw.java
index 63f1343..2c192cf 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/SendLogToChainsaw.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/SendLogToChainsaw.java
@@ -254,7 +254,7 @@ public class SendLogToChainsaw extends XMLLayout {
   }
 
   /**
-   * 
+   *
    * @param args
    *          <ol>
    *          <li>path to log directory</li>

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/SystemPropUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/SystemPropUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/SystemPropUtil.java
index a376ed6..49a6971 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/SystemPropUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/SystemPropUtil.java
@@ -38,14 +38,14 @@ public class SystemPropUtil {
       log.warn("Ignoring property {} it is null, an invalid format, or not capable of being changed in zookeeper", property);
       return false;
     }
-    
+
     // create the zk node for this property and set it's data to the specified value
     String zPath = ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZCONFIG + "/" + property;
     ZooReaderWriter.getInstance().putPersistentData(zPath, value.getBytes(UTF_8), NodeExistsPolicy.OVERWRITE);
-    
+
     return true;
   }
-  
+
   public static void removeSystemProperty(String property) throws InterruptedException, KeeperException {
     String zPath = ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZCONFIG + "/" + property;
     ZooReaderWriter.getInstance().recursiveDelete(zPath, NodeMissingPolicy.FAIL);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/TableInfoUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/TableInfoUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/TableInfoUtil.java
index cc91ef3..6aa937f 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/TableInfoUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/TableInfoUtil.java
@@ -25,10 +25,10 @@ import org.apache.accumulo.core.master.thrift.TableInfo;
 import org.apache.accumulo.core.master.thrift.TabletServerStatus;
 
 /**
- * 
+ *
  */
 public class TableInfoUtil {
-  
+
   public static void add(TableInfo total, TableInfo more) {
     if (total.minors == null)
       total.minors = new Compacting();
@@ -58,7 +58,7 @@ public class TableInfoUtil {
     total.queryByteRate += more.queryByteRate;
     total.scanRate += more.scanRate;
   }
-  
+
   public static TableInfo summarizeTableStats(TabletServerStatus status) {
     TableInfo summary = new TableInfo();
     summary.majors = new Compacting();
@@ -69,7 +69,7 @@ public class TableInfoUtil {
     }
     return summary;
   }
-  
+
   public static Map<String,Double> summarizeTableStats(MasterMonitorInfo mmi) {
     Map<String,Double> compactingByTable = new HashMap<String,Double>();
     if (mmi != null && mmi.tServerInfo != null) {
@@ -84,5 +84,5 @@ public class TableInfoUtil {
     }
     return compactingByTable;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/TablePropUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/TablePropUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/TablePropUtil.java
index 5e6542d..ab14311 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/TablePropUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/TablePropUtil.java
@@ -31,31 +31,31 @@ public class TablePropUtil {
   public static boolean setTableProperty(String tableId, String property, String value) throws KeeperException, InterruptedException {
     if (!isPropertyValid(property, value))
       return false;
-    
+
     // create the zk node for per-table properties for this table if it doesn't already exist
     String zkTablePath = getTablePath(tableId);
     ZooReaderWriter.getInstance().putPersistentData(zkTablePath, new byte[0], NodeExistsPolicy.SKIP);
-    
+
     // create the zk node for this property and set it's data to the specified value
     String zPath = zkTablePath + "/" + property;
     ZooReaderWriter.getInstance().putPersistentData(zPath, value.getBytes(UTF_8), NodeExistsPolicy.OVERWRITE);
-    
+
     return true;
   }
-  
+
   public static boolean isPropertyValid(String property, String value) {
     Property p = Property.getPropertyByKey(property);
     if ((p != null && !p.getType().isValidFormat(value)) || !Property.isValidTablePropertyKey(property))
       return false;
-    
+
     return true;
   }
-  
+
   public static void removeTableProperty(String tableId, String property) throws InterruptedException, KeeperException {
     String zPath = getTablePath(tableId) + "/" + property;
     ZooReaderWriter.getInstance().recursiveDelete(zPath, NodeMissingPolicy.SKIP);
   }
-  
+
   private static String getTablePath(String tablename) {
     return ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZTABLES + "/" + tablename + Constants.ZTABLE_CONF;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/TabletIterator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/TabletIterator.java b/server/base/src/main/java/org/apache/accumulo/server/util/TabletIterator.java
index 8dd414b..3bc6c96 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/TabletIterator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/TabletIterator.java
@@ -40,56 +40,52 @@ import org.apache.log4j.Logger;
  * This class iterates over the metadata table returning all key values for a tablet in one chunk. As it scans the metadata table it checks the correctness of
  * the metadata table, and rescans if needed. So the tablet key/values returned by this iterator should satisfy the sorted linked list property of the metadata
  * table.
- * 
+ *
  * The purpose of this is to hide inconsistencies caused by splits and detect anomalies in the metadata table.
- * 
+ *
  * If a tablet that was returned by this iterator is subsequently deleted from the metadata table, then this iterator will throw a TabletDeletedException. This
  * could occur when a table is merged.
- * 
- * 
+ *
+ *
  */
 public class TabletIterator implements Iterator<Map<Key,Value>> {
-  
+
   private static final Logger log = Logger.getLogger(TabletIterator.class);
-  
+
   private SortedMap<Key,Value> currentTabletKeys;
-  
+
   private Text lastTablet;
-  
+
   private Scanner scanner;
   private Iterator<Entry<Key,Value>> iter;
-  
+
   private boolean returnPrevEndRow;
-  
+
   private boolean returnDir;
-  
+
   private Range range;
-  
+
   public static class TabletDeletedException extends RuntimeException {
-    
-    /**
-		 * 
-		 */
-    
+
     private static final long serialVersionUID = 1L;
-    
+
     public TabletDeletedException(String msg) {
       super(msg);
     }
   }
-  
+
   /*
    * public TabletIterator(String table, boolean returnPrevEndRow){
-   * 
+   *
    * }
    */
-  
+
   /**
-   * 
+   *
    * @param s
    *          A scanner over the entire metadata table configure to fetch needed columns.
    */
-  
+
   public TabletIterator(Scanner s, Range range, boolean returnPrevEndRow, boolean returnDir) {
     this.scanner = s;
     this.range = range;
@@ -100,129 +96,129 @@ public class TabletIterator implements Iterator<Map<Key,Value>> {
     this.returnPrevEndRow = returnPrevEndRow;
     this.returnDir = returnDir;
   }
-  
+
   @Override
   public boolean hasNext() {
     while (currentTabletKeys == null) {
-      
+
       currentTabletKeys = scanToPrevEndRow();
       if (currentTabletKeys.size() == 0) {
         break;
       }
-      
+
       Key prevEndRowKey = currentTabletKeys.lastKey();
       Value prevEndRowValue = currentTabletKeys.get(prevEndRowKey);
-      
+
       if (!TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.hasColumns(prevEndRowKey)) {
         log.debug(currentTabletKeys);
         throw new RuntimeException("Unexpected key " + prevEndRowKey);
       }
-      
+
       Text per = KeyExtent.decodePrevEndRow(prevEndRowValue);
       Text lastEndRow;
-      
+
       if (lastTablet == null) {
         lastEndRow = null;
       } else {
         lastEndRow = new KeyExtent(lastTablet, (Text) null).getEndRow();
-        
+
         // do table transition sanity check
         String lastTable = new KeyExtent(lastTablet, (Text) null).getTableId().toString();
         String currentTable = new KeyExtent(prevEndRowKey.getRow(), (Text) null).getTableId().toString();
-        
+
         if (!lastTable.equals(currentTable) && (per != null || lastEndRow != null)) {
           log.info("Metadata inconsistency on table transition : " + lastTable + " " + currentTable + " " + per + " " + lastEndRow);
-          
+
           currentTabletKeys = null;
           resetScanner();
-          
+
           UtilWaitThread.sleep(250);
-          
+
           continue;
         }
       }
-      
+
       boolean perEqual = (per == null && lastEndRow == null) || (per != null && lastEndRow != null && per.equals(lastEndRow));
-      
+
       if (!perEqual) {
-        
+
         log.info("Metadata inconsistency : " + per + " != " + lastEndRow + " metadataKey = " + prevEndRowKey);
-        
+
         currentTabletKeys = null;
         resetScanner();
-        
+
         UtilWaitThread.sleep(250);
-        
+
         continue;
-        
+
       }
       // this tablet is good, so set it as the last tablet
       lastTablet = prevEndRowKey.getRow();
     }
-    
+
     return currentTabletKeys.size() > 0;
   }
-  
+
   @Override
   public Map<Key,Value> next() {
-    
+
     if (!hasNext())
       throw new NoSuchElementException();
-    
+
     Map<Key,Value> tmp = currentTabletKeys;
     currentTabletKeys = null;
-    
+
     Set<Entry<Key,Value>> es = tmp.entrySet();
     Iterator<Entry<Key,Value>> esIter = es.iterator();
-    
+
     while (esIter.hasNext()) {
       Map.Entry<Key,Value> entry = esIter.next();
       if (!returnPrevEndRow && TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.hasColumns(entry.getKey())) {
         esIter.remove();
       }
-      
+
       if (!returnDir && TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.hasColumns(entry.getKey())) {
         esIter.remove();
       }
     }
-    
+
     return tmp;
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
   private SortedMap<Key,Value> scanToPrevEndRow() {
-    
+
     Text curMetaDataRow = null;
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     boolean sawPrevEndRow = false;
-    
+
     while (true) {
       while (iter.hasNext()) {
         Entry<Key,Value> entry = iter.next();
-        
+
         if (curMetaDataRow == null) {
           curMetaDataRow = entry.getKey().getRow();
         }
-        
+
         if (!curMetaDataRow.equals(entry.getKey().getRow())) {
           // tablet must not have a prev end row, try scanning again
           break;
         }
-        
+
         tm.put(entry.getKey(), entry.getValue());
-        
+
         if (TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.hasColumns(entry.getKey())) {
           sawPrevEndRow = true;
           break;
         }
       }
-      
+
       if (!sawPrevEndRow && tm.size() > 0) {
         log.warn("Metadata problem : tablet " + curMetaDataRow + " has no prev end row");
         resetScanner();
@@ -233,14 +229,14 @@ public class TabletIterator implements Iterator<Map<Key,Value>> {
         break;
       }
     }
-    
+
     return tm;
   }
-  
+
   protected void resetScanner() {
-    
+
     Range range;
-    
+
     if (lastTablet == null) {
       range = this.range;
     } else {
@@ -252,19 +248,19 @@ public class TabletIterator implements Iterator<Map<Key,Value>> {
       Entry<Key,Value> entry : scanner) {
         count++;
       }
-      
+
       if (count == 0)
         throw new TabletDeletedException("Tablet " + lastTablet + " was deleted while iterating");
-      
+
       // start right after the last good tablet
       range = new Range(new Key(lastTablet).followingKey(PartialKey.ROW), true, this.range.getEndKey(), this.range.isEndKeyInclusive());
     }
-    
+
     log.info("Resetting " + MetadataTable.NAME + " scanner to " + range);
-    
+
     scanner.setRange(range);
     iter = scanner.iterator();
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/TabletOperations.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/TabletOperations.java b/server/base/src/main/java/org/apache/accumulo/server/util/TabletOperations.java
index c0e1a9b..b1e2d35 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/TabletOperations.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/TabletOperations.java
@@ -24,8 +24,8 @@ import org.apache.accumulo.core.util.UtilWaitThread;
 import org.apache.accumulo.server.ServerConstants;
 import org.apache.accumulo.server.fs.VolumeManager;
 import org.apache.accumulo.server.fs.VolumeManagerImpl;
-import org.apache.hadoop.fs.FileSystem;
 import org.apache.accumulo.server.tablets.UniqueNameAllocator;
+import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 import org.apache.log4j.Logger;
@@ -33,15 +33,15 @@ import org.apache.log4j.Logger;
 import com.google.common.base.Optional;
 
 public class TabletOperations {
-  
+
   private static final Logger log = Logger.getLogger(TabletOperations.class);
-  
+
   public static String createTabletDirectory(VolumeManager fs, String tableId, Text endRow) {
     String lowDirectory;
-    
+
     UniqueNameAllocator namer = UniqueNameAllocator.getInstance();
     String volume = fs.choose(Optional.of(tableId), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR;
-    
+
     while (true) {
       try {
         if (endRow == null) {
@@ -54,7 +54,7 @@ public class TabletOperations {
           log.warn("Failed to create " + lowDirectoryPath + " for unknown reason");
         } else {
           lowDirectory = "/" + Constants.GENERATED_TABLET_DIRECTORY_PREFIX + namer.getNextName();
-          Path lowDirectoryPath = new Path(volume + "/" + tableId + "/" +  lowDirectory);
+          Path lowDirectoryPath = new Path(volume + "/" + tableId + "/" + lowDirectory);
           if (fs.exists(lowDirectoryPath))
             throw new IllegalStateException("Dir exist when it should not " + lowDirectoryPath);
           if (fs.mkdirs(lowDirectoryPath)) {
@@ -65,13 +65,13 @@ public class TabletOperations {
       } catch (IOException e) {
         log.warn(e);
       }
-      
+
       log.warn("Failed to create dir for tablet in table " + tableId + " in volume " + volume + " + will retry ...");
       UtilWaitThread.sleep(3000);
-      
+
     }
   }
-  
+
   public static String createTabletDirectory(String tableDir, Text endRow) {
     while (true) {
       try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/TabletServerLocks.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/TabletServerLocks.java b/server/base/src/main/java/org/apache/accumulo/server/util/TabletServerLocks.java
index 307bb0c..01bb926 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/TabletServerLocks.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/TabletServerLocks.java
@@ -33,34 +33,35 @@ import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import com.beust.jcommander.Parameter;
 
 public class TabletServerLocks {
-  
+
   static class Opts extends Help {
-    @Parameter(names="-list")
+    @Parameter(names = "-list")
     boolean list = false;
-    @Parameter(names="-delete")
+    @Parameter(names = "-delete")
     String delete = null;
   }
+
   public static void main(String[] args) throws Exception {
-    
+
     Instance instance = HdfsZooInstance.getInstance();
     String tserverPath = ZooUtil.getRoot(instance) + Constants.ZTSERVERS;
     Opts opts = new Opts();
     opts.parseArgs(TabletServerLocks.class.getName(), args);
-    
+
     ZooCache cache = new ZooCache(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
-    
+
     if (opts.list) {
       IZooReaderWriter zoo = ZooReaderWriter.getInstance();
-      
+
       List<String> tabletServers = zoo.getChildren(tserverPath);
-      
+
       for (String tabletServer : tabletServers) {
         byte[] lockData = ZooLock.getLockData(cache, tserverPath + "/" + tabletServer, null);
         String holder = null;
         if (lockData != null) {
           holder = new String(lockData, UTF_8);
         }
-        
+
         System.out.printf("%32s %16s%n", tabletServer, holder);
       }
     } else if (opts.delete != null) {
@@ -68,7 +69,7 @@ public class TabletServerLocks {
     } else {
       System.out.println("Usage : " + TabletServerLocks.class.getName() + " -list|-delete <tserver lock>");
     }
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/VerifyTabletAssignments.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/VerifyTabletAssignments.java b/server/base/src/main/java/org/apache/accumulo/server/util/VerifyTabletAssignments.java
index 8e6b339..98de6be 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/VerifyTabletAssignments.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/VerifyTabletAssignments.java
@@ -61,36 +61,36 @@ import com.google.common.net.HostAndPort;
 
 public class VerifyTabletAssignments {
   private static final Logger log = Logger.getLogger(VerifyTabletAssignments.class);
-  
+
   static class Opts extends ClientOpts {
     @Parameter(names = {"-v", "--verbose"}, description = "verbose mode (prints locations of tablets)")
     boolean verbose = false;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(VerifyTabletAssignments.class.getName(), args);
-    
+
     ClientContext context = new ClientContext(opts.getInstance(), new Credentials(opts.principal, opts.getToken()), opts.getClientConfiguration());
     Connector conn = opts.getConnector();
     for (String table : conn.tableOperations().list())
       checkTable(context, opts, table, null);
-    
+
   }
-  
+
   private static void checkTable(final ClientContext context, final Opts opts, String tableName, HashSet<KeyExtent> check) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException, InterruptedException {
-    
+
     if (check == null)
       System.out.println("Checking table " + tableName);
     else
       System.out.println("Checking table " + tableName + " again, failures " + check.size());
-    
+
     TreeMap<KeyExtent,String> tabletLocations = new TreeMap<KeyExtent,String>();
-    
+
     String tableId = Tables.getNameToIdMap(context.getInstance()).get(tableName);
     MetadataServicer.forTableId(context, tableId).getTabletLocations(tabletLocations);
-    
+
     final HashSet<KeyExtent> failures = new HashSet<KeyExtent>();
 
     Map<HostAndPort,List<KeyExtent>> extentsPerServer = new TreeMap<HostAndPort,List<KeyExtent>>();
@@ -102,7 +102,7 @@ public class VerifyTabletAssignments {
         System.out.println(" Tablet " + keyExtent + " has no location");
       else if (opts.verbose)
         System.out.println(" Tablet " + keyExtent + " is located at " + loc);
-      
+
       if (loc != null) {
         final HostAndPort parsedLoc = HostAndPort.fromString(loc);
         List<KeyExtent> extentList = extentsPerServer.get(parsedLoc);
@@ -110,35 +110,35 @@ public class VerifyTabletAssignments {
           extentList = new ArrayList<KeyExtent>();
           extentsPerServer.put(parsedLoc, extentList);
         }
-        
+
         if (check == null || check.contains(keyExtent))
           extentList.add(keyExtent);
       }
     }
-    
+
     ExecutorService tp = Executors.newFixedThreadPool(20);
     for (final Entry<HostAndPort,List<KeyExtent>> entry : extentsPerServer.entrySet()) {
       Runnable r = new Runnable() {
-        
+
         @Override
         public void run() {
           try {
             checkTabletServer(context, entry, failures);
           } catch (Exception e) {
-            log.error("Failure on tablet server '"+entry.getKey()+".", e);
+            log.error("Failure on tablet server '" + entry.getKey() + ".", e);
             failures.addAll(entry.getValue());
           }
         }
-        
+
       };
-      
+
       tp.execute(r);
     }
-    
+
     tp.shutdown();
-    
+
     while (!tp.awaitTermination(1, TimeUnit.HOURS)) {}
-    
+
     if (failures.size() > 0)
       checkTable(context, opts, tableName, failures);
   }
@@ -154,32 +154,32 @@ public class VerifyTabletAssignments {
   private static void checkTabletServer(ClientContext context, Entry<HostAndPort,List<KeyExtent>> entry, HashSet<KeyExtent> failures)
       throws ThriftSecurityException, TException, NoSuchScanIDException {
     TabletClientService.Iface client = ThriftUtil.getTServerClient(entry.getKey(), context);
-    
+
     Map<TKeyExtent,List<TRange>> batch = new TreeMap<TKeyExtent,List<TRange>>();
-    
+
     for (KeyExtent keyExtent : entry.getValue()) {
       Text row = keyExtent.getEndRow();
       Text row2 = null;
-      
+
       if (row == null) {
         row = keyExtent.getPrevEndRow();
-        
+
         if (row != null) {
           row = new Text(row);
           row.append(new byte[] {'a'}, 0, 1);
         } else {
           row = new Text("1234567890");
         }
-        
+
         row2 = new Text(row);
         row2.append(new byte[] {'!'}, 0, 1);
       } else {
         row = new Text(row);
         row2 = new Text(row);
-        
+
         row.getBytes()[row.getLength() - 1] = (byte) (row.getBytes()[row.getLength() - 1] - 1);
       }
-      
+
       Range r = new Range(row, true, row2, false);
       batch.put(keyExtent.toThrift(), Collections.singletonList(r.toThrift()));
     }
@@ -192,15 +192,15 @@ public class VerifyTabletAssignments {
     if (is.result.more) {
       MultiScanResult result = client.continueMultiScan(tinfo, is.scanID);
       checkFailures(entry.getKey(), failures, result);
-      
+
       while (result.more) {
         result = client.continueMultiScan(tinfo, is.scanID);
         checkFailures(entry.getKey(), failures, result);
       }
     }
-    
+
     client.closeMultiScan(tinfo, is.scanID);
-    
+
     ThriftUtil.returnClient((TServiceClient) client);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/ZooKeeperMain.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/ZooKeeperMain.java b/server/base/src/main/java/org/apache/accumulo/server/util/ZooKeeperMain.java
index efcefde..0edcf71 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/ZooKeeperMain.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/ZooKeeperMain.java
@@ -27,16 +27,16 @@ import org.apache.hadoop.fs.Path;
 import com.beust.jcommander.Parameter;
 
 public class ZooKeeperMain {
-  
+
   static class Opts extends Help {
-    
+
     @Parameter(names = {"-z", "--keepers"}, description = "Comma separated list of zookeeper hosts (host:port,host:port)")
     String servers = null;
-    
+
     @Parameter(names = {"-t", "--timeout"}, description = "timeout, in seconds to timeout the zookeeper connection")
     long timeout = 30;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(ZooKeeperMain.class.getName(), args);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/ZooZap.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/ZooZap.java b/server/base/src/main/java/org/apache/accumulo/server/util/ZooZap.java
index 7fdbf13..ef182f1 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/ZooZap.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/ZooZap.java
@@ -20,67 +20,66 @@ import java.util.List;
 
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
-import org.apache.accumulo.server.cli.ClientOpts;
 import org.apache.accumulo.fate.zookeeper.IZooReaderWriter;
 import org.apache.accumulo.fate.zookeeper.ZooUtil.NodeMissingPolicy;
+import org.apache.accumulo.server.cli.ClientOpts;
 import org.apache.accumulo.server.zookeeper.ZooLock;
 import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
+import org.apache.log4j.Logger;
 
 import com.beust.jcommander.JCommander;
 import com.beust.jcommander.Parameter;
-import org.apache.log4j.Logger;
 
 public class ZooZap {
   private static final Logger log = Logger.getLogger(ZooZap.class);
-  
+
   static boolean verbose = false;
-  
+
   private static void message(String msg) {
     if (verbose)
       System.out.println(msg);
   }
-  
+
   static class Opts extends ClientOpts {
-    @Parameter(names="-master", description="remove master locks")
+    @Parameter(names = "-master", description = "remove master locks")
     boolean zapMaster = false;
-    @Parameter(names="-tservers", description="remove tablet server locks")
+    @Parameter(names = "-tservers", description = "remove tablet server locks")
     boolean zapTservers = false;
-    @Parameter(names="-tracers", description="remove tracer locks")
+    @Parameter(names = "-tracers", description = "remove tracer locks")
     boolean zapTracers = false;
-    @Parameter(names="-verbose", description="print out messages about progress")
+    @Parameter(names = "-verbose", description = "print out messages about progress")
     boolean verbose = false;
 
     String getTraceZKPath() {
       return super.getClientConfiguration().get(ClientProperty.TRACE_ZK_PATH);
     }
   }
-  
+
   public static void main(String[] args) {
     Opts opts = new Opts();
     opts.parseArgs(ZooZap.class.getName(), args);
-    
-    if (!opts.zapMaster && !opts.zapTservers && !opts.zapTracers)
-    {
-        new JCommander(opts).usage();
-        return;
+
+    if (!opts.zapMaster && !opts.zapTservers && !opts.zapTracers) {
+      new JCommander(opts).usage();
+      return;
     }
-    
+
     String iid = opts.getInstance().getInstanceID();
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
-    
+
     if (opts.zapMaster) {
       String masterLockPath = Constants.ZROOT + "/" + iid + Constants.ZMASTER_LOCK;
-      
+
       zapDirectory(zoo, masterLockPath);
     }
-    
+
     if (opts.zapTservers) {
       String tserversPath = Constants.ZROOT + "/" + iid + Constants.ZTSERVERS;
       try {
         List<String> children = zoo.getChildren(tserversPath);
         for (String child : children) {
           message("Deleting " + tserversPath + "/" + child + " from zookeeper");
-          
+
           if (opts.zapMaster)
             ZooReaderWriter.getInstance().recursiveDelete(tserversPath + "/" + child, NodeMissingPolicy.SKIP);
           else {
@@ -96,14 +95,14 @@ public class ZooZap {
         log.error(e);
       }
     }
-    
+
     if (opts.zapTracers) {
       String path = opts.getTraceZKPath();
       zapDirectory(zoo, path);
     }
-    
+
   }
-  
+
   private static void zapDirectory(IZooReaderWriter zoo, String path) {
     try {
       List<String> children = zoo.getChildren(path);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/time/BaseRelativeTime.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/time/BaseRelativeTime.java b/server/base/src/main/java/org/apache/accumulo/server/util/time/BaseRelativeTime.java
index 393c6d2..73afc7e 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/time/BaseRelativeTime.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/time/BaseRelativeTime.java
@@ -18,25 +18,25 @@ package org.apache.accumulo.server.util.time;
 
 /**
  * Provide time from a local source and a hint from a time source.
- * 
+ *
  * RelativeTime and BaseRelativeTime are separated to provide unit tests of the core functionality of Relative timekeeping.
- * 
+ *
  */
 public class BaseRelativeTime implements ProvidesTime {
-  
+
   private long diff = 0;
   private long lastReportedTime = 0;
   ProvidesTime local;
-  
+
   BaseRelativeTime(ProvidesTime real, long lastReportedTime) {
     this.local = real;
     this.lastReportedTime = lastReportedTime;
   }
-  
+
   BaseRelativeTime(ProvidesTime real) {
     this(real, 0);
   }
-  
+
   @Override
   synchronized public long currentTime() {
     long localNow = local.currentTime();
@@ -46,12 +46,12 @@ public class BaseRelativeTime implements ProvidesTime {
     lastReportedTime = result;
     return result;
   }
-  
+
   synchronized public void updateTime(long advice) {
     long localNow = local.currentTime();
     long diff = advice - localNow;
     // smooth in 20% of the change, not the whole thing.
     this.diff = (this.diff * 4 / 5) + diff / 5;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/time/ProvidesTime.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/time/ProvidesTime.java b/server/base/src/main/java/org/apache/accumulo/server/util/time/ProvidesTime.java
index 117fa5f..1042c32 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/time/ProvidesTime.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/time/ProvidesTime.java
@@ -18,7 +18,7 @@ package org.apache.accumulo.server.util.time;
 
 /**
  * An interface for anything that returns the time in the same format as System.currentTimeMillis().
- * 
+ *
  */
 public interface ProvidesTime {
   long currentTime();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/time/RelativeTime.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/time/RelativeTime.java b/server/base/src/main/java/org/apache/accumulo/server/util/time/RelativeTime.java
index 99581e9..bc48b10 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/time/RelativeTime.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/time/RelativeTime.java
@@ -18,27 +18,27 @@ package org.apache.accumulo.server.util.time;
 
 /**
  * Provide time from System time and hints from another time source.
- * 
+ *
  * Provides a convenient static replacement for System.currentTimeMillis()
  */
 public class RelativeTime extends BaseRelativeTime {
-  
+
   private RelativeTime() {
     super(new SystemTime());
   }
-  
+
   private static BaseRelativeTime instance = new RelativeTime();
-  
+
   public static BaseRelativeTime getInstance() {
     return instance;
   }
-  
+
   public static void setInstance(BaseRelativeTime newInstance) {
     instance = newInstance;
   }
-  
+
   public static long currentTimeMillis() {
     return getInstance().currentTime();
   }
-  
+
 }


[06/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/performance/metadata/MetadataBatchScanTest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/performance/metadata/MetadataBatchScanTest.java b/test/src/main/java/org/apache/accumulo/test/performance/metadata/MetadataBatchScanTest.java
index a8b219c..979a6b0 100644
--- a/test/src/main/java/org/apache/accumulo/test/performance/metadata/MetadataBatchScanTest.java
+++ b/test/src/main/java/org/apache/accumulo/test/performance/metadata/MetadataBatchScanTest.java
@@ -52,96 +52,96 @@ import org.apache.log4j.Logger;
 import com.google.common.net.HostAndPort;
 
 /**
- * This little program can be used to write a lot of metadata entries and measure the performance of varying numbers of threads doing metadata
- * lookups using the batch scanner.
- * 
- * 
+ * This little program can be used to write a lot of metadata entries and measure the performance of varying numbers of threads doing metadata lookups using the
+ * batch scanner.
+ *
+ *
  */
 
 public class MetadataBatchScanTest {
-  
+
   private static final Logger log = Logger.getLogger(MetadataBatchScanTest.class);
-  
+
   public static void main(String[] args) throws Exception {
-    
+
     ClientOpts opts = new ClientOpts();
     opts.parseArgs(MetadataBatchScanTest.class.getName(), args);
     Instance inst = new ZooKeeperInstance(new ClientConfiguration().withInstance("acu14").withZkHosts("localhost"));
     final Connector connector = inst.getConnector(opts.principal, opts.getToken());
-    
+
     TreeSet<Long> splits = new TreeSet<Long>();
     Random r = new Random(42);
-    
+
     while (splits.size() < 99999) {
       splits.add((r.nextLong() & 0x7fffffffffffffffl) % 1000000000000l);
     }
-    
+
     Text tid = new Text("8");
     Text per = null;
-    
+
     ArrayList<KeyExtent> extents = new ArrayList<KeyExtent>();
-    
+
     for (Long split : splits) {
       Text er = new Text(String.format("%012d", split));
       KeyExtent ke = new KeyExtent(tid, er, per);
       per = er;
-      
+
       extents.add(ke);
     }
-    
+
     extents.add(new KeyExtent(tid, null, per));
-    
+
     if (args[0].equals("write")) {
-      
+
       BatchWriter bw = connector.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-      
+
       for (KeyExtent extent : extents) {
         Mutation mut = extent.getPrevRowUpdateMutation();
         new TServerInstance(HostAndPort.fromParts("192.168.1.100", 4567), "DEADBEEF").putLocation(mut);
         bw.addMutation(mut);
       }
-      
+
       bw.close();
     } else if (args[0].equals("writeFiles")) {
       BatchWriter bw = connector.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-      
+
       for (KeyExtent extent : extents) {
-        
+
         Mutation mut = new Mutation(extent.getMetadataEntry());
-        
+
         String dir = "/t-" + UUID.randomUUID();
-        
+
         TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(mut, new Value(dir.getBytes(UTF_8)));
-        
+
         for (int i = 0; i < 5; i++) {
           mut.put(DataFileColumnFamily.NAME, new Text(dir + "/00000_0000" + i + ".map"), new Value("10000,1000000".getBytes(UTF_8)));
         }
-        
+
         bw.addMutation(mut);
       }
-      
+
       bw.close();
     } else if (args[0].equals("scan")) {
-      
+
       int numThreads = Integer.parseInt(args[1]);
       final int numLoop = Integer.parseInt(args[2]);
       int numLookups = Integer.parseInt(args[3]);
-      
+
       HashSet<Integer> indexes = new HashSet<Integer>();
       while (indexes.size() < numLookups) {
         indexes.add(r.nextInt(extents.size()));
       }
-      
+
       final List<Range> ranges = new ArrayList<Range>();
       for (Integer i : indexes) {
         ranges.add(extents.get(i).toMetadataRange());
       }
-      
+
       Thread threads[] = new Thread[numThreads];
-      
+
       for (int i = 0; i < threads.length; i++) {
         threads[i] = new Thread(new Runnable() {
-          
+
           @Override
           public void run() {
             try {
@@ -152,79 +152,79 @@ public class MetadataBatchScanTest {
           }
         });
       }
-      
+
       long t1 = System.currentTimeMillis();
-      
+
       for (int i = 0; i < threads.length; i++) {
         threads[i].start();
       }
-      
+
       for (int i = 0; i < threads.length; i++) {
         threads[i].join();
       }
-      
+
       long t2 = System.currentTimeMillis();
-      
+
       System.out.printf("tt : %6.2f%n", (t2 - t1) / 1000.0);
-      
+
     } else {
       throw new IllegalArgumentException();
     }
-    
+
   }
-  
+
   private static ScanStats runScanTest(Connector connector, int numLoop, List<Range> ranges) throws Exception {
     Scanner scanner = null;
-    
+
     BatchScanner bs = connector.createBatchScanner(MetadataTable.NAME, Authorizations.EMPTY, 1);
     bs.fetchColumnFamily(TabletsSection.CurrentLocationColumnFamily.NAME);
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.fetch(bs);
-    
+
     bs.setRanges(ranges);
-    
+
     // System.out.println(ranges);
-    
+
     ScanStats stats = new ScanStats();
     for (int i = 0; i < numLoop; i++) {
       ScanStat ss = scan(bs, ranges, scanner);
       stats.merge(ss);
     }
-    
+
     return stats;
   }
-  
+
   private static class ScanStat {
     long delta1;
     long delta2;
     int count1;
     int count2;
   }
-  
+
   private static class ScanStats {
     Stat delta1 = new Stat();
     Stat delta2 = new Stat();
     Stat count1 = new Stat();
     Stat count2 = new Stat();
-    
+
     void merge(ScanStat ss) {
       delta1.addStat(ss.delta1);
       delta2.addStat(ss.delta2);
       count1.addStat(ss.count1);
       count2.addStat(ss.count2);
     }
-    
+
     @Override
     public String toString() {
       return "[" + delta1 + "] [" + delta2 + "]";
     }
   }
-  
+
   private static ScanStat scan(BatchScanner bs, List<Range> ranges, Scanner scanner) {
-    
+
     // System.out.println("ranges : "+ranges);
-    
+
     ScanStat ss = new ScanStat();
-    
+
     long t1 = System.currentTimeMillis();
     int count = 0;
     for (@SuppressWarnings("unused")
@@ -233,22 +233,22 @@ public class MetadataBatchScanTest {
     }
     bs.close();
     long t2 = System.currentTimeMillis();
-    
+
     ss.delta1 = t2 - t1;
     ss.count1 = count;
-    
+
     count = 0;
     t1 = System.currentTimeMillis();
     /*
      * for (Range range : ranges) { scanner.setRange(range); for (Entry<Key, Value> entry : scanner) { count++; } }
      */
-    
+
     t2 = System.currentTimeMillis();
-    
+
     ss.delta2 = t2 - t1;
     ss.count2 = count;
-    
+
     return ss;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/performance/scan/CollectTabletStats.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/performance/scan/CollectTabletStats.java b/test/src/main/java/org/apache/accumulo/test/performance/scan/CollectTabletStats.java
index 15c9861..3a45215 100644
--- a/test/src/main/java/org/apache/accumulo/test/performance/scan/CollectTabletStats.java
+++ b/test/src/main/java/org/apache/accumulo/test/performance/scan/CollectTabletStats.java
@@ -82,7 +82,7 @@ import com.google.common.net.HostAndPort;
 
 public class CollectTabletStats {
   private static final Logger log = Logger.getLogger(CollectTabletStats.class);
-  
+
   static class CollectOptions extends ClientOnRequiredTable {
     @Parameter(names = "--iterations", description = "number of iterations")
     int iterations = 3;
@@ -93,69 +93,69 @@ public class CollectTabletStats {
     @Parameter(names = "-c", description = "comma separated list of columns")
     String columns;
   }
-  
+
   public static void main(String[] args) throws Exception {
-    
+
     final CollectOptions opts = new CollectOptions();
     final ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(CollectTabletStats.class.getName(), args, scanOpts);
-    
+
     String columnsTmp[] = new String[] {};
     if (opts.columns != null)
       columnsTmp = opts.columns.split(",");
     final String columns[] = columnsTmp;
-    
+
     final VolumeManager fs = VolumeManagerImpl.get();
-    
+
     Instance instance = opts.getInstance();
     final ServerConfigurationFactory sconf = new ServerConfigurationFactory(instance);
     Credentials creds = new Credentials(opts.principal, opts.getToken());
     ClientContext context = new ClientContext(instance, creds, sconf.getConfiguration());
-    
+
     String tableId = Tables.getNameToIdMap(instance).get(opts.getTableName());
     if (tableId == null) {
       log.error("Unable to find table named " + opts.getTableName());
       System.exit(-1);
     }
-    
+
     TreeMap<KeyExtent,String> tabletLocations = new TreeMap<KeyExtent,String>();
     List<KeyExtent> candidates = findTablets(context, !opts.selectFarTablets, opts.getTableName(), tabletLocations);
-    
+
     if (candidates.size() < opts.numThreads) {
       System.err.println("ERROR : Unable to find " + opts.numThreads + " " + (opts.selectFarTablets ? "far" : "local") + " tablets");
       System.exit(-1);
     }
-    
+
     List<KeyExtent> tabletsToTest = selectRandomTablets(opts.numThreads, candidates);
-    
+
     Map<KeyExtent,List<FileRef>> tabletFiles = new HashMap<KeyExtent,List<FileRef>>();
-    
+
     for (KeyExtent ke : tabletsToTest) {
       List<FileRef> files = getTabletFiles(context, tableId, ke);
       tabletFiles.put(ke, files);
     }
-    
+
     System.out.println();
     System.out.println("run location      : " + InetAddress.getLocalHost().getHostName() + "/" + InetAddress.getLocalHost().getHostAddress());
     System.out.println("num threads       : " + opts.numThreads);
     System.out.println("table             : " + opts.getTableName());
     System.out.println("table id          : " + tableId);
-    
+
     for (KeyExtent ke : tabletsToTest) {
       System.out.println("\t *** Information about tablet " + ke.getUUID() + " *** ");
       System.out.println("\t\t# files in tablet : " + tabletFiles.get(ke).size());
       System.out.println("\t\ttablet location   : " + tabletLocations.get(ke));
       reportHdfsBlockLocations(tabletFiles.get(ke));
     }
-    
+
     System.out.println("%n*** RUNNING TEST ***%n");
-    
+
     ExecutorService threadPool = Executors.newFixedThreadPool(opts.numThreads);
-    
+
     for (int i = 0; i < opts.iterations; i++) {
-      
+
       ArrayList<Test> tests = new ArrayList<Test>();
-      
+
       for (final KeyExtent ke : tabletsToTest) {
         final List<FileRef> files = tabletFiles.get(ke);
         Test test = new Test(ke) {
@@ -163,19 +163,19 @@ public class CollectTabletStats {
           public int runTest() throws Exception {
             return readFiles(fs, sconf.getConfiguration(), files, ke, columns);
           }
-          
+
         };
-        
+
         tests.add(test);
       }
-      
+
       runTest("read files", tests, opts.numThreads, threadPool);
     }
-    
+
     for (int i = 0; i < opts.iterations; i++) {
-      
+
       ArrayList<Test> tests = new ArrayList<Test>();
-      
+
       for (final KeyExtent ke : tabletsToTest) {
         final List<FileRef> files = tabletFiles.get(ke);
         Test test = new Test(ke) {
@@ -184,16 +184,16 @@ public class CollectTabletStats {
             return readFilesUsingIterStack(fs, sconf, files, opts.auths, ke, columns, false);
           }
         };
-        
+
         tests.add(test);
       }
-      
+
       runTest("read tablet files w/ system iter stack", tests, opts.numThreads, threadPool);
     }
-    
+
     for (int i = 0; i < opts.iterations; i++) {
       ArrayList<Test> tests = new ArrayList<Test>();
-      
+
       for (final KeyExtent ke : tabletsToTest) {
         final List<FileRef> files = tabletFiles.get(ke);
         Test test = new Test(ke) {
@@ -202,19 +202,19 @@ public class CollectTabletStats {
             return readFilesUsingIterStack(fs, sconf, files, opts.auths, ke, columns, true);
           }
         };
-        
+
         tests.add(test);
       }
-      
+
       runTest("read tablet files w/ table iter stack", tests, opts.numThreads, threadPool);
     }
-    
+
     for (int i = 0; i < opts.iterations; i++) {
-      
+
       ArrayList<Test> tests = new ArrayList<Test>();
-      
+
       final Connector conn = opts.getConnector();
-      
+
       for (final KeyExtent ke : tabletsToTest) {
         Test test = new Test(ke) {
           @Override
@@ -222,16 +222,16 @@ public class CollectTabletStats {
             return scanTablet(conn, opts.getTableName(), opts.auths, scanOpts.scanBatchSize, ke.getPrevEndRow(), ke.getEndRow(), columns);
           }
         };
-        
+
         tests.add(test);
       }
-      
+
       runTest("read tablet data through accumulo", tests, opts.numThreads, threadPool);
     }
-    
+
     for (final KeyExtent ke : tabletsToTest) {
       final Connector conn = opts.getConnector();
-      
+
       threadPool.submit(new Runnable() {
         @Override
         public void run() {
@@ -243,123 +243,123 @@ public class CollectTabletStats {
         }
       });
     }
-    
+
     threadPool.shutdown();
   }
-  
+
   private static abstract class Test implements Runnable {
-    
+
     private int count;
     private long t1;
     private long t2;
     private CountDownLatch startCdl, finishCdl;
     private KeyExtent ke;
-    
+
     Test(KeyExtent ke) {
       this.ke = ke;
     }
-    
+
     public abstract int runTest() throws Exception;
-    
+
     void setSignals(CountDownLatch scdl, CountDownLatch fcdl) {
       this.startCdl = scdl;
       this.finishCdl = fcdl;
     }
-    
+
     @Override
     public void run() {
-      
+
       try {
         startCdl.await();
       } catch (InterruptedException e) {
         log.error("startCdl.await() failed.", e);
       }
-      
+
       t1 = System.currentTimeMillis();
-      
+
       try {
         count = runTest();
       } catch (Exception e) {
         log.error("runTest() failed.", e);
       }
-      
+
       t2 = System.currentTimeMillis();
-      
+
       double time = (t2 - t1) / 1000.0;
-      
+
       System.out.printf("\t\ttablet: " + ke.getUUID() + "  thread: " + Thread.currentThread().getId()
           + " count: %,d cells  time: %6.2f  rate: %,6.2f cells/sec%n", count, time, count / time);
-      
+
       finishCdl.countDown();
     }
-    
+
     int getCount() {
       return count;
     }
-    
+
     long getStartTime() {
       return t1;
     }
-    
+
     long getFinishTime() {
       return t2;
     }
-    
+
   }
-  
+
   private static void runTest(String desc, List<Test> tests, int numThreads, ExecutorService threadPool) throws Exception {
-    
+
     System.out.println("\tRunning test : " + desc);
-    
+
     CountDownLatch startSignal = new CountDownLatch(1);
     CountDownLatch finishedSignal = new CountDownLatch(numThreads);
-    
+
     for (Test test : tests) {
       threadPool.submit(test);
       test.setSignals(startSignal, finishedSignal);
     }
-    
+
     startSignal.countDown();
-    
+
     finishedSignal.await();
-    
+
     long minTime = Long.MAX_VALUE;
     long maxTime = Long.MIN_VALUE;
     long count = 0;
-    
+
     for (Test test : tests) {
       minTime = Math.min(test.getStartTime(), minTime);
       maxTime = Math.max(test.getFinishTime(), maxTime);
       count += test.getCount();
     }
-    
+
     double time = (maxTime - minTime) / 1000.0;
     System.out.printf("\tAggregate stats  count: %,d cells  time: %6.2f  rate: %,6.2f cells/sec%n", count, time, count / time);
     System.out.println();
-    
+
     // run the gc between test so that object created during previous test are not
     // collected in following test
     System.gc();
     System.gc();
     System.gc();
-    
+
   }
-  
-  private static List<KeyExtent> findTablets(ClientContext context, boolean selectLocalTablets, String tableName,
-      SortedMap<KeyExtent,String> tabletLocations) throws Exception {
-    
+
+  private static List<KeyExtent> findTablets(ClientContext context, boolean selectLocalTablets, String tableName, SortedMap<KeyExtent,String> tabletLocations)
+      throws Exception {
+
     String tableId = Tables.getNameToIdMap(context.getInstance()).get(tableName);
     MetadataServicer.forTableId(context, tableId).getTabletLocations(tabletLocations);
-    
+
     InetAddress localaddress = InetAddress.getLocalHost();
-    
+
     List<KeyExtent> candidates = new ArrayList<KeyExtent>();
-    
+
     for (Entry<KeyExtent,String> entry : tabletLocations.entrySet()) {
       String loc = entry.getValue();
       if (loc != null) {
         boolean isLocal = HostAndPort.fromString(entry.getValue()).getHostText().equals(localaddress.getHostName());
-        
+
         if (selectLocalTablets && isLocal) {
           candidates.add(entry.getKey());
         } else if (!selectLocalTablets && !isLocal) {
@@ -369,10 +369,10 @@ public class CollectTabletStats {
     }
     return candidates;
   }
-  
+
   private static List<KeyExtent> selectRandomTablets(int numThreads, List<KeyExtent> candidates) {
     List<KeyExtent> tabletsToTest = new ArrayList<KeyExtent>();
-    
+
     Random rand = new Random();
     for (int i = 0; i < numThreads; i++) {
       int rindex = rand.nextInt(candidates.size());
@@ -382,29 +382,29 @@ public class CollectTabletStats {
     }
     return tabletsToTest;
   }
-  
+
   private static List<FileRef> getTabletFiles(ClientContext context, String tableId, KeyExtent ke) throws IOException {
     return new ArrayList<FileRef>(MetadataTableUtil.getDataFileSizes(ke, context).keySet());
   }
 
-  //TODO Remove deprecation warning suppression when Hadoop1 support is dropped
+  // TODO Remove deprecation warning suppression when Hadoop1 support is dropped
   @SuppressWarnings("deprecation")
   private static void reportHdfsBlockLocations(List<FileRef> files) throws Exception {
     VolumeManager fs = VolumeManagerImpl.get();
-    
+
     System.out.println("\t\tFile block report : ");
     for (FileRef file : files) {
       FileStatus status = fs.getFileStatus(file.path());
-      
+
       if (status.isDir()) {
         // assume it is a map file
         status = fs.getFileStatus(new Path(file + "/data"));
       }
       FileSystem ns = fs.getVolumeByPath(file.path()).getFileSystem();
       BlockLocation[] locs = ns.getFileBlockLocations(status, 0, status.getLen());
-      
+
       System.out.println("\t\t\tBlocks for : " + file);
-      
+
       for (BlockLocation blockLocation : locs) {
         System.out.printf("\t\t\t\t offset : %,13d  hosts :", blockLocation.getOffset());
         for (String host : blockLocation.getHosts()) {
@@ -413,39 +413,39 @@ public class CollectTabletStats {
         System.out.println();
       }
     }
-    
+
     System.out.println();
-    
+
   }
-  
+
   private static SortedKeyValueIterator<Key,Value> createScanIterator(KeyExtent ke, Collection<SortedKeyValueIterator<Key,Value>> mapfiles,
       Authorizations authorizations, byte[] defaultLabels, HashSet<Column> columnSet, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio,
       boolean useTableIterators, TableConfiguration conf) throws IOException {
-    
+
     SortedMapIterator smi = new SortedMapIterator(new TreeMap<Key,Value>());
-    
+
     List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(mapfiles.size() + 1);
-    
+
     iters.addAll(mapfiles);
     iters.add(smi);
-    
+
     MultiIterator multiIter = new MultiIterator(iters, ke);
     DeletingIterator delIter = new DeletingIterator(multiIter, false);
     ColumnFamilySkippingIterator cfsi = new ColumnFamilySkippingIterator(delIter);
     ColumnQualifierFilter colFilter = new ColumnQualifierFilter(cfsi, columnSet);
     VisibilityFilter visFilter = new VisibilityFilter(colFilter, authorizations, defaultLabels);
-    
+
     if (useTableIterators)
       return IteratorUtil.loadIterators(IteratorScope.scan, visFilter, ke, conf, ssiList, ssio, null);
     return visFilter;
   }
-  
+
   private static int readFiles(VolumeManager fs, AccumuloConfiguration aconf, List<FileRef> files, KeyExtent ke, String[] columns) throws Exception {
-    
+
     int count = 0;
-    
+
     HashSet<ByteSequence> columnSet = createColumnBSS(columns);
-    
+
     for (FileRef file : files) {
       FileSystem ns = fs.getVolumeByPath(file.path()).getFileSystem();
       FileSKVIterator reader = FileOperations.getInstance().openReader(file.path().toString(), false, ns, ns.getConf(), aconf);
@@ -457,10 +457,10 @@ public class CollectTabletStats {
       }
       reader.close();
     }
-    
+
     return count;
   }
-  
+
   private static HashSet<ByteSequence> createColumnBSS(String[] columns) {
     HashSet<ByteSequence> columnSet = new HashSet<ByteSequence>();
     for (String c : columns) {
@@ -468,106 +468,106 @@ public class CollectTabletStats {
     }
     return columnSet;
   }
-  
+
   private static int readFilesUsingIterStack(VolumeManager fs, ServerConfigurationFactory aconf, List<FileRef> files, Authorizations auths, KeyExtent ke,
       String[] columns, boolean useTableIterators) throws Exception {
-    
+
     SortedKeyValueIterator<Key,Value> reader;
-    
+
     List<SortedKeyValueIterator<Key,Value>> readers = new ArrayList<SortedKeyValueIterator<Key,Value>>(files.size());
-    
+
     for (FileRef file : files) {
       FileSystem ns = fs.getVolumeByPath(file.path()).getFileSystem();
       readers.add(FileOperations.getInstance().openReader(file.path().toString(), false, ns, ns.getConf(), aconf.getConfiguration()));
     }
-    
+
     List<IterInfo> emptyIterinfo = Collections.emptyList();
     Map<String,Map<String,String>> emptySsio = Collections.emptyMap();
     TableConfiguration tconf = aconf.getTableConfiguration(ke.getTableId().toString());
     reader = createScanIterator(ke, readers, auths, new byte[] {}, new HashSet<Column>(), emptyIterinfo, emptySsio, useTableIterators, tconf);
-    
+
     HashSet<ByteSequence> columnSet = createColumnBSS(columns);
-    
+
     reader.seek(new Range(ke.getPrevEndRow(), false, ke.getEndRow(), true), columnSet, columnSet.size() == 0 ? false : true);
-    
+
     int count = 0;
-    
+
     while (reader.hasTop()) {
       count++;
       reader.next();
     }
-    
+
     return count;
-    
+
   }
-  
+
   private static int scanTablet(Connector conn, String table, Authorizations auths, int batchSize, Text prevEndRow, Text endRow, String[] columns)
       throws Exception {
-    
+
     Scanner scanner = conn.createScanner(table, auths);
     scanner.setBatchSize(batchSize);
     scanner.setRange(new Range(prevEndRow, false, endRow, true));
-    
+
     for (String c : columns) {
       scanner.fetchColumnFamily(new Text(c));
     }
-    
+
     int count = 0;
-    
+
     for (Entry<Key,Value> entry : scanner) {
       if (entry != null)
         count++;
     }
-    
+
     return count;
   }
-  
+
   private static void calcTabletStats(Connector conn, String table, Authorizations auths, int batchSize, KeyExtent ke, String[] columns) throws Exception {
-    
+
     // long t1 = System.currentTimeMillis();
-    
+
     Scanner scanner = conn.createScanner(table, auths);
     scanner.setBatchSize(batchSize);
     scanner.setRange(new Range(ke.getPrevEndRow(), false, ke.getEndRow(), true));
-    
+
     for (String c : columns) {
       scanner.fetchColumnFamily(new Text(c));
     }
-    
+
     Stat rowLen = new Stat();
     Stat cfLen = new Stat();
     Stat cqLen = new Stat();
     Stat cvLen = new Stat();
     Stat valLen = new Stat();
     Stat colsPerRow = new Stat();
-    
+
     Text lastRow = null;
     int colsPerRowCount = 0;
-    
+
     for (Entry<Key,Value> entry : scanner) {
-      
+
       Key key = entry.getKey();
       Text row = key.getRow();
-      
+
       if (lastRow == null) {
         lastRow = row;
       }
-      
+
       if (!lastRow.equals(row)) {
         colsPerRow.addStat(colsPerRowCount);
         lastRow = row;
         colsPerRowCount = 0;
       }
-      
+
       colsPerRowCount++;
-      
+
       rowLen.addStat(row.getLength());
       cfLen.addStat(key.getColumnFamilyData().length());
       cqLen.addStat(key.getColumnQualifierData().length());
       cvLen.addStat(key.getColumnVisibilityData().length());
       valLen.addStat(entry.getValue().get().length);
     }
-    
+
     synchronized (System.out) {
       System.out.println("");
       System.out.println("\tTablet " + ke.getUUID() + " statistics : ");
@@ -579,13 +579,13 @@ public class CollectTabletStats {
       printStat("Columns per row", colsPerRow);
       System.out.println("");
     }
-    
+
   }
-  
+
   private static void printStat(String desc, Stat s) {
     System.out.printf("\t\tDescription: [%30s]  average: %,6.2f  std dev: %,6.2f  min: %,d  max: %,d %n", desc, s.getAverage(), s.getStdDev(), s.getMin(),
         s.getMax());
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/performance/thrift/NullTserver.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/performance/thrift/NullTserver.java b/test/src/main/java/org/apache/accumulo/test/performance/thrift/NullTserver.java
index 2ebc2e3..0afa243 100644
--- a/test/src/main/java/org/apache/accumulo/test/performance/thrift/NullTserver.java
+++ b/test/src/main/java/org/apache/accumulo/test/performance/thrift/NullTserver.java
@@ -254,7 +254,8 @@ public class NullTserver {
     TransactionWatcher watcher = new TransactionWatcher();
     ThriftClientHandler tch = new ThriftClientHandler(new AccumuloServerContext(new ServerConfigurationFactory(HdfsZooInstance.getInstance())), watcher);
     Processor<Iface> processor = new Processor<Iface>(tch);
-    TServerUtils.startTServer(context.getConfiguration(), HostAndPort.fromParts("0.0.0.0", opts.port), processor, "NullTServer", "null tserver", 2, 1, 1000, 10 * 1024 * 1024, null, -1);
+    TServerUtils.startTServer(context.getConfiguration(), HostAndPort.fromParts("0.0.0.0", opts.port), processor, "NullTServer", "null tserver", 2, 1, 1000,
+        10 * 1024 * 1024, null, -1);
 
     HostAndPort addr = HostAndPort.fromParts(InetAddress.getLocalHost().getHostName(), opts.port);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/Environment.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/Environment.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/Environment.java
index 72792ef..92f6427 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/Environment.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/Environment.java
@@ -77,7 +77,7 @@ public class Environment {
 
   /**
    * Creates a new test environment.
-   * 
+   *
    * @param p
    *          configuration properties
    * @throws NullPointerException
@@ -90,7 +90,7 @@ public class Environment {
 
   /**
    * Gets a copy of the configuration properties.
-   * 
+   *
    * @return a copy of the configuration properties
    */
   Properties copyConfigProperties() {
@@ -99,7 +99,7 @@ public class Environment {
 
   /**
    * Gets a configuration property.
-   * 
+   *
    * @param key
    *          key
    * @return property value
@@ -110,7 +110,7 @@ public class Environment {
 
   /**
    * Gets the configured username.
-   * 
+   *
    * @return username
    */
   public String getUserName() {
@@ -119,7 +119,7 @@ public class Environment {
 
   /**
    * Gets the configured password.
-   * 
+   *
    * @return password
    */
   public String getPassword() {
@@ -137,7 +137,7 @@ public class Environment {
 
   /**
    * Gets an authentication token based on the configured password.
-   * 
+   *
    * @return authentication token
    */
   public AuthenticationToken getToken() {
@@ -146,7 +146,7 @@ public class Environment {
 
   /**
    * Gets an Accumulo instance object. The same instance is reused after the first call.
-   * 
+   *
    * @return instance
    */
   public Instance getInstance() {
@@ -160,7 +160,7 @@ public class Environment {
 
   /**
    * Gets an Accumulo connector. The same connector is reused after the first call.
-   * 
+   *
    * @return connector
    */
   public Connector getConnector() throws AccumuloException, AccumuloSecurityException {
@@ -172,7 +172,7 @@ public class Environment {
 
   /**
    * Gets a multitable batch writer. The same object is reused after the first call unless it is reset.
-   * 
+   *
    * @return multitable batch writer
    * @throws NumberFormatException
    *           if any of the numeric batch writer configuration properties cannot be parsed
@@ -192,7 +192,7 @@ public class Environment {
 
   /**
    * Checks if a multitable batch writer has been created by this wrapper.
-   * 
+   *
    * @return true if multitable batch writer is already created
    */
   public boolean isMultiTableBatchWriterInitialized() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/Fixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/Fixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/Fixture.java
index 3f18201..f8d01d9 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/Fixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/Fixture.java
@@ -19,10 +19,10 @@ package org.apache.accumulo.test.randomwalk;
 import org.apache.log4j.Logger;
 
 public abstract class Fixture {
-  
+
   protected final Logger log = Logger.getLogger(this.getClass());
-  
+
   public abstract void setUp(State state, Environment env) throws Exception;
-  
+
   public abstract void tearDown(State state, Environment env) throws Exception;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/Framework.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/Framework.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/Framework.java
index 0e54d90..f5b721b 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/Framework.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/Framework.java
@@ -27,31 +27,31 @@ import org.apache.log4j.xml.DOMConfigurator;
 import com.beust.jcommander.Parameter;
 
 public class Framework {
-  
+
   private static final Logger log = Logger.getLogger(Framework.class);
   private HashMap<String,Node> nodes = new HashMap<String,Node>();
   private String configDir = null;
   private static final Framework INSTANCE = new Framework();
-  
+
   /**
    * @return Singleton instance of Framework
    */
   public static Framework getInstance() {
     return INSTANCE;
   }
-  
+
   public void setConfigDir(String confDir) {
     configDir = confDir;
   }
-  
+
   /**
    * Run random walk framework
-   * 
+   *
    * @param startName
    *          Full name of starting graph or test
    */
   public int run(String startName, State state, Environment env, String confDir) {
-    
+
     try {
       System.out.println("confDir " + confDir);
       setConfigDir(confDir);
@@ -63,21 +63,21 @@ public class Framework {
     }
     return 0;
   }
-  
+
   /**
    * Creates node (if it does not already exist) and inserts into map
-   * 
+   *
    * @param id
    *          Name of node
    * @return Node specified by id
    */
   public Node getNode(String id) throws Exception {
-    
+
     // check for node in nodes
     if (nodes.containsKey(id)) {
       return nodes.get(id);
     }
-    
+
     // otherwise create and put in nodes
     Node node = null;
     if (id.endsWith(".xml")) {
@@ -88,18 +88,18 @@ public class Framework {
     nodes.put(id, node);
     return node;
   }
-  
+
   static class Opts extends org.apache.accumulo.core.cli.Help {
-    @Parameter(names="--configDir", required=true, description="directory containing the test configuration")
+    @Parameter(names = "--configDir", required = true, description = "directory containing the test configuration")
     String configDir;
-    @Parameter(names="--logDir", required=true, description="location of the local logging directory")
+    @Parameter(names = "--logDir", required = true, description = "location of the local logging directory")
     String localLogPath;
-    @Parameter(names="--logId", required=true, description="a unique log identifier (like a hostname, or pid)")
+    @Parameter(names = "--logId", required = true, description = "a unique log identifier (like a hostname, or pid)")
     String logId;
-    @Parameter(names="--module", required=true, description="the name of the module to run")
+    @Parameter(names = "--module", required = true, description = "the name of the module to run")
     String module;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(Framework.class.getName(), args);
@@ -108,16 +108,16 @@ public class Framework {
     FileInputStream fis = new FileInputStream(opts.configDir + "/randomwalk.conf");
     props.load(fis);
     fis.close();
-    
+
     System.setProperty("localLog", opts.localLogPath + "/" + opts.logId);
     System.setProperty("nfsLog", props.getProperty("NFS_LOGPATH") + "/" + opts.logId);
-    
+
     DOMConfigurator.configure(opts.configDir + "logger.xml");
-    
+
     State state = new State();
     Environment env = new Environment(props);
     int retval = getInstance().run(opts.module, state, env, opts.configDir);
-    
+
     System.exit(retval);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/Module.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/Module.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/Module.java
index 93a8f61..e5af8e6 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/Module.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/Module.java
@@ -58,7 +58,6 @@ public class Module extends Node {
 
   private static final Logger log = Logger.getLogger(Module.class);
 
-
   private class Dummy extends Node {
 
     String name;
@@ -406,7 +405,7 @@ public class Module extends Node {
       try {
         timer.join();
       } catch (InterruptedException e) {
-        log.error("Failed to join timer '"+timer.getName()+"'.", e);
+        log.error("Failed to join timer '" + timer.getName() + "'.", e);
       }
     }
     if (runningLong.get())

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/Node.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/Node.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/Node.java
index 1588d5a..cb0a468 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/Node.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/Node.java
@@ -25,41 +25,41 @@ import org.apache.log4j.Logger;
  * Represents a point in graph of RandomFramework
  */
 public abstract class Node {
-  
+
   protected final Logger log = Logger.getLogger(this.getClass());
   long progress = System.currentTimeMillis();
-  
+
   /**
    * Visits node
-   * 
+   *
    * @param state
    *          Random walk state passed between nodes
    * @param env
    *          test environment
    */
   public abstract void visit(State state, Environment env, Properties props) throws Exception;
-  
+
   @Override
   public boolean equals(Object o) {
     if (o == null)
       return false;
     return toString().equals(o.toString());
   }
-  
+
   @Override
   public String toString() {
     return this.getClass().getName();
   }
-  
+
   @Override
   public int hashCode() {
     return toString().hashCode();
   }
-  
+
   synchronized public void makingProgress() {
     progress = System.currentTimeMillis();
   }
-  
+
   synchronized public long lastProgress() {
     return progress;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/State.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/State.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/State.java
index 6eb2568..18e21e2 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/State.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/State.java
@@ -19,8 +19,7 @@ package org.apache.accumulo.test.randomwalk;
 import java.util.HashMap;
 
 /**
- * A structure for storing state kept during a test. This class is not
- * thread-safe.
+ * A structure for storing state kept during a test. This class is not thread-safe.
  */
 public class State {
 
@@ -29,14 +28,15 @@ public class State {
   /**
    * Creates new empty state.
    */
-  State() {
-  }
+  State() {}
 
   /**
    * Sets a state object.
    *
-   * @param key key for object
-   * @param value object
+   * @param key
+   *          key for object
+   * @param value
+   *          object
    */
   public void set(String key, Object value) {
     stateMap.put(key, value);
@@ -45,7 +45,8 @@ public class State {
   /**
    * Removes a state object.
    *
-   * @param key key for object
+   * @param key
+   *          key for object
    */
   public void remove(String key) {
     stateMap.remove(key);
@@ -54,9 +55,11 @@ public class State {
   /**
    * Gets a state object.
    *
-   * @param key key for object
+   * @param key
+   *          key for object
    * @return value object
-   * @throws RuntimeException if state object is not present
+   * @throws RuntimeException
+   *           if state object is not present
    */
   public Object get(String key) {
     if (stateMap.containsKey(key) == false) {
@@ -68,7 +71,8 @@ public class State {
   /**
    * Gets a state object, returning null if it is absent.
    *
-   * @param key key for object
+   * @param key
+   *          key for object
    * @return value object, or null if not present
    */
   public Object getOkIfAbsent(String key) {
@@ -76,8 +80,7 @@ public class State {
   }
 
   /**
-   * Gets the map of state objects. The backing map for state is returned, so
-   * changes to it affect the state.
+   * Gets the map of state objects. The backing map for state is returned, so changes to it affect the state.
    *
    * @return state map
    */
@@ -88,9 +91,11 @@ public class State {
   /**
    * Gets a state object as a string.
    *
-   * @param key key for object
+   * @param key
+   *          key for object
    * @return value as string
-   * @throws ClassCastException if the value object is not a string
+   * @throws ClassCastException
+   *           if the value object is not a string
    */
   public String getString(String key) {
     return (String) stateMap.get(key);
@@ -99,9 +104,11 @@ public class State {
   /**
    * Gets a state object as an integer.
    *
-   * @param key key for object
+   * @param key
+   *          key for object
    * @return value as integer
-   * @throws ClassCastException if the value object is not an integer
+   * @throws ClassCastException
+   *           if the value object is not an integer
    */
   public Integer getInteger(String key) {
     return (Integer) stateMap.get(key);
@@ -110,9 +117,11 @@ public class State {
   /**
    * Gets a state object as a long.
    *
-   * @param key key for object
+   * @param key
+   *          key for object
    * @return value as long
-   * @throws ClassCastException if the value object is not a long
+   * @throws ClassCastException
+   *           if the value object is not a long
    */
   public Long getLong(String key) {
     return (Long) stateMap.get(key);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/Test.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/Test.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/Test.java
index a7db7dd..f781c9a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/Test.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/Test.java
@@ -20,7 +20,7 @@ package org.apache.accumulo.test.randomwalk;
  * Tests are extended by users to perform actions on accumulo and are a node of the graph
  */
 public abstract class Test extends Node {
-  
+
   @Override
   public String toString() {
     return getClass().getName();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Setup.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Setup.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Setup.java
index a2af0bc..b95c141 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Setup.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Setup.java
@@ -34,10 +34,10 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.fs.FileSystem;
 
 public class Setup extends Test {
-  
+
   private static final int MAX_POOL_SIZE = 8;
   static String tableName = null;
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Random rand = new Random();
@@ -45,7 +45,7 @@ public class Setup extends Test {
     String pid = env.getPid();
     tableName = String.format("bulk_%s_%s_%d", hostname, pid, System.currentTimeMillis());
     log.info("Starting bulk test on " + tableName);
-    
+
     TableOperations tableOps = env.getConnector().tableOperations();
     try {
       if (!tableOps.exists(getTableName())) {
@@ -62,21 +62,21 @@ public class Setup extends Test {
     state.set("fs", FileSystem.get(CachedConfiguration.getInstance()));
     state.set("bulkImportSuccess", "true");
     BulkPlusOne.counter.set(0l);
-    
+
     ThreadPoolExecutor e = new SimpleThreadPool(MAX_POOL_SIZE, "bulkImportPool");
     state.set("pool", e);
   }
-  
+
   public static String getTableName() {
     return tableName;
   }
-  
+
   public static ThreadPoolExecutor getThreadPool(State state) {
     return (ThreadPoolExecutor) state.get("pool");
   }
-  
+
   public static void run(State state, Runnable r) {
     getThreadPool(state).submit(r);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Verify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Verify.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Verify.java
index f7a727a..f92c31d 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Verify.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/bulk/Verify.java
@@ -37,9 +37,9 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Verify extends Test {
-  
+
   static byte[] zero = new byte[] {'0'};
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     ThreadPoolExecutor threadPool = Setup.getThreadPool(state);
@@ -57,7 +57,7 @@ public class Verify extends Test {
       log.info("Not verifying bulk import test due to import failures");
       return;
     }
-    
+
     String user = env.getConnector().whoami();
     Authorizations auths = env.getConnector().securityOperations().getUserAuthorizations(user);
     Scanner scanner = env.getConnector().createScanner(Setup.getTableName(), auths);
@@ -68,18 +68,18 @@ public class Verify extends Test {
         throw new Exception("Bad key at " + entry);
       }
     }
-    
+
     scanner.clearColumns();
     scanner.fetchColumnFamily(BulkPlusOne.MARKER_CF);
     RowIterator rowIter = new RowIterator(scanner);
-    
+
     while (rowIter.hasNext()) {
       Iterator<Entry<Key,Value>> row = rowIter.next();
       long prev = 0;
       Text rowText = null;
       while (row.hasNext()) {
         Entry<Key,Value> entry = row.next();
-        
+
         if (rowText == null)
           rowText = entry.getKey().getRow();
 
@@ -87,13 +87,13 @@ public class Verify extends Test {
 
         if (curr - 1 != prev)
           throw new Exception("Bad marker count " + entry.getKey() + " " + entry.getValue() + " " + prev);
-        
+
         if (!entry.getValue().toString().equals("1"))
           throw new Exception("Bad marker value " + entry.getKey() + " " + entry.getValue());
-        
+
         prev = curr;
       }
-      
+
       if (BulkPlusOne.counter.get() != prev) {
         throw new Exception("Row " + rowText + " does not have all markers " + BulkPlusOne.counter.get() + " " + prev);
       }
@@ -102,7 +102,7 @@ public class Verify extends Test {
     log.info("Test successful on table " + Setup.getTableName());
     env.getConnector().tableOperations().delete(Setup.getTableName());
   }
-    
+
   public static void main(String args[]) throws Exception {
     ClientOnRequiredTable opts = new ClientOnRequiredTable();
     opts.parseArgs(Verify.class.getName(), args);
@@ -139,10 +139,10 @@ public class Verify extends Test {
       report(startBadRow, lastBadRow, currentBadValue);
     }
   }
-  
+
   private static void report(Text startBadRow, Text lastBadRow, Value value) {
     System.out.println("Bad value " + new String(value.get(), UTF_8));
     System.out.println(" Range [" + startBadRow + " -> " + lastBadRow + "]");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/AddSplits.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/AddSplits.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/AddSplits.java
index 8fc4fb4..2727e62 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/AddSplits.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/AddSplits.java
@@ -32,24 +32,24 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class AddSplits extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
     tableNames = new ArrayList<String>(tableNames);
     tableNames.add(MetadataTable.NAME);
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     TreeSet<Text> splits = new TreeSet<Text>();
-    
+
     for (int i = 0; i < rand.nextInt(10) + 1; i++)
       splits.add(new Text(String.format("%016x", rand.nextLong() & 0x7fffffffffffffffl)));
-    
+
     try {
       conn.tableOperations().addSplits(tableName, splits);
       log.debug("Added " + splits.size() + " splits " + tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Apocalypse.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Apocalypse.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Apocalypse.java
index 512cb1d..b2d3d50 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Apocalypse.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Apocalypse.java
@@ -23,12 +23,12 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Apocalypse extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Process exec = Runtime.getRuntime().exec(new String[] {System.getenv("ACCUMULO_HOME") + "/test/system/randomwalk/bin/apocalypse.sh"});
     if (exec.waitFor() != 0)
       throw new RuntimeException("apocalypse.sh returned a non-zero response: " + exec.exitValue());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchScan.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchScan.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchScan.java
index 6afc7c8..187199f 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchScan.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchScan.java
@@ -38,26 +38,26 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class BatchScan extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       BatchScanner bs = conn.createBatchScanner(tableName, Authorizations.EMPTY, 3);
       List<Range> ranges = new ArrayList<Range>();
       for (int i = 0; i < rand.nextInt(2000) + 1; i++)
         ranges.add(new Range(String.format("%016x", rand.nextLong() & 0x7fffffffffffffffl)));
-      
+
       bs.setRanges(ranges);
-      
+
       try {
         Iterator<Entry<Key,Value>> iter = bs.iterator();
         while (iter.hasNext())
@@ -65,7 +65,7 @@ public class BatchScan extends Test {
       } finally {
         bs.close();
       }
-      
+
       log.debug("Wrote to " + tableName);
     } catch (TableNotFoundException e) {
       log.debug("BatchScan " + tableName + " failed, doesnt exist");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchWrite.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchWrite.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchWrite.java
index 09bf883..76f5cbd 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchWrite.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BatchWrite.java
@@ -36,18 +36,18 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class BatchWrite extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       BatchWriter bw = conn.createBatchWriter(tableName, new BatchWriterConfig());
       try {
@@ -58,13 +58,13 @@ public class BatchWrite extends Test {
           for (int j = 0; j < 10; j++) {
             m.put("cf", "cq" + j, new Value(String.format("%016x", val).getBytes(UTF_8)));
           }
-          
+
           bw.addMutation(m);
         }
       } finally {
         bw.close();
       }
-      
+
       log.debug("Wrote to " + tableName);
     } catch (TableNotFoundException e) {
       log.debug("BatchWrite " + tableName + " failed, doesnt exist");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BulkImport.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BulkImport.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BulkImport.java
index 0a9d3b9..5af08ec 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BulkImport.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/BulkImport.java
@@ -45,11 +45,11 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
 public class BulkImport extends Test {
-  
+
   public static class RFileBatchWriter implements BatchWriter {
-    
+
     RFile.Writer writer;
-    
+
     public RFileBatchWriter(Configuration conf, FileSystem fs, String file) throws IOException {
       AccumuloConfiguration aconf = AccumuloConfiguration.getDefaultConfiguration();
       CachableBlockFile.Writer cbw = new CachableBlockFile.Writer(fs.create(new Path(file), false, conf.getInt("io.file.buffer.size", 4096),
@@ -57,14 +57,14 @@ public class BulkImport extends Test {
       writer = new RFile.Writer(cbw, 100000);
       writer.startDefaultLocalityGroup();
     }
-    
+
     @Override
     public void addMutation(Mutation m) throws MutationsRejectedException {
       List<ColumnUpdate> updates = m.getUpdates();
       for (ColumnUpdate cu : updates) {
         Key key = new Key(m.getRow(), cu.getColumnFamily(), cu.getColumnQualifier(), cu.getColumnVisibility(), 42, false, false);
         Value val = new Value(cu.getValue(), false);
-        
+
         try {
           writer.append(key, val);
         } catch (IOException e) {
@@ -72,16 +72,16 @@ public class BulkImport extends Test {
         }
       }
     }
-    
+
     @Override
     public void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException {
       for (Mutation mutation : iterable)
         addMutation(mutation);
     }
-    
+
     @Override
     public void flush() throws MutationsRejectedException {}
-    
+
     @Override
     public void close() throws MutationsRejectedException {
       try {
@@ -90,28 +90,28 @@ public class BulkImport extends Test {
         throw new RuntimeException(e);
       }
     }
-    
+
   }
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     Configuration conf = CachedConfiguration.getInstance();
     FileSystem fs = FileSystem.get(conf);
-    
+
     String bulkDir = "/tmp/concurrent_bulk/b_" + String.format("%016x", rand.nextLong() & 0x7fffffffffffffffl);
-    
+
     fs.mkdirs(new Path(bulkDir));
     fs.mkdirs(new Path(bulkDir + "_f"));
-    
+
     try {
       BatchWriter bw = new RFileBatchWriter(conf, fs, bulkDir + "/file01.rf");
       try {
@@ -120,22 +120,22 @@ public class BulkImport extends Test {
         for (int i = 0; i < numRows; i++) {
           rows.add(rand.nextLong() & 0x7fffffffffffffffl);
         }
-        
+
         for (Long row : rows) {
           Mutation m = new Mutation(String.format("%016x", row));
           long val = rand.nextLong() & 0x7fffffffffffffffl;
           for (int j = 0; j < 10; j++) {
             m.put("cf", "cq" + j, new Value(String.format("%016x", val).getBytes(UTF_8)));
           }
-          
+
           bw.addMutation(m);
         }
       } finally {
         bw.close();
       }
-      
+
       conn.tableOperations().importDirectory(tableName, bulkDir, bulkDir + "_f", rand.nextBoolean());
-      
+
       log.debug("BulkImported to " + tableName);
     } catch (TableNotFoundException e) {
       log.debug("BulkImport " + tableName + " failed, doesnt exist");
@@ -145,6 +145,6 @@ public class BulkImport extends Test {
       fs.delete(new Path(bulkDir), true);
       fs.delete(new Path(bulkDir + "_f"), true);
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangeAuthorizations.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangeAuthorizations.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangeAuthorizations.java
index 03f2f39..65502c3 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangeAuthorizations.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangeAuthorizations.java
@@ -31,20 +31,20 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class ChangeAuthorizations extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> userNames = (List<String>) state.get("users");
-    
+
     String userName = userNames.get(rand.nextInt(userNames.size()));
     try {
       List<byte[]> auths = new ArrayList<byte[]>(conn.securityOperations().getUserAuthorizations(userName).getAuthorizations());
-      
+
       if (rand.nextBoolean()) {
         String authorization = String.format("a%d", rand.nextInt(5000));
         log.debug("adding authorization " + authorization);
@@ -59,5 +59,5 @@ public class ChangeAuthorizations extends Test {
       log.debug("Unable to change user authorizations: " + ex.getCause());
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangePermissions.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangePermissions.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangePermissions.java
index 5df1e21..680750a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangePermissions.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ChangePermissions.java
@@ -27,33 +27,33 @@ import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.impl.thrift.TableOperationExceptionType;
 import org.apache.accumulo.core.client.impl.thrift.ThriftTableOperationException;
-import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.NamespacePermission;
+import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
 import org.apache.accumulo.test.randomwalk.Environment;
 import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class ChangePermissions extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> userNames = (List<String>) state.get("users");
     String userName = userNames.get(rand.nextInt(userNames.size()));
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     @SuppressWarnings("unchecked")
     List<String> namespaces = (List<String>) state.get("namespaces");
     String namespace = namespaces.get(rand.nextInt(namespaces.size()));
-    
+
     try {
       int dice = rand.nextInt(3);
       if (dice == 0)
@@ -67,7 +67,7 @@ public class ChangePermissions extends Test {
     } catch (AccumuloException ex) {
       Throwable cause = ex.getCause();
       if (cause != null && cause instanceof ThriftTableOperationException) {
-        ThriftTableOperationException toe = (ThriftTableOperationException)cause.getCause();
+        ThriftTableOperationException toe = (ThriftTableOperationException) cause.getCause();
         if (toe.type == TableOperationExceptionType.NAMESPACE_NOTFOUND) {
           log.debug("Unable to change user permissions: " + toe);
           return;
@@ -75,18 +75,18 @@ public class ChangePermissions extends Test {
       }
     }
   }
-  
+
   private void changeTablePermission(Connector conn, Random rand, String userName, String tableName) throws AccumuloException, AccumuloSecurityException {
-    
+
     EnumSet<TablePermission> perms = EnumSet.noneOf(TablePermission.class);
     for (TablePermission p : TablePermission.values()) {
       if (conn.securityOperations().hasTablePermission(userName, tableName, p))
         perms.add(p);
     }
-    
+
     EnumSet<TablePermission> more = EnumSet.allOf(TablePermission.class);
     more.removeAll(perms);
-    
+
     if (rand.nextBoolean() && more.size() > 0) {
       List<TablePermission> moreList = new ArrayList<TablePermission>(more);
       TablePermission choice = moreList.get(rand.nextInt(moreList.size()));
@@ -101,18 +101,18 @@ public class ChangePermissions extends Test {
       }
     }
   }
-  
+
   private void changeSystemPermission(Connector conn, Random rand, String userName) throws AccumuloException, AccumuloSecurityException {
     EnumSet<SystemPermission> perms = EnumSet.noneOf(SystemPermission.class);
     for (SystemPermission p : SystemPermission.values()) {
       if (conn.securityOperations().hasSystemPermission(userName, p))
         perms.add(p);
     }
-    
+
     EnumSet<SystemPermission> more = EnumSet.allOf(SystemPermission.class);
     more.removeAll(perms);
     more.remove(SystemPermission.GRANT);
-    
+
     if (rand.nextBoolean() && more.size() > 0) {
       List<SystemPermission> moreList = new ArrayList<SystemPermission>(more);
       SystemPermission choice = moreList.get(rand.nextInt(moreList.size()));
@@ -127,18 +127,18 @@ public class ChangePermissions extends Test {
       }
     }
   }
-  
+
   private void changeNamespacePermission(Connector conn, Random rand, String userName, String namespace) throws AccumuloException, AccumuloSecurityException {
-    
+
     EnumSet<NamespacePermission> perms = EnumSet.noneOf(NamespacePermission.class);
     for (NamespacePermission p : NamespacePermission.values()) {
       if (conn.securityOperations().hasNamespacePermission(userName, namespace, p))
         perms.add(p);
     }
-    
+
     EnumSet<NamespacePermission> more = EnumSet.allOf(NamespacePermission.class);
     more.removeAll(perms);
-    
+
     if (rand.nextBoolean() && more.size() > 0) {
       List<NamespacePermission> moreList = new ArrayList<NamespacePermission>(more);
       NamespacePermission choice = moreList.get(rand.nextInt(moreList.size()));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CheckBalance.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CheckBalance.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CheckBalance.java
index df246d4..c113091 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CheckBalance.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CheckBalance.java
@@ -33,10 +33,10 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 /**
- * 
+ *
  */
 public class CheckBalance extends Test {
-  
+
   static final String LAST_UNBALANCED_TIME = "lastUnbalancedTime";
   static final String UNBALANCED_COUNT = "unbalancedCount";
 
@@ -75,7 +75,7 @@ public class CheckBalance extends Test {
         lastCount = thisCount;
       }
     }
-    
+
     // It is expected that the number of tablets will be uneven for short
     // periods of time. Don't complain unless we've seen it only unbalanced
     // over a 15 minute period and it's been at least three checks.
@@ -97,7 +97,7 @@ public class CheckBalance extends Test {
       state.remove(UNBALANCED_COUNT);
     }
   }
-  
+
   private static double stddev(Collection<Long> samples, double avg) {
     int num = samples.size();
     double sqrtotal = 0.0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Compact.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Compact.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Compact.java
index 30476a4..d0f1010 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Compact.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Compact.java
@@ -29,20 +29,20 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Compact extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     List<Text> range = ConcurrentFixture.generateRange(rand);
-    
+
     try {
       boolean wait = rand.nextBoolean();
       conn.tableOperations().compact(tableName, range.get(0), range.get(1), false, wait);
@@ -52,6 +52,6 @@ public class Compact extends Test {
     } catch (TableOfflineException toe) {
       log.debug("compact " + tableName + " from " + range.get(0) + " to " + range.get(1) + " failed, offline");
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ConcurrentFixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ConcurrentFixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ConcurrentFixture.java
index 9c51e81..a32e463 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ConcurrentFixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ConcurrentFixture.java
@@ -20,44 +20,42 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Random;
 
-import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.Environment;
+import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.State;
 import org.apache.hadoop.io.Text;
 
 /**
  * When multiple instance of this test suite are run, all instances will operate on the same set of table names.
- * 
- * 
+ *
+ *
  */
 
 public class ConcurrentFixture extends Fixture {
-  
+
   @Override
   public void setUp(State state, Environment env) throws Exception {}
-  
+
   @Override
   public void tearDown(State state, Environment env) throws Exception {
     state.remove(CheckBalance.LAST_UNBALANCED_TIME);
     state.remove(CheckBalance.UNBALANCED_COUNT);
   }
-  
+
   /**
-   * 
+   *
    * @param rand
-   *  A Random to use
-   * @return
-   *  A two element list with first being smaller than the second, but either value (or both) can be null
+   *          A Random to use
+   * @return A two element list with first being smaller than the second, but either value (or both) can be null
    */
   public static List<Text> generateRange(Random rand) {
     ArrayList<Text> toRet = new ArrayList<Text>(2);
 
     long firstLong = rand.nextLong();
-    
-    
+
     long secondLong = rand.nextLong();
     Text first = null, second = null;
-    
+
     // Having all negative values = null might be too frequent
     if (firstLong >= 0)
       first = new Text(String.format("%016x", firstLong & 0x7fffffffffffffffl));
@@ -69,10 +67,10 @@ public class ConcurrentFixture extends Fixture {
       first = second;
       second = swap;
     }
-    
+
     toRet.add(first);
     toRet.add(second);
-    
+
     return toRet;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateTable.java
index b9e1ece..30d49f0 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateTable.java
@@ -30,18 +30,18 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class CreateTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       conn.tableOperations().create(tableName);
       log.debug("Created table " + tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateUser.java
index 8f265ad..e73e80a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/CreateUser.java
@@ -31,14 +31,14 @@ public class CreateUser extends Test {
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> userNames = (List<String>) state.get("users");
-    
+
     String userName = userNames.get(rand.nextInt(userNames.size()));
-    
+
     try {
       log.debug("Creating user " + userName);
       conn.securityOperations().createLocalUser(userName, new PasswordToken(userName + "pass"));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteRange.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteRange.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteRange.java
index ced8011..280f620 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteRange.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteRange.java
@@ -31,18 +31,18 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class DeleteRange extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     List<Text> range = new ArrayList<Text>();
     do {
       range.add(new Text(String.format("%016x", rand.nextLong() & 0x7fffffffffffffffl)));
@@ -53,7 +53,7 @@ public class DeleteRange extends Test {
       range.set(0, null);
     if (rand.nextInt(20) == 0)
       range.set(1, null);
-    
+
     try {
       conn.tableOperations().deleteRows(tableName, range.get(0), range.get(1));
       log.debug("deleted rows (" + range.get(0) + " -> " + range.get(1) + "] in " + tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteTable.java
index 6fb9f7f..4bee7f1 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DeleteTable.java
@@ -27,18 +27,18 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class DeleteTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       conn.tableOperations().delete(tableName);
       log.debug("Deleted table " + tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DropUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DropUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DropUser.java
index 13d3c05..a4442c6 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DropUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/DropUser.java
@@ -30,14 +30,14 @@ public class DropUser extends Test {
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> userNames = (List<String>) state.get("users");
-    
+
     String userName = userNames.get(rand.nextInt(userNames.size()));
-    
+
     try {
       log.debug("Dropping user " + userName);
       conn.securityOperations().dropLocalUser(userName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/IsolatedScan.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/IsolatedScan.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/IsolatedScan.java
index 78f73b4..1bb51bb 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/IsolatedScan.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/IsolatedScan.java
@@ -36,21 +36,21 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class IsolatedScan extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       RowIterator iter = new RowIterator(new IsolatedScanner(conn.createScanner(tableName, Authorizations.EMPTY)));
-      
+
       while (iter.hasNext()) {
         PeekingIterator<Entry<Key,Value>> row = new PeekingIterator<Entry<Key,Value>>(iter.next());
         Entry<Key,Value> kv = null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ListSplits.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ListSplits.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ListSplits.java
index 1f82fc0..6944092 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ListSplits.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ListSplits.java
@@ -30,18 +30,18 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class ListSplits extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       Collection<Text> splits = conn.tableOperations().listSplits(tableName);
       log.debug("Table " + tableName + " had " + splits.size() + " splits");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Merge.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Merge.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Merge.java
index 84a4665..a997c2b 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Merge.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Merge.java
@@ -31,21 +31,21 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Merge extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
     tableNames = new ArrayList<String>(tableNames);
     tableNames.add(MetadataTable.NAME);
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     List<Text> range = ConcurrentFixture.generateRange(rand);
-    
+
     try {
       conn.tableOperations().merge(tableName, range.get(0), range.get(1));
       log.debug("merged " + tableName + " from " + range.get(0) + " to " + range.get(1));
@@ -54,6 +54,6 @@ public class Merge extends Test {
     } catch (TableNotFoundException tne) {
       log.debug("merge " + tableName + " from " + range.get(0) + " to " + range.get(1) + " failed, doesnt exist");
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/OfflineTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/OfflineTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/OfflineTable.java
index 1d725bc..ba6389f 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/OfflineTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/OfflineTable.java
@@ -28,18 +28,18 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class OfflineTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       conn.tableOperations().offline(tableName, rand.nextBoolean());
       log.debug("Offlined " + tableName);
@@ -49,6 +49,6 @@ public class OfflineTable extends Test {
     } catch (TableNotFoundException tne) {
       log.debug("offline or online failed " + tableName + ", doesnt exist");
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/RenameTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/RenameTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/RenameTable.java
index d468614..57119eb 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/RenameTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/RenameTable.java
@@ -30,31 +30,31 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class RenameTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String srcTableName = tableNames.get(rand.nextInt(tableNames.size()));
     String newTableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     String srcNamespace = "", newNamespace = "";
-    
+
     int index = srcTableName.indexOf('.');
     if (-1 != index) {
       srcNamespace = srcTableName.substring(0, index);
     }
-    
+
     index = newTableName.indexOf('.');
     if (-1 != index) {
       newNamespace = newTableName.substring(0, index);
     }
-    
+
     try {
       conn.tableOperations().rename(srcTableName, newTableName);
       log.debug("Renamed table " + srcTableName + " " + newTableName);
@@ -71,7 +71,7 @@ public class RenameTable extends Test {
           return;
         }
       }
-      
+
       log.debug("Rename " + srcTableName + " failed, doesnt exist");
     } catch (IllegalArgumentException e) {
       log.debug("Rename: " + e.toString());
@@ -82,7 +82,7 @@ public class RenameTable extends Test {
       }
       log.debug("Rename " + srcTableName + " failed.", e);
     }
-    
+
     if (!srcNamespace.equals(newNamespace)) {
       log.error("RenameTable operation should have failed when renaming across namespaces.");
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ScanTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ScanTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ScanTable.java
index 8dd24f7..e0bec38 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ScanTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/ScanTable.java
@@ -36,18 +36,18 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class ScanTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     @SuppressWarnings("unchecked")
     List<String> tableNames = (List<String>) state.get("tables");
-    
+
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
-    
+
     try {
       Scanner scanner = conn.createScanner(tableName, Authorizations.EMPTY);
       Iterator<Entry<Key,Value>> iter = scanner.iterator();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Setup.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Setup.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Setup.java
index 142287d..c19fcbd 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Setup.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Setup.java
@@ -26,29 +26,29 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Setup extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Random rand = new Random();
     state.set("rand", rand);
-    
+
     int numTables = Integer.parseInt(props.getProperty("numTables", "9"));
     int numNamespaces = Integer.parseInt(props.getProperty("numNamespaces", "2"));
     log.debug("numTables = " + numTables);
     log.debug("numNamespaces = " + numNamespaces);
     List<String> tables = new ArrayList<String>();
     List<String> namespaces = new ArrayList<String>();
-    
+
     for (int i = 0; i < numNamespaces; i++) {
       namespaces.add(String.format("nspc_%03d", i));
     }
-    
+
     // Make tables in the default namespace
-    double tableCeil = Math.ceil((double)numTables / (numNamespaces + 1));
+    double tableCeil = Math.ceil((double) numTables / (numNamespaces + 1));
     for (int i = 0; i < tableCeil; i++) {
       tables.add(String.format("ctt_%03d", i));
     }
-    
+
     // Make tables in each namespace
     double tableFloor = Math.floor(numTables / (numNamespaces + 1));
     for (String n : namespaces) {
@@ -56,10 +56,10 @@ public class Setup extends Test {
         tables.add(String.format(n + ".ctt_%03d", i));
       }
     }
-    
+
     state.set("tables", tables);
     state.set("namespaces", namespaces);
-    
+
     int numUsers = Integer.parseInt(props.getProperty("numUsers", "5"));
     log.debug("numUsers = " + numUsers);
     List<String> users = new ArrayList<String>();
@@ -67,5 +67,5 @@ public class Setup extends Test {
       users.add(String.format("user%03d", i));
     state.set("users", users);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Shutdown.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Shutdown.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Shutdown.java
index 6715edb..6cc8312 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Shutdown.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/Shutdown.java
@@ -32,16 +32,16 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Shutdown extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     log.info("shutting down");
     SetGoalState.main(new String[] {MasterGoalState.CLEAN_STOP.name()});
-    
+
     while (!env.getConnector().instanceOperations().getTabletServers().isEmpty()) {
       UtilWaitThread.sleep(1000);
     }
-    
+
     while (true) {
       try {
         AccumuloServerContext context = new AccumuloServerContext(new ServerConfigurationFactory(HdfsZooInstance.getInstance()));
@@ -53,9 +53,9 @@ public class Shutdown extends Test {
       }
       UtilWaitThread.sleep(1000);
     }
-    
+
     log.info("servers stopped");
     UtilWaitThread.sleep(10000);
   }
-  
+
 }


[62/66] [abbrv] accumulo git commit: Merge branch '1.5' into 1.6

Posted by ct...@apache.org.
Merge branch '1.5' into 1.6

Conflicts:
	core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
	core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
	core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
	core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
	core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
	core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
	core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
	server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
	start/src/main/java/org/apache/accumulo/start/Main.java


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/1368d09c
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/1368d09c
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/1368d09c

Branch: refs/heads/1.6
Commit: 1368d09c198c898a856e9fe3e25150ade33fdb97
Parents: 627e525 901d60e
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 21:29:28 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 21:29:28 2015 -0500

----------------------------------------------------------------------
 .../org/apache/accumulo/core/cli/ClientOpts.java     |  6 +++---
 .../core/client/admin/InstanceOperations.java        |  6 +++---
 .../apache/accumulo/core/file/rfile/CreateEmpty.java |  4 ++--
 .../accumulo/core/file/rfile/bcfile/Utils.java       |  1 +
 .../org/apache/accumulo/core/iterators/Combiner.java |  2 +-
 .../apache/accumulo/core/iterators/LongCombiner.java |  4 ++--
 .../accumulo/core/iterators/OptionDescriber.java     |  4 ++--
 .../accumulo/core/iterators/TypedValueCombiner.java  | 10 +++++-----
 .../core/iterators/user/SummingArrayCombiner.java    |  5 +++--
 .../accumulo/core/util/shell/commands/DUCommand.java |  3 ++-
 .../core/util/shell/commands/EGrepCommand.java       |  3 ++-
 .../util/shell/commands/QuotedStringTokenizer.java   | 15 ++++++---------
 .../examples/simple/mapreduce/TableToFile.java       |  4 ++--
 .../java/org/apache/accumulo/server/Accumulo.java    |  4 ++--
 .../java/org/apache/accumulo/tserver/Compactor.java  |  5 +++--
 .../main/java/org/apache/accumulo/start/Main.java    |  4 ++--
 .../start/classloader/AccumuloClassLoader.java       |  4 ++--
 .../test/randomwalk/security/CreateTable.java        |  5 ++---
 .../test/randomwalk/security/CreateUser.java         |  5 ++---
 19 files changed, 47 insertions(+), 47 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
----------------------------------------------------------------------
diff --cc core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
index f1465bc,82ac73d..8bb8b3f
--- a/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
@@@ -166,14 -163,6 +166,14 @@@ public class ClientOpts extends Help 
    @Parameter(names = "--site-file", description = "Read the given accumulo site file to find the accumulo instance")
    public String siteFile = null;
  
 +  @Parameter(names = "--ssl", description = "Connect to accumulo over SSL")
 +  public boolean sslEnabled = false;
 +
-   @Parameter(
-       names = "--config-file",
-       description = "Read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
++  @Parameter(names = "--config-file", description = "Read the given client config file. "
++      + "If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, "
++      + "which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
 +  public String clientConfigFile = null;
 +
    public void startDebugLogging() {
      if (debug)
        Logger.getLogger(Constants.CORE_PACKAGE_NAME).setLevel(Level.TRACE);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------
diff --cc core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
index 49dd1d5,8cb656c..cc9d282
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
@@@ -103,10 -103,10 +103,10 @@@ public interface InstanceOperations 
     * Throws an exception if a tablet server can not be contacted.
     *
     * @param tserver
-    *          The tablet server address should be of the form <ip address>:<port>
+    *          The tablet server address should be of the form {@code <ip address>:<port>}
     * @since 1.5.0
     */
 -  public void ping(String tserver) throws AccumuloException;
 +  void ping(String tserver) throws AccumuloException;
  
    /**
     * Test to see if the instance can load the given class as the given type. This check does not consider per table classpaths, see

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
----------------------------------------------------------------------
diff --cc core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
index cd6bff8,8d54088..dcbaf9e
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
@@@ -60,9 -58,9 +60,9 @@@ public class CreateEmpty 
  
    static class Opts extends Help {
      @Parameter(names = {"-c", "--codec"}, description = "the compression codec to use.", validateWith = IsSupportedCompressionAlgorithm.class)
 -    String codec = TFile.COMPRESSION_NONE;
 +    String codec = Compression.COMPRESSION_NONE;
-     @Parameter(
-         description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
+     @Parameter(description = " <path> { <path> ... } Each path given is a URL. "
+         + "Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
          required = true, validateWith = NamedLikeRFile.class)
      List<String> files = new ArrayList<String>();
    }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------
diff --cc server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
index 1b93c05,0000000..147a7c2
mode 100644,000000..100644
--- a/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
@@@ -1,336 -1,0 +1,336 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.server;
 +
 +import static com.google.common.base.Charsets.UTF_8;
 +
 +import java.io.File;
 +import java.io.FileInputStream;
 +import java.io.IOException;
 +import java.io.InputStream;
 +import java.net.InetAddress;
 +import java.net.UnknownHostException;
 +import java.util.Arrays;
 +import java.util.Map.Entry;
 +import java.util.TreeMap;
 +
 +import org.apache.accumulo.core.Constants;
 +import org.apache.accumulo.core.client.AccumuloException;
 +import org.apache.accumulo.core.client.Instance;
 +import org.apache.accumulo.core.conf.AccumuloConfiguration;
 +import org.apache.accumulo.core.conf.Property;
 +import org.apache.accumulo.core.trace.DistributedTrace;
 +import org.apache.accumulo.core.util.AddressUtil;
 +import org.apache.accumulo.core.util.UtilWaitThread;
 +import org.apache.accumulo.core.util.Version;
 +import org.apache.accumulo.core.volume.Volume;
 +import org.apache.accumulo.core.zookeeper.ZooUtil;
 +import org.apache.accumulo.fate.ReadOnlyStore;
 +import org.apache.accumulo.fate.ReadOnlyTStore;
 +import org.apache.accumulo.fate.ZooStore;
 +import org.apache.accumulo.server.client.HdfsZooInstance;
 +import org.apache.accumulo.server.conf.ServerConfiguration;
 +import org.apache.accumulo.server.fs.VolumeManager;
 +import org.apache.accumulo.server.util.time.SimpleTimer;
 +import org.apache.accumulo.server.watcher.Log4jConfiguration;
 +import org.apache.accumulo.server.watcher.MonitorLog4jWatcher;
 +import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 +import org.apache.hadoop.fs.FileStatus;
 +import org.apache.hadoop.fs.FileSystem;
 +import org.apache.hadoop.fs.Path;
 +import org.apache.log4j.Logger;
 +import org.apache.log4j.helpers.LogLog;
 +import org.apache.zookeeper.KeeperException;
 +
 +public class Accumulo {
 +
 +  private static final Logger log = Logger.getLogger(Accumulo.class);
 +
 +  public static synchronized void updateAccumuloVersion(VolumeManager fs, int oldVersion) {
 +    for (Volume volume : fs.getVolumes()) {
 +      try {
 +        if (getAccumuloPersistentVersion(fs) == oldVersion) {
 +          log.debug("Attempting to upgrade " + volume);
 +          Path dataVersionLocation = ServerConstants.getDataVersionLocation(volume);
 +          fs.create(new Path(dataVersionLocation, Integer.toString(ServerConstants.DATA_VERSION))).close();
 +          // TODO document failure mode & recovery if FS permissions cause above to work and below to fail ACCUMULO-2596
 +          Path prevDataVersionLoc = new Path(dataVersionLocation, Integer.toString(oldVersion));
 +          if (!fs.delete(prevDataVersionLoc)) {
 +            throw new RuntimeException("Could not delete previous data version location (" + prevDataVersionLoc + ") for " + volume);
 +          }
 +        }
 +      } catch (IOException e) {
 +        throw new RuntimeException("Unable to set accumulo version: an error occurred.", e);
 +      }
 +    }
 +  }
 +
 +  public static synchronized int getAccumuloPersistentVersion(FileSystem fs, Path path) {
 +    int dataVersion;
 +    try {
 +      FileStatus[] files = fs.listStatus(path);
 +      if (files == null || files.length == 0) {
 +        dataVersion = -1; // assume it is 0.5 or earlier
 +      } else {
 +        dataVersion = Integer.parseInt(files[0].getPath().getName());
 +      }
 +      return dataVersion;
 +    } catch (IOException e) {
 +      throw new RuntimeException("Unable to read accumulo version: an error occurred.", e);
 +    }
 +  }
 +
 +  public static synchronized int getAccumuloPersistentVersion(VolumeManager fs) {
 +    // It doesn't matter which Volume is used as they should all have the data version stored
 +    Volume v = fs.getVolumes().iterator().next();
 +    Path path = ServerConstants.getDataVersionLocation(v);
 +    return getAccumuloPersistentVersion(v.getFileSystem(), path);
 +  }
 +
 +  public static synchronized Path getAccumuloInstanceIdPath(VolumeManager fs) {
 +    // It doesn't matter which Volume is used as they should all have the instance ID stored
 +    Volume v = fs.getVolumes().iterator().next();
 +    return ServerConstants.getInstanceIdLocation(v);
 +  }
 +
 +  public static void enableTracing(String address, String application) {
 +    try {
 +      DistributedTrace.enable(HdfsZooInstance.getInstance(), ZooReaderWriter.getInstance(), application, address);
 +    } catch (Exception ex) {
 +      log.error("creating remote sink for trace spans", ex);
 +    }
 +  }
 +
 +  /**
 +   * Finds the best log4j configuration file. A generic file is used only if an application-specific file is not available. An XML file is preferred over a
 +   * properties file, if possible.
 +   *
 +   * @param confDir
 +   *          directory where configuration files should reside
 +   * @param application
 +   *          application name for configuration file name
 +   * @return configuration file name
 +   */
 +  static String locateLogConfig(String confDir, String application) {
 +    String explicitConfigFile = System.getProperty("log4j.configuration");
 +    if (explicitConfigFile != null) {
 +      return explicitConfigFile;
 +    }
 +    String[] configFiles = {String.format("%s/%s_logger.xml", confDir, application), String.format("%s/%s_logger.properties", confDir, application),
 +        String.format("%s/generic_logger.xml", confDir), String.format("%s/generic_logger.properties", confDir)};
 +    String defaultConfigFile = configFiles[2]; // generic_logger.xml
 +    for (String f : configFiles) {
 +      if (new File(f).exists()) {
 +        return f;
 +      }
 +    }
 +    return defaultConfigFile;
 +  }
 +
 +  public static void setupLogging(String application) throws UnknownHostException {
 +    System.setProperty("org.apache.accumulo.core.application", application);
 +
 +    if (System.getenv("ACCUMULO_LOG_DIR") != null)
 +      System.setProperty("org.apache.accumulo.core.dir.log", System.getenv("ACCUMULO_LOG_DIR"));
 +    else
 +      System.setProperty("org.apache.accumulo.core.dir.log", System.getenv("ACCUMULO_HOME") + "/logs/");
 +
 +    String localhost = InetAddress.getLocalHost().getHostName();
 +    System.setProperty("org.apache.accumulo.core.ip.localhost.hostname", localhost);
 +
 +    // Use a specific log config, if it exists
 +    String logConfigFile = locateLogConfig(System.getenv("ACCUMULO_CONF_DIR"), application);
 +    // Turn off messages about not being able to reach the remote logger... we protect against that.
 +    LogLog.setQuietMode(true);
 +
 +    // Set up local file-based logging right away
 +    Log4jConfiguration logConf = new Log4jConfiguration(logConfigFile);
 +    logConf.resetLogger();
 +  }
 +
 +  public static void init(VolumeManager fs, ServerConfiguration serverConfig, String application) throws IOException {
 +    final AccumuloConfiguration conf = serverConfig.getConfiguration();
 +    final Instance instance = serverConfig.getInstance();
 +
 +    // Use a specific log config, if it exists
 +    final String logConfigFile = locateLogConfig(System.getenv("ACCUMULO_CONF_DIR"), application);
 +
 +    // Set up polling log4j updates and log-forwarding using information advertised in zookeeper by the monitor
 +    MonitorLog4jWatcher logConfigWatcher = new MonitorLog4jWatcher(instance.getInstanceID(), logConfigFile);
 +    logConfigWatcher.setDelay(5000L);
 +    logConfigWatcher.start();
 +
 +    // Makes sure the log-forwarding to the monitor is configured
 +    int logPort = conf.getPort(Property.MONITOR_LOG4J_PORT);
 +    System.setProperty("org.apache.accumulo.core.host.log.port", Integer.toString(logPort));
 +
 +    log.info(application + " starting");
 +    log.info("Instance " + serverConfig.getInstance().getInstanceID());
 +    int dataVersion = Accumulo.getAccumuloPersistentVersion(fs);
 +    log.info("Data Version " + dataVersion);
 +    Accumulo.waitForZookeeperAndHdfs(fs);
 +
 +    Version codeVersion = new Version(Constants.VERSION);
 +    if (!(canUpgradeFromDataVersion(dataVersion))) {
 +      throw new RuntimeException("This version of accumulo (" + codeVersion + ") is not compatible with files stored using data version " + dataVersion);
 +    }
 +
 +    TreeMap<String,String> sortedProps = new TreeMap<String,String>();
 +    for (Entry<String,String> entry : conf)
 +      sortedProps.put(entry.getKey(), entry.getValue());
 +
 +    for (Entry<String,String> entry : sortedProps.entrySet()) {
 +      String key = entry.getKey();
 +      log.info(key + " = " + (Property.isSensitive(key) ? "<hidden>" : entry.getValue()));
 +    }
 +
 +    monitorSwappiness();
 +
 +    // Encourage users to configure TLS
 +    final String SSL = "SSL";
 +    for (Property sslProtocolProperty : Arrays.asList(Property.RPC_SSL_CLIENT_PROTOCOL, Property.RPC_SSL_ENABLED_PROTOCOLS,
 +        Property.MONITOR_SSL_INCLUDE_PROTOCOLS)) {
 +      String value = conf.get(sslProtocolProperty);
 +      if (value.contains(SSL)) {
 +        log.warn("It is recommended that " + sslProtocolProperty + " only allow TLS");
 +      }
 +    }
 +
 +  }
 +
 +  /**
 +   * Sanity check that the current persistent version is allowed to upgrade to the version of Accumulo running.
 +   *
 +   * @param dataVersion
 +   *          the version that is persisted in the backing Volumes
 +   */
 +  public static boolean canUpgradeFromDataVersion(final int dataVersion) {
 +    return dataVersion == ServerConstants.DATA_VERSION || dataVersion == ServerConstants.PREV_DATA_VERSION
 +        || dataVersion == ServerConstants.TWO_DATA_VERSIONS_AGO;
 +  }
 +
 +  /**
 +   * Does the data version number stored in the backing Volumes indicate we need to upgrade something?
 +   */
 +  public static boolean persistentVersionNeedsUpgrade(final int accumuloPersistentVersion) {
 +    return accumuloPersistentVersion == ServerConstants.TWO_DATA_VERSIONS_AGO || accumuloPersistentVersion == ServerConstants.PREV_DATA_VERSION;
 +  }
 +
 +  /**
 +   *
 +   */
 +  public static void monitorSwappiness() {
 +    SimpleTimer.getInstance().schedule(new Runnable() {
 +      @Override
 +      public void run() {
 +        try {
 +          String procFile = "/proc/sys/vm/swappiness";
 +          File swappiness = new File(procFile);
 +          if (swappiness.exists() && swappiness.canRead()) {
 +            InputStream is = new FileInputStream(procFile);
 +            try {
 +              byte[] buffer = new byte[10];
 +              int bytes = is.read(buffer);
 +              String setting = new String(buffer, 0, bytes, UTF_8);
 +              setting = setting.trim();
 +              if (bytes > 0 && Integer.parseInt(setting) > 10) {
 +                log.warn("System swappiness setting is greater than ten (" + setting + ") which can cause time-sensitive operations to be delayed. "
 +                    + " Accumulo is time sensitive because it needs to maintain distributed lock agreement.");
 +              }
 +            } finally {
 +              is.close();
 +            }
 +          }
 +        } catch (Throwable t) {
 +          log.error(t, t);
 +        }
 +      }
 +    }, 1000, 10 * 60 * 1000);
 +  }
 +
 +  public static void waitForZookeeperAndHdfs(VolumeManager fs) {
 +    log.info("Attempting to talk to zookeeper");
 +    while (true) {
 +      try {
 +        ZooReaderWriter.getInstance().getChildren(Constants.ZROOT);
 +        break;
 +      } catch (InterruptedException e) {
 +        // ignored
 +      } catch (KeeperException ex) {
 +        log.info("Waiting for accumulo to be initialized");
 +        UtilWaitThread.sleep(1000);
 +      }
 +    }
 +    log.info("Zookeeper connected and initialized, attemping to talk to HDFS");
 +    long sleep = 1000;
 +    int unknownHostTries = 3;
 +    while (true) {
 +      try {
 +        if (fs.isReady())
 +          break;
 +        log.warn("Waiting for the NameNode to leave safemode");
 +      } catch (IOException ex) {
 +        log.warn("Unable to connect to HDFS", ex);
 +      } catch (IllegalArgumentException exception) {
 +        /* Unwrap the UnknownHostException so we can deal with it directly */
 +        if (exception.getCause() instanceof UnknownHostException) {
 +          if (unknownHostTries > 0) {
 +            log.warn("Unable to connect to HDFS, will retry. cause: " + exception.getCause());
 +            /* We need to make sure our sleep period is long enough to avoid getting a cached failure of the host lookup. */
 +            sleep = Math.max(sleep, (AddressUtil.getAddressCacheNegativeTtl((UnknownHostException) (exception.getCause())) + 1) * 1000);
 +          } else {
 +            log.error("Unable to connect to HDFS and have exceeded max number of retries.", exception);
 +            throw exception;
 +          }
 +          unknownHostTries--;
 +        } else {
 +          throw exception;
 +        }
 +      }
 +      log.info("Backing off due to failure; current sleep period is " + sleep / 1000. + " seconds");
 +      UtilWaitThread.sleep(sleep);
 +      /* Back off to give transient failures more time to clear. */
 +      sleep = Math.min(60 * 1000, sleep * 2);
 +    }
 +    log.info("Connected to HDFS");
 +  }
 +
 +  /**
 +   * Exit loudly if there are outstanding Fate operations. Since Fate serializes class names, we need to make sure there are no queued transactions from a
 +   * previous version before continuing an upgrade. The status of the operations is irrelevant; those in SUCCESSFUL status cause the same problem as those just
 +   * queued.
 +   *
 +   * Note that the Master should not allow write access to Fate until after all upgrade steps are complete.
 +   *
 +   * Should be called as a guard before performing any upgrade steps, after determining that an upgrade is needed.
 +   *
 +   * see ACCUMULO-2519
 +   */
 +  public static void abortIfFateTransactions() {
 +    try {
 +      final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(
 +          ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE, ZooReaderWriter.getInstance()));
 +      if (!(fate.list().isEmpty())) {
-         throw new AccumuloException(
-             "Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
++        throw new AccumuloException("Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. "
++            + "Please see the README document for instructions on what to do under your previous version.");
 +      }
 +    } catch (Exception exception) {
 +      log.fatal("Problem verifying Fate readiness", exception);
 +      System.exit(1);
 +    }
 +  }
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
----------------------------------------------------------------------
diff --cc server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
index 822171c,0000000..381f75c
mode 100644,000000..100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
@@@ -1,548 -1,0 +1,549 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.tserver;
 +
 +import java.io.IOException;
 +import java.text.DateFormat;
 +import java.text.SimpleDateFormat;
 +import java.util.ArrayList;
 +import java.util.Collections;
 +import java.util.Date;
 +import java.util.HashMap;
 +import java.util.HashSet;
 +import java.util.List;
 +import java.util.Map;
 +import java.util.Map.Entry;
 +import java.util.Set;
 +import java.util.concurrent.Callable;
 +import java.util.concurrent.atomic.AtomicLong;
 +
 +import org.apache.accumulo.core.client.IteratorSetting;
 +import org.apache.accumulo.core.conf.AccumuloConfiguration;
 +import org.apache.accumulo.core.data.ByteSequence;
 +import org.apache.accumulo.core.data.Key;
 +import org.apache.accumulo.core.data.KeyExtent;
 +import org.apache.accumulo.core.data.Value;
 +import org.apache.accumulo.core.data.thrift.IterInfo;
 +import org.apache.accumulo.core.file.FileOperations;
 +import org.apache.accumulo.core.file.FileSKVIterator;
 +import org.apache.accumulo.core.file.FileSKVWriter;
 +import org.apache.accumulo.core.iterators.IteratorEnvironment;
 +import org.apache.accumulo.core.iterators.IteratorUtil;
 +import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 +import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 +import org.apache.accumulo.core.iterators.WrappingIterator;
 +import org.apache.accumulo.core.iterators.system.ColumnFamilySkippingIterator;
 +import org.apache.accumulo.core.iterators.system.DeletingIterator;
 +import org.apache.accumulo.core.iterators.system.MultiIterator;
 +import org.apache.accumulo.core.iterators.system.TimeSettingIterator;
 +import org.apache.accumulo.core.metadata.schema.DataFileValue;
 +import org.apache.accumulo.core.tabletserver.thrift.ActiveCompaction;
 +import org.apache.accumulo.core.tabletserver.thrift.CompactionReason;
 +import org.apache.accumulo.core.tabletserver.thrift.CompactionType;
 +import org.apache.accumulo.core.util.LocalityGroupUtil;
 +import org.apache.accumulo.core.util.LocalityGroupUtil.LocalityGroupConfigurationError;
 +import org.apache.accumulo.server.fs.FileRef;
 +import org.apache.accumulo.server.fs.VolumeManager;
 +import org.apache.accumulo.server.problems.ProblemReport;
 +import org.apache.accumulo.server.problems.ProblemReportingIterator;
 +import org.apache.accumulo.server.problems.ProblemReports;
 +import org.apache.accumulo.server.problems.ProblemType;
 +import org.apache.accumulo.trace.instrument.Span;
 +import org.apache.accumulo.trace.instrument.Trace;
 +import org.apache.accumulo.tserver.Tablet.MinorCompactionReason;
 +import org.apache.accumulo.tserver.compaction.MajorCompactionReason;
 +import org.apache.hadoop.conf.Configuration;
 +import org.apache.hadoop.fs.FileSystem;
 +import org.apache.log4j.Logger;
 +
 +public class Compactor implements Callable<CompactionStats> {
 +
 +  public static class CountingIterator extends WrappingIterator {
 +
 +    private long count;
 +    private ArrayList<CountingIterator> deepCopies;
 +    private AtomicLong entriesRead;
 +
 +    @Override
 +    public CountingIterator deepCopy(IteratorEnvironment env) {
 +      return new CountingIterator(this, env);
 +    }
 +
 +    private CountingIterator(CountingIterator other, IteratorEnvironment env) {
 +      setSource(other.getSource().deepCopy(env));
 +      count = 0;
 +      this.deepCopies = other.deepCopies;
 +      this.entriesRead = other.entriesRead;
 +      deepCopies.add(this);
 +    }
 +
 +    public CountingIterator(SortedKeyValueIterator<Key,Value> source, AtomicLong entriesRead) {
 +      deepCopies = new ArrayList<Compactor.CountingIterator>();
 +      this.setSource(source);
 +      count = 0;
 +      this.entriesRead = entriesRead;
 +    }
 +
 +    @Override
 +    public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) {
 +      throw new UnsupportedOperationException();
 +    }
 +
 +    @Override
 +    public void next() throws IOException {
 +      super.next();
 +      count++;
 +      if (count % 1024 == 0) {
 +        entriesRead.addAndGet(1024);
 +      }
 +    }
 +
 +    public long getCount() {
 +      long sum = 0;
 +      for (CountingIterator dc : deepCopies) {
 +        sum += dc.count;
 +      }
 +
 +      return count + sum;
 +    }
 +  }
 +
 +  private static final Logger log = Logger.getLogger(Compactor.class);
 +
 +  static class CompactionCanceledException extends Exception {
 +    private static final long serialVersionUID = 1L;
 +  }
 +
 +  interface CompactionEnv {
 +    boolean isCompactionEnabled();
 +
 +    IteratorScope getIteratorScope();
 +  }
 +
 +  private Map<FileRef,DataFileValue> filesToCompact;
 +  private InMemoryMap imm;
 +  private FileRef outputFile;
 +  private boolean propogateDeletes;
 +  private AccumuloConfiguration acuTableConf;
 +  private CompactionEnv env;
 +  private Configuration conf;
 +  private VolumeManager fs;
 +  protected KeyExtent extent;
 +  private List<IteratorSetting> iterators;
 +
 +  // things to report
 +  private String currentLocalityGroup = "";
 +  private long startTime;
 +
 +  private MajorCompactionReason reason;
 +  protected MinorCompactionReason mincReason;
 +
 +  private AtomicLong entriesRead = new AtomicLong(0);
 +  private AtomicLong entriesWritten = new AtomicLong(0);
 +  private DateFormat dateFormatter = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss.SSS");
 +
 +  private static AtomicLong nextCompactorID = new AtomicLong(0);
 +
 +  // a unique id to identify a compactor
 +  private long compactorID = nextCompactorID.getAndIncrement();
 +
 +  protected volatile Thread thread;
 +
 +  private synchronized void setLocalityGroup(String name) {
 +    this.currentLocalityGroup = name;
 +  }
 +
 +  private void clearStats() {
 +    entriesRead.set(0);
 +    entriesWritten.set(0);
 +  }
 +
 +  protected static final Set<Compactor> runningCompactions = Collections.synchronizedSet(new HashSet<Compactor>());
 +
 +  public static class CompactionInfo {
 +
 +    private Compactor compactor;
 +    private String localityGroup;
 +    private long entriesRead;
 +    private long entriesWritten;
 +
 +    CompactionInfo(Compactor compactor) {
 +      this.localityGroup = compactor.currentLocalityGroup;
 +      this.entriesRead = compactor.entriesRead.get();
 +      this.entriesWritten = compactor.entriesWritten.get();
 +      this.compactor = compactor;
 +    }
 +
 +    public long getID() {
 +      return compactor.compactorID;
 +    }
 +
 +    public KeyExtent getExtent() {
 +      return compactor.getExtent();
 +    }
 +
 +    public long getEntriesRead() {
 +      return entriesRead;
 +    }
 +
 +    public long getEntriesWritten() {
 +      return entriesWritten;
 +    }
 +
 +    public Thread getThread() {
 +      return compactor.thread;
 +    }
 +
 +    public String getOutputFile() {
 +      return compactor.getOutputFile();
 +    }
 +
 +    public ActiveCompaction toThrift() {
 +
 +      CompactionType type;
 +
 +      if (compactor.imm != null)
 +        if (compactor.filesToCompact.size() > 0)
 +          type = CompactionType.MERGE;
 +        else
 +          type = CompactionType.MINOR;
 +      else if (!compactor.propogateDeletes)
 +        type = CompactionType.FULL;
 +      else
 +        type = CompactionType.MAJOR;
 +
 +      CompactionReason reason;
 +
-       if (compactor.imm != null)
++      if (compactor.imm != null) {
 +        switch (compactor.mincReason) {
 +          case USER:
 +            reason = CompactionReason.USER;
 +            break;
 +          case CLOSE:
 +            reason = CompactionReason.CLOSE;
 +            break;
 +          case SYSTEM:
 +          default:
 +            reason = CompactionReason.SYSTEM;
 +            break;
 +        }
-       else
++      } else {
 +        switch (compactor.reason) {
 +          case USER:
 +            reason = CompactionReason.USER;
 +            break;
 +          case CHOP:
 +            reason = CompactionReason.CHOP;
 +            break;
 +          case IDLE:
 +            reason = CompactionReason.IDLE;
 +            break;
 +          case NORMAL:
 +          default:
 +            reason = CompactionReason.SYSTEM;
 +            break;
 +        }
++      }
 +
 +      List<IterInfo> iiList = new ArrayList<IterInfo>();
 +      Map<String,Map<String,String>> iterOptions = new HashMap<String,Map<String,String>>();
 +
 +      for (IteratorSetting iterSetting : compactor.iterators) {
 +        iiList.add(new IterInfo(iterSetting.getPriority(), iterSetting.getIteratorClass(), iterSetting.getName()));
 +        iterOptions.put(iterSetting.getName(), iterSetting.getOptions());
 +      }
 +      List<String> filesToCompact = new ArrayList<String>();
 +      for (FileRef ref : compactor.filesToCompact.keySet())
 +        filesToCompact.add(ref.toString());
 +      return new ActiveCompaction(compactor.extent.toThrift(), System.currentTimeMillis() - compactor.startTime, filesToCompact,
 +          compactor.outputFile.toString(), type, reason, localityGroup, entriesRead, entriesWritten, iiList, iterOptions);
 +    }
 +  }
 +
 +  public static List<CompactionInfo> getRunningCompactions() {
 +    ArrayList<CompactionInfo> compactions = new ArrayList<Compactor.CompactionInfo>();
 +
 +    synchronized (runningCompactions) {
 +      for (Compactor compactor : runningCompactions) {
 +        compactions.add(new CompactionInfo(compactor));
 +      }
 +    }
 +
 +    return compactions;
 +  }
 +
 +  Compactor(Configuration conf, VolumeManager fs, Map<FileRef,DataFileValue> files, InMemoryMap imm, FileRef outputFile, boolean propogateDeletes,
 +      AccumuloConfiguration acuTableConf, KeyExtent extent, CompactionEnv env, List<IteratorSetting> iterators, MajorCompactionReason reason) {
 +    this.extent = extent;
 +    this.conf = conf;
 +    this.fs = fs;
 +    this.filesToCompact = files;
 +    this.imm = imm;
 +    this.outputFile = outputFile;
 +    this.propogateDeletes = propogateDeletes;
 +    this.acuTableConf = acuTableConf;
 +    this.env = env;
 +    this.iterators = iterators;
 +    this.reason = reason;
 +
 +    startTime = System.currentTimeMillis();
 +  }
 +
 +  Compactor(Configuration conf, VolumeManager fs, Map<FileRef,DataFileValue> files, InMemoryMap imm, FileRef outputFile, boolean propogateDeletes,
 +      AccumuloConfiguration acuTableConf, KeyExtent extent, CompactionEnv env) {
 +    this(conf, fs, files, imm, outputFile, propogateDeletes, acuTableConf, extent, env, new ArrayList<IteratorSetting>(), null);
 +  }
 +
 +  public VolumeManager getFileSystem() {
 +    return fs;
 +  }
 +
 +  KeyExtent getExtent() {
 +    return extent;
 +  }
 +
 +  String getOutputFile() {
 +    return outputFile.toString();
 +  }
 +
 +  @Override
 +  public CompactionStats call() throws IOException, CompactionCanceledException {
 +
 +    FileSKVWriter mfw = null;
 +
 +    CompactionStats majCStats = new CompactionStats();
 +
 +    boolean remove = runningCompactions.add(this);
 +
 +    clearStats();
 +
 +    String oldThreadName = Thread.currentThread().getName();
 +    String newThreadName = "MajC compacting " + extent.toString() + " started " + dateFormatter.format(new Date()) + " file: " + outputFile;
 +    Thread.currentThread().setName(newThreadName);
 +    thread = Thread.currentThread();
 +    try {
 +      FileOperations fileFactory = FileOperations.getInstance();
 +      FileSystem ns = this.fs.getVolumeByPath(outputFile.path()).getFileSystem();
 +      mfw = fileFactory.openWriter(outputFile.path().toString(), ns, ns.getConf(), acuTableConf);
 +
 +      Map<String,Set<ByteSequence>> lGroups;
 +      try {
 +        lGroups = LocalityGroupUtil.getLocalityGroups(acuTableConf);
 +      } catch (LocalityGroupConfigurationError e) {
 +        throw new IOException(e);
 +      }
 +
 +      long t1 = System.currentTimeMillis();
 +
 +      HashSet<ByteSequence> allColumnFamilies = new HashSet<ByteSequence>();
 +
 +      if (mfw.supportsLocalityGroups()) {
 +        for (Entry<String,Set<ByteSequence>> entry : lGroups.entrySet()) {
 +          setLocalityGroup(entry.getKey());
 +          compactLocalityGroup(entry.getKey(), entry.getValue(), true, mfw, majCStats);
 +          allColumnFamilies.addAll(entry.getValue());
 +        }
 +      }
 +
 +      setLocalityGroup("");
 +      compactLocalityGroup(null, allColumnFamilies, false, mfw, majCStats);
 +
 +      long t2 = System.currentTimeMillis();
 +
 +      FileSKVWriter mfwTmp = mfw;
 +      mfw = null; // set this to null so we do not try to close it again in finally if the close fails
 +      mfwTmp.close(); // if the close fails it will cause the compaction to fail
 +
 +      // Verify the file, since hadoop 0.20.2 sometimes lies about the success of close()
 +      try {
 +        FileSKVIterator openReader = fileFactory.openReader(outputFile.path().toString(), false, ns, ns.getConf(), acuTableConf);
 +        openReader.close();
 +      } catch (IOException ex) {
 +        log.error("Verification of successful compaction fails!!! " + extent + " " + outputFile, ex);
 +        throw ex;
 +      }
 +
 +      log.debug(String.format("Compaction %s %,d read | %,d written | %,6d entries/sec | %6.3f secs", extent, majCStats.getEntriesRead(),
 +          majCStats.getEntriesWritten(), (int) (majCStats.getEntriesRead() / ((t2 - t1) / 1000.0)), (t2 - t1) / 1000.0));
 +
 +      majCStats.setFileSize(fileFactory.getFileSize(outputFile.path().toString(), ns, ns.getConf(), acuTableConf));
 +      return majCStats;
 +    } catch (IOException e) {
 +      log.error(e, e);
 +      throw e;
 +    } catch (RuntimeException e) {
 +      log.error(e, e);
 +      throw e;
 +    } finally {
 +      Thread.currentThread().setName(oldThreadName);
 +      if (remove) {
 +        thread = null;
 +        runningCompactions.remove(this);
 +      }
 +
 +      try {
 +        if (mfw != null) {
 +          // compaction must not have finished successfully, so close its output file
 +          try {
 +            mfw.close();
 +          } finally {
 +            if (!fs.deleteRecursively(outputFile.path()))
 +              if (fs.exists(outputFile.path()))
 +                log.error("Unable to delete " + outputFile);
 +          }
 +        }
 +      } catch (IOException e) {
 +        log.warn(e, e);
 +      } catch (RuntimeException exception) {
 +        log.warn(exception, exception);
 +      }
 +    }
 +  }
 +
 +  private List<SortedKeyValueIterator<Key,Value>> openMapDataFiles(String lgName, ArrayList<FileSKVIterator> readers) throws IOException {
 +
 +    List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(filesToCompact.size());
 +
 +    for (FileRef mapFile : filesToCompact.keySet()) {
 +      try {
 +
 +        FileOperations fileFactory = FileOperations.getInstance();
 +        FileSystem fs = this.fs.getVolumeByPath(mapFile.path()).getFileSystem();
 +        FileSKVIterator reader;
 +
 +        reader = fileFactory.openReader(mapFile.path().toString(), false, fs, conf, acuTableConf);
 +
 +        readers.add(reader);
 +
 +        SortedKeyValueIterator<Key,Value> iter = new ProblemReportingIterator(extent.getTableId().toString(), mapFile.path().toString(), false, reader);
 +
 +        if (filesToCompact.get(mapFile).isTimeSet()) {
 +          iter = new TimeSettingIterator(iter, filesToCompact.get(mapFile).getTime());
 +        }
 +
 +        iters.add(iter);
 +
 +      } catch (Throwable e) {
 +
 +        ProblemReports.getInstance().report(new ProblemReport(extent.getTableId().toString(), ProblemType.FILE_READ, mapFile.path().toString(), e));
 +
 +        log.warn("Some problem opening map file " + mapFile + " " + e.getMessage(), e);
 +        // failed to open some map file... close the ones that were opened
 +        for (FileSKVIterator reader : readers) {
 +          try {
 +            reader.close();
 +          } catch (Throwable e2) {
 +            log.warn("Failed to close map file", e2);
 +          }
 +        }
 +
 +        readers.clear();
 +
 +        if (e instanceof IOException)
 +          throw (IOException) e;
 +        throw new IOException("Failed to open map data files", e);
 +      }
 +    }
 +
 +    return iters;
 +  }
 +
 +  private void compactLocalityGroup(String lgName, Set<ByteSequence> columnFamilies, boolean inclusive, FileSKVWriter mfw, CompactionStats majCStats)
 +      throws IOException, CompactionCanceledException {
 +    ArrayList<FileSKVIterator> readers = new ArrayList<FileSKVIterator>(filesToCompact.size());
 +    Span span = Trace.start("compact");
 +    try {
 +      long entriesCompacted = 0;
 +      List<SortedKeyValueIterator<Key,Value>> iters = openMapDataFiles(lgName, readers);
 +
 +      if (imm != null) {
 +        iters.add(imm.compactionIterator());
 +      }
 +
 +      CountingIterator citr = new CountingIterator(new MultiIterator(iters, extent.toDataRange()), entriesRead);
 +      DeletingIterator delIter = new DeletingIterator(citr, propogateDeletes);
 +      ColumnFamilySkippingIterator cfsi = new ColumnFamilySkippingIterator(delIter);
 +
 +      // if(env.getIteratorScope() )
 +
 +      TabletIteratorEnvironment iterEnv;
 +      if (env.getIteratorScope() == IteratorScope.majc)
 +        iterEnv = new TabletIteratorEnvironment(IteratorScope.majc, !propogateDeletes, acuTableConf);
 +      else if (env.getIteratorScope() == IteratorScope.minc)
 +        iterEnv = new TabletIteratorEnvironment(IteratorScope.minc, acuTableConf);
 +      else
 +        throw new IllegalArgumentException();
 +
 +      SortedKeyValueIterator<Key,Value> itr = iterEnv.getTopLevelIterator(IteratorUtil.loadIterators(env.getIteratorScope(), cfsi, extent, acuTableConf,
 +          iterators, iterEnv));
 +
 +      itr.seek(extent.toDataRange(), columnFamilies, inclusive);
 +
 +      if (!inclusive) {
 +        mfw.startDefaultLocalityGroup();
 +      } else {
 +        mfw.startNewLocalityGroup(lgName, columnFamilies);
 +      }
 +
 +      Span write = Trace.start("write");
 +      try {
 +        while (itr.hasTop() && env.isCompactionEnabled()) {
 +          mfw.append(itr.getTopKey(), itr.getTopValue());
 +          itr.next();
 +          entriesCompacted++;
 +
 +          if (entriesCompacted % 1024 == 0) {
 +            // Periodically update stats, do not want to do this too often since its volatile
 +            entriesWritten.addAndGet(1024);
 +          }
 +        }
 +
 +        if (itr.hasTop() && !env.isCompactionEnabled()) {
 +          // cancel major compaction operation
 +          try {
 +            try {
 +              mfw.close();
 +            } catch (IOException e) {
 +              log.error(e, e);
 +            }
 +            fs.deleteRecursively(outputFile.path());
 +          } catch (Exception e) {
 +            log.warn("Failed to delete Canceled compaction output file " + outputFile, e);
 +          }
 +          throw new CompactionCanceledException();
 +        }
 +
 +      } finally {
 +        CompactionStats lgMajcStats = new CompactionStats(citr.getCount(), entriesCompacted);
 +        majCStats.add(lgMajcStats);
 +        write.stop();
 +      }
 +
 +    } finally {
 +      // close sequence files opened
 +      for (FileSKVIterator reader : readers) {
 +        try {
 +          reader.close();
 +        } catch (Throwable e) {
 +          log.warn("Failed to close map file", e);
 +        }
 +      }
 +      span.stop();
 +    }
 +  }
 +
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------
diff --cc start/src/main/java/org/apache/accumulo/start/Main.java
index c4aaea1,a80ebe6..b9753bf
--- a/start/src/main/java/org/apache/accumulo/start/Main.java
+++ b/start/src/main/java/org/apache/accumulo/start/Main.java
@@@ -175,28 -141,7 +175,28 @@@ public class Main 
    }
  
    private static void printUsage() {
-     System.out
-         .println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info | tracer | minicluster | proxy | zookeeper | create-token | info | version | help | jar <jar> [<main class>] args | <accumulo class> args");
+     System.out.println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info "
 -        + "| tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
++        + "| tracer | minicluster | proxy | zookeeper | create-token | info | version | help | jar <jar> [<main class>] args | <accumulo class> args");
 +  }
 +
 +  // feature: will work even if main class isn't in the JAR
 +  static Class<?> loadClassFromJar(String[] args, JarFile f, ClassLoader cl) throws IOException, ClassNotFoundException {
 +    ClassNotFoundException explicitNotFound = null;
 +    if (args.length >= 3) {
 +      try {
 +        return cl.loadClass(args[2]); // jar jar-file main-class
 +      } catch (ClassNotFoundException cnfe) {
 +        // assume this is the first argument, look for main class in JAR manifest
 +        explicitNotFound = cnfe;
 +      }
 +    }
 +    String mainClass = f.getManifest().getMainAttributes().getValue(Attributes.Name.MAIN_CLASS);
 +    if (mainClass == null) {
 +      if (explicitNotFound != null) {
 +        throw explicitNotFound;
 +      }
 +      throw new ClassNotFoundException("No main class was specified, and the JAR manifest does not specify one");
 +    }
 +    return cl.loadClass(mainClass);
    }
  }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------


[37/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/RelativeKey.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/RelativeKey.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/RelativeKey.java
index 07bf6d3..aeba4e2 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/RelativeKey.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/RelativeKey.java
@@ -28,15 +28,15 @@ import org.apache.hadoop.io.Writable;
 import org.apache.hadoop.io.WritableUtils;
 
 public class RelativeKey implements Writable {
-  
+
   private static final byte BIT = 0x01;
-  
+
   private Key key;
   private Key prevKey;
-  
+
   private byte fieldsSame;
   private byte fieldsPrefixed;
-  
+
   // Exact match compression options (first byte) and flag for further
   private static final byte ROW_SAME = BIT << 0;
   private static final byte CF_SAME = BIT << 1;
@@ -46,47 +46,47 @@ public class RelativeKey implements Writable {
   private static final byte DELETED = BIT << 5;
   // private static final byte UNUSED_1_6 = BIT << 6;
   private static final byte PREFIX_COMPRESSION_ENABLED = (byte) (BIT << 7);
-  
+
   // Prefix compression (second byte)
   private static final byte ROW_COMMON_PREFIX = BIT << 0;
   private static final byte CF_COMMON_PREFIX = BIT << 1;
   private static final byte CQ_COMMON_PREFIX = BIT << 2;
   private static final byte CV_COMMON_PREFIX = BIT << 3;
   private static final byte TS_DIFF = BIT << 4;
-  
+
   // private static final byte UNUSED_2_5 = BIT << 5;
   // private static final byte UNUSED_2_6 = BIT << 6;
   // private static final byte UNUSED_2_7 = (byte) (BIT << 7);
-  
+
   // Values for prefix compression
   int rowCommonPrefixLen;
   int cfCommonPrefixLen;
   int cqCommonPrefixLen;
   int cvCommonPrefixLen;
   long tsDiff;
-  
+
   /**
    * This constructor is used when one needs to read from an input stream
    */
   public RelativeKey() {
-    
+
   }
-  
+
   /**
    * This constructor is used when constructing a key for writing to an output stream
    */
   public RelativeKey(Key prevKey, Key key) {
-    
+
     this.key = key;
-    
+
     fieldsSame = 0;
     fieldsPrefixed = 0;
-    
+
     ByteSequence prevKeyScratch;
     ByteSequence keyScratch;
-    
+
     if (prevKey != null) {
-      
+
       prevKeyScratch = prevKey.getRowData();
       keyScratch = key.getRowData();
       rowCommonPrefixLen = getCommonPrefix(prevKeyScratch, keyScratch);
@@ -94,7 +94,7 @@ public class RelativeKey implements Writable {
         fieldsSame |= ROW_SAME;
       else if (rowCommonPrefixLen > 1)
         fieldsPrefixed |= ROW_COMMON_PREFIX;
-      
+
       prevKeyScratch = prevKey.getColumnFamilyData();
       keyScratch = key.getColumnFamilyData();
       cfCommonPrefixLen = getCommonPrefix(prevKeyScratch, keyScratch);
@@ -102,7 +102,7 @@ public class RelativeKey implements Writable {
         fieldsSame |= CF_SAME;
       else if (cfCommonPrefixLen > 1)
         fieldsPrefixed |= CF_COMMON_PREFIX;
-      
+
       prevKeyScratch = prevKey.getColumnQualifierData();
       keyScratch = key.getColumnQualifierData();
       cqCommonPrefixLen = getCommonPrefix(prevKeyScratch, keyScratch);
@@ -110,7 +110,7 @@ public class RelativeKey implements Writable {
         fieldsSame |= CQ_SAME;
       else if (cqCommonPrefixLen > 1)
         fieldsPrefixed |= CQ_COMMON_PREFIX;
-      
+
       prevKeyScratch = prevKey.getColumnVisibilityData();
       keyScratch = key.getColumnVisibilityData();
       cvCommonPrefixLen = getCommonPrefix(prevKeyScratch, keyScratch);
@@ -118,29 +118,29 @@ public class RelativeKey implements Writable {
         fieldsSame |= CV_SAME;
       else if (cvCommonPrefixLen > 1)
         fieldsPrefixed |= CV_COMMON_PREFIX;
-      
+
       tsDiff = key.getTimestamp() - prevKey.getTimestamp();
       if (tsDiff == 0)
         fieldsSame |= TS_SAME;
       else
         fieldsPrefixed |= TS_DIFF;
-      
+
       fieldsSame |= fieldsPrefixed == 0 ? 0 : PREFIX_COMPRESSION_ENABLED;
     }
-    
+
     // stored deleted information in bit vector instead of its own byte
     if (key.isDeleted())
       fieldsSame |= DELETED;
   }
-  
+
   /**
-   * 
+   *
    * @return -1 (exact match) or the number of bytes in common
    */
   static int getCommonPrefix(ByteSequence prev, ByteSequence cur) {
     if (prev == cur)
       return -1; // infinite... exact match
-      
+
     int prevLen = prev.length();
     int curLen = cur.length();
     int maxChecks = Math.min(prevLen, curLen);
@@ -157,11 +157,11 @@ public class RelativeKey implements Writable {
     // and if not, then they have a common prefix over all the checks we've done
     return prevLen == curLen ? -1 : maxChecks;
   }
-  
+
   public void setPrevKey(Key pk) {
     this.prevKey = pk;
   }
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     fieldsSame = in.readByte();
@@ -170,10 +170,10 @@ public class RelativeKey implements Writable {
     } else {
       fieldsPrefixed = 0;
     }
-    
+
     byte[] row, cf, cq, cv;
     long ts;
-    
+
     if ((fieldsSame & ROW_SAME) == ROW_SAME) {
       row = prevKey.getRowData().toArray();
     } else if ((fieldsPrefixed & ROW_COMMON_PREFIX) == ROW_COMMON_PREFIX) {
@@ -181,7 +181,7 @@ public class RelativeKey implements Writable {
     } else {
       row = read(in);
     }
-    
+
     if ((fieldsSame & CF_SAME) == CF_SAME) {
       cf = prevKey.getColumnFamilyData().toArray();
     } else if ((fieldsPrefixed & CF_COMMON_PREFIX) == CF_COMMON_PREFIX) {
@@ -189,7 +189,7 @@ public class RelativeKey implements Writable {
     } else {
       cf = read(in);
     }
-    
+
     if ((fieldsSame & CQ_SAME) == CQ_SAME) {
       cq = prevKey.getColumnQualifierData().toArray();
     } else if ((fieldsPrefixed & CQ_COMMON_PREFIX) == CQ_COMMON_PREFIX) {
@@ -197,7 +197,7 @@ public class RelativeKey implements Writable {
     } else {
       cq = read(in);
     }
-    
+
     if ((fieldsSame & CV_SAME) == CV_SAME) {
       cv = prevKey.getColumnVisibilityData().toArray();
     } else if ((fieldsPrefixed & CV_COMMON_PREFIX) == CV_COMMON_PREFIX) {
@@ -205,7 +205,7 @@ public class RelativeKey implements Writable {
     } else {
       cv = read(in);
     }
-    
+
     if ((fieldsSame & TS_SAME) == TS_SAME) {
       ts = prevKey.getTimestamp();
     } else if ((fieldsPrefixed & TS_DIFF) == TS_DIFF) {
@@ -213,75 +213,75 @@ public class RelativeKey implements Writable {
     } else {
       ts = WritableUtils.readVLong(in);
     }
-    
+
     this.key = new Key(row, cf, cq, cv, ts, (fieldsSame & DELETED) == DELETED, false);
     this.prevKey = this.key;
   }
-  
+
   public static class SkippR {
     RelativeKey rk;
     int skipped;
     Key prevKey;
-    
+
     SkippR(RelativeKey rk, int skipped, Key prevKey) {
       this.rk = rk;
       this.skipped = skipped;
       this.prevKey = prevKey;
     }
   }
-  
+
   public static SkippR fastSkip(DataInput in, Key seekKey, MutableByteSequence value, Key prevKey, Key currKey) throws IOException {
     // this method assumes that fast skip is being called on a compressed block where the last key
     // in the compressed block is >= seekKey... therefore this method shouldn't go past the end of the
     // compressed block... if it does, there is probably an error in the caller's logic
-    
+
     // this method mostly avoids object allocation and only does compares when the row changes
-    
+
     MutableByteSequence row, cf, cq, cv;
     MutableByteSequence prow, pcf, pcq, pcv;
-    
+
     ByteSequence stopRow = seekKey.getRowData();
     ByteSequence stopCF = seekKey.getColumnFamilyData();
     ByteSequence stopCQ = seekKey.getColumnQualifierData();
-    
+
     long ts = -1;
     long pts = -1;
     boolean pdel = false;
-    
+
     int rowCmp = -1, cfCmp = -1, cqCmp = -1;
-    
+
     if (currKey != null) {
-      
+
       prow = new MutableByteSequence(currKey.getRowData());
       pcf = new MutableByteSequence(currKey.getColumnFamilyData());
       pcq = new MutableByteSequence(currKey.getColumnQualifierData());
       pcv = new MutableByteSequence(currKey.getColumnVisibilityData());
       pts = currKey.getTimestamp();
-      
+
       row = new MutableByteSequence(currKey.getRowData());
       cf = new MutableByteSequence(currKey.getColumnFamilyData());
       cq = new MutableByteSequence(currKey.getColumnQualifierData());
       cv = new MutableByteSequence(currKey.getColumnVisibilityData());
       ts = currKey.getTimestamp();
-      
+
       rowCmp = row.compareTo(stopRow);
       cfCmp = cf.compareTo(stopCF);
       cqCmp = cq.compareTo(stopCQ);
-      
+
       if (rowCmp >= 0) {
         if (rowCmp > 0) {
           RelativeKey rk = new RelativeKey();
           rk.key = rk.prevKey = new Key(currKey);
           return new SkippR(rk, 0, prevKey);
         }
-        
+
         if (cfCmp >= 0) {
           if (cfCmp > 0) {
             RelativeKey rk = new RelativeKey();
             rk.key = rk.prevKey = new Key(currKey);
             return new SkippR(rk, 0, prevKey);
           }
-          
+
           if (cqCmp >= 0) {
             RelativeKey rk = new RelativeKey();
             rk.key = rk.prevKey = new Key(currKey);
@@ -289,126 +289,126 @@ public class RelativeKey implements Writable {
           }
         }
       }
-      
+
     } else {
       row = new MutableByteSequence(new byte[64], 0, 0);
       cf = new MutableByteSequence(new byte[64], 0, 0);
       cq = new MutableByteSequence(new byte[64], 0, 0);
       cv = new MutableByteSequence(new byte[64], 0, 0);
-      
+
       prow = new MutableByteSequence(new byte[64], 0, 0);
       pcf = new MutableByteSequence(new byte[64], 0, 0);
       pcq = new MutableByteSequence(new byte[64], 0, 0);
       pcv = new MutableByteSequence(new byte[64], 0, 0);
     }
-    
+
     byte fieldsSame = -1;
     byte fieldsPrefixed = 0;
     int count = 0;
     Key newPrevKey = null;
-    
+
     while (true) {
-      
+
       pdel = (fieldsSame & DELETED) == DELETED;
-      
+
       fieldsSame = in.readByte();
       if ((fieldsSame & PREFIX_COMPRESSION_ENABLED) == PREFIX_COMPRESSION_ENABLED)
         fieldsPrefixed = in.readByte();
       else
         fieldsPrefixed = 0;
-      
+
       boolean changed = false;
-      
+
       if ((fieldsSame & ROW_SAME) != ROW_SAME) {
-        
+
         MutableByteSequence tmp = prow;
         prow = row;
         row = tmp;
-        
+
         if ((fieldsPrefixed & ROW_COMMON_PREFIX) == ROW_COMMON_PREFIX)
           readPrefix(in, row, prow);
         else
-        read(in, row);
-        
+          read(in, row);
+
         // read a new row, so need to compare...
         rowCmp = row.compareTo(stopRow);
         changed = true;
       }// else the row is the same as the last, so no need to compare
-      
+
       if ((fieldsSame & CF_SAME) != CF_SAME) {
-        
+
         MutableByteSequence tmp = pcf;
         pcf = cf;
         cf = tmp;
-        
+
         if ((fieldsPrefixed & CF_COMMON_PREFIX) == CF_COMMON_PREFIX)
           readPrefix(in, cf, pcf);
         else
-        read(in, cf);
-        
+          read(in, cf);
+
         cfCmp = cf.compareTo(stopCF);
         changed = true;
       }
-      
+
       if ((fieldsSame & CQ_SAME) != CQ_SAME) {
-        
+
         MutableByteSequence tmp = pcq;
         pcq = cq;
         cq = tmp;
-        
+
         if ((fieldsPrefixed & CQ_COMMON_PREFIX) == CQ_COMMON_PREFIX)
           readPrefix(in, cq, pcq);
         else
-        read(in, cq);
-        
+          read(in, cq);
+
         cqCmp = cq.compareTo(stopCQ);
         changed = true;
       }
-      
+
       if ((fieldsSame & CV_SAME) != CV_SAME) {
-        
+
         MutableByteSequence tmp = pcv;
         pcv = cv;
         cv = tmp;
-        
+
         if ((fieldsPrefixed & CV_COMMON_PREFIX) == CV_COMMON_PREFIX)
           readPrefix(in, cv, pcv);
         else
-        read(in, cv);
+          read(in, cv);
       }
-      
+
       if ((fieldsSame & TS_SAME) != TS_SAME) {
         pts = ts;
-        
+
         if ((fieldsPrefixed & TS_DIFF) == TS_DIFF)
           ts = WritableUtils.readVLong(in) + pts;
         else
-        ts = WritableUtils.readVLong(in);
+          ts = WritableUtils.readVLong(in);
       }
-      
+
       readValue(in, value);
-      
+
       count++;
-      
+
       if (changed && rowCmp >= 0) {
         if (rowCmp > 0)
           break;
-        
+
         if (cfCmp >= 0) {
           if (cfCmp > 0)
             break;
-          
+
           if (cqCmp >= 0)
             break;
         }
       }
-      
+
     }
-    
+
     if (count > 1) {
       MutableByteSequence trow, tcf, tcq, tcv;
       long tts;
-      
+
       // when the current keys field is same as the last, then
       // set the prev keys field the same as the current key
       trow = (fieldsSame & ROW_SAME) == ROW_SAME ? row : prow;
@@ -416,7 +416,7 @@ public class RelativeKey implements Writable {
       tcq = (fieldsSame & CQ_SAME) == CQ_SAME ? cq : pcq;
       tcv = (fieldsSame & CV_SAME) == CV_SAME ? cv : pcv;
       tts = (fieldsSame & TS_SAME) == TS_SAME ? ts : pts;
-      
+
       newPrevKey = new Key(trow.getBackingArray(), trow.offset(), trow.length(), tcf.getBackingArray(), tcf.offset(), tcf.length(), tcq.getBackingArray(),
           tcq.offset(), tcq.length(), tcv.getBackingArray(), tcv.offset(), tcv.length(), tts);
       newPrevKey.setDeleted(pdel);
@@ -428,35 +428,35 @@ public class RelativeKey implements Writable {
     } else {
       throw new IllegalStateException();
     }
-    
+
     RelativeKey result = new RelativeKey();
     result.key = new Key(row.getBackingArray(), row.offset(), row.length(), cf.getBackingArray(), cf.offset(), cf.length(), cq.getBackingArray(), cq.offset(),
         cq.length(), cv.getBackingArray(), cv.offset(), cv.length(), ts);
     result.key.setDeleted((fieldsSame & DELETED) != 0);
     result.prevKey = result.key;
-    
+
     return new SkippR(result, count, newPrevKey);
   }
-  
+
   private static void read(DataInput in, MutableByteSequence mbseq) throws IOException {
     int len = WritableUtils.readVInt(in);
     read(in, mbseq, len);
   }
-  
+
   private static void readValue(DataInput in, MutableByteSequence mbseq) throws IOException {
     int len = in.readInt();
     read(in, mbseq, len);
   }
-  
+
   private static void read(DataInput in, MutableByteSequence mbseqDestination, int len) throws IOException {
     if (mbseqDestination.getBackingArray().length < len) {
       mbseqDestination.setArray(new byte[UnsynchronizedBuffer.nextArraySize(len)], 0, 0);
     }
-    
+
     in.readFully(mbseqDestination.getBackingArray(), 0, len);
     mbseqDestination.setLength(len);
   }
-  
+
   private static byte[] readPrefix(DataInput in, ByteSequence prefixSource) throws IOException {
     int prefixLen = WritableUtils.readVInt(in);
     int remainingLen = WritableUtils.readVInt(in);
@@ -470,8 +470,8 @@ public class RelativeKey implements Writable {
     // read remaining
     in.readFully(data, prefixLen, remainingLen);
     return data;
-    }
-    
+  }
+
   private static void readPrefix(DataInput in, MutableByteSequence dest, ByteSequence prefixSource) throws IOException {
     int prefixLen = WritableUtils.readVInt(in);
     int remainingLen = WritableUtils.readVInt(in);
@@ -489,38 +489,38 @@ public class RelativeKey implements Writable {
     in.readFully(dest.getBackingArray(), prefixLen, remainingLen);
     dest.setLength(len);
   }
-  
+
   private static byte[] read(DataInput in) throws IOException {
     int len = WritableUtils.readVInt(in);
     byte[] data = new byte[len];
     in.readFully(data);
     return data;
   }
-  
+
   public Key getKey() {
     return key;
   }
-  
+
   private static void write(DataOutput out, ByteSequence bs) throws IOException {
     WritableUtils.writeVInt(out, bs.length());
     out.write(bs.getBackingArray(), bs.offset(), bs.length());
   }
-  
+
   private static void writePrefix(DataOutput out, ByteSequence bs, int commonPrefixLength) throws IOException {
     WritableUtils.writeVInt(out, commonPrefixLength);
     WritableUtils.writeVInt(out, bs.length() - commonPrefixLength);
     out.write(bs.getBackingArray(), bs.offset() + commonPrefixLength, bs.length() - commonPrefixLength);
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
-    
+
     out.writeByte(fieldsSame);
-    
+
     if ((fieldsSame & PREFIX_COMPRESSION_ENABLED) == PREFIX_COMPRESSION_ENABLED) {
       out.write(fieldsPrefixed);
     }
-    
+
     if ((fieldsSame & ROW_SAME) == ROW_SAME) {
       // same, write nothing
     } else if ((fieldsPrefixed & ROW_COMMON_PREFIX) == ROW_COMMON_PREFIX) {
@@ -530,7 +530,7 @@ public class RelativeKey implements Writable {
       // write it all
       write(out, key.getRowData());
     }
-    
+
     if ((fieldsSame & CF_SAME) == CF_SAME) {
       // same, write nothing
     } else if ((fieldsPrefixed & CF_COMMON_PREFIX) == CF_COMMON_PREFIX) {
@@ -540,7 +540,7 @@ public class RelativeKey implements Writable {
       // write it all
       write(out, key.getColumnFamilyData());
     }
-    
+
     if ((fieldsSame & CQ_SAME) == CQ_SAME) {
       // same, write nothing
     } else if ((fieldsPrefixed & CQ_COMMON_PREFIX) == CQ_COMMON_PREFIX) {
@@ -550,7 +550,7 @@ public class RelativeKey implements Writable {
       // write it all
       write(out, key.getColumnQualifierData());
     }
-    
+
     if ((fieldsSame & CV_SAME) == CV_SAME) {
       // same, write nothing
     } else if ((fieldsPrefixed & CV_COMMON_PREFIX) == CV_COMMON_PREFIX) {
@@ -560,7 +560,7 @@ public class RelativeKey implements Writable {
       // write it all
       write(out, key.getColumnVisibilityData());
     }
-    
+
     if ((fieldsSame & TS_SAME) == TS_SAME) {
       // same, write nothing
     } else if ((fieldsPrefixed & TS_DIFF) == TS_DIFF) {
@@ -571,5 +571,5 @@ public class RelativeKey implements Writable {
       WritableUtils.writeVLong(out, key.getTimestamp());
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/SplitLarge.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/SplitLarge.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/SplitLarge.java
index 53e4aaa..b87705c 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/SplitLarge.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/SplitLarge.java
@@ -39,37 +39,36 @@ import com.beust.jcommander.Parameter;
 
 /**
  * Split an RFile into large and small key/value files.
- * 
+ *
  */
 public class SplitLarge {
-  
+
   static class Opts extends Help {
-    @Parameter(names="-m", description="the maximum size of the key/value pair to shunt to the small file")
+    @Parameter(names = "-m", description = "the maximum size of the key/value pair to shunt to the small file")
     long maxSize = 10 * 1024 * 1024;
-    @Parameter(description="<file.rf> { <file.rf> ... }")
+    @Parameter(description = "<file.rf> { <file.rf> ... }")
     List<String> files = new ArrayList<String>();
   }
-  
-  
+
   public static void main(String[] args) throws Exception {
     Configuration conf = CachedConfiguration.getInstance();
     FileSystem fs = FileSystem.get(conf);
     long maxSize = 10 * 1024 * 1024;
     Opts opts = new Opts();
     opts.parseArgs(SplitLarge.class.getName(), args);
-    
+
     for (String file : opts.files) {
-      AccumuloConfiguration aconf = DefaultConfiguration.getDefaultConfiguration(); 
+      AccumuloConfiguration aconf = DefaultConfiguration.getDefaultConfiguration();
       Path path = new Path(file);
       CachableBlockFile.Reader rdr = new CachableBlockFile.Reader(fs, path, conf, null, null, aconf);
       Reader iter = new RFile.Reader(rdr);
-      
+
       if (!file.endsWith(".rf")) {
         throw new IllegalArgumentException("File must end with .rf");
       }
       String smallName = file.substring(0, file.length() - 3) + "_small.rf";
       String largeName = file.substring(0, file.length() - 3) + "_large.rf";
-      
+
       int blockSize = (int) aconf.getMemoryInBytes(Property.TABLE_FILE_BLOCK_SIZE);
       Writer small = new RFile.Writer(new CachableBlockFile.Writer(fs, new Path(smallName), "gz", conf, aconf), blockSize);
       small.startDefaultLocalityGroup();
@@ -93,5 +92,5 @@ public class SplitLarge {
       small.close();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BCFile.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BCFile.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BCFile.java
index 044989d..ecc0b90 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BCFile.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BCFile.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -109,7 +109,7 @@ public final class BCFile {
     private interface BlockRegister {
       /**
        * Register a block that is fully closed.
-       * 
+       *
        * @param raw
        *          The size of block in terms of uncompressed bytes.
        * @param offsetStart
@@ -198,7 +198,7 @@ public final class BCFile {
 
       /**
        * Get the output stream for BlockAppender's consumption.
-       * 
+       *
        * @return the output stream suitable for writing block data.
        */
       OutputStream getOutputStream() {
@@ -207,7 +207,7 @@ public final class BCFile {
 
       /**
        * Get the current position in file.
-       * 
+       *
        * @return The current byte offset in underlying file.
        */
       long getCurrentPos() throws IOException {
@@ -256,7 +256,7 @@ public final class BCFile {
 
     /**
      * Access point to stuff data into a block.
-     * 
+     *
      */
     public class BlockAppender extends DataOutputStream {
       private final BlockRegister blockRegister;
@@ -265,7 +265,7 @@ public final class BCFile {
 
       /**
        * Constructor
-       * 
+       *
        * @param register
        *          the block register, which is called when the block is closed.
        * @param wbs
@@ -279,7 +279,7 @@ public final class BCFile {
 
       /**
        * Get the raw size of the block.
-       * 
+       *
        * @return the number of uncompressed bytes written through the BlockAppender so far.
        */
       public long getRawSize() throws IOException {
@@ -291,7 +291,7 @@ public final class BCFile {
 
       /**
        * Get the compressed size of the block in progress.
-       * 
+       *
        * @return the number of compressed bytes written to the underlying FS file. The size may be smaller than actual need to compress the all data written due
        *         to internal buffering inside the compressor.
        */
@@ -331,14 +331,15 @@ public final class BCFile {
 
     /**
      * Constructor
-     * 
+     *
      * @param fout
      *          FS output stream.
      * @param compressionName
      *          Name of the compression algorithm, which will be used for all data blocks.
      * @see Compression#getSupportedAlgorithms
      */
-    public Writer(FSDataOutputStream fout, String compressionName, Configuration conf, boolean trackDataBlocks, AccumuloConfiguration accumuloConfiguration) throws IOException {
+    public Writer(FSDataOutputStream fout, String compressionName, Configuration conf, boolean trackDataBlocks, AccumuloConfiguration accumuloConfiguration)
+        throws IOException {
       if (fout.getPos() != 0) {
         throw new IOException("Output file not at zero offset.");
       }
@@ -435,7 +436,7 @@ public final class BCFile {
     /**
      * Create a Meta Block and obtain an output stream for adding data into the block. There can only be one BlockAppender stream active at any time. Regular
      * Blocks may not be created after the first Meta Blocks. The caller must call BlockAppender.close() to conclude the block creation.
-     * 
+     *
      * @param name
      *          The name of the Meta Block. The name must not conflict with existing Meta Blocks.
      * @param compressionName
@@ -452,7 +453,7 @@ public final class BCFile {
      * Create a Meta Block and obtain an output stream for adding data into the block. The Meta Block will be compressed with the same compression algorithm as
      * data blocks. There can only be one BlockAppender stream active at any time. Regular Blocks may not be created after the first Meta Blocks. The caller
      * must call BlockAppender.close() to conclude the block creation.
-     * 
+     *
      * @param name
      *          The name of the Meta Block. The name must not conflict with existing Meta Blocks.
      * @return The BlockAppender stream
@@ -466,7 +467,7 @@ public final class BCFile {
     /**
      * Create a Data Block and obtain an output stream for adding data into the block. There can only be one BlockAppender stream active at any time. Data
      * Blocks may not be created after the first Meta Blocks. The caller must call BlockAppender.close() to conclude the block creation.
-     * 
+     *
      * @return The BlockAppender stream
      */
     public BlockAppender prepareDataBlock() throws IOException {
@@ -506,7 +507,7 @@ public final class BCFile {
 
     /**
      * Callback to make sure a data block is added to the internal list when it's being closed.
-     * 
+     *
      */
     private class DataBlockRegister implements BlockRegister {
       DataBlockRegister() {
@@ -628,7 +629,7 @@ public final class BCFile {
 
       /**
        * Get the output stream for BlockAppender's consumption.
-       * 
+       *
        * @return the output stream suitable for writing block data.
        */
       public InputStream getInputStream() {
@@ -684,7 +685,7 @@ public final class BCFile {
 
       /**
        * Get the name of the compression algorithm used to compress the block.
-       * 
+       *
        * @return name of the compression algorithm.
        */
       public String getCompressionName() {
@@ -693,7 +694,7 @@ public final class BCFile {
 
       /**
        * Get the uncompressed size of the block.
-       * 
+       *
        * @return uncompressed size of the block.
        */
       public long getRawSize() {
@@ -702,7 +703,7 @@ public final class BCFile {
 
       /**
        * Get the compressed size of the block.
-       * 
+       *
        * @return compressed size of the block.
        */
       public long getCompressedSize() {
@@ -711,7 +712,7 @@ public final class BCFile {
 
       /**
        * Get the starting position of the block in the file.
-       * 
+       *
        * @return the starting position of the block in the file.
        */
       public long getStartPos() {
@@ -721,7 +722,7 @@ public final class BCFile {
 
     /**
      * Constructor
-     * 
+     *
      * @param fin
      *          FS input stream.
      * @param fileLength
@@ -807,7 +808,8 @@ public final class BCFile {
       }
     }
 
-    public Reader(CachableBlockFile.Reader cache, FSDataInputStream fin, long fileLength, Configuration conf, AccumuloConfiguration accumuloConfiguration) throws IOException {
+    public Reader(CachableBlockFile.Reader cache, FSDataInputStream fin, long fileLength, Configuration conf, AccumuloConfiguration accumuloConfiguration)
+        throws IOException {
       this.in = fin;
       this.conf = conf;
 
@@ -943,7 +945,7 @@ public final class BCFile {
 
     /**
      * Get the name of the default compression algorithm.
-     * 
+     *
      * @return the name of the default compression algorithm.
      */
     public String getDefaultCompressionName() {
@@ -952,7 +954,7 @@ public final class BCFile {
 
     /**
      * Get version of BCFile file being read.
-     * 
+     *
      * @return version of BCFile file being read.
      */
     public Version getBCFileVersion() {
@@ -961,7 +963,7 @@ public final class BCFile {
 
     /**
      * Get version of BCFile API.
-     * 
+     *
      * @return version of BCFile API.
      */
     public Version getAPIVersion() {
@@ -978,7 +980,7 @@ public final class BCFile {
 
     /**
      * Get the number of data blocks.
-     * 
+     *
      * @return the number of data blocks.
      */
     public int getBlockCount() {
@@ -987,7 +989,7 @@ public final class BCFile {
 
     /**
      * Stream access to a Meta Block.
-     * 
+     *
      * @param name
      *          meta block name
      * @return BlockReader input stream for reading the meta block.
@@ -1006,7 +1008,7 @@ public final class BCFile {
 
     /**
      * Stream access to a Data Block.
-     * 
+     *
      * @param blockIndex
      *          0-based data block index.
      * @return BlockReader input stream for reading the data block.
@@ -1032,7 +1034,7 @@ public final class BCFile {
 
     /**
      * Find the smallest Block index whose starting offset is greater than or equal to the specified offset.
-     * 
+     *
      * @param offset
      *          User-specific offset.
      * @return the index to the data Block if such block exists; or -1 otherwise.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BoundedRangeFileInputStream.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BoundedRangeFileInputStream.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BoundedRangeFileInputStream.java
index e6c2a66..f93bb84 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BoundedRangeFileInputStream.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/BoundedRangeFileInputStream.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -30,46 +30,46 @@ import org.apache.hadoop.fs.FSDataInputStream;
  * BoundedRangeFileInputStream on top of the same FSDataInputStream and they would not interfere with each other.
  */
 class BoundedRangeFileInputStream extends InputStream {
-  
+
   private FSDataInputStream in;
   private long pos;
   private long end;
   private long mark;
   private final byte[] oneByte = new byte[1];
-  
+
   /**
    * Constructor
-   * 
+   *
    * @param in
    *          The FSDataInputStream we connect to.
    * @param offset
    *          Beginning offset of the region.
    * @param length
    *          Length of the region.
-   * 
+   *
    *          The actual length of the region may be smaller if (off_begin + length) goes beyond the end of FS input stream.
    */
   public BoundedRangeFileInputStream(FSDataInputStream in, long offset, long length) {
     if (offset < 0 || length < 0) {
       throw new IndexOutOfBoundsException("Invalid offset/length: " + offset + "/" + length);
     }
-    
+
     this.in = in;
     this.pos = offset;
     this.end = offset + length;
     this.mark = -1;
   }
-  
+
   @Override
   public int available() throws IOException {
     int avail = in.available();
     if (pos + avail > end) {
       avail = (int) (end - pos);
     }
-    
+
     return avail;
   }
-  
+
   @Override
   public int read() throws IOException {
     int ret = read(oneByte);
@@ -77,25 +77,25 @@ class BoundedRangeFileInputStream extends InputStream {
       return oneByte[0] & 0xff;
     return -1;
   }
-  
+
   @Override
   public int read(byte[] b) throws IOException {
     return read(b, 0, b.length);
   }
-  
+
   @Override
   public int read(final byte[] b, final int off, int len) throws IOException {
     if ((off | len | (off + len) | (b.length - (off + len))) < 0) {
       throw new IndexOutOfBoundsException();
     }
-    
+
     final int n = (int) Math.min(Integer.MAX_VALUE, Math.min(len, (end - pos)));
     if (n == 0)
       return -1;
     Integer ret = 0;
     final FSDataInputStream inLocal = in;
     synchronized (inLocal) {
-    	inLocal.seek(pos);
+      inLocal.seek(pos);
       try {
         ret = AccessController.doPrivileged(new PrivilegedExceptionAction<Integer>() {
           @Override
@@ -116,7 +116,7 @@ class BoundedRangeFileInputStream extends InputStream {
     pos += ret;
     return ret;
   }
-  
+
   @Override
   /*
    * We may skip beyond the end of the file.
@@ -126,24 +126,24 @@ class BoundedRangeFileInputStream extends InputStream {
     pos += len;
     return len;
   }
-  
+
   @Override
   public void mark(int readlimit) {
     mark = pos;
   }
-  
+
   @Override
   public void reset() throws IOException {
     if (mark < 0)
       throw new IOException("Resetting to invalid mark");
     pos = mark;
   }
-  
+
   @Override
   public boolean markSupported() {
     return true;
   }
-  
+
   @Override
   public void close() {
     // Invalidate the state of the stream.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/CompareUtils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/CompareUtils.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/CompareUtils.java
index fe45bad..d7651e8 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/CompareUtils.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/CompareUtils.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -29,46 +29,46 @@ class CompareUtils {
   private CompareUtils() {
     // nothing
   }
-  
+
   /**
    * A comparator to compare anything that implements {@link RawComparable} using a customized comparator.
    */
   public static final class BytesComparator implements Comparator<RawComparable> {
     private RawComparator<Object> cmp;
-    
+
     public BytesComparator(RawComparator<Object> cmp) {
       this.cmp = cmp;
     }
-    
+
     @Override
     public int compare(RawComparable o1, RawComparable o2) {
       return compare(o1.buffer(), o1.offset(), o1.size(), o2.buffer(), o2.offset(), o2.size());
     }
-    
+
     public int compare(byte[] a, int off1, int len1, byte[] b, int off2, int len2) {
       return cmp.compare(a, off1, len1, b, off2, len2);
     }
   }
-  
+
   /**
    * Interface for all objects that has a single integer magnitude.
    */
   interface Scalar {
     long magnitude();
   }
-  
+
   static final class ScalarLong implements Scalar {
     private long magnitude;
-    
+
     public ScalarLong(long m) {
       magnitude = m;
     }
-    
+
     public long magnitude() {
       return magnitude;
     }
   }
-  
+
   public static final class ScalarComparator implements Comparator<Scalar>, Serializable {
     private static final long serialVersionUID = 1L;
 
@@ -82,7 +82,7 @@ class CompareUtils {
       return 0;
     }
   }
-  
+
   public static final class MemcmpRawComparator implements RawComparator<Object>, Serializable {
     private static final long serialVersionUID = 1L;
 
@@ -90,7 +90,7 @@ class CompareUtils {
     public int compare(byte[] b1, int s1, int l1, byte[] b2, int s2, int l2) {
       return WritableComparator.compareBytes(b1, s1, l1, b2, s2, l2);
     }
-    
+
     @Override
     public int compare(Object o1, Object o2) {
       throw new RuntimeException("Object comparison not supported");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Compression.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Compression.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Compression.java
index 5288bbb..9defa1c 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Compression.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Compression.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -41,24 +41,24 @@ import org.apache.hadoop.util.ReflectionUtils;
  */
 public final class Compression {
   static final Log LOG = LogFactory.getLog(Compression.class);
-  
+
   /**
    * Prevent the instantiation of class.
    */
   private Compression() {
     // nothing
   }
-  
+
   static class FinishOnFlushCompressionStream extends FilterOutputStream {
     public FinishOnFlushCompressionStream(CompressionOutputStream cout) {
       super(cout);
     }
-    
+
     @Override
     public void write(byte b[], int off, int len) throws IOException {
       out.write(b, off, len);
     }
-    
+
     @Override
     public void flush() throws IOException {
       CompressionOutputStream cout = (CompressionOutputStream) out;
@@ -67,7 +67,7 @@ public final class Compression {
       cout.resetState();
     }
   }
-  
+
   /** snappy codec **/
   public static final String COMPRESSION_SNAPPY = "snappy";
   /** compression: gzip */
@@ -76,7 +76,7 @@ public final class Compression {
   public static final String COMPRESSION_LZO = "lzo";
   /** compression: none */
   public static final String COMPRESSION_NONE = "none";
-  
+
   /**
    * Compression algorithms.
    */
@@ -85,7 +85,7 @@ public final class Compression {
       private transient boolean checked = false;
       private static final String defaultClazz = "org.apache.hadoop.io.compress.LzoCodec";
       private transient CompressionCodec codec = null;
-      
+
       @Override
       public synchronized boolean isSupported() {
         if (!checked) {
@@ -101,16 +101,16 @@ public final class Compression {
         }
         return codec != null;
       }
-      
+
       @Override
       CompressionCodec getCodec() throws IOException {
         if (!isSupported()) {
           throw new IOException("LZO codec class not specified. Did you forget to set property " + CONF_LZO_CLASS + "?");
         }
-        
+
         return codec;
       }
-      
+
       @Override
       public synchronized InputStream createDecompressionStream(InputStream downStream, Decompressor decompressor, int downStreamBufferSize) throws IOException {
         if (!isSupported()) {
@@ -127,7 +127,7 @@ public final class Compression {
         BufferedInputStream bis2 = new BufferedInputStream(cis, DATA_IBUF_SIZE);
         return bis2;
       }
-      
+
       @Override
       public synchronized OutputStream createCompressionStream(OutputStream downStream, Compressor compressor, int downStreamBufferSize) throws IOException {
         if (!isSupported()) {
@@ -145,20 +145,20 @@ public final class Compression {
         return bos2;
       }
     },
-    
+
     GZ(COMPRESSION_GZ) {
       private transient DefaultCodec codec;
-      
+
       @Override
       synchronized CompressionCodec getCodec() {
         if (codec == null) {
           codec = new DefaultCodec();
           codec.setConf(conf);
         }
-        
+
         return codec;
       }
-      
+
       @Override
       public synchronized InputStream createDecompressionStream(InputStream downStream, Decompressor decompressor, int downStreamBufferSize) throws IOException {
         // Set the internal buffer size to read from down stream.
@@ -169,7 +169,7 @@ public final class Compression {
         BufferedInputStream bis2 = new BufferedInputStream(cis, DATA_IBUF_SIZE);
         return bis2;
       }
-      
+
       @Override
       public synchronized OutputStream createCompressionStream(OutputStream downStream, Compressor compressor, int downStreamBufferSize) throws IOException {
         OutputStream bos1 = null;
@@ -183,19 +183,19 @@ public final class Compression {
         BufferedOutputStream bos2 = new BufferedOutputStream(new FinishOnFlushCompressionStream(cos), DATA_OBUF_SIZE);
         return bos2;
       }
-      
+
       @Override
       public boolean isSupported() {
         return true;
       }
     },
-    
+
     NONE(COMPRESSION_NONE) {
       @Override
       CompressionCodec getCodec() {
         return null;
       }
-      
+
       @Override
       public synchronized InputStream createDecompressionStream(InputStream downStream, Decompressor decompressor, int downStreamBufferSize) throws IOException {
         if (downStreamBufferSize > 0) {
@@ -203,38 +203,38 @@ public final class Compression {
         }
         return downStream;
       }
-      
+
       @Override
       public synchronized OutputStream createCompressionStream(OutputStream downStream, Compressor compressor, int downStreamBufferSize) throws IOException {
         if (downStreamBufferSize > 0) {
           return new BufferedOutputStream(downStream, downStreamBufferSize);
         }
-        
+
         return downStream;
       }
-      
+
       @Override
       public boolean isSupported() {
         return true;
       }
     },
-    
+
     SNAPPY(COMPRESSION_SNAPPY) {
       // Use base type to avoid compile-time dependencies.
       private transient CompressionCodec snappyCodec = null;
       private transient boolean checked = false;
       private static final String defaultClazz = "org.apache.hadoop.io.compress.SnappyCodec";
-      
+
       public CompressionCodec getCodec() throws IOException {
         if (!isSupported()) {
           throw new IOException("SNAPPY codec class not specified. Did you forget to set property " + CONF_SNAPPY_CLASS + "?");
         }
         return snappyCodec;
       }
-      
+
       @Override
       public synchronized OutputStream createCompressionStream(OutputStream downStream, Compressor compressor, int downStreamBufferSize) throws IOException {
-        
+
         if (!isSupported()) {
           throw new IOException("SNAPPY codec class not specified. Did you forget to set property " + CONF_SNAPPY_CLASS + "?");
         }
@@ -249,7 +249,7 @@ public final class Compression {
         BufferedOutputStream bos2 = new BufferedOutputStream(new FinishOnFlushCompressionStream(cos), DATA_OBUF_SIZE);
         return bos2;
       }
-      
+
       @Override
       public synchronized InputStream createDecompressionStream(InputStream downStream, Decompressor decompressor, int downStreamBufferSize) throws IOException {
         if (!isSupported()) {
@@ -262,7 +262,7 @@ public final class Compression {
         BufferedInputStream bis2 = new BufferedInputStream(cis, DATA_IBUF_SIZE);
         return bis2;
       }
-      
+
       @Override
       public synchronized boolean isSupported() {
         if (!checked) {
@@ -289,19 +289,19 @@ public final class Compression {
     private static final int DATA_OBUF_SIZE = 4 * 1024;
     public static final String CONF_LZO_CLASS = "io.compression.codec.lzo.class";
     public static final String CONF_SNAPPY_CLASS = "io.compression.codec.snappy.class";
-    
+
     Algorithm(String name) {
       this.compressName = name;
     }
-    
+
     abstract CompressionCodec getCodec() throws IOException;
-    
+
     public abstract InputStream createDecompressionStream(InputStream downStream, Decompressor decompressor, int downStreamBufferSize) throws IOException;
-    
+
     public abstract OutputStream createCompressionStream(OutputStream downStream, Compressor compressor, int downStreamBufferSize) throws IOException;
-    
+
     public abstract boolean isSupported();
-    
+
     public Compressor getCompressor() throws IOException {
       CompressionCodec codec = getCodec();
       if (codec != null) {
@@ -323,14 +323,14 @@ public final class Compression {
       }
       return null;
     }
-    
+
     public void returnCompressor(Compressor compressor) {
       if (compressor != null) {
         LOG.debug("Return a compressor: " + compressor.hashCode());
         CodecPool.returnCompressor(compressor);
       }
     }
-    
+
     public Decompressor getDecompressor() throws IOException {
       CompressionCodec codec = getCodec();
       if (codec != null) {
@@ -350,37 +350,37 @@ public final class Compression {
         }
         return decompressor;
       }
-      
+
       return null;
     }
-    
+
     public void returnDecompressor(Decompressor decompressor) {
       if (decompressor != null) {
         LOG.debug("Returned a decompressor: " + decompressor.hashCode());
         CodecPool.returnDecompressor(decompressor);
       }
     }
-    
+
     public String getName() {
       return compressName;
     }
   }
-  
+
   static Algorithm getCompressionAlgorithmByName(String compressName) {
     Algorithm[] algos = Algorithm.class.getEnumConstants();
-    
+
     for (Algorithm a : algos) {
       if (a.getName().equals(compressName)) {
         return a;
       }
     }
-    
+
     throw new IllegalArgumentException("Unsupported compression algorithm name: " + compressName);
   }
-  
+
   public static String[] getSupportedAlgorithms() {
     Algorithm[] algos = Algorithm.class.getEnumConstants();
-    
+
     ArrayList<String> ret = new ArrayList<String>();
     for (Algorithm a : algos) {
       if (a.isSupported()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockAlreadyExists.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockAlreadyExists.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockAlreadyExists.java
index 612c738..617da34 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockAlreadyExists.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockAlreadyExists.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -26,7 +26,7 @@ import java.io.IOException;
 public class MetaBlockAlreadyExists extends IOException {
   /**
    * Constructor
-   * 
+   *
    * @param s
    *          message.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockDoesNotExist.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockDoesNotExist.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockDoesNotExist.java
index 566075e..bf9319d 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockDoesNotExist.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/MetaBlockDoesNotExist.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -26,7 +26,7 @@ import java.io.IOException;
 public class MetaBlockDoesNotExist extends IOException {
   /**
    * Constructor
-   * 
+   *
    * @param s
    *          message.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/PrintInfo.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/PrintInfo.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/PrintInfo.java
index a67a242..9ac1e96 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/PrintInfo.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/PrintInfo.java
@@ -37,9 +37,9 @@ public class PrintInfo {
     BCFile.Reader bcfr = null;
     try {
       bcfr = new BCFile.Reader(fsin, fs.getFileStatus(path).getLen(), conf, accumuloConfiguration);
-      
+
       Set<Entry<String,MetaIndexEntry>> es = bcfr.metaIndex.index.entrySet();
-      
+
       for (Entry<String,MetaIndexEntry> entry : es) {
         PrintStream out = System.out;
         out.println("Meta block     : " + entry.getKey());
@@ -54,7 +54,7 @@ public class PrintInfo {
       }
     }
   }
-  
+
   public static void main(String[] args) throws Exception {
     Configuration conf = new Configuration();
     AccumuloConfiguration siteConf = SiteConfiguration.getInstance(DefaultConfiguration.getInstance());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/RawComparable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/RawComparable.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/RawComparable.java
index 48c0bfe..a624223 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/RawComparable.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/RawComparable.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -25,28 +25,28 @@ import org.apache.hadoop.io.RawComparator;
 /**
  * Interface for objects that can be compared through {@link RawComparator}. This is useful in places where we need a single object reference to specify a range
  * of bytes in a byte array, such as {@link Comparable} or {@link Collections#binarySearch(java.util.List, Object, Comparator)}
- * 
+ *
  * The actual comparison among RawComparable's requires an external RawComparator and it is applications' responsibility to ensure two RawComparable are
  * supposed to be semantically comparable with the same RawComparator.
  */
 public interface RawComparable {
   /**
    * Get the underlying byte array.
-   * 
+   *
    * @return The underlying byte array.
    */
   byte[] buffer();
-  
+
   /**
    * Get the offset of the first byte in the byte array.
-   * 
+   *
    * @return The offset of the first byte in the byte array.
    */
   int offset();
-  
+
   /**
    * Get the size of the byte range in the byte array.
-   * 
+   *
    * @return The size of the byte range in the byte array.
    */
   int size();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/SimpleBufferedOutputStream.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/SimpleBufferedOutputStream.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/SimpleBufferedOutputStream.java
index 39a853c..3662208 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/SimpleBufferedOutputStream.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/SimpleBufferedOutputStream.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -27,20 +27,20 @@ import java.io.OutputStream;
 class SimpleBufferedOutputStream extends FilterOutputStream {
   protected byte buf[]; // the borrowed buffer
   protected int count = 0; // bytes used in buffer.
-  
+
   // Constructor
   public SimpleBufferedOutputStream(OutputStream out, byte[] buf) {
     super(out);
     this.buf = buf;
   }
-  
+
   private void flushBuffer() throws IOException {
     if (count > 0) {
       out.write(buf, 0, count);
       count = 0;
     }
   }
-  
+
   @Override
   public void write(int b) throws IOException {
     if (count >= buf.length) {
@@ -48,7 +48,7 @@ class SimpleBufferedOutputStream extends FilterOutputStream {
     }
     buf[count++] = (byte) b;
   }
-  
+
   @Override
   public void write(byte b[], int off, int len) throws IOException {
     if (len >= buf.length) {
@@ -62,13 +62,13 @@ class SimpleBufferedOutputStream extends FilterOutputStream {
     System.arraycopy(b, off, buf, count, len);
     count += len;
   }
-  
+
   @Override
   public synchronized void flush() throws IOException {
     flushBuffer();
     out.flush();
   }
-  
+
   // Get the size of internal buffer being used.
   public int size() {
     return count;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
index 9131d30..fb0277a 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
@@ -5,9 +5,9 @@
  * licenses this file to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance with the License.
  * You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
  * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
@@ -29,17 +29,17 @@ import org.apache.hadoop.io.Text;
  * Supporting Utility classes used by TFile, and shared by users of TFile.
  */
 public final class Utils {
-  
+
   /**
    * Prevent the instantiation of Utils.
    */
   private Utils() {
     // nothing
   }
-  
+
   /**
    * Encoding an integer into a variable-length encoding format. Synonymous to <code>Utils#writeVLong(out, n)</code>.
-   * 
+   *
    * @param out
    *          output stream
    * @param n
@@ -49,7 +49,7 @@ public final class Utils {
   public static void writeVInt(DataOutput out, int n) throws IOException {
     writeVLong(out, n);
   }
-  
+
   /**
    * Encoding a Long integer into a variable-length encoding format.
    * <ul>
@@ -67,7 +67,7 @@ public final class Utils {
    * <li>if n in [-2^63, 2^63): encode in nine bytes: byte[0]=-121; byte[1] = (n>>54)&0xff; byte[2] = (n>>48)&0xff; byte[3] = (n>>40)&0xff;
    * byte[4]=(n>>32)&0xff; byte[5]=(n>>24)&0xff; byte[6]=(n>>16)&0xff; byte[7]=(n>>8)&0xff; byte[8]=n&0xff;
    * </ul>
-   * 
+   *
    * @param out
    *          output stream
    * @param n
@@ -79,7 +79,7 @@ public final class Utils {
       out.writeByte((int) n);
       return;
     }
-    
+
     long un = (n < 0) ? ~n : n;
     // how many bytes do we need to represent the number with sign bit?
     int len = (Long.SIZE - Long.numberOfLeadingZeros(un)) / 8 + 1;
@@ -138,14 +138,14 @@ public final class Utils {
         throw new RuntimeException("Internel error");
     }
   }
-  
+
   /**
    * Decoding the variable-length integer. Synonymous to <code>(int)Utils#readVLong(in)</code>.
-   * 
+   *
    * @param in
    *          input stream
    * @return the decoded integer
-   * 
+   *
    * @see Utils#readVLong(DataInput)
    */
   public static int readVInt(DataInput in) throws IOException {
@@ -155,7 +155,7 @@ public final class Utils {
     }
     return (int) ret;
   }
-  
+
   /**
    * Decoding the variable-length integer. Suppose the value of the first byte is FB, and the following bytes are NB[*].
    * <ul>
@@ -164,18 +164,18 @@ public final class Utils {
    * <li>if (FB in [-104, -73]), return (FB+88)<<16 + (NB[0]&0xff)<<8 + NB[1]&0xff;
    * <li>if (FB in [-120, -105]), return (FB+112)<<24 + (NB[0]&0xff)<<16 + (NB[1]&0xff)<<8 + NB[2]&0xff;
    * <li>if (FB in [-128, -121]), return interpret NB[FB+129] as a signed big-endian integer.
-   * 
+   *
    * @param in
    *          input stream
    * @return the decoded long integer.
    */
-  
+
   public static long readVLong(DataInput in) throws IOException {
     int firstByte = in.readByte();
     if (firstByte >= -32) {
       return firstByte;
     }
-    
+
     switch ((firstByte + 128) / 8) {
       case 11:
       case 10:
@@ -211,7 +211,7 @@ public final class Utils {
         throw new RuntimeException("Internal error");
     }
   }
-  
+
   /**
    * Write a String as a VInt n, followed by n Bytes as in Text format.
    */
@@ -226,10 +226,10 @@ public final class Utils {
       writeVInt(out, -1);
     }
   }
-  
+
   /**
    * Read a String as a VInt n, followed by n Bytes in Text format.
-   * 
+   *
    * @param in
    *          The input stream.
    * @return The string
@@ -242,20 +242,20 @@ public final class Utils {
     in.readFully(buffer);
     return Text.decode(buffer);
   }
-  
+
   /**
    * A generic Version class. We suggest applications built on top of TFile use this class to maintain version information in their meta blocks.
-   * 
+   *
    * A version number consists of a major version and a minor version. The suggested usage of major and minor version number is to increment major version
    * number when the new storage format is not backward compatible, and increment the minor version otherwise.
    */
   public static final class Version implements Comparable<Version> {
     private final short major;
     private final short minor;
-    
+
     /**
      * Construct the Version object by reading from the input stream.
-     * 
+     *
      * @param in
      *          input stream
      */
@@ -263,10 +263,10 @@ public final class Utils {
       major = in.readShort();
       minor = in.readShort();
     }
-    
+
     /**
      * Constructor.
-     * 
+     *
      * @param major
      *          major version.
      * @param minor
@@ -276,10 +276,10 @@ public final class Utils {
       this.major = major;
       this.minor = minor;
     }
-    
+
     /**
      * Write the object to a DataOutput. The serialized format of the Version is major version followed by minor version, both as big-endian short integers.
-     * 
+     *
      * @param out
      *          The DataOutput object.
      */
@@ -287,34 +287,34 @@ public final class Utils {
       out.writeShort(major);
       out.writeShort(minor);
     }
-    
+
     /**
      * Get the major version.
-     * 
+     *
      * @return Major version.
      */
     public int getMajor() {
       return major;
     }
-    
+
     /**
      * Get the minor version.
-     * 
+     *
      * @return The minor version.
      */
     public int getMinor() {
       return minor;
     }
-    
+
     /**
      * Get the size of the serialized Version object.
-     * 
+     *
      * @return serialized size of the version object.
      */
     public static int size() {
       return (Short.SIZE + Short.SIZE) / Byte.SIZE;
     }
-    
+
     /**
      * Return a string representation of the version.
      */
@@ -322,10 +322,10 @@ public final class Utils {
     public String toString() {
       return new StringBuilder("v").append(major).append(".").append(minor).toString();
     }
-    
+
     /**
      * Test compatibility.
-     * 
+     *
      * @param other
      *          The Version object to test compatibility with.
      * @return true if both versions have the same major version number; false otherwise.
@@ -333,7 +333,7 @@ public final class Utils {
     public boolean compatibleWith(Version other) {
       return major == other.major;
     }
-    
+
     /**
      * Compare this version with another version.
      */
@@ -344,7 +344,7 @@ public final class Utils {
       }
       return minor - that.minor;
     }
-    
+
     @Override
     public boolean equals(Object other) {
       if (this == other)
@@ -353,16 +353,16 @@ public final class Utils {
         return false;
       return compareTo((Version) other) == 0;
     }
-    
+
     @Override
     public int hashCode() {
       return (major << 16 + minor);
     }
   }
-  
+
   /**
    * Lower bound binary search. Find the index to the first element in the list that compares greater than or equal to key.
-   * 
+   *
    * @param <T>
    *          Type of the input key.
    * @param list
@@ -376,7 +376,7 @@ public final class Utils {
   public static <T> int lowerBound(List<? extends T> list, T key, Comparator<? super T> cmp) {
     int low = 0;
     int high = list.size();
-    
+
     while (low < high) {
       int mid = (low + high) >>> 1;
       T midVal = list.get(mid);
@@ -388,10 +388,10 @@ public final class Utils {
     }
     return low;
   }
-  
+
   /**
    * Upper bound binary search. Find the index to the first element in the list that compares greater than the input key.
-   * 
+   *
    * @param <T>
    *          Type of the input key.
    * @param list
@@ -405,7 +405,7 @@ public final class Utils {
   public static <T> int upperBound(List<? extends T> list, T key, Comparator<? super T> cmp) {
     int low = 0;
     int high = list.size();
-    
+
     while (low < high) {
       int mid = (low + high) >>> 1;
       T midVal = list.get(mid);
@@ -417,10 +417,10 @@ public final class Utils {
     }
     return low;
   }
-  
+
   /**
    * Lower bound binary search. Find the index to the first element in the list that compares greater than or equal to key.
-   * 
+   *
    * @param <T>
    *          Type of the input key.
    * @param list
@@ -432,7 +432,7 @@ public final class Utils {
   public static <T> int lowerBound(List<? extends Comparable<? super T>> list, T key) {
     int low = 0;
     int high = list.size();
-    
+
     while (low < high) {
       int mid = (low + high) >>> 1;
       Comparable<? super T> midVal = list.get(mid);
@@ -444,10 +444,10 @@ public final class Utils {
     }
     return low;
   }
-  
+
   /**
    * Upper bound binary search. Find the index to the first element in the list that compares greater than the input key.
-   * 
+   *
    * @param <T>
    *          Type of the input key.
    * @param list
@@ -459,7 +459,7 @@ public final class Utils {
   public static <T> int upperBound(List<? extends Comparable<? super T>> list, T key) {
     int low = 0;
     int high = list.size();
-    
+
     while (low < high) {
       int mid = (low + high) >>> 1;
       Comparable<? super T> midVal = list.get(mid);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/AggregatingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/AggregatingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/AggregatingIterator.java
index 9b89b47..fd5525b 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/AggregatingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/AggregatingIterator.java
@@ -35,59 +35,59 @@ import org.apache.log4j.Logger;
 
 /**
  * This iterator wraps another iterator. It automatically aggregates.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.Combiner}
  */
 
 @Deprecated
 public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, OptionDescriber {
-  
+
   private SortedKeyValueIterator<Key,Value> iterator;
   private ColumnToClassMapping<Aggregator> aggregators;
-  
+
   private Key workKey = new Key();
-  
+
   private Key aggrKey;
   private Value aggrValue;
   // private boolean propogateDeletes;
   private static final Logger log = Logger.getLogger(AggregatingIterator.class);
-  
+
   public AggregatingIterator deepCopy(IteratorEnvironment env) {
     return new AggregatingIterator(this, env);
   }
-  
+
   private AggregatingIterator(AggregatingIterator other, IteratorEnvironment env) {
     iterator = other.iterator.deepCopy(env);
     aggregators = other.aggregators;
   }
-  
+
   public AggregatingIterator() {}
-  
+
   private void aggregateRowColumn(Aggregator aggr) throws IOException {
     // this function assumes that first value is not delete
-    
+
     if (iterator.getTopKey().isDeleted())
       return;
-    
+
     workKey.set(iterator.getTopKey());
-    
+
     Key keyToAggregate = workKey;
-    
+
     aggr.reset();
-    
+
     aggr.collect(iterator.getTopValue());
     iterator.next();
-    
+
     while (iterator.hasTop() && !iterator.getTopKey().isDeleted() && iterator.getTopKey().equals(keyToAggregate, PartialKey.ROW_COLFAM_COLQUAL_COLVIS)) {
       aggr.collect(iterator.getTopValue());
       iterator.next();
     }
-    
+
     aggrKey = workKey;
     aggrValue = aggr.aggregate();
-    
+
   }
-  
+
   private void findTop() throws IOException {
     // check if aggregation is needed
     if (iterator.hasTop()) {
@@ -97,12 +97,12 @@ public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, O
       }
     }
   }
-  
+
   public AggregatingIterator(SortedKeyValueIterator<Key,Value> iterator, ColumnToClassMapping<Aggregator> aggregators) throws IOException {
     this.iterator = iterator;
     this.aggregators = aggregators;
   }
-  
+
   @Override
   public Key getTopKey() {
     if (aggrKey != null) {
@@ -110,7 +110,7 @@ public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, O
     }
     return iterator.getTopKey();
   }
-  
+
   @Override
   public Value getTopValue() {
     if (aggrKey != null) {
@@ -118,12 +118,12 @@ public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, O
     }
     return iterator.getTopValue();
   }
-  
+
   @Override
   public boolean hasTop() {
     return aggrKey != null || iterator.hasTop();
   }
-  
+
   @Override
   public void next() throws IOException {
     if (aggrKey != null) {
@@ -132,20 +132,20 @@ public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, O
     } else {
       iterator.next();
     }
-    
+
     findTop();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     // do not want to seek to the middle of a value that should be
     // aggregated...
-    
+
     Range seekRange = IteratorUtil.maximizeStartKeyTimeStamp(range);
-    
+
     iterator.seek(seekRange, columnFamilies, inclusive);
     findTop();
-    
+
     if (range.getStartKey() != null) {
       while (hasTop() && getTopKey().equals(range.getStartKey(), PartialKey.ROW_COLFAM_COLQUAL_COLVIS)
           && getTopKey().getTimestamp() > range.getStartKey().getTimestamp()) {
@@ -154,19 +154,19 @@ public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, O
         // log.debug("skipping "+getTopKey());
         next();
       }
-      
+
       while (hasTop() && range.beforeStartKey(getTopKey())) {
         next();
       }
     }
-    
+
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
-    
+
     this.iterator = source;
-    
+
     try {
       String context = null;
       if (null != env)
@@ -183,13 +183,13 @@ public class AggregatingIterator implements SortedKeyValueIterator<Key,Value>, O
       throw new IllegalArgumentException(e);
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     return new IteratorOptions("agg", "Aggregators apply aggregating functions to values with identical keys", null,
         Collections.singletonList("<columnName> <aggregatorClass>"));
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     for (Entry<String,String> entry : options.entrySet()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/ColumnFamilyCounter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/ColumnFamilyCounter.java b/core/src/main/java/org/apache/accumulo/core/iterators/ColumnFamilyCounter.java
index b3607ef..934658e 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/ColumnFamilyCounter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/ColumnFamilyCounter.java
@@ -28,65 +28,65 @@ import org.apache.accumulo.core.data.Range;
 import org.apache.accumulo.core.data.Value;
 
 public class ColumnFamilyCounter implements SortedKeyValueIterator<Key,Value> {
-  
+
   private SortedKeyValueIterator<Key,Value> source;
   private Key key;
   private Value value;
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     this.source = source;
   }
-  
+
   @Override
   public boolean hasTop() {
     return key != null;
   }
-  
+
   @Override
   public void next() throws IOException {
     if (source.hasTop()) {
       ByteSequence currentRow = source.getTopKey().getRowData();
       ByteSequence currentColf = source.getTopKey().getColumnFamilyData();
       long ts = source.getTopKey().getTimestamp();
-      
+
       source.next();
-      
+
       int count = 1;
-      
+
       while (source.hasTop() && source.getTopKey().getRowData().equals(currentRow) && source.getTopKey().getColumnFamilyData().equals(currentColf)) {
         count++;
         source.next();
       }
-      
+
       this.key = new Key(currentRow.toArray(), currentColf.toArray(), new byte[0], new byte[0], ts);
       this.value = new Value(Integer.toString(count).getBytes(UTF_8));
-      
+
     } else {
       this.key = null;
       this.value = null;
     }
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     source.seek(range, columnFamilies, inclusive);
     next();
   }
-  
+
   @Override
   public Key getTopKey() {
     return key;
   }
-  
+
   @Override
   public Value getTopValue() {
     return value;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
index 683e3f7..6d7cc7e 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
@@ -23,8 +23,6 @@ import java.util.List;
 import java.util.Map;
 import java.util.NoSuchElementException;
 
-import com.google.common.base.Splitter;
-import com.google.common.collect.Lists;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.IteratorSetting.Column;
 import org.apache.accumulo.core.client.ScannerBase;
@@ -36,18 +34,21 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.conf.ColumnSet;
 import org.apache.hadoop.io.Text;
 
+import com.google.common.base.Splitter;
+import com.google.common.collect.Lists;
+
 /**
  * A SortedKeyValueIterator that combines the Values for different versions (timestamp) of a Key within a row into a single Value. Combiner will replace one or
- * more versions of a Key and their Values with the most recent Key and a Value which is the result of the reduce method. An {@link Column}
- * which only specifies a column family will combine all Keys in that column family individually. Similarly, a {@link Column} which specifies a
- * column family and column qualifier will combine all Keys in column family and qualifier individually. Combination is only ever performed on multiple versions
- * and not across column qualifiers or column visibilities.
- * 
+ * more versions of a Key and their Values with the most recent Key and a Value which is the result of the reduce method. An {@link Column} which only specifies
+ * a column family will combine all Keys in that column family individually. Similarly, a {@link Column} which specifies a column family and column qualifier
+ * will combine all Keys in column family and qualifier individually. Combination is only ever performed on multiple versions and not across column qualifiers
+ * or column visibilities.
+ *
  * Implementations must provide a reduce method: {@code public Value reduce(Key key, Iterator<Value> iter)}.
- * 
+ *
  * This reduce method will be passed the most recent Key and an iterator over the Values for all non-deleted versions of that Key. A combiner will not combine
  * keys that differ by more than the timestamp.
- * 
+ *
  * This class and its implementations do not automatically filter out unwanted columns from those being combined, thus it is generally recommended to use a
  * {@link Combiner} implementation with the {@link ScannerBase#fetchColumnFamily(Text)} or {@link ScannerBase#fetchColumn(Text, Text)} methods.
  */
@@ -65,7 +66,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
 
     /**
      * Constructs an iterator over Values whose Keys are versions of the current topKey of the source SortedKeyValueIterator.
-     * 
+     *
      * @param source
      *          The SortedKeyValueIterator<Key,Value> from which to read data.
      */
@@ -100,7 +101,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
 
     /**
      * This method is unsupported in this iterator.
-     * 
+     *
      * @throws UnsupportedOperationException
      *           when called
      */
@@ -192,13 +193,13 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
 
   /**
    * Reduces a list of Values into a single Value.
-   * 
+   *
    * @param key
    *          The most recent version of the Key being reduced.
-   * 
+   *
    * @param iter
    *          An iterator over the Values for different versions of the key.
-   * 
+   *
    * @return The combined Value.
    */
   public abstract Value reduce(Key key, Iterator<Value> iter);
@@ -279,7 +280,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
    * A convenience method to set which columns a combiner should be applied to. For each column specified, all versions of a Key which match that @{link
    * IteratorSetting.Column} will be combined individually in each row. This method is likely to be used in conjunction with
    * {@link ScannerBase#fetchColumnFamily(Text)} or {@link ScannerBase#fetchColumn(Text,Text)}.
-   * 
+   *
    * @param is
    *          iterator settings object to configure
    * @param columns
@@ -301,7 +302,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
 
   /**
    * A convenience method to set the "all columns" option on a Combiner. This will combine all columns individually within each row.
-   * 
+   *
    * @param is
    *          iterator settings object to configure
    * @param combineAllColumns

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/DebugIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/DebugIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/DebugIterator.java
index 83265cb..92f49f9 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/DebugIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/DebugIterator.java
@@ -27,48 +27,48 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.log4j.Logger;
 
 public class DebugIterator extends WrappingIterator implements OptionDescriber {
-  
+
   private String prefix;
-  
+
   private static final Logger log = Logger.getLogger(DebugIterator.class);
-  
+
   public DebugIterator() {}
-  
+
   public DebugIterator deepCopy(IteratorEnvironment env) {
     return new DebugIterator(this, env);
   }
-  
+
   private DebugIterator(DebugIterator other, IteratorEnvironment env) {
     setSource(other.getSource().deepCopy(env));
     prefix = other.prefix;
   }
-  
+
   public DebugIterator(String prefix, SortedKeyValueIterator<Key,Value> source) {
     this.prefix = prefix;
     this.setSource(source);
   }
-  
+
   @Override
   public Key getTopKey() {
     Key wc = super.getTopKey();
     log.debug(prefix + " getTopKey() --> " + wc);
     return wc;
   }
-  
+
   @Override
   public Value getTopValue() {
     Value w = super.getTopValue();
     log.debug(prefix + " getTopValue() --> " + w);
     return w;
   }
-  
+
   @Override
   public boolean hasTop() {
     boolean b = super.hasTop();
     log.debug(prefix + " hasTop() --> " + b);
     return b;
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     log.debug(prefix + " seek(" + range + ", " + columnFamilies + ", " + inclusive + ")");
@@ -80,7 +80,7 @@ public class DebugIterator extends WrappingIterator implements OptionDescriber {
     log.debug(prefix + " next()");
     super.next();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     log.debug("init(" + source + ", " + options + ", " + env + ")");
@@ -94,7 +94,8 @@ public class DebugIterator extends WrappingIterator implements OptionDescriber {
 
   @Override
   public IteratorOptions describeOptions() {
-    return new IteratorOptions("debug", DebugIterator.class.getSimpleName() + " prints debug information on each SortedKeyValueIterator method invocation", null, null);
+    return new IteratorOptions("debug", DebugIterator.class.getSimpleName() + " prints debug information on each SortedKeyValueIterator method invocation",
+        null, null);
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/DevNull.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/DevNull.java b/core/src/main/java/org/apache/accumulo/core/iterators/DevNull.java
index e7de5e1..d9a7b8d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/DevNull.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/DevNull.java
@@ -28,54 +28,54 @@ import org.apache.accumulo.core.data.Value;
 /**
  * An iterator that is useful testing... for example if you want to test ingest performance w/o writing data to disk, insert this iterator for scan as follows
  * using the accumulo shell.
- * 
+ *
  * config -t ci -s table.iterator.minc.devnull=21,org.apache.accumulo.core.iterators.DevNull
- * 
+ *
  * Could also make scans never return anything very quickly by adding it to the scan stack
- * 
+ *
  * config -t ci -s table.iterator.scan.devnull=21,org.apache.accumulo.core.iterators.DevNull
- * 
+ *
  * And to make major compactions never write anything
- * 
+ *
  * config -t ci -s table.iterator.majc.devnull=21,org.apache.accumulo.core.iterators.DevNull
- * 
+ *
  */
 
 public class DevNull implements SortedKeyValueIterator<Key,Value> {
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public Key getTopKey() {
     return null;
   }
-  
+
   @Override
   public Value getTopValue() {
     return null;
   }
-  
+
   @Override
   public boolean hasTop() {
     return false;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
-    
+
   }
-  
+
   @Override
   public void next() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/FamilyIntersectingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/FamilyIntersectingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/FamilyIntersectingIterator.java
index e0623e2..04102b8 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/FamilyIntersectingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/FamilyIntersectingIterator.java
@@ -20,10 +20,10 @@ import org.apache.accumulo.core.iterators.user.IndexedDocIterator;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.IndexedDocIterator}
  */
 @Deprecated
 public class FamilyIntersectingIterator extends IndexedDocIterator {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/Filter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/Filter.java b/core/src/main/java/org/apache/accumulo/core/iterators/Filter.java
index 25653eb..8b135c7 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/Filter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/Filter.java
@@ -29,9 +29,9 @@ import org.apache.accumulo.core.data.Value;
 
 /**
  * A SortedKeyValueIterator that filters entries from its source iterator.
- * 
+ *
  * Subclasses must implement an accept method: public boolean accept(Key k, Value v);
- * 
+ *
  * Key/Value pairs for which the accept method returns true are said to match the filter. By default, this class iterates over entries that match its filter.
  * This iterator takes an optional "negate" boolean parameter that defaults to false. If negate is set to true, this class instead omits entries that match its
  * filter, thus iterating over entries that do not match its filter.
@@ -49,22 +49,22 @@ public abstract class Filter extends WrappingIterator implements OptionDescriber
     newInstance.negate = negate;
     return newInstance;
   }
-  
+
   protected static final String NEGATE = "negate";
   boolean negate = false;
-  
+
   @Override
   public void next() throws IOException {
     super.next();
     findTop();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     super.seek(range, columnFamilies, inclusive);
     findTop();
   }
-  
+
   /**
    * Iterates over the source until an acceptable key/value pair is found.
    */
@@ -77,12 +77,12 @@ public abstract class Filter extends WrappingIterator implements OptionDescriber
       }
     }
   }
-  
+
   /**
    * @return <tt>true</tt> if the key/value pair is accepted by the filter.
    */
   public abstract boolean accept(Key k, Value v);
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
@@ -91,13 +91,13 @@ public abstract class Filter extends WrappingIterator implements OptionDescriber
       negate = Boolean.parseBoolean(options.get(NEGATE));
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     return new IteratorOptions("filter", "Filter accepts or rejects each Key/Value pair", Collections.singletonMap("negate",
         "default false keeps k/v that pass accept method, true rejects k/v that pass accept method"), null);
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (options.get(NEGATE) != null) {
@@ -109,10 +109,10 @@ public abstract class Filter extends WrappingIterator implements OptionDescriber
     }
     return true;
   }
-  
+
   /**
    * A convenience method for setting the negation option on a filter.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param negate


[56/66] [abbrv] accumulo git commit: ACCUMULO-3451 comply with checkstyle rules

Posted by ct...@apache.org.
ACCUMULO-3451 comply with checkstyle rules

  Adds javadoc and style changes to pass checkstyle enforcement rules.


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/901d60ef
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/901d60ef
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/901d60ef

Branch: refs/heads/1.6
Commit: 901d60ef1cf72c2d55c90746fce94e108a992d3b
Parents: d2c116f
Author: Christopher Tubbs <ct...@apache.org>
Authored: Wed Dec 24 15:22:30 2014 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:25:40 2015 -0500

----------------------------------------------------------------------
 .../core/client/admin/InstanceOperations.java        |  6 +++---
 .../java/org/apache/accumulo/core/data/Value.java    |  7 +++----
 .../apache/accumulo/core/file/rfile/CreateEmpty.java |  4 ++--
 .../accumulo/core/file/rfile/bcfile/TFile.java       |  4 ++--
 .../accumulo/core/file/rfile/bcfile/Utils.java       |  1 +
 .../org/apache/accumulo/core/iterators/Combiner.java |  2 +-
 .../apache/accumulo/core/iterators/LongCombiner.java |  4 ++--
 .../accumulo/core/iterators/OptionDescriber.java     |  4 ++--
 .../accumulo/core/iterators/TypedValueCombiner.java  | 10 +++++-----
 .../core/iterators/user/SummingArrayCombiner.java    |  5 +++--
 .../accumulo/core/security/ColumnVisibility.java     | 10 +++++-----
 .../crypto/DefaultSecretKeyEncryptionStrategy.java   |  3 ++-
 .../accumulo/core/util/format/BinaryFormatter.java   |  7 ++++---
 .../accumulo/core/util/shell/commands/DUCommand.java |  4 +++-
 .../core/util/shell/commands/EGrepCommand.java       |  3 ++-
 .../core/util/shell/commands/HistoryCommand.java     |  4 +---
 .../util/shell/commands/QuotedStringTokenizer.java   | 15 ++++++---------
 .../examples/simple/mapreduce/TableToFile.java       |  4 ++--
 .../java/org/apache/accumulo/server/Accumulo.java    |  4 ++--
 .../accumulo/server/tabletserver/Compactor.java      |  6 ++++--
 .../org/apache/accumulo/server/util/Initialize.java  |  3 ++-
 .../main/java/org/apache/accumulo/start/Main.java    |  4 ++--
 .../start/classloader/AccumuloClassLoader.java       |  4 ++--
 .../test/randomwalk/security/CreateTable.java        |  5 ++---
 .../test/randomwalk/security/CreateUser.java         |  5 ++---
 25 files changed, 65 insertions(+), 63 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
index 049dd85..8cb656c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
@@ -82,7 +82,7 @@ public interface InstanceOperations {
    * List the active scans on tablet server.
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @return A list of active scans on tablet server.
    */
 
@@ -92,7 +92,7 @@ public interface InstanceOperations {
    * List the active compaction running on a tablet server
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @return the list of active compactions
    * @since 1.5.0
    */
@@ -103,7 +103,7 @@ public interface InstanceOperations {
    * Throws an exception if a tablet server can not be contacted.
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @since 1.5.0
    */
   public void ping(String tserver) throws AccumuloException;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/data/Value.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Value.java b/core/src/main/java/org/apache/accumulo/core/data/Value.java
index ba89d6c..b937203 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Value.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Value.java
@@ -139,12 +139,13 @@ public class Value implements WritableComparable<Object> {
     return this.value.length;
   }
 
+  @Override
   public void readFields(final DataInput in) throws IOException {
     this.value = new byte[in.readInt()];
     in.readFully(this.value, 0, this.value.length);
   }
 
-  /** {@inheritDoc} */
+  @Override
   public void write(final DataOutput out) throws IOException {
     out.writeInt(this.value.length);
     out.write(this.value, 0, this.value.length);
@@ -152,7 +153,6 @@ public class Value implements WritableComparable<Object> {
 
   // Below methods copied from BytesWritable
 
-  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     return WritableComparator.hashBytes(value, this.value.length);
@@ -165,6 +165,7 @@ public class Value implements WritableComparable<Object> {
    *          The other bytes writable
    * @return Positive if left is bigger than right, 0 if they are equal, and negative if left is smaller than right.
    */
+  @Override
   public int compareTo(Object right_obj) {
     return compareTo(((Value) right_obj).get());
   }
@@ -179,7 +180,6 @@ public class Value implements WritableComparable<Object> {
     return (diff != 0) ? diff : WritableComparator.compareBytes(this.value, 0, this.value.length, that, 0, that.length);
   }
 
-  /** {@inheritDoc} */
   @Override
   public boolean equals(Object right_obj) {
     if (right_obj instanceof byte[]) {
@@ -207,7 +207,6 @@ public class Value implements WritableComparable<Object> {
       super(Value.class);
     }
 
-    /** {@inheritDoc} */
     @Override
     public int compare(byte[] b1, int s1, int l1, byte[] b2, int s2, int l2) {
       return comparator.compare(b1, s1, l1, b2, s2, l2);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
index bd9fa43..8d54088 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
@@ -59,8 +59,8 @@ public class CreateEmpty {
   static class Opts extends Help {
     @Parameter(names = {"-c", "--codec"}, description = "the compression codec to use.", validateWith = IsSupportedCompressionAlgorithm.class)
     String codec = TFile.COMPRESSION_NONE;
-    @Parameter(
-        description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
+    @Parameter(description = " <path> { <path> ... } Each path given is a URL. "
+        + "Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
         required = true, validateWith = NamedLikeRFile.class)
     List<String> files = new ArrayList<String>();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
index e21598a..400695b 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
@@ -1065,9 +1065,9 @@ public class TFile {
        * @param reader
        *          The TFile reader object.
        * @param beginKey
-       *          Begin key of the scan. If null, scan from the first <K,V> entry of the TFile.
+       *          Begin key of the scan. If null, scan from the first {@code <K,V>} entry of the TFile.
        * @param endKey
-       *          End key of the scan. If null, scan up to the last <K, V> entry of the TFile.
+       *          End key of the scan. If null, scan up to the last {@code <K, V>} entry of the TFile.
        */
       protected Scanner(Reader reader, RawComparable beginKey, RawComparable endKey) throws IOException {
         this(reader, (beginKey == null) ? reader.begin() : reader.getBlockContainsKey(beginKey, false), reader.end());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
index fb0277a..84b861b 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
@@ -164,6 +164,7 @@ public final class Utils {
    * <li>if (FB in [-104, -73]), return (FB+88)<<16 + (NB[0]&0xff)<<8 + NB[1]&0xff;
    * <li>if (FB in [-120, -105]), return (FB+112)<<24 + (NB[0]&0xff)<<16 + (NB[1]&0xff)<<8 + NB[2]&0xff;
    * <li>if (FB in [-128, -121]), return interpret NB[FB+129] as a signed big-endian integer.
+   * </ul>
    *
    * @param in
    *          input stream

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
index 41e4e1e..f75076d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
@@ -70,7 +70,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
      * Constructs an iterator over Values whose Keys are versions of the current topKey of the source SortedKeyValueIterator.
      *
      * @param source
-     *          The SortedKeyValueIterator<Key,Value> from which to read data.
+     *          The {@code SortedKeyValueIterator<Key,Value>} from which to read data.
      */
     public ValueIterator(SortedKeyValueIterator<Key,Value> source) {
       this.source = source;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
index 0bffbf7..cfdfd6e 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.io.WritableUtils;
 /**
  * A TypedValueCombiner that translates each Value to a Long before reducing, then encodes the reduced Long back to a Value.
  *
- * Subclasses must implement a typedReduce method: public Long typedReduce(Key key, Iterator<Long> iter);
+ * Subclasses must implement a typedReduce method: {@code public Long typedReduce(Key key, Iterator<Long> iter);}
  *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Longs) for all non-deleted versions of that Key.
  *
@@ -226,7 +226,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
-   *          Class<? extends Encoder<Long>> specifying the encoding type.
+   *          {@code Class<? extends Encoder<Long>>} specifying the encoding type.
    */
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<Long>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
index 6158bfc..6593ecd 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
@@ -50,8 +50,8 @@ public interface OptionDescriber {
      * @param unnamedOptionDescriptions
      *          is a list of descriptions of additional options that don't have fixed names (null if unused). The descriptions are intended to describe a
      *          category, and the user will provide parameter names and values in that category; e.g., the FilteringIterator needs a list of Filters intended to
-     *          be named by their priority numbers, so its unnamedOptionDescriptions =
-     *          Collections.singletonList("<filterPriorityNumber> <ageoff|regex|filterClass>")
+     *          be named by their priority numbers, so its<br>
+     *          {@code unnamedOptionDescriptions = Collections.singletonList("<filterPriorityNumber> <ageoff|regex|filterClass>")}
      */
     public IteratorOptions(String name, String description, Map<String,String> namedOptions, List<String> unnamedOptionDescriptions) {
       this.name = name;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
index d1ae9f5..0f26fab 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
@@ -29,7 +29,7 @@ import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
 /**
  * A Combiner that decodes each Value to type V before reducing, then encodes the result of typedReduce back to Value.
  *
- * Subclasses must implement a typedReduce method: public V typedReduce(Key key, Iterator<V> iter);
+ * Subclasses must implement a typedReduce method: {@code public V typedReduce(Key key, Iterator<V> iter);}
  *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Vs) for all non-deleted versions of that Key.
  *
@@ -42,7 +42,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   protected static final String LOSSY = "lossy";
 
   /**
-   * A Java Iterator that translates an Iterator<Value> to an Iterator<V> using the decode method of an Encoder.
+   * A Java Iterator that translates an {@code Iterator<Value>} to an {@code Iterator<V>} using the decode method of an Encoder.
    */
   private static class VIterator<V> implements Iterator<V> {
     private Iterator<Value> source;
@@ -50,7 +50,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     private boolean lossy;
 
     /**
-     * Constructs an Iterator<V> from an Iterator<Value>
+     * Constructs an {@code Iterator<V>} from an {@code Iterator<Value>}
      *
      * @param iter
      *          The source iterator
@@ -114,14 +114,14 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   }
 
   /**
-   * Sets the Encoder<V> used to translate Values to V and back.
+   * Sets the {@code Encoder<V>} used to translate Values to V and back.
    */
   protected void setEncoder(Encoder<V> encoder) {
     this.encoder = encoder;
   }
 
   /**
-   * Instantiates and sets the Encoder<V> used to translate Values to V and back.
+   * Instantiates and sets the {@code Encoder<V>} used to translate Values to V and back.
    *
    * @throws IllegalArgumentException
    *           if ClassNotFoundException, InstantiationException, or IllegalAccessException occurs

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
index efced19..c2023c2 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
@@ -122,7 +122,8 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
     io.setName("sumarray");
-    io.setDescription("SummingArrayCombiner can interpret Values as arrays of Longs using a variety of encodings (arrays of variable length longs or fixed length longs, or comma-separated strings) before summing element-wise.");
+    io.setDescription("SummingArrayCombiner can interpret Values as arrays of Longs using a variety of encodings "
+        + "(arrays of variable length longs or fixed length longs, or comma-separated strings) before summing element-wise.");
     io.addNamedOption(TYPE, "<VARLEN|FIXEDLEN|STRING|fullClassName>");
     return io;
   }
@@ -248,7 +249,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
-   *          Class<? extends Encoder<List<Long>>> specifying the encoding type.
+   *          {@code Class<? extends Encoder<List<Long>>>} specifying the encoding type.
    */
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<List<Long>>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
index e0c8e17..78ad4c9 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
@@ -38,24 +38,24 @@ import org.apache.hadoop.io.WritableComparator;
  * definition of an expression.
  *
  * <P>
- * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&" and "|" indicating that both operands are
+ * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&amp;" and "|" indicating that both operands are
  * necessary, or the either is necessary. The following are valid expressions for visibility:
  *
  * <pre>
  * A
  * A|B
- * (A|B)&(C|D)
- * orange|(red&yellow)
+ * (A|B)&amp;(C|D)
+ * orange|(red&amp;yellow)
  * </pre>
  *
  * <P>
  * The following are not valid expressions for visibility:
  *
  * <pre>
- * A|B&C
+ * A|B&amp;C
  * A=B
  * A|B|
- * A&|B
+ * A&amp;|B
  * ()
  * )
  * dog|!cat

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
index bb72ce5..4fd367c 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
@@ -204,7 +204,8 @@ public class DefaultSecretKeyEncryptionStrategy implements SecretKeyEncryptionSt
       if (!fs.exists(pathToKey)) {
 
         if (encryptionMode == Cipher.DECRYPT_MODE) {
-          log.error("There was a call to decrypt the session key but no key encryption key exists.  Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
+          log.error("There was a call to decrypt the session key but no key encryption key exists.  "
+              + "Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
           throw new RuntimeException("Could not find key encryption key file in configured location in HDFS (" + pathToKeyName + ")");
         } else {
           initializeKeyEncryptingKey(fs, pathToKey, context);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
index d60d076..89a380f 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
@@ -36,16 +36,19 @@ public class BinaryFormatter implements Formatter {
     doTimestamps = printTimestamps;
   }
 
+  @Override
   public boolean hasNext() {
     checkState(si, true);
     return si.hasNext();
   }
 
+  @Override
   public String next() {
     checkState(si, true);
     return formatEntry(si.next(), doTimestamps);
   }
 
+  @Override
   public void remove() {
     checkState(si, true);
     si.remove();
@@ -108,9 +111,7 @@ public class BinaryFormatter implements Formatter {
           sb.append("\\x").append(String.format("%02X", c));
       }
       return sb;
-    }
-
-    else {
+    } else {
       for (int i = 0; i < len; i++) {
 
         int c = 0xff & ba[offset + i];

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
index ca80e37..8215a5a 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
@@ -38,6 +38,7 @@ public class DUCommand extends Command {
 
   private Option optTablePattern, optHumanReadble;
 
+  @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException, TableNotFoundException {
 
     final SortedSet<String> tablesToFlush = new TreeSet<String>(Arrays.asList(cl.getArgs()));
@@ -79,7 +80,8 @@ public class DUCommand extends Command {
 
   @Override
   public String description() {
-    return "prints how much space, in bytes, is used by files referenced by a table.  When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
+    return "prints how much space, in bytes, is used by files referenced by a table.  "
+        + "When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
index b2b2663..16cc5d1 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
@@ -41,7 +41,8 @@ public class EGrepCommand extends GrepCommand {
 
   @Override
   public String description() {
-    return "searches each row, column family, column qualifier and value, in parallel, on the server side (using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
+    return "searches each row, column family, column qualifier and value, in parallel, on the server side "
+        + "(using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
index 0ad94f1..145bb75 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
@@ -61,9 +61,7 @@ public class HistoryCommand extends Command {
           out.close();
         }
       }
-    }
-
-    else {
+    } else {
       BufferedReader in = null;
       try {
         in = new BufferedReader(new InputStreamReader(new FileInputStream(histDir + "/shell_history.txt"), UTF_8));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
index 18de460..f2b78cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
@@ -73,9 +73,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
         } else {
           throw new BadArgumentException("can only escape single quotes, double quotes, the space character, the backslash, and hex input", input, i);
         }
-      }
-      // in a hex escape sequence
-      else if (hexChars != null) {
+      } else if (hexChars != null) {
+        // in a hex escape sequence
         final int digit = Character.digit(ch, 16);
         if (digit < 0) {
           throw new BadArgumentException("expected hex character", input, i);
@@ -93,9 +92,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
           token[tokenLength++] = b;
           hexChars = null;
         }
-      }
-      // in a quote, either end the quote, start escape, or continue a token
-      else if (inQuote) {
+      } else if (inQuote) {
+        // in a quote, either end the quote, start escape, or continue a token
         if (ch == inQuoteChar) {
           inQuote = false;
           tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
@@ -105,9 +103,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
         } else {
           token[tokenLength++] = inputBytes[i];
         }
-      }
-      // not in a quote, either enter a quote, end a token, start escape, or continue a token
-      else {
+      } else {
+        // not in a quote, either enter a quote, end a token, start escape, or continue a token
         if (ch == '\'' || ch == '"') {
           if (tokenLength > 0) {
             tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
index 79aae12..a785727 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
@@ -41,8 +41,8 @@ import org.apache.hadoop.util.ToolRunner;
 import com.beust.jcommander.Parameter;
 
 /**
- * Takes a table and outputs the specified column to a set of part files on hdfs accumulo accumulo.examples.mapreduce.TableToFile <username> <password>
- * <tablename> <column> <hdfs-output-path>
+ * Takes a table and outputs the specified column to a set of part files on hdfs
+ * {@code accumulo accumulo.examples.mapreduce.TableToFile <username> <password> <tablename> <column> <hdfs-output-path>}
  */
 public class TableToFile extends Configured implements Tool {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/Accumulo.java b/server/src/main/java/org/apache/accumulo/server/Accumulo.java
index 0509601..1727bec 100644
--- a/server/src/main/java/org/apache/accumulo/server/Accumulo.java
+++ b/server/src/main/java/org/apache/accumulo/server/Accumulo.java
@@ -299,8 +299,8 @@ public class Accumulo {
       final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(
           ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE, ZooReaderWriter.getRetryingInstance()));
       if (!(fate.list().isEmpty())) {
-        throw new AccumuloException(
-            "Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
+        throw new AccumuloException("Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. "
+            + "Please see the README document for instructions on what to do under your previous version.");
       }
     } catch (Exception exception) {
       log.fatal("Problem verifying Fate readiness", exception);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java b/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
index c0c1e4b..9569e9a 100644
--- a/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
+++ b/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
@@ -75,6 +75,7 @@ public class Compactor implements Callable<CompactionStats> {
 
     private long count;
 
+    @Override
     public CountingIterator deepCopy(IteratorEnvironment env) {
       return new CountingIterator(this, env);
     }
@@ -183,7 +184,7 @@ public class Compactor implements Callable<CompactionStats> {
 
       CompactionReason reason;
 
-      if (compactor.imm != null)
+      if (compactor.imm != null) {
         switch (compactor.mincReason) {
           case USER:
             reason = CompactionReason.USER;
@@ -196,7 +197,7 @@ public class Compactor implements Callable<CompactionStats> {
             reason = CompactionReason.SYSTEM;
             break;
         }
-      else
+      } else {
         switch (compactor.reason) {
           case USER:
             reason = CompactionReason.USER;
@@ -212,6 +213,7 @@ public class Compactor implements Callable<CompactionStats> {
             reason = CompactionReason.SYSTEM;
             break;
         }
+      }
 
       List<IterInfo> iiList = new ArrayList<IterInfo>();
       Map<String,Map<String,String>> iterOptions = new HashMap<String,Map<String,String>>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/util/Initialize.java b/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
index e626fd8..31d53eb 100644
--- a/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
+++ b/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
@@ -157,7 +157,8 @@ public class Initialize {
       c.printNewline();
       c.printString("   bin/accumulo " + org.apache.accumulo.server.util.ChangeSecret.class.getName() + " oldPassword newPassword.");
       c.printNewline();
-      c.printString("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. Without this accumulo will not operate correctly");
+      c.printString("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. "
+          + "Without this accumulo will not operate correctly");
       c.printNewline();
     }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/Main.java b/start/src/main/java/org/apache/accumulo/start/Main.java
index f9b1ab9..a80ebe6 100644
--- a/start/src/main/java/org/apache/accumulo/start/Main.java
+++ b/start/src/main/java/org/apache/accumulo/start/Main.java
@@ -141,7 +141,7 @@ public class Main {
   }
 
   private static void printUsage() {
-    System.out
-        .println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info | tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
+    System.out.println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info "
+        + "| tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
index 2673b5a..d9e2821 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
@@ -71,8 +71,8 @@ public class AccumuloClassLoader {
   }
 
   /**
-   * Parses and XML Document for a property node for a <name> with the value propertyName if it finds one the function return that property's value for its
-   * <value> node. If not found the function will return null
+   * Parses and XML Document for a property node for a &lt;name&gt; with the value propertyName if it finds one the function return that property's value for its
+   * &lt;value&gt; node. If not found the function will return null
    *
    * @param d
    *          XMLDocument to search through

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
index 90d7735..61b146a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
@@ -44,9 +44,8 @@ public class CreateTable extends Test {
       if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
         if (hasPermission)
           throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-        else
-        // create table anyway for sake of state
-        {
+        else {
+          // create table anyway for sake of state
           try {
             state.getConnector().tableOperations().create(tableName);
             WalkingSecurity.get(state).initTable(tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
index 4ec9d22..1f539ff 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
@@ -43,9 +43,8 @@ public class CreateUser extends Test {
         case PERMISSION_DENIED:
           if (hasPermission)
             throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-          else
-          // create user anyway for sake of state
-          {
+          else {
+            // create user anyway for sake of state
             if (!exists) {
               state.getConnector().securityOperations().createLocalUser(tableUserName, tabUserPass);
               WalkingSecurity.get(state).createUser(tableUserName, tabUserPass);


[64/66] [abbrv] accumulo git commit: ACCUMULO-3451 fixes checkstyle violations for 1.6

Posted by ct...@apache.org.
ACCUMULO-3451 fixes checkstyle violations for 1.6

  fixes additional checkstyle rule violations introduced in the 1.6 branch


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/9ca1ff02
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/9ca1ff02
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/9ca1ff02

Branch: refs/heads/master
Commit: 9ca1ff02ef732a8c3727049615ea005bffb7778a
Parents: 1368d09
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 21:29:50 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 21:29:50 2015 -0500

----------------------------------------------------------------------
 .../accumulo/core/security/crypto/CryptoModuleParameters.java  | 2 +-
 .../security/crypto/NonCachingSecretKeyEncryptionStrategy.java | 3 ++-
 .../org/apache/accumulo/core/util/shell/ShellOptionsJC.java    | 6 +++---
 pom.xml                                                        | 2 +-
 .../main/java/org/apache/accumulo/server/init/Initialize.java  | 3 ++-
 .../apache/accumulo/server/util/CustomNonBlockingServer.java   | 4 ++--
 6 files changed, 11 insertions(+), 9 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
index 573d64b..b9bf253 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
@@ -280,7 +280,7 @@ public class CryptoModuleParameters {
    * <li>the code reading an encrypted stream and coming across the encrypted version of one of these keys, OR
    * <li>the {@link CryptoModuleParameters#getKeyEncryptionStrategyClass()} that encrypted the plaintext key (see
    * {@link CryptoModuleParameters#getPlaintextKey()}).
-   * <ul>
+   * </ul>
    * <p>
    * For <b>encryption</b>, this value is generally not required, but is usually set by the underlying module during encryption. <br>
    * For <b>decryption</b>, this value is <b>usually required</b>.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
index f42f9ff..500627c 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
@@ -47,7 +47,8 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
       if (!fs.exists(pathToKey)) {
 
         if (encryptionMode == Cipher.UNWRAP_MODE) {
-          log.error("There was a call to decrypt the session key but no key encryption key exists.  Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
+          log.error("There was a call to decrypt the session key but no key encryption key exists. "
+              + "Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
           throw new RuntimeException("Could not find key encryption key file in configured location in HDFS (" + pathToKeyName + ")");
         } else {
           DataOutputStream out = null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java b/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
index 4787693..dfa24c5 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
@@ -169,9 +169,9 @@ public class ShellOptionsJC {
   @Parameter(names = {"--ssl"}, description = "use ssl to connect to accumulo")
   private boolean useSsl = false;
 
-  @Parameter(
-      names = "--config-file",
-      description = "read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
+  @Parameter(names = "--config-file", description = "read the given client config file. "
+      + "If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, "
+      + "which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
   private String clientConfigFile = null;
 
   @Parameter(names = {"-zi", "--zooKeeperInstanceName"}, description = "use a zookeeper instance with the given instance name")

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 6a90918..8ada8fc 100644
--- a/pom.xml
+++ b/pom.xml
@@ -954,7 +954,7 @@
           </checkstyleRules>
           <violationSeverity>warning</violationSeverity>
           <includeTestSourceDirectory>true</includeTestSourceDirectory>
-          <excludes>**/thrift/*.java</excludes>
+          <excludes>**/thrift/*.java,**/HelpMojo.java</excludes>
         </configuration>
         <dependencies>
           <dependency>

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java b/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
index fae9397..e0a3797 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
@@ -170,7 +170,8 @@ public class Initialize {
       c.println();
       c.println("You can change the instance secret in accumulo by using:");
       c.println("   bin/accumulo " + org.apache.accumulo.server.util.ChangeSecret.class.getName() + " oldPassword newPassword.");
-      c.println("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. Without this accumulo will not operate correctly");
+      c.println("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. "
+          + "Without this accumulo will not operate correctly");
     }
     try {
       if (isInitialized(fs)) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java b/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
index ca53399..472bcb9 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
@@ -250,8 +250,8 @@ public class CustomNonBlockingServer extends THsHaServer {
         clientKey = client.registerSelector(selector, SelectionKey.OP_READ);
 
         // add this key to the map
-          FrameBuffer frameBuffer = processorFactory_.isAsyncProcessor() ?
-                  new CustomAsyncFrameBuffer(client, clientKey,SelectAcceptThread.this) :
+          FrameBuffer frameBuffer =
+              processorFactory_.isAsyncProcessor() ? new CustomAsyncFrameBuffer(client, clientKey,SelectAcceptThread.this) :
                   new CustomFrameBuffer(client, clientKey,SelectAcceptThread.this);
 
           clientKey.attach(frameBuffer);


[57/66] [abbrv] accumulo git commit: ACCUMULO-3451 comply with checkstyle rules

Posted by ct...@apache.org.
ACCUMULO-3451 comply with checkstyle rules

  Adds javadoc and style changes to pass checkstyle enforcement rules.


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/901d60ef
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/901d60ef
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/901d60ef

Branch: refs/heads/1.5
Commit: 901d60ef1cf72c2d55c90746fce94e108a992d3b
Parents: d2c116f
Author: Christopher Tubbs <ct...@apache.org>
Authored: Wed Dec 24 15:22:30 2014 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:25:40 2015 -0500

----------------------------------------------------------------------
 .../core/client/admin/InstanceOperations.java        |  6 +++---
 .../java/org/apache/accumulo/core/data/Value.java    |  7 +++----
 .../apache/accumulo/core/file/rfile/CreateEmpty.java |  4 ++--
 .../accumulo/core/file/rfile/bcfile/TFile.java       |  4 ++--
 .../accumulo/core/file/rfile/bcfile/Utils.java       |  1 +
 .../org/apache/accumulo/core/iterators/Combiner.java |  2 +-
 .../apache/accumulo/core/iterators/LongCombiner.java |  4 ++--
 .../accumulo/core/iterators/OptionDescriber.java     |  4 ++--
 .../accumulo/core/iterators/TypedValueCombiner.java  | 10 +++++-----
 .../core/iterators/user/SummingArrayCombiner.java    |  5 +++--
 .../accumulo/core/security/ColumnVisibility.java     | 10 +++++-----
 .../crypto/DefaultSecretKeyEncryptionStrategy.java   |  3 ++-
 .../accumulo/core/util/format/BinaryFormatter.java   |  7 ++++---
 .../accumulo/core/util/shell/commands/DUCommand.java |  4 +++-
 .../core/util/shell/commands/EGrepCommand.java       |  3 ++-
 .../core/util/shell/commands/HistoryCommand.java     |  4 +---
 .../util/shell/commands/QuotedStringTokenizer.java   | 15 ++++++---------
 .../examples/simple/mapreduce/TableToFile.java       |  4 ++--
 .../java/org/apache/accumulo/server/Accumulo.java    |  4 ++--
 .../accumulo/server/tabletserver/Compactor.java      |  6 ++++--
 .../org/apache/accumulo/server/util/Initialize.java  |  3 ++-
 .../main/java/org/apache/accumulo/start/Main.java    |  4 ++--
 .../start/classloader/AccumuloClassLoader.java       |  4 ++--
 .../test/randomwalk/security/CreateTable.java        |  5 ++---
 .../test/randomwalk/security/CreateUser.java         |  5 ++---
 25 files changed, 65 insertions(+), 63 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
index 049dd85..8cb656c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
@@ -82,7 +82,7 @@ public interface InstanceOperations {
    * List the active scans on tablet server.
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @return A list of active scans on tablet server.
    */
 
@@ -92,7 +92,7 @@ public interface InstanceOperations {
    * List the active compaction running on a tablet server
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @return the list of active compactions
    * @since 1.5.0
    */
@@ -103,7 +103,7 @@ public interface InstanceOperations {
    * Throws an exception if a tablet server can not be contacted.
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @since 1.5.0
    */
   public void ping(String tserver) throws AccumuloException;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/data/Value.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Value.java b/core/src/main/java/org/apache/accumulo/core/data/Value.java
index ba89d6c..b937203 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Value.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Value.java
@@ -139,12 +139,13 @@ public class Value implements WritableComparable<Object> {
     return this.value.length;
   }
 
+  @Override
   public void readFields(final DataInput in) throws IOException {
     this.value = new byte[in.readInt()];
     in.readFully(this.value, 0, this.value.length);
   }
 
-  /** {@inheritDoc} */
+  @Override
   public void write(final DataOutput out) throws IOException {
     out.writeInt(this.value.length);
     out.write(this.value, 0, this.value.length);
@@ -152,7 +153,6 @@ public class Value implements WritableComparable<Object> {
 
   // Below methods copied from BytesWritable
 
-  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     return WritableComparator.hashBytes(value, this.value.length);
@@ -165,6 +165,7 @@ public class Value implements WritableComparable<Object> {
    *          The other bytes writable
    * @return Positive if left is bigger than right, 0 if they are equal, and negative if left is smaller than right.
    */
+  @Override
   public int compareTo(Object right_obj) {
     return compareTo(((Value) right_obj).get());
   }
@@ -179,7 +180,6 @@ public class Value implements WritableComparable<Object> {
     return (diff != 0) ? diff : WritableComparator.compareBytes(this.value, 0, this.value.length, that, 0, that.length);
   }
 
-  /** {@inheritDoc} */
   @Override
   public boolean equals(Object right_obj) {
     if (right_obj instanceof byte[]) {
@@ -207,7 +207,6 @@ public class Value implements WritableComparable<Object> {
       super(Value.class);
     }
 
-    /** {@inheritDoc} */
     @Override
     public int compare(byte[] b1, int s1, int l1, byte[] b2, int s2, int l2) {
       return comparator.compare(b1, s1, l1, b2, s2, l2);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
index bd9fa43..8d54088 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
@@ -59,8 +59,8 @@ public class CreateEmpty {
   static class Opts extends Help {
     @Parameter(names = {"-c", "--codec"}, description = "the compression codec to use.", validateWith = IsSupportedCompressionAlgorithm.class)
     String codec = TFile.COMPRESSION_NONE;
-    @Parameter(
-        description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
+    @Parameter(description = " <path> { <path> ... } Each path given is a URL. "
+        + "Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
         required = true, validateWith = NamedLikeRFile.class)
     List<String> files = new ArrayList<String>();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
index e21598a..400695b 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
@@ -1065,9 +1065,9 @@ public class TFile {
        * @param reader
        *          The TFile reader object.
        * @param beginKey
-       *          Begin key of the scan. If null, scan from the first <K,V> entry of the TFile.
+       *          Begin key of the scan. If null, scan from the first {@code <K,V>} entry of the TFile.
        * @param endKey
-       *          End key of the scan. If null, scan up to the last <K, V> entry of the TFile.
+       *          End key of the scan. If null, scan up to the last {@code <K, V>} entry of the TFile.
        */
       protected Scanner(Reader reader, RawComparable beginKey, RawComparable endKey) throws IOException {
         this(reader, (beginKey == null) ? reader.begin() : reader.getBlockContainsKey(beginKey, false), reader.end());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
index fb0277a..84b861b 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
@@ -164,6 +164,7 @@ public final class Utils {
    * <li>if (FB in [-104, -73]), return (FB+88)<<16 + (NB[0]&0xff)<<8 + NB[1]&0xff;
    * <li>if (FB in [-120, -105]), return (FB+112)<<24 + (NB[0]&0xff)<<16 + (NB[1]&0xff)<<8 + NB[2]&0xff;
    * <li>if (FB in [-128, -121]), return interpret NB[FB+129] as a signed big-endian integer.
+   * </ul>
    *
    * @param in
    *          input stream

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
index 41e4e1e..f75076d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
@@ -70,7 +70,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
      * Constructs an iterator over Values whose Keys are versions of the current topKey of the source SortedKeyValueIterator.
      *
      * @param source
-     *          The SortedKeyValueIterator<Key,Value> from which to read data.
+     *          The {@code SortedKeyValueIterator<Key,Value>} from which to read data.
      */
     public ValueIterator(SortedKeyValueIterator<Key,Value> source) {
       this.source = source;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
index 0bffbf7..cfdfd6e 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.io.WritableUtils;
 /**
  * A TypedValueCombiner that translates each Value to a Long before reducing, then encodes the reduced Long back to a Value.
  *
- * Subclasses must implement a typedReduce method: public Long typedReduce(Key key, Iterator<Long> iter);
+ * Subclasses must implement a typedReduce method: {@code public Long typedReduce(Key key, Iterator<Long> iter);}
  *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Longs) for all non-deleted versions of that Key.
  *
@@ -226,7 +226,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
-   *          Class<? extends Encoder<Long>> specifying the encoding type.
+   *          {@code Class<? extends Encoder<Long>>} specifying the encoding type.
    */
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<Long>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
index 6158bfc..6593ecd 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
@@ -50,8 +50,8 @@ public interface OptionDescriber {
      * @param unnamedOptionDescriptions
      *          is a list of descriptions of additional options that don't have fixed names (null if unused). The descriptions are intended to describe a
      *          category, and the user will provide parameter names and values in that category; e.g., the FilteringIterator needs a list of Filters intended to
-     *          be named by their priority numbers, so its unnamedOptionDescriptions =
-     *          Collections.singletonList("<filterPriorityNumber> <ageoff|regex|filterClass>")
+     *          be named by their priority numbers, so its<br>
+     *          {@code unnamedOptionDescriptions = Collections.singletonList("<filterPriorityNumber> <ageoff|regex|filterClass>")}
      */
     public IteratorOptions(String name, String description, Map<String,String> namedOptions, List<String> unnamedOptionDescriptions) {
       this.name = name;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
index d1ae9f5..0f26fab 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
@@ -29,7 +29,7 @@ import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
 /**
  * A Combiner that decodes each Value to type V before reducing, then encodes the result of typedReduce back to Value.
  *
- * Subclasses must implement a typedReduce method: public V typedReduce(Key key, Iterator<V> iter);
+ * Subclasses must implement a typedReduce method: {@code public V typedReduce(Key key, Iterator<V> iter);}
  *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Vs) for all non-deleted versions of that Key.
  *
@@ -42,7 +42,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   protected static final String LOSSY = "lossy";
 
   /**
-   * A Java Iterator that translates an Iterator<Value> to an Iterator<V> using the decode method of an Encoder.
+   * A Java Iterator that translates an {@code Iterator<Value>} to an {@code Iterator<V>} using the decode method of an Encoder.
    */
   private static class VIterator<V> implements Iterator<V> {
     private Iterator<Value> source;
@@ -50,7 +50,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     private boolean lossy;
 
     /**
-     * Constructs an Iterator<V> from an Iterator<Value>
+     * Constructs an {@code Iterator<V>} from an {@code Iterator<Value>}
      *
      * @param iter
      *          The source iterator
@@ -114,14 +114,14 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   }
 
   /**
-   * Sets the Encoder<V> used to translate Values to V and back.
+   * Sets the {@code Encoder<V>} used to translate Values to V and back.
    */
   protected void setEncoder(Encoder<V> encoder) {
     this.encoder = encoder;
   }
 
   /**
-   * Instantiates and sets the Encoder<V> used to translate Values to V and back.
+   * Instantiates and sets the {@code Encoder<V>} used to translate Values to V and back.
    *
    * @throws IllegalArgumentException
    *           if ClassNotFoundException, InstantiationException, or IllegalAccessException occurs

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
index efced19..c2023c2 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
@@ -122,7 +122,8 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
     io.setName("sumarray");
-    io.setDescription("SummingArrayCombiner can interpret Values as arrays of Longs using a variety of encodings (arrays of variable length longs or fixed length longs, or comma-separated strings) before summing element-wise.");
+    io.setDescription("SummingArrayCombiner can interpret Values as arrays of Longs using a variety of encodings "
+        + "(arrays of variable length longs or fixed length longs, or comma-separated strings) before summing element-wise.");
     io.addNamedOption(TYPE, "<VARLEN|FIXEDLEN|STRING|fullClassName>");
     return io;
   }
@@ -248,7 +249,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
-   *          Class<? extends Encoder<List<Long>>> specifying the encoding type.
+   *          {@code Class<? extends Encoder<List<Long>>>} specifying the encoding type.
    */
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<List<Long>>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
index e0c8e17..78ad4c9 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
@@ -38,24 +38,24 @@ import org.apache.hadoop.io.WritableComparator;
  * definition of an expression.
  *
  * <P>
- * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&" and "|" indicating that both operands are
+ * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&amp;" and "|" indicating that both operands are
  * necessary, or the either is necessary. The following are valid expressions for visibility:
  *
  * <pre>
  * A
  * A|B
- * (A|B)&(C|D)
- * orange|(red&yellow)
+ * (A|B)&amp;(C|D)
+ * orange|(red&amp;yellow)
  * </pre>
  *
  * <P>
  * The following are not valid expressions for visibility:
  *
  * <pre>
- * A|B&C
+ * A|B&amp;C
  * A=B
  * A|B|
- * A&|B
+ * A&amp;|B
  * ()
  * )
  * dog|!cat

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
index bb72ce5..4fd367c 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
@@ -204,7 +204,8 @@ public class DefaultSecretKeyEncryptionStrategy implements SecretKeyEncryptionSt
       if (!fs.exists(pathToKey)) {
 
         if (encryptionMode == Cipher.DECRYPT_MODE) {
-          log.error("There was a call to decrypt the session key but no key encryption key exists.  Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
+          log.error("There was a call to decrypt the session key but no key encryption key exists.  "
+              + "Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
           throw new RuntimeException("Could not find key encryption key file in configured location in HDFS (" + pathToKeyName + ")");
         } else {
           initializeKeyEncryptingKey(fs, pathToKey, context);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
index d60d076..89a380f 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
@@ -36,16 +36,19 @@ public class BinaryFormatter implements Formatter {
     doTimestamps = printTimestamps;
   }
 
+  @Override
   public boolean hasNext() {
     checkState(si, true);
     return si.hasNext();
   }
 
+  @Override
   public String next() {
     checkState(si, true);
     return formatEntry(si.next(), doTimestamps);
   }
 
+  @Override
   public void remove() {
     checkState(si, true);
     si.remove();
@@ -108,9 +111,7 @@ public class BinaryFormatter implements Formatter {
           sb.append("\\x").append(String.format("%02X", c));
       }
       return sb;
-    }
-
-    else {
+    } else {
       for (int i = 0; i < len; i++) {
 
         int c = 0xff & ba[offset + i];

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
index ca80e37..8215a5a 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
@@ -38,6 +38,7 @@ public class DUCommand extends Command {
 
   private Option optTablePattern, optHumanReadble;
 
+  @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException, TableNotFoundException {
 
     final SortedSet<String> tablesToFlush = new TreeSet<String>(Arrays.asList(cl.getArgs()));
@@ -79,7 +80,8 @@ public class DUCommand extends Command {
 
   @Override
   public String description() {
-    return "prints how much space, in bytes, is used by files referenced by a table.  When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
+    return "prints how much space, in bytes, is used by files referenced by a table.  "
+        + "When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
index b2b2663..16cc5d1 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
@@ -41,7 +41,8 @@ public class EGrepCommand extends GrepCommand {
 
   @Override
   public String description() {
-    return "searches each row, column family, column qualifier and value, in parallel, on the server side (using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
+    return "searches each row, column family, column qualifier and value, in parallel, on the server side "
+        + "(using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
index 0ad94f1..145bb75 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
@@ -61,9 +61,7 @@ public class HistoryCommand extends Command {
           out.close();
         }
       }
-    }
-
-    else {
+    } else {
       BufferedReader in = null;
       try {
         in = new BufferedReader(new InputStreamReader(new FileInputStream(histDir + "/shell_history.txt"), UTF_8));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
index 18de460..f2b78cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
@@ -73,9 +73,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
         } else {
           throw new BadArgumentException("can only escape single quotes, double quotes, the space character, the backslash, and hex input", input, i);
         }
-      }
-      // in a hex escape sequence
-      else if (hexChars != null) {
+      } else if (hexChars != null) {
+        // in a hex escape sequence
         final int digit = Character.digit(ch, 16);
         if (digit < 0) {
           throw new BadArgumentException("expected hex character", input, i);
@@ -93,9 +92,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
           token[tokenLength++] = b;
           hexChars = null;
         }
-      }
-      // in a quote, either end the quote, start escape, or continue a token
-      else if (inQuote) {
+      } else if (inQuote) {
+        // in a quote, either end the quote, start escape, or continue a token
         if (ch == inQuoteChar) {
           inQuote = false;
           tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
@@ -105,9 +103,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
         } else {
           token[tokenLength++] = inputBytes[i];
         }
-      }
-      // not in a quote, either enter a quote, end a token, start escape, or continue a token
-      else {
+      } else {
+        // not in a quote, either enter a quote, end a token, start escape, or continue a token
         if (ch == '\'' || ch == '"') {
           if (tokenLength > 0) {
             tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
index 79aae12..a785727 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
@@ -41,8 +41,8 @@ import org.apache.hadoop.util.ToolRunner;
 import com.beust.jcommander.Parameter;
 
 /**
- * Takes a table and outputs the specified column to a set of part files on hdfs accumulo accumulo.examples.mapreduce.TableToFile <username> <password>
- * <tablename> <column> <hdfs-output-path>
+ * Takes a table and outputs the specified column to a set of part files on hdfs
+ * {@code accumulo accumulo.examples.mapreduce.TableToFile <username> <password> <tablename> <column> <hdfs-output-path>}
  */
 public class TableToFile extends Configured implements Tool {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/Accumulo.java b/server/src/main/java/org/apache/accumulo/server/Accumulo.java
index 0509601..1727bec 100644
--- a/server/src/main/java/org/apache/accumulo/server/Accumulo.java
+++ b/server/src/main/java/org/apache/accumulo/server/Accumulo.java
@@ -299,8 +299,8 @@ public class Accumulo {
       final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(
           ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE, ZooReaderWriter.getRetryingInstance()));
       if (!(fate.list().isEmpty())) {
-        throw new AccumuloException(
-            "Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
+        throw new AccumuloException("Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. "
+            + "Please see the README document for instructions on what to do under your previous version.");
       }
     } catch (Exception exception) {
       log.fatal("Problem verifying Fate readiness", exception);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java b/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
index c0c1e4b..9569e9a 100644
--- a/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
+++ b/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
@@ -75,6 +75,7 @@ public class Compactor implements Callable<CompactionStats> {
 
     private long count;
 
+    @Override
     public CountingIterator deepCopy(IteratorEnvironment env) {
       return new CountingIterator(this, env);
     }
@@ -183,7 +184,7 @@ public class Compactor implements Callable<CompactionStats> {
 
       CompactionReason reason;
 
-      if (compactor.imm != null)
+      if (compactor.imm != null) {
         switch (compactor.mincReason) {
           case USER:
             reason = CompactionReason.USER;
@@ -196,7 +197,7 @@ public class Compactor implements Callable<CompactionStats> {
             reason = CompactionReason.SYSTEM;
             break;
         }
-      else
+      } else {
         switch (compactor.reason) {
           case USER:
             reason = CompactionReason.USER;
@@ -212,6 +213,7 @@ public class Compactor implements Callable<CompactionStats> {
             reason = CompactionReason.SYSTEM;
             break;
         }
+      }
 
       List<IterInfo> iiList = new ArrayList<IterInfo>();
       Map<String,Map<String,String>> iterOptions = new HashMap<String,Map<String,String>>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/util/Initialize.java b/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
index e626fd8..31d53eb 100644
--- a/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
+++ b/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
@@ -157,7 +157,8 @@ public class Initialize {
       c.printNewline();
       c.printString("   bin/accumulo " + org.apache.accumulo.server.util.ChangeSecret.class.getName() + " oldPassword newPassword.");
       c.printNewline();
-      c.printString("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. Without this accumulo will not operate correctly");
+      c.printString("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. "
+          + "Without this accumulo will not operate correctly");
       c.printNewline();
     }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/Main.java b/start/src/main/java/org/apache/accumulo/start/Main.java
index f9b1ab9..a80ebe6 100644
--- a/start/src/main/java/org/apache/accumulo/start/Main.java
+++ b/start/src/main/java/org/apache/accumulo/start/Main.java
@@ -141,7 +141,7 @@ public class Main {
   }
 
   private static void printUsage() {
-    System.out
-        .println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info | tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
+    System.out.println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info "
+        + "| tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
index 2673b5a..d9e2821 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
@@ -71,8 +71,8 @@ public class AccumuloClassLoader {
   }
 
   /**
-   * Parses and XML Document for a property node for a <name> with the value propertyName if it finds one the function return that property's value for its
-   * <value> node. If not found the function will return null
+   * Parses and XML Document for a property node for a &lt;name&gt; with the value propertyName if it finds one the function return that property's value for its
+   * &lt;value&gt; node. If not found the function will return null
    *
    * @param d
    *          XMLDocument to search through

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
index 90d7735..61b146a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
@@ -44,9 +44,8 @@ public class CreateTable extends Test {
       if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
         if (hasPermission)
           throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-        else
-        // create table anyway for sake of state
-        {
+        else {
+          // create table anyway for sake of state
           try {
             state.getConnector().tableOperations().create(tableName);
             WalkingSecurity.get(state).initTable(tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
index 4ec9d22..1f539ff 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
@@ -43,9 +43,8 @@ public class CreateUser extends Test {
         case PERMISSION_DENIED:
           if (hasPermission)
             throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-          else
-          // create user anyway for sake of state
-          {
+          else {
+            // create user anyway for sake of state
             if (!exists) {
               state.getConnector().securityOperations().createLocalUser(tableUserName, tabUserPass);
               WalkingSecurity.get(state).createUser(tableUserName, tabUserPass);


[43/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
index 37caf15..77081bf 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
@@ -36,16 +36,16 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
 /**
  * This class allows MapReduce jobs to use Accumulo as the source of data. This {@link InputFormat} provides row names as {@link Text} as keys, and a
  * corresponding {@link PeekingIterator} as a value, which in turn makes the {@link Key}/{@link Value} pairs for that row available to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloRowInputFormat#setConnectorInfo(Job, String, AuthenticationToken)}
  * <li>{@link AccumuloRowInputFormat#setInputTableName(Job, String)}
  * <li>{@link AccumuloRowInputFormat#setScanAuthorizations(Job, Authorizations)}
  * <li>{@link AccumuloRowInputFormat#setZooKeeperInstance(Job, ClientConfiguration)} OR {@link AccumuloRowInputFormat#setMockInstance(Job, String)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 public class AccumuloRowInputFormat extends InputFormatBase<Text,PeekingIterator<Entry<Key,Value>>> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
index e58e350..a60cb80 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
@@ -54,7 +54,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the table name from the configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return the table name
@@ -67,7 +67,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param tableName
@@ -80,7 +80,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Sets the input ranges to scan for the single input table associated with this job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param ranges
@@ -93,7 +93,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the ranges to scan over from a job.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return the ranges
@@ -106,7 +106,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Restricts the columns that will be mapped over for this job for the default input table.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param columnFamilyColumnQualifierPairs
@@ -120,7 +120,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the columns to be mapped over from this job.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return a set of columns
@@ -133,7 +133,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Encode an iterator on the single input table for this job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param cfg
@@ -146,7 +146,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets a list of the iterator settings (for iterators to apply to a scanner) from this configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return a list of iterators
@@ -160,10 +160,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -177,7 +177,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return false if the feature is disabled, true otherwise
@@ -190,10 +190,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Controls the use of the {@link IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -206,7 +206,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -220,10 +220,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
   /**
    * Controls the use of the {@link ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack to be constructed within the Map
    * task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -236,7 +236,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -252,26 +252,26 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -284,7 +284,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -297,7 +297,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Initializes an Accumulo {@link org.apache.accumulo.core.client.impl.TabletLocator} based on the configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return an Accumulo tablet locator
@@ -315,7 +315,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
     /**
      * Apply the configured iterators from the configuration to the scanner for the specified table name
-     * 
+     *
      * @param context
      *          the Hadoop context for the configured job
      * @param scanner
@@ -329,7 +329,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
     /**
      * Apply the configured iterators from the configuration to the scanner.
-     * 
+     *
      * @param context
      *          the Hadoop context for the configured job
      * @param scanner

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
index fa3b7eb..03473f2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
@@ -48,7 +48,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Creates a batch scan config object out of a previously serialized batch scan config object.
-   * 
+   *
    * @param input
    *          the data input of the serialized batch scan config
    */
@@ -58,7 +58,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Sets the input ranges to scan for all tables associated with this job. This will be added to any per-table ranges that have been set using
-   * 
+   *
    * @param ranges
    *          the ranges that will be mapped over
    * @since 1.6.0
@@ -77,7 +77,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Restricts the columns that will be mapped over for this job for the default input table.
-   * 
+   *
    * @param columns
    *          a pair of {@link Text} objects corresponding to column family and column qualifier. If the column qualifier is null, the entire column family is
    *          selected. An empty set is the default and is equivalent to scanning the all columns.
@@ -97,7 +97,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Set iterators on to be used in the query.
-   * 
+   *
    * @param iterators
    *          the configurations for the iterators
    * @since 1.6.0
@@ -117,10 +117,10 @@ public class InputTableConfig implements Writable {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param autoAdjustRanges
    *          the feature is enabled if true, disabled otherwise
    * @see #setRanges(java.util.List)
@@ -133,7 +133,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @return false if the feature is disabled, true otherwise
    * @since 1.6.0
    * @see #setAutoAdjustRanges(boolean)
@@ -146,10 +146,10 @@ public class InputTableConfig implements Writable {
    * Controls the use of the {@link org.apache.accumulo.core.client.ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack
    * to be constructed within the Map task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be
    * available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param useLocalIterators
    *          the feature is enabled if true, disabled otherwise
    * @since 1.6.0
@@ -161,7 +161,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @return true if the feature is enabled, false otherwise
    * @since 1.6.0
    * @see #setUseLocalIterators(boolean)
@@ -175,26 +175,26 @@ public class InputTableConfig implements Writable {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath. The accumulo-site.xml may need to be on the mapper's classpath if HDFS or the Accumulo directory in HDFS are non-standard.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param offlineScan
    *          the feature is enabled if true, disabled otherwise
    * @since 1.6.0
@@ -206,7 +206,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @return true if the feature is enabled, false otherwise
    * @since 1.6.0
    * @see #setOfflineScan(boolean)
@@ -217,10 +217,10 @@ public class InputTableConfig implements Writable {
 
   /**
    * Controls the use of the {@link org.apache.accumulo.core.client.IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param useIsolatedScanners
    *          the feature is enabled if true, disabled otherwise
    * @since 1.6.0
@@ -232,7 +232,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @return true if the feature is enabled, false otherwise
    * @since 1.6.0
    * @see #setUseIsolatedScanners(boolean)
@@ -243,7 +243,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Writes the state for the current object out to the specified {@link DataOutput}
-   * 
+   *
    * @param dataOutput
    *          the output for which to write the object's state
    */
@@ -286,7 +286,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Reads the fields in the {@link DataInput} into the current object
-   * 
+   *
    * @param dataInput
    *          the input fields to read into the current object
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
index 29cf95d..fe27b01 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
@@ -33,8 +33,8 @@ import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.ZooKeeperInstance;
-import org.apache.accumulo.core.client.mapreduce.lib.impl.InputConfigurator;
 import org.apache.accumulo.core.client.mapreduce.lib.impl.ConfiguratorBase.TokenSource;
+import org.apache.accumulo.core.client.mapreduce.lib.impl.InputConfigurator;
 import org.apache.accumulo.core.client.mock.MockInstance;
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken;
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken.AuthenticationTokenSerializer;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
index ae1d46f..b2b5150 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
@@ -22,6 +22,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 import java.io.IOException;
 import java.net.URI;
 import java.net.URISyntaxException;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.ClientConfiguration;
@@ -47,7 +48,7 @@ public class ConfiguratorBase {
 
   /**
    * Configuration keys for {@link Instance#getConnector(String, AuthenticationToken)}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum ConnectorInfo {
@@ -70,7 +71,7 @@ public class ConfiguratorBase {
 
   /**
    * Configuration keys for {@link Instance}, {@link ZooKeeperInstance}, and {@link MockInstance}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum InstanceOpts {
@@ -79,17 +80,16 @@ public class ConfiguratorBase {
 
   /**
    * Configuration keys for general configuration options.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum GeneralOpts {
-    LOG_LEVEL,
-    VISIBILITY_CACHE_SIZE
+    LOG_LEVEL, VISIBILITY_CACHE_SIZE
   }
 
   /**
    * Provides a configuration key for a given feature enum, prefixed by the implementingClass
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param e
@@ -102,23 +102,23 @@ public class ConfiguratorBase {
   }
 
   /**
-  * Provides a configuration key for a given feature enum.
-  * 
-  * @param e
-  *          the enum used to provide the unique part of the configuration key
-  * @return the configuration key
-  */
+   * Provides a configuration key for a given feature enum.
+   *
+   * @param e
+   *          the enum used to provide the unique part of the configuration key
+   * @return the configuration key
+   */
   protected static String enumToConfKey(Enum<?> e) {
-	  return  e.getDeclaringClass().getSimpleName() + "." +  StringUtils.camelize(e.name().toLowerCase());
+    return e.getDeclaringClass().getSimpleName() + "." + StringUtils.camelize(e.name().toLowerCase());
   }
 
   /**
    * Sets the connector information needed to communicate with Accumulo in this job.
-   * 
+   *
    * <p>
    * <b>WARNING:</b> The serialized token is stored in the configuration and shared with all MapReduce tasks. It is BASE64 encoded to provide a charset safe
    * conversion to a string, and is not intended to be secure.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -144,11 +144,11 @@ public class ConfiguratorBase {
 
   /**
    * Sets the connector information needed to communicate with Accumulo in this job.
-   * 
+   *
    * <p>
    * Pulls a token file into the Distributed Cache that contains the authentication token in an attempt to be more secure than storing the password in the
    * Configuration. Token file created with "bin/accumulo create-token".
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -179,7 +179,7 @@ public class ConfiguratorBase {
 
   /**
    * Determines if the connector info has already been set for this instance.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -194,7 +194,7 @@ public class ConfiguratorBase {
 
   /**
    * Gets the user name from the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -209,7 +209,7 @@ public class ConfiguratorBase {
 
   /**
    * Gets the authenticated token from either the specified token file or directly from the configuration, whichever was used when the job was configured.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -237,7 +237,7 @@ public class ConfiguratorBase {
 
   /**
    * Reads from the token file in distributed cache. Currently, the token file stores data separated by colons e.g. principal:token_class:token
-   * 
+   *
    * @param conf
    *          the Hadoop context for the configured job
    * @return path to the token file as a String
@@ -275,7 +275,7 @@ public class ConfiguratorBase {
 
   /**
    * Configures a {@link ZooKeeperInstance} for this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -296,7 +296,7 @@ public class ConfiguratorBase {
 
   /**
    * Configures a {@link MockInstance} for this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -317,7 +317,7 @@ public class ConfiguratorBase {
 
   /**
    * Initializes an Accumulo {@link Instance} based on the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -348,7 +348,7 @@ public class ConfiguratorBase {
 
   /**
    * Sets the log level for this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -365,7 +365,7 @@ public class ConfiguratorBase {
 
   /**
    * Gets the log level from this configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -380,7 +380,7 @@ public class ConfiguratorBase {
 
   /**
    * Sets the valid visibility count for this job.
-   * 
+   *
    * @param conf
    *          the Hadoop configuration object to configure
    * @param visibilityCacheSize
@@ -392,13 +392,13 @@ public class ConfiguratorBase {
 
   /**
    * Gets the valid visibility count for this job.
-   * 
+   *
    * @param conf
    *          the Hadoop configuration object to configure
    * @return the valid visibility count
    */
   public static int getVisibilityCacheSize(Configuration conf) {
-    return conf.getInt(enumToConfKey(GeneralOpts.VISIBILITY_CACHE_SIZE),Constants.DEFAULT_VISIBILITY_CACHE_SIZE);
+    return conf.getInt(enumToConfKey(GeneralOpts.VISIBILITY_CACHE_SIZE), Constants.DEFAULT_VISIBILITY_CACHE_SIZE);
   }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
index ce84209..882c6d3 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
@@ -31,7 +31,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for {@link AccumuloConfiguration}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum Opts {
@@ -41,7 +41,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * The supported Accumulo properties we set in this OutputFormat, that change the behavior of the RecordWriter.<br />
    * These properties correspond to the supported public static setter methods available to this class.
-   * 
+   *
    * @param property
    *          the Accumulo property to check
    * @since 1.6.0
@@ -61,7 +61,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Helper for transforming Accumulo configuration properties into something that can be stored safely inside the Hadoop Job configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -86,7 +86,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * This helper method provides an AccumuloConfiguration object constructed from the Accumulo defaults, and overridden with Accumulo properties that have been
    * stored in the Job's configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -104,7 +104,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the compression type to use for data blocks. Specifying a compression may require additional libraries to be available to your Job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -122,10 +122,10 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the size for data blocks within each file.<br />
    * Data blocks are a span of key/value pairs stored in the file that are compressed and indexed as a group.
-   * 
+   *
    * <p>
    * Making this value smaller may increase seek performance, but at the cost of increasing the size of the indexes (which can also affect seek performance).
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -140,7 +140,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the size for file blocks in the file system; file blocks are managed, and replicated, by the underlying file system.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -156,7 +156,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the size for index blocks within each file; smaller blocks means a deeper index hierarchy within the file, while larger blocks mean a more shallow
    * index hierarchy within the file. This can affect the performance of queries.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -171,7 +171,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the file system replication factor for the resulting file, overriding the file system default.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
index af84bb4..5405ac0 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
@@ -81,7 +81,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for {@link Scanner}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum ScanOpts {
@@ -90,7 +90,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for various features.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum Features {
@@ -99,7 +99,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -115,7 +115,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -128,7 +128,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the {@link Authorizations} used to scan. Must be a subset of the user's authorization. Defaults to the empty set.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -144,7 +144,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the authorizations to set for the scans from the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -160,7 +160,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the input ranges to scan on all input tables for this job. If not set, the entire table will be scanned.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -189,7 +189,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the ranges to scan over from a job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -215,7 +215,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets a list of the iterator settings (for iterators to apply to a scanner) from this configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -249,7 +249,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Restricts the columns that will be mapped over for the single input table on this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -286,7 +286,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the columns to be mapped over from this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -326,7 +326,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Encode an iterator on the input for the single input table associated with this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -364,10 +364,10 @@ public class InputConfigurator extends ConfiguratorBase {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -383,7 +383,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -398,10 +398,10 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Controls the use of the {@link IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -416,7 +416,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -432,10 +432,10 @@ public class InputConfigurator extends ConfiguratorBase {
   /**
    * Controls the use of the {@link ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack to be constructed within the Map
    * task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -450,7 +450,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -468,26 +468,26 @@ public class InputConfigurator extends ConfiguratorBase {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -502,7 +502,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -517,7 +517,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets configurations for multiple tables at a time.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -544,7 +544,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Returns all {@link InputTableConfig} objects associated with this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -577,7 +577,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Returns the {@link InputTableConfig} for the given table
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -594,7 +594,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Initializes an Accumulo {@link TabletLocator} based on the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -619,7 +619,7 @@ public class InputConfigurator extends ConfiguratorBase {
   // InputFormat doesn't have the equivalent of OutputFormat's checkOutputSpecs(JobContext job)
   /**
    * Check whether a configuration is fully configured to be used with an Accumulo {@link org.apache.hadoop.mapreduce.InputFormat}.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -674,7 +674,7 @@ public class InputConfigurator extends ConfiguratorBase {
   /**
    * Returns the {@link org.apache.accumulo.core.client.mapreduce.InputTableConfig} for the configuration based on the properties set using the single-table
    * input methods.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
index 13b67d5..55e980c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
@@ -23,6 +23,7 @@ import java.io.ByteArrayOutputStream;
 import java.io.DataInputStream;
 import java.io.DataOutputStream;
 import java.io.IOException;
+
 import org.apache.accumulo.core.client.BatchWriter;
 import org.apache.accumulo.core.client.BatchWriterConfig;
 import org.apache.hadoop.conf.Configuration;
@@ -34,7 +35,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for {@link BatchWriter}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum WriteOpts {
@@ -43,7 +44,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for various features.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum Features {
@@ -53,7 +54,7 @@ public class OutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the default table name to use if one emits a null in place of a table name for a given mutation. Table names can only be alpha-numeric and
    * underscores.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -69,7 +70,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the default table name from the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -85,7 +86,7 @@ public class OutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the configuration for for the job's {@link BatchWriter} instances. If not set, a new {@link BatchWriterConfig}, with sensible built-in defaults is
    * used. Setting the configuration multiple times overwrites any previous configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -109,7 +110,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the {@link BatchWriterConfig} settings.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -137,10 +138,10 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the directive to create new tables, as necessary. Table names can only be alpha-numeric and underscores.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -155,7 +156,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether tables are permitted to be created as needed.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -170,10 +171,10 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the directive to use simulation mode for this job. In simulation mode, no output is produced. This is useful for testing.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -188,7 +189,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether this feature is enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
index 243160d..34ea7d2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
@@ -21,13 +21,13 @@
  * InputFormats/OutputFormats, so as not to clutter their API with methods that don't match the conventions for that framework. These classes may be useful to
  * input/output plugins for other frameworks, so they can reuse the same configuration options and/or serialize them into a
  * {@link org.apache.hadoop.conf.Configuration} instance in a standard way.
- * 
+ *
  * <p>
  * It is not expected these will change much (except when new features are added), but end users should not use these classes. They should use the static
  * configurators on the {@link org.apache.hadoop.mapreduce.InputFormat} or {@link org.apache.hadoop.mapreduce.OutputFormat} they are configuring, which in turn
  * may use these classes to implement their own static configurators. Once again, these classes are intended for internal use, but may be useful to developers
  * of plugins for other frameworks that read/write to Accumulo.
- * 
+ *
  * @since 1.6.0
  */
 package org.apache.accumulo.core.client.mapreduce.lib.impl;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
index c59841d..bd4857e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
@@ -28,29 +28,29 @@ import org.apache.hadoop.mapreduce.Partitioner;
  */
 public class KeyRangePartitioner extends Partitioner<Key,Writable> implements Configurable {
   private RangePartitioner rp = new RangePartitioner();
-  
+
   @Override
   public int getPartition(Key key, Writable value, int numPartitions) {
     return rp.getPartition(key.getRow(), value, numPartitions);
   }
-  
+
   @Override
   public Configuration getConf() {
     return rp.getConf();
   }
-  
+
   @Override
   public void setConf(Configuration conf) {
     rp.setConf(conf);
   }
-  
+
   /**
    * Sets the hdfs file name to use, containing a newline separated list of Base64 encoded split points that represent ranges for partitioning
    */
   public static void setSplitFile(Job job, String file) {
     RangePartitioner.setSplitFile(job, file);
   }
-  
+
   /**
    * Sets the number of random sub-bins per range
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
index 840db41..d4d4004 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
@@ -27,18 +27,18 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 public class IteratorAdapter implements Iterator<Entry<Key,Value>> {
-  
+
   SortedKeyValueIterator<Key,Value> inner;
-  
+
   public IteratorAdapter(SortedKeyValueIterator<Key,Value> inner) {
     this.inner = inner;
   }
-  
+
   @Override
   public boolean hasNext() {
     return inner.hasTop();
   }
-  
+
   @Override
   public Entry<Key,Value> next() {
     try {
@@ -49,7 +49,7 @@ public class IteratorAdapter implements Iterator<Entry<Key,Value>> {
       throw new NoSuchElementException();
     }
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
index c55c378..f171889 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
@@ -110,7 +110,7 @@ public class MockAccumulo {
 
   public void createTable(String username, String tableName, TimeType timeType, Map<String,String> properties) {
     String namespace = Tables.qualify(tableName).getFirst();
-    HashMap<String, String> props = new HashMap<>(properties);
+    HashMap<String,String> props = new HashMap<>(properties);
 
     if (!namespaceExists(namespace)) {
       return;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
index 6f321ff..bb9f2c8 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
@@ -35,14 +35,14 @@ import org.apache.accumulo.core.security.ColumnVisibility;
  * <li>There is no waiting for memory to fill before flushing</li>
  * <li>Only one thread is used for writing</li>
  * </ol>
- * 
+ *
  * Otherwise, it behaves as expected.
  */
 public class MockBatchDeleter extends MockBatchScanner implements BatchDeleter {
-  
+
   private final MockAccumulo acc;
   private final String tableName;
-  
+
   /**
    * Create a {@link BatchDeleter} for the specified instance on the specified table where the writer uses the specified {@link Authorizations}.
    */
@@ -51,10 +51,10 @@ public class MockBatchDeleter extends MockBatchScanner implements BatchDeleter {
     this.acc = acc;
     this.tableName = tableName;
   }
-  
+
   @Override
   public void delete() throws MutationsRejectedException, TableNotFoundException {
-    
+
     BatchWriter writer = new MockBatchWriter(acc, tableName);
     try {
       Iterator<Entry<Key,Value>> iter = super.iterator();
@@ -69,5 +69,5 @@ public class MockBatchDeleter extends MockBatchScanner implements BatchDeleter {
       writer.close();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
index 4512006..4034271 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
@@ -33,22 +33,22 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.commons.collections.iterators.IteratorChain;
 
 public class MockBatchScanner extends MockScannerBase implements BatchScanner {
-  
+
   List<Range> ranges = null;
-  
+
   public MockBatchScanner(MockTable mockTable, Authorizations authorizations) {
     super(mockTable, authorizations);
   }
-  
+
   @Override
   public void setRanges(Collection<Range> ranges) {
     if (ranges == null || ranges.size() == 0) {
       throw new IllegalArgumentException("ranges must be non null and contain at least 1 range");
     }
-    
+
     this.ranges = new ArrayList<Range>(ranges);
   }
-  
+
   @SuppressWarnings("unchecked")
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
@@ -69,7 +69,7 @@ public class MockBatchScanner extends MockScannerBase implements BatchScanner {
     }
     return chain;
   }
-  
+
   @Override
   public void close() {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
index f2c5c85..163587f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
@@ -23,21 +23,21 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.data.Mutation;
 
 public class MockBatchWriter implements BatchWriter {
-  
+
   final String tablename;
   final MockAccumulo acu;
-  
+
   MockBatchWriter(MockAccumulo acu, String tablename) {
     this.acu = acu;
     this.tablename = tablename;
   }
-  
+
   @Override
   public void addMutation(Mutation m) throws MutationsRejectedException {
     checkArgument(m != null, "m is null");
     acu.addMutation(tablename, m);
   }
-  
+
   @Override
   public void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException {
     checkArgument(iterable != null, "iterable is null");
@@ -45,11 +45,11 @@ public class MockBatchWriter implements BatchWriter {
       acu.addMutation(tablename, m);
     }
   }
-  
+
   @Override
   public void flush() throws MutationsRejectedException {}
-  
+
   @Override
   public void close() throws MutationsRejectedException {}
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
index ce262a2..8c57c5e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
@@ -24,15 +24,15 @@ import org.apache.accumulo.core.conf.Property;
 
 class MockConfiguration extends AccumuloConfiguration {
   Map<String,String> map;
-  
+
   MockConfiguration(Map<String,String> settings) {
     map = settings;
   }
-  
+
   public void put(String k, String v) {
     map.put(k, v);
   }
-  
+
   @Override
   public String get(Property property) {
     return map.get(property.getKey());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
index 8613602..4d32093 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
@@ -31,9 +31,9 @@ import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.Scanner;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.admin.InstanceOperations;
+import org.apache.accumulo.core.client.admin.NamespaceOperations;
 import org.apache.accumulo.core.client.admin.ReplicationOperations;
 import org.apache.accumulo.core.client.admin.SecurityOperations;
-import org.apache.accumulo.core.client.admin.NamespaceOperations;
 import org.apache.accumulo.core.client.admin.TableOperations;
 import org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode;
 import org.apache.accumulo.core.client.security.tokens.NullToken;
@@ -41,15 +41,15 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.Credentials;
 
 public class MockConnector extends Connector {
-  
+
   String username;
   private final MockAccumulo acu;
   private final Instance instance;
-  
+
   MockConnector(String username, MockInstance instance) throws AccumuloSecurityException {
     this(new Credentials(username, new NullToken()), new MockAccumulo(MockInstance.getDefaultFileSystem()), instance);
   }
-  
+
   MockConnector(Credentials credentials, MockAccumulo acu, MockInstance instance) throws AccumuloSecurityException {
     if (credentials.getToken().isDestroyed())
       throw new AccumuloSecurityException(credentials.getPrincipal(), SecurityErrorCode.TOKEN_EXPIRED);
@@ -57,14 +57,14 @@ public class MockConnector extends Connector {
     this.acu = acu;
     this.instance = instance;
   }
-  
+
   @Override
   public BatchScanner createBatchScanner(String tableName, Authorizations authorizations, int numQueryThreads) throws TableNotFoundException {
     if (acu.tables.get(tableName) == null)
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return acu.createBatchScanner(tableName, authorizations);
   }
-  
+
   @Deprecated
   @Override
   public BatchDeleter createBatchDeleter(String tableName, Authorizations authorizations, int numQueryThreads, long maxMemory, long maxLatency,
@@ -73,14 +73,14 @@ public class MockConnector extends Connector {
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return new MockBatchDeleter(acu, tableName, authorizations);
   }
-  
+
   @Override
   public BatchDeleter createBatchDeleter(String tableName, Authorizations authorizations, int numQueryThreads, BatchWriterConfig config)
       throws TableNotFoundException {
     return createBatchDeleter(tableName, authorizations, numQueryThreads, config.getMaxMemory(), config.getMaxLatency(TimeUnit.MILLISECONDS),
         config.getMaxWriteThreads());
   }
-  
+
   @Deprecated
   @Override
   public BatchWriter createBatchWriter(String tableName, long maxMemory, long maxLatency, int maxWriteThreads) throws TableNotFoundException {
@@ -88,23 +88,23 @@ public class MockConnector extends Connector {
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return new MockBatchWriter(acu, tableName);
   }
-  
+
   @Override
   public BatchWriter createBatchWriter(String tableName, BatchWriterConfig config) throws TableNotFoundException {
     return createBatchWriter(tableName, config.getMaxMemory(), config.getMaxLatency(TimeUnit.MILLISECONDS), config.getMaxWriteThreads());
   }
-  
+
   @Deprecated
   @Override
   public MultiTableBatchWriter createMultiTableBatchWriter(long maxMemory, long maxLatency, int maxWriteThreads) {
     return new MockMultiTableBatchWriter(acu);
   }
-  
+
   @Override
   public MultiTableBatchWriter createMultiTableBatchWriter(BatchWriterConfig config) {
     return createMultiTableBatchWriter(config.getMaxMemory(), config.getMaxLatency(TimeUnit.MILLISECONDS), config.getMaxWriteThreads());
   }
-  
+
   @Override
   public Scanner createScanner(String tableName, Authorizations authorizations) throws TableNotFoundException {
     MockTable table = acu.tables.get(tableName);
@@ -112,27 +112,27 @@ public class MockConnector extends Connector {
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return new MockScanner(table, authorizations);
   }
-  
+
   @Override
   public Instance getInstance() {
     return instance;
   }
-  
+
   @Override
   public String whoami() {
     return username;
   }
-  
+
   @Override
   public TableOperations tableOperations() {
     return new MockTableOperations(acu, username);
   }
-  
+
   @Override
   public SecurityOperations securityOperations() {
     return new MockSecurityOperations(acu);
   }
-  
+
   @Override
   public InstanceOperations instanceOperations() {
     return new MockInstanceOperations(acu);
@@ -142,7 +142,7 @@ public class MockConnector extends Connector {
   public NamespaceOperations namespaceOperations() {
     return new MockNamespaceOperations(acu, username);
   }
-  
+
   @Override
   public ConditionalWriter createConditionalWriter(String tableName, ConditionalWriterConfig config) throws TableNotFoundException {
     // TODO add implementation
@@ -154,5 +154,5 @@ public class MockConnector extends Connector {
     // TODO add implementation
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
index 9b07d49..67435d2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
@@ -42,11 +42,11 @@ import org.apache.hadoop.io.Text;
  * Mock Accumulo provides an in memory implementation of the Accumulo client API. It is possible that the behavior of this implementation may differ subtly from
  * the behavior of Accumulo. This could result in unit tests that pass on Mock Accumulo and fail on Accumulo or visa-versa. Documenting the differences would be
  * difficult and is not done.
- * 
+ *
  * <p>
  * An alternative to Mock Accumulo called MiniAccumuloCluster was introduced in Accumulo 1.5. MiniAccumuloCluster spins up actual Accumulo server processes, can
  * be used for unit testing, and its behavior should match Accumulo. The drawback of MiniAccumuloCluster is that it starts more slowly than Mock Accumulo.
- * 
+ *
  */
 
 public class MockInstance implements Instance {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
index 87359bc..48122b7 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
@@ -71,7 +71,7 @@ class MockInstanceOperations implements InstanceOperations {
     try {
       AccumuloVFSClassLoader.loadClass(className, Class.forName(asTypeName));
     } catch (ClassNotFoundException e) {
-      log.warn("Could not find class named '"+className+"' in testClassLoad.", e);
+      log.warn("Could not find class named '" + className + "' in testClassLoad.", e);
       return false;
     }
     return true;
@@ -88,6 +88,5 @@ class MockInstanceOperations implements InstanceOperations {
   }
 
   @Override
-  public void waitForBalance() throws AccumuloException {
-  }
+  public void waitForBalance() throws AccumuloException {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
index b4a7068..9cc3dfb 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
@@ -29,12 +29,12 @@ import org.apache.accumulo.core.client.TableNotFoundException;
 public class MockMultiTableBatchWriter implements MultiTableBatchWriter {
   MockAccumulo acu = null;
   Map<String,MockBatchWriter> bws = null;
-  
+
   public MockMultiTableBatchWriter(MockAccumulo acu) {
     this.acu = acu;
     bws = new HashMap<String,MockBatchWriter>();
   }
-  
+
   @Override
   public BatchWriter getBatchWriter(String table) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     if (!bws.containsKey(table)) {
@@ -42,13 +42,13 @@ public class MockMultiTableBatchWriter implements MultiTableBatchWriter {
     }
     return bws.get(table);
   }
-  
+
   @Override
   public void flush() throws MutationsRejectedException {}
-  
+
   @Override
   public void close() throws MutationsRejectedException {}
-  
+
   @Override
   public boolean isClosed() {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
index 7e7eecb..ac581ab 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
@@ -36,7 +36,7 @@ import org.apache.log4j.Logger;
 class MockNamespaceOperations extends NamespaceOperationsHelper {
 
   private static final Logger log = Logger.getLogger(MockNamespaceOperations.class);
-  
+
   final private MockAccumulo acu;
   final private String username;
 
@@ -125,7 +125,7 @@ class MockNamespaceOperations extends NamespaceOperationsHelper {
     try {
       AccumuloVFSClassLoader.loadClass(className, Class.forName(asTypeName));
     } catch (ClassNotFoundException e) {
-      log.warn("Could not load class '"+className+"' with type name '"+asTypeName+"' in testClassLoad()", e);
+      log.warn("Could not load class '" + className + "' with type name '" + asTypeName + "' in testClassLoad()", e);
       return false;
     }
     return true;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
index e7c0ee0..a9b6fd5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
@@ -31,14 +31,14 @@ import org.apache.accumulo.core.iterators.SortedMapIterator;
 import org.apache.accumulo.core.security.Authorizations;
 
 public class MockScanner extends MockScannerBase implements Scanner {
-  
+
   int batchSize = 0;
   Range range = new Range();
-  
+
   MockScanner(MockTable table, Authorizations auths) {
     super(table, auths);
   }
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {
@@ -47,7 +47,7 @@ public class MockScanner extends MockScannerBase implements Scanner {
     else
       setTimeout(timeOut, TimeUnit.SECONDS);
   }
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
@@ -56,47 +56,47 @@ public class MockScanner extends MockScannerBase implements Scanner {
       return Integer.MAX_VALUE;
     return (int) timeout;
   }
-  
+
   @Override
   public void setRange(Range range) {
     this.range = range;
   }
-  
+
   @Override
   public Range getRange() {
     return this.range;
   }
-  
+
   @Override
   public void setBatchSize(int size) {
     this.batchSize = size;
   }
-  
+
   @Override
   public int getBatchSize() {
     return this.batchSize;
   }
-  
+
   @Override
   public void enableIsolation() {}
-  
+
   @Override
   public void disableIsolation() {}
-  
+
   static class RangeFilter extends Filter {
     Range range;
-    
+
     RangeFilter(SortedKeyValueIterator<Key,Value> i, Range range) {
       setSource(i);
       this.range = range;
     }
-    
+
     @Override
     public boolean accept(Key k, Value v) {
       return range.contains(k);
     }
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     SortedKeyValueIterator<Key,Value> i = new SortedMapIterator(table.table);
@@ -107,7 +107,7 @@ public class MockScanner extends MockScannerBase implements Scanner {
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
   }
 
   @Override
@@ -117,7 +117,7 @@ public class MockScanner extends MockScannerBase implements Scanner {
 
   @Override
   public void setReadaheadThreshold(long batches) {
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
index 72cb863..d88c30a 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
@@ -44,15 +44,15 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.commons.lang.NotImplementedException;
 
 public class MockScannerBase extends ScannerOptions implements ScannerBase {
-  
+
   protected final MockTable table;
   protected final Authorizations auths;
-  
+
   MockScannerBase(MockTable mockTable, Authorizations authorizations) {
     this.table = mockTable;
     this.auths = authorizations;
   }
-  
+
   static HashSet<ByteSequence> createColumnBSS(Collection<Column> columns) {
     HashSet<ByteSequence> columnSet = new HashSet<ByteSequence>();
     for (Column c : columns) {
@@ -60,35 +60,35 @@ public class MockScannerBase extends ScannerOptions implements ScannerBase {
     }
     return columnSet;
   }
-  
+
   static class MockIteratorEnvironment implements IteratorEnvironment {
     @Override
     public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public AccumuloConfiguration getConfig() {
       return AccumuloConfiguration.getDefaultConfiguration();
     }
-    
+
     @Override
     public IteratorScope getIteratorScope() {
       return IteratorScope.scan;
     }
-    
+
     @Override
     public boolean isFullMajorCompaction() {
       return false;
     }
-    
+
     private ArrayList<SortedKeyValueIterator<Key,Value>> topLevelIterators = new ArrayList<SortedKeyValueIterator<Key,Value>>();
-    
+
     @Override
     public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
       topLevelIterators.add(iter);
     }
-    
+
     SortedKeyValueIterator<Key,Value> getTopLevelIterator(SortedKeyValueIterator<Key,Value> iter) {
       if (topLevelIterators.isEmpty())
         return iter;
@@ -97,7 +97,7 @@ public class MockScannerBase extends ScannerOptions implements ScannerBase {
       return new MultiIterator(allIters, false);
     }
   }
-  
+
   public SortedKeyValueIterator<Key,Value> createFilter(SortedKeyValueIterator<Key,Value> inner) throws IOException {
     byte[] defaultLabels = {};
     inner = new ColumnFamilySkippingIterator(new DeletingIterator(inner, false));
@@ -109,7 +109,7 @@ public class MockScannerBase extends ScannerOptions implements ScannerBase {
         serverSideIteratorList, serverSideIteratorOptions, iterEnv, false));
     return result;
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
index ee9244b..2244d20 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
@@ -41,30 +41,30 @@ import org.apache.accumulo.core.security.TablePermission;
 import org.apache.hadoop.io.Text;
 
 public class MockTable {
-  
+
   static class MockMemKey extends Key {
     private int count;
-    
+
     MockMemKey(Key key, int count) {
       super(key);
       this.count = count;
     }
-    
+
     @Override
     public int hashCode() {
       return super.hashCode() + count;
     }
-    
+
     @Override
     public boolean equals(Object other) {
       return (other instanceof MockMemKey) && super.equals((MockMemKey) other) && count == ((MockMemKey) other).count;
     }
-    
+
     @Override
     public String toString() {
       return super.toString() + " count=" + count;
     }
-    
+
     @Override
     public int compareTo(Key o) {
       int compare = super.compareTo(o);
@@ -82,7 +82,7 @@ public class MockTable {
       return 0;
     }
   };
-  
+
   final SortedMap<Key,Value> table = new ConcurrentSkipListMap<Key,Value>();
   int mutationCount = 0;
   final Map<String,String> settings;
@@ -93,7 +93,7 @@ public class MockTable {
   private MockNamespace namespace;
   private String namespaceName;
   private String tableId;
-  
+
   MockTable(boolean limitVersion, TimeType timeType, String tableId) {
     this.timeType = timeType;
     this.tableId = tableId;
@@ -160,27 +160,27 @@ public class MockTable {
           key.setTimestamp(mutationCount);
         else
           key.setTimestamp(now);
-      
+
       table.put(new MockMemKey(key, mutationCount), new Value(u.getValue()));
     }
   }
-  
+
   public void addSplits(SortedSet<Text> partitionKeys) {
     splits.addAll(partitionKeys);
   }
-  
+
   public Collection<Text> getSplits() {
     return splits;
   }
-  
+
   public void setLocalityGroups(Map<String,Set<Text>> groups) {
     localityGroups = groups;
   }
-  
+
   public Map<String,Set<Text>> getLocalityGroups() {
     return localityGroups;
   }
-  
+
   public void merge(Text start, Text end) {
     boolean reAdd = false;
     if (splits.contains(start))
@@ -189,19 +189,19 @@ public class MockTable {
     if (reAdd)
       splits.add(start);
   }
-  
+
   public void setNamespaceName(String n) {
     this.namespaceName = n;
   }
-  
+
   public void setNamespace(MockNamespace n) {
     this.namespace = n;
   }
-  
+
   public String getNamespaceName() {
     return this.namespaceName;
   }
-  
+
   public MockNamespace getNamespace() {
     return this.namespace;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
index b39791d..efc896e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
@@ -27,7 +27,7 @@ public class MockUser {
   final String name;
   AuthenticationToken token;
   Authorizations authorizations;
-  
+
   MockUser(String principal, AuthenticationToken token, Authorizations auths) {
     this.name = principal;
     this.token = token.clone();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
index 900ae5a..bdcc652 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
@@ -28,10 +28,15 @@ public interface ReplicaSystem {
 
   /**
    * Replicate the given status to the target peer
-   * @param p Path to the resource we're reading from
-   * @param status Information to replicate
-   * @param target The peer
-   * @param helper Instance of ReplicaSystemHelper
+   *
+   * @param p
+   *          Path to the resource we're reading from
+   * @param status
+   *          Information to replicate
+   * @param target
+   *          The peer
+   * @param helper
+   *          Instance of ReplicaSystemHelper
    * @return A new Status for the progress that was made
    */
   public Status replicate(Path p, Status status, ReplicationTarget target, ReplicaSystemHelper helper);
@@ -39,8 +44,7 @@ public interface ReplicaSystem {
   /**
    * Configure the implementation with necessary information from the system configuration
    * <p>
-   * For example, we only need one implementation for Accumulo, but, for each peer,
-   * we have a ZK quorum and instance name
+   * For example, we only need one implementation for Accumulo, but, for each peer, we have a ZK quorum and instance name
    */
   public void configure(String configuration);
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
index e721278..d76b3d8 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
@@ -22,7 +22,7 @@ import org.slf4j.LoggerFactory;
 import com.google.common.base.Preconditions;
 
 /**
- * 
+ *
  */
 public class ReplicaSystemFactory {
   private static final Logger log = LoggerFactory.getLogger(ReplicaSystemFactory.class);
@@ -64,7 +64,7 @@ public class ReplicaSystemFactory {
 
   /**
    * Generate the configuration value for a {@link ReplicaSystem} in the instance properties
-   * 
+   *
    * @param system
    *          The desired ReplicaSystem to use
    * @param configuration

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java b/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
index 30c60d5..b4027b2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.core.client.security;
 
 /**
- * 
+ *
  */
 public enum SecurityErrorCode {
   DEFAULT_SECURITY_ERROR,

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
index 99cc721..7836ea5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
@@ -33,15 +33,15 @@ import javax.security.auth.Destroyable;
 import org.apache.hadoop.io.Writable;
 
 /**
- * 
+ *
  * @since 1.5.0
  */
 public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
-  
+
   /**
    * A utility class to serialize/deserialize {@link AuthenticationToken} objects.<br/>
    * Unfortunately, these methods are provided in an inner-class, to avoid breaking the interface API.
-   * 
+   *
    * @since 1.6.0
    */
   public static final class AuthenticationTokenSerializer {
@@ -49,7 +49,7 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
      * A convenience method to create tokens from serialized bytes, created by {@link #serialize(AuthenticationToken)}
      * <p>
      * The specified tokenType will be instantiated, and used to deserialize the decoded bytes. The resulting object will then be returned to the caller.
-     * 
+     *
      * @param tokenType
      *          the token class to use to deserialize the bytes
      * @param tokenBytes
@@ -78,10 +78,10 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       }
       return type;
     }
-    
+
     /**
      * An alternate version of {@link #deserialize(Class, byte[])} that accepts a token class name rather than a token class.
-     * 
+     *
      * @param tokenClassName
      *          the fully-qualified class name to be returned
      * @see #serialize(AuthenticationToken)
@@ -97,12 +97,12 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       }
       return deserialize(tokenType, tokenBytes);
     }
-    
+
     /**
      * A convenience method to serialize tokens.
      * <p>
      * The provided {@link AuthenticationToken} will be serialized to bytes by its own implementation and returned to the caller.
-     * 
+     *
      * @param token
      *          the token to serialize
      * @return a serialized representation of the provided {@link AuthenticationToken}
@@ -125,17 +125,17 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       return bytes;
     }
   }
-  
+
   class Properties implements Destroyable, Map<String,char[]> {
-    
+
     private boolean destroyed = false;
     private HashMap<String,char[]> map = new HashMap<String,char[]>();
-    
+
     private void checkDestroyed() {
       if (destroyed)
         throw new IllegalStateException();
     }
-    
+
     public char[] put(String key, CharSequence value) {
       checkDestroyed();
       char[] toPut = new char[value.length()];
@@ -143,14 +143,14 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
         toPut[i] = value.charAt(i);
       return map.put(key, toPut);
     }
-    
+
     public void putAllStrings(Map<String,? extends CharSequence> map) {
       checkDestroyed();
       for (Map.Entry<String,? extends CharSequence> entry : map.entrySet()) {
         put(entry.getKey(), entry.getValue());
       }
     }
-    
+
     @Override
     public void destroy() throws DestroyFailedException {
       for (String key : this.keySet()) {
@@ -160,133 +160,133 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       this.clear();
       destroyed = true;
     }
-    
+
     @Override
     public boolean isDestroyed() {
       return destroyed;
     }
-    
+
     @Override
     public int size() {
       checkDestroyed();
       return map.size();
     }
-    
+
     @Override
     public boolean isEmpty() {
       checkDestroyed();
       return map.isEmpty();
     }
-    
+
     @Override
     public boolean containsKey(Object key) {
       checkDestroyed();
       return map.containsKey(key);
     }
-    
+
     @Override
     public boolean containsValue(Object value) {
       checkDestroyed();
       return map.containsValue(value);
     }
-    
+
     @Override
     public char[] get(Object key) {
       checkDestroyed();
       return map.get(key);
     }
-    
+
     @Override
     public char[] put(String key, char[] value) {
       checkDestroyed();
       return map.put(key, value);
     }
-    
+
     @Override
     public char[] remove(Object key) {
       checkDestroyed();
       return map.remove(key);
     }
-    
+
     @Override
     public void putAll(Map<? extends String,? extends char[]> m) {
       checkDestroyed();
       map.putAll(m);
     }
-    
+
     @Override
     public void clear() {
       checkDestroyed();
       map.clear();
     }
-    
+
     @Override
     public Set<String> keySet() {
       checkDestroyed();
       return map.keySet();
     }
-    
+
     @Override
     public Collection<char[]> values() {
       checkDestroyed();
       return map.values();
     }
-    
+
     @Override
     public Set<Map.Entry<String,char[]>> entrySet() {
       checkDestroyed();
       return map.entrySet();
     }
   }
-  
+
   static class TokenProperty implements Comparable<TokenProperty> {
     private String key, description;
     private boolean masked;
-    
+
     public TokenProperty(String name, String description, boolean mask) {
       this.key = name;
       this.description = description;
       this.masked = mask;
     }
-    
+
     @Override
     public String toString() {
       return this.key + " - " + description;
     }
-    
+
     public String getKey() {
       return this.key;
     }
-    
+
     public String getDescription() {
       return this.description;
     }
-    
+
     public boolean getMask() {
       return this.masked;
     }
-    
+
     @Override
     public int hashCode() {
       return key.hashCode();
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof TokenProperty)
         return ((TokenProperty) o).key.equals(key);
       return false;
     }
-    
+
     @Override
     public int compareTo(TokenProperty o) {
       return key.compareTo(o.key);
     }
   }
-  
+
   public void init(Properties properties);
-  
+
   public Set<TokenProperty> getProperties();
-  
+
   public AuthenticationToken clone();
 }


[17/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/time/SimpleTimer.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/time/SimpleTimer.java b/server/base/src/main/java/org/apache/accumulo/server/util/time/SimpleTimer.java
index bc1d652..556e6b9 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/time/SimpleTimer.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/time/SimpleTimer.java
@@ -16,24 +16,26 @@
  */
 package org.apache.accumulo.server.util.time;
 
-import com.google.common.annotations.VisibleForTesting;
-import com.google.common.util.concurrent.ThreadFactoryBuilder;
 import java.util.concurrent.Executors;
 import java.util.concurrent.ScheduledExecutorService;
 import java.util.concurrent.ScheduledFuture;
 import java.util.concurrent.TimeUnit;
+
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.core.conf.Property;
 import org.apache.log4j.Logger;
 
+import com.google.common.annotations.VisibleForTesting;
+import com.google.common.util.concurrent.ThreadFactoryBuilder;
+
 /**
  * Generic singleton timer. Don't use this if you are going to do anything that will take very long. Please use it to reduce the number of threads dedicated to
  * simple events.
- * 
+ *
  */
 public class SimpleTimer {
   private static final Logger log = Logger.getLogger(SimpleTimer.class);
-  
+
   private static class ExceptionHandler implements Thread.UncaughtExceptionHandler {
     public void uncaughtException(Thread t, Throwable e) {
       log.warn("SimpleTimer task failed", e);
@@ -43,24 +45,25 @@ public class SimpleTimer {
   private static int instanceThreadPoolSize = -1;
   private static SimpleTimer instance;
   private ScheduledExecutorService executor;
-  
+
   private static final int DEFAULT_THREAD_POOL_SIZE = 1;
+
   /**
    * Gets the timer instance.
    *
-   * @deprecated Use {@link #getInstance(AccumuloConfiguration)} instead to
-   * get the configured number of threads.
+   * @deprecated Use {@link #getInstance(AccumuloConfiguration)} instead to get the configured number of threads.
    */
   @Deprecated
   public static synchronized SimpleTimer getInstance() {
     return getInstance(null);
   }
+
   /**
-   * Gets the timer instance. If an instance has already been created, it will
-   * have the number of threads supplied when it was constructed, and the size
+   * Gets the timer instance. If an instance has already been created, it will have the number of threads supplied when it was constructed, and the size
    * provided here is ignored.
    *
-   * @param threadPoolSize number of threads
+   * @param threadPoolSize
+   *          number of threads
    */
   public static synchronized SimpleTimer getInstance(int threadPoolSize) {
     if (instance == null) {
@@ -68,20 +71,18 @@ public class SimpleTimer {
       SimpleTimer.instanceThreadPoolSize = threadPoolSize;
     } else {
       if (SimpleTimer.instanceThreadPoolSize != threadPoolSize) {
-        log.warn("Asked to create SimpleTimer with thread pool size " +
-                 threadPoolSize + ", existing instance has " +
-                 instanceThreadPoolSize);
+        log.warn("Asked to create SimpleTimer with thread pool size " + threadPoolSize + ", existing instance has " + instanceThreadPoolSize);
       }
     }
     return instance;
   }
+
   /**
-   * Gets the timer instance. If an instance has already been created, it will
-   * have the number of threads supplied when it was constructed, and the size
-   * provided by the configuration here is ignored. If a null configuration is
-   * supplied, the number of threads defaults to 1.
+   * Gets the timer instance. If an instance has already been created, it will have the number of threads supplied when it was constructed, and the size
+   * provided by the configuration here is ignored. If a null configuration is supplied, the number of threads defaults to 1.
    *
-   * @param conf configuration from which to get the number of threads
+   * @param conf
+   *          configuration from which to get the number of threads
    * @see Property#GENERAL_SIMPLETIMER_THREADPOOL_SIZE
    */
   public static synchronized SimpleTimer getInstance(AccumuloConfiguration conf) {
@@ -103,30 +104,34 @@ public class SimpleTimer {
   static int getInstanceThreadPoolSize() {
     return instanceThreadPoolSize;
   }
-  
+
   private SimpleTimer(int threadPoolSize) {
     executor = Executors.newScheduledThreadPool(threadPoolSize, new ThreadFactoryBuilder().setNameFormat("SimpleTimer-%d").setDaemon(true)
-      .setUncaughtExceptionHandler(new ExceptionHandler()).build());
+        .setUncaughtExceptionHandler(new ExceptionHandler()).build());
   }
-  
+
   /**
    * Schedules a task to run in the future.
    *
-   * @param task task to run
-   * @param delay number of milliseconds to wait before execution
+   * @param task
+   *          task to run
+   * @param delay
+   *          number of milliseconds to wait before execution
    * @return future for scheduled task
    */
   public ScheduledFuture<?> schedule(Runnable task, long delay) {
     return executor.schedule(task, delay, TimeUnit.MILLISECONDS);
   }
-  
+
   /**
-   * Schedules a task to run in the future with a fixed delay between repeated
-   * executions.
+   * Schedules a task to run in the future with a fixed delay between repeated executions.
    *
-   * @param task task to run
-   * @param delay number of milliseconds to wait before first execution
-   * @param period number of milliseconds to wait between executions
+   * @param task
+   *          task to run
+   * @param delay
+   *          number of milliseconds to wait before first execution
+   * @param period
+   *          number of milliseconds to wait between executions
    * @return future for scheduled task
    */
   public ScheduledFuture<?> schedule(Runnable task, long delay, long period) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/time/SystemTime.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/time/SystemTime.java b/server/base/src/main/java/org/apache/accumulo/server/util/time/SystemTime.java
index c421f5f..b8fe371 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/time/SystemTime.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/time/SystemTime.java
@@ -18,13 +18,13 @@ package org.apache.accumulo.server.util.time;
 
 /**
  * The most obvious implementation of ProvidesTime.
- * 
+ *
  */
 public class SystemTime implements ProvidesTime {
-  
+
   @Override
   public long currentTime() {
     return System.currentTimeMillis();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/zookeeper/DistributedWorkQueue.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/DistributedWorkQueue.java b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/DistributedWorkQueue.java
index 1452aa7..412a97a 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/DistributedWorkQueue.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/DistributedWorkQueue.java
@@ -38,15 +38,15 @@ import org.apache.zookeeper.Watcher;
 
 /**
  * Provides a way to push work out to tablet servers via zookeeper and wait for that work to be done. Any tablet server can pick up a work item and process it.
- * 
+ *
  * Worker processes watch a zookeeper node for tasks to be performed. After getting an exclusive lock on the node, the worker will perform the task.
  */
 public class DistributedWorkQueue {
-  
+
   private static final String LOCKS_NODE = "locks";
 
   private static final Logger log = Logger.getLogger(DistributedWorkQueue.class);
-  
+
   private ThreadPoolExecutor threadPool;
   private ZooReaderWriter zoo = ZooReaderWriter.getInstance();
   private String path;
@@ -58,15 +58,15 @@ public class DistributedWorkQueue {
   private void lookForWork(final Processor processor, List<String> children) {
     if (children.size() == 0)
       return;
-    
+
     if (numTask.get() >= threadPool.getCorePoolSize())
       return;
-    
+
     Random random = new Random();
     Collections.shuffle(children, random);
     try {
       for (final String child : children) {
-        
+
         if (child.equals(LOCKS_NODE))
           continue;
 
@@ -82,7 +82,7 @@ public class DistributedWorkQueue {
         }
 
         final String childPath = path + "/" + child;
-        
+
         // check to see if another node processed it already
         if (!zoo.exists(childPath)) {
           zoo.recursiveDelete(lockPath, NodeMissingPolicy.SKIP);
@@ -94,9 +94,9 @@ public class DistributedWorkQueue {
           zoo.recursiveDelete(lockPath, NodeMissingPolicy.SKIP);
           break;
         }
-        
+
         log.debug("got lock for " + child);
-        
+
         Runnable task = new Runnable() {
 
           @Override
@@ -104,18 +104,18 @@ public class DistributedWorkQueue {
             try {
               try {
                 processor.newProcessor().process(child, zoo.getData(childPath, null));
-                
+
                 // if the task fails, then its entry in the Q is not deleted... so it will be retried
                 try {
                   zoo.recursiveDelete(childPath, NodeMissingPolicy.SKIP);
                 } catch (Exception e) {
                   log.error("Error received when trying to delete entry in zookeeper " + childPath, e);
                 }
-                
+
               } catch (Exception e) {
                 log.warn("Failed to process work " + child, e);
               }
-              
+
               try {
                 zoo.recursiveDelete(lockPath, NodeMissingPolicy.SKIP);
               } catch (Exception e) {
@@ -125,7 +125,7 @@ public class DistributedWorkQueue {
             } finally {
               numTask.decrementAndGet();
             }
-            
+
             try {
               // its important that this is called after numTask is decremented
               lookForWork(processor, zoo.getChildren(path));
@@ -136,7 +136,7 @@ public class DistributedWorkQueue {
             }
           }
         };
-        
+
         numTask.incrementAndGet();
         threadPool.execute(task);
 
@@ -151,10 +151,10 @@ public class DistributedWorkQueue {
 
     void process(String workID, byte[] data);
   }
-  
+
   public DistributedWorkQueue(String path, AccumuloConfiguration config) {
     // Preserve the old delay and period
-    this(path, config, new Random().nextInt(60*1000), 60*1000);
+    this(path, config, new Random().nextInt(60 * 1000), 60 * 1000);
   }
 
   public DistributedWorkQueue(String path, AccumuloConfiguration config, long timerInitialDelay, long timerPeriod) {
@@ -163,9 +163,9 @@ public class DistributedWorkQueue {
     this.timerInitialDelay = timerInitialDelay;
     this.timerPeriod = timerPeriod;
   }
-  
+
   public void startProcessing(final Processor processor, ThreadPoolExecutor executorService) throws KeeperException, InterruptedException {
-    
+
     threadPool = executorService;
 
     zoo.mkdirs(path);
@@ -193,13 +193,13 @@ public class DistributedWorkQueue {
           case None:
             log.info("Got unexpected zookeeper event: " + event.getType() + " for " + path);
             break;
-        
+
         }
       }
     });
-    
+
     lookForWork(processor, children);
-    
+
     // Add a little jitter to avoid all the tservers slamming zookeeper at once
     SimpleTimer.getInstance(config).schedule(new Runnable() {
       @Override
@@ -222,7 +222,7 @@ public class DistributedWorkQueue {
   public void addWork(String workId, String data) throws KeeperException, InterruptedException {
     addWork(workId, data.getBytes(UTF_8));
   }
-  
+
   public void addWork(String workId, byte[] data) throws KeeperException, InterruptedException {
     if (workId.equalsIgnoreCase(LOCKS_NODE))
       throw new IllegalArgumentException("locks is reserved work id");
@@ -230,7 +230,7 @@ public class DistributedWorkQueue {
     zoo.mkdirs(path);
     zoo.putPersistentData(path + "/" + workId, data, NodeExistsPolicy.SKIP);
   }
-  
+
   public List<String> getWorkQueued() throws KeeperException, InterruptedException {
     ArrayList<String> children = new ArrayList<String>(zoo.getChildren(path));
     children.remove(LOCKS_NODE);
@@ -238,9 +238,9 @@ public class DistributedWorkQueue {
   }
 
   public void waitUntilDone(Set<String> workIDs) throws KeeperException, InterruptedException {
-    
+
     final Object condVar = new Object();
-    
+
     Watcher watcher = new Watcher() {
       @Override
       public void process(WatchedEvent event) {
@@ -256,13 +256,13 @@ public class DistributedWorkQueue {
           case None:
             log.info("Got unexpected zookeeper event: " + event.getType() + " for " + path);
             break;
-        
+
         }
       }
     };
-    
+
     List<String> children = zoo.getChildren(path, watcher);
-    
+
     while (!Collections.disjoint(children, workIDs)) {
       synchronized (condVar) {
         condVar.wait(10000);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/zookeeper/TransactionWatcher.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/TransactionWatcher.java b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/TransactionWatcher.java
index 4e0e977..0e1cdfd 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/TransactionWatcher.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/TransactionWatcher.java
@@ -27,17 +27,17 @@ import org.apache.zookeeper.KeeperException;
 
 public class TransactionWatcher extends org.apache.accumulo.fate.zookeeper.TransactionWatcher {
   public static class ZooArbitrator implements Arbitrator {
-    
+
     Instance instance = HdfsZooInstance.getInstance();
     ZooReader rdr = new ZooReader(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
-    
+
     @Override
     public boolean transactionAlive(String type, long tid) throws Exception {
       String path = ZooUtil.getRoot(instance) + "/" + type + "/" + tid;
       rdr.sync(path);
       return rdr.exists(path);
     }
-    
+
     public static void start(String type, long tid) throws KeeperException, InterruptedException {
       Instance instance = HdfsZooInstance.getInstance();
       IZooReaderWriter writer = ZooReaderWriter.getInstance();
@@ -45,13 +45,13 @@ public class TransactionWatcher extends org.apache.accumulo.fate.zookeeper.Trans
       writer.putPersistentData(ZooUtil.getRoot(instance) + "/" + type + "/" + tid, new byte[] {}, NodeExistsPolicy.OVERWRITE);
       writer.putPersistentData(ZooUtil.getRoot(instance) + "/" + type + "/" + tid + "-running", new byte[] {}, NodeExistsPolicy.OVERWRITE);
     }
-    
+
     public static void stop(String type, long tid) throws KeeperException, InterruptedException {
       Instance instance = HdfsZooInstance.getInstance();
       IZooReaderWriter writer = ZooReaderWriter.getInstance();
       writer.recursiveDelete(ZooUtil.getRoot(instance) + "/" + type + "/" + tid, NodeMissingPolicy.SKIP);
     }
-    
+
     public static void cleanup(String type, long tid) throws KeeperException, InterruptedException {
       Instance instance = HdfsZooInstance.getInstance();
       IZooReaderWriter writer = ZooReaderWriter.getInstance();
@@ -66,7 +66,7 @@ public class TransactionWatcher extends org.apache.accumulo.fate.zookeeper.Trans
       return !rdr.exists(path);
     }
   }
-  
+
   public TransactionWatcher() {
     super(new ZooArbitrator());
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooCache.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooCache.java b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooCache.java
index bf34ef6..aca9c82 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooCache.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooCache.java
@@ -24,11 +24,11 @@ public class ZooCache extends org.apache.accumulo.fate.zookeeper.ZooCache {
   public ZooCache() {
     this(null);
   }
-  
+
   public ZooCache(Watcher watcher) {
     super(ZooReaderWriter.getInstance(), watcher);
   }
-  
+
   public ZooCache(AccumuloConfiguration conf, Watcher watcher) {
     super(conf.get(Property.INSTANCE_ZK_HOST), (int) conf.getTimeInMillis(Property.INSTANCE_ZK_TIMEOUT), watcher);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooLock.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooLock.java b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooLock.java
index dce6d38..b67b82b 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooLock.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooLock.java
@@ -19,15 +19,15 @@ package org.apache.accumulo.server.zookeeper;
 import org.apache.zookeeper.KeeperException;
 
 public class ZooLock extends org.apache.accumulo.fate.zookeeper.ZooLock {
-  
+
   public ZooLock(String path) {
     super(new ZooCache(), ZooReaderWriter.getInstance(), path);
   }
-  
+
   public static void deleteLock(String path) throws InterruptedException, KeeperException {
     deleteLock(ZooReaderWriter.getInstance(), path);
   }
-  
+
   public static boolean deleteLock(String path, String lockData) throws InterruptedException, KeeperException {
     return deleteLock(ZooReaderWriter.getInstance(), path, lockData);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooReaderWriter.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooReaderWriter.java b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooReaderWriter.java
index 34c9070..8a3383c 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooReaderWriter.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/zookeeper/ZooReaderWriter.java
@@ -26,11 +26,11 @@ public class ZooReaderWriter extends org.apache.accumulo.fate.zookeeper.ZooReade
   private static final String SCHEME = "digest";
   private static final String USER = "accumulo";
   private static ZooReaderWriter instance = null;
-  
+
   public ZooReaderWriter(String string, int timeInMillis, String secret) {
     super(string, timeInMillis, SCHEME, (USER + ":" + secret).getBytes(UTF_8));
   }
-  
+
   public static synchronized ZooReaderWriter getInstance() {
     if (instance == null) {
       AccumuloConfiguration conf = SiteConfiguration.getInstance();
@@ -39,5 +39,5 @@ public class ZooReaderWriter extends org.apache.accumulo.fate.zookeeper.ZooReade
     }
     return instance;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/AccumuloTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/AccumuloTest.java b/server/base/src/test/java/org/apache/accumulo/server/AccumuloTest.java
index 52929ed..19b0a9b 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/AccumuloTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/AccumuloTest.java
@@ -108,6 +108,7 @@ public class AccumuloTest {
       FileUtils.deleteDirectory(confDir);
     }
   }
+
   @Test
   public void testLocateLogConfig_Default() throws Exception {
     File confDir = new File(FileUtils.getTempDirectory(), "AccumuloTest" + System.currentTimeMillis());
@@ -121,6 +122,7 @@ public class AccumuloTest {
       FileUtils.deleteDirectory(confDir);
     }
   }
+
   @Test
   public void testLocateLogConfig_Explicit() throws Exception {
     File confDir = new File(FileUtils.getTempDirectory(), "AccumuloTest" + System.currentTimeMillis());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/ServerConstantsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/ServerConstantsTest.java b/server/base/src/test/java/org/apache/accumulo/server/ServerConstantsTest.java
index 8c8b2f2..40d5eb8 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/ServerConstantsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/ServerConstantsTest.java
@@ -33,7 +33,7 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
 /**
- * 
+ *
  */
 public class ServerConstantsTest {
   @Rule

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/ServerOptsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/ServerOptsTest.java b/server/base/src/test/java/org/apache/accumulo/server/ServerOptsTest.java
index d4420aa..2c6889b 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/ServerOptsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/ServerOptsTest.java
@@ -16,9 +16,10 @@
  */
 package org.apache.accumulo.server;
 
+import static org.junit.Assert.assertEquals;
+
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class ServerOptsTest {
   private ServerOpts opts;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/client/BulkImporterTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/client/BulkImporterTest.java b/server/base/src/test/java/org/apache/accumulo/server/client/BulkImporterTest.java
index d12483c..fad6398 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/client/BulkImporterTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/client/BulkImporterTest.java
@@ -52,7 +52,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 public class BulkImporterTest {
-  
+
   static final SortedSet<KeyExtent> fakeMetaData = new TreeSet<KeyExtent>();
   static final Text tableId = new Text("1");
   static {
@@ -62,49 +62,49 @@ public class BulkImporterTest {
     }
     fakeMetaData.add(new KeyExtent(tableId, null, fakeMetaData.last().getEndRow()));
   }
-  
+
   class MockTabletLocator extends TabletLocator {
     int invalidated = 0;
-    
+
     @Override
     public TabletLocation locateTablet(ClientContext context, Text row, boolean skipRow, boolean retry) throws AccumuloException, AccumuloSecurityException,
         TableNotFoundException {
       return new TabletLocation(fakeMetaData.tailSet(new KeyExtent(tableId, row, null)).first(), "localhost", "1");
     }
-    
+
     @Override
     public <T extends Mutation> void binMutations(ClientContext context, List<T> mutations, Map<String,TabletServerMutations<T>> binnedMutations,
         List<T> failures) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public List<Range> binRanges(ClientContext context, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges) throws AccumuloException,
         AccumuloSecurityException, TableNotFoundException {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public void invalidateCache(KeyExtent failedExtent) {
       invalidated++;
     }
-    
+
     @Override
     public void invalidateCache(Collection<KeyExtent> keySet) {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public void invalidateCache() {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public void invalidateCache(Instance instance, String server) {
       throw new NotImplementedException();
     }
   }
-  
+
   @Test
   public void testFindOverlappingTablets() throws Exception {
     MockTabletLocator locator = new MockTabletLocator();
@@ -145,7 +145,7 @@ public class BulkImporterTest {
     Assert.assertEquals(new KeyExtent(tableId, new Text("dm"), new Text("d")), overlaps.get(2).tablet_extent);
     Assert.assertEquals(new KeyExtent(tableId, new Text("j"), new Text("i")), overlaps.get(3).tablet_extent);
     Assert.assertEquals(new KeyExtent(tableId, null, new Text("l")), overlaps.get(4).tablet_extent);
-    
+
     List<TabletLocation> overlaps2 = BulkImporter.findOverlappingTablets(context, vm, locator, new Path(file), new KeyExtent(tableId, new Text("h"), new Text(
         "b")));
     Assert.assertEquals(3, overlaps2.size());
@@ -154,5 +154,5 @@ public class BulkImporterTest {
     Assert.assertEquals(new KeyExtent(tableId, new Text("j"), new Text("i")), overlaps2.get(2).tablet_extent);
     Assert.assertEquals(locator.invalidated, 1);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/conf/NamespaceConfigurationTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/conf/NamespaceConfigurationTest.java b/server/base/src/test/java/org/apache/accumulo/server/conf/NamespaceConfigurationTest.java
index 628b981..6bd6424 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/conf/NamespaceConfigurationTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/conf/NamespaceConfigurationTest.java
@@ -17,9 +17,6 @@
 package org.apache.accumulo.server.conf;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertNull;
-import static org.junit.Assert.assertTrue;
 import static org.easymock.EasyMock.anyObject;
 import static org.easymock.EasyMock.createMock;
 import static org.easymock.EasyMock.eq;
@@ -27,11 +24,15 @@ import static org.easymock.EasyMock.expect;
 import static org.easymock.EasyMock.expectLastCall;
 import static org.easymock.EasyMock.replay;
 import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertTrue;
 
 import java.util.Collection;
 import java.util.List;
 import java.util.Map;
 import java.util.UUID;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.client.impl.Namespaces;
@@ -86,8 +87,8 @@ public class NamespaceConfigurationTest {
   @Test
   public void testGet_InZK() {
     Property p = Property.INSTANCE_SECRET;
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + p.getKey())).andReturn(
-        "sekrit".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + p.getKey()))
+        .andReturn("sekrit".getBytes(UTF_8));
     replay(zc);
     assertEquals("sekrit", c.get(Property.INSTANCE_SECRET));
   }
@@ -123,10 +124,8 @@ public class NamespaceConfigurationTest {
     children.add("foo");
     children.add("ding");
     expect(zc.getChildren(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF)).andReturn(children);
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + "foo")).andReturn(
-        "bar".getBytes(UTF_8));
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + "ding")).andReturn(
-        "dong".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + "foo")).andReturn("bar".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + "ding")).andReturn("dong".getBytes(UTF_8));
     replay(zc);
     c.getProperties(props, filter);
     assertEquals(2, props.size());
@@ -150,8 +149,8 @@ public class NamespaceConfigurationTest {
   public void testInvalidateCache() {
     // need to do a get so the accessor is created
     Property p = Property.INSTANCE_SECRET;
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + p.getKey())).andReturn(
-        "sekrit".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZNAMESPACES + "/" + NSID + Constants.ZNAMESPACE_CONF + "/" + p.getKey()))
+        .andReturn("sekrit".getBytes(UTF_8));
     zc.clear();
     replay(zc);
     c.get(Property.INSTANCE_SECRET);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/conf/TableConfigurationTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/conf/TableConfigurationTest.java b/server/base/src/test/java/org/apache/accumulo/server/conf/TableConfigurationTest.java
index 0ede571..68ee2b9 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/conf/TableConfigurationTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/conf/TableConfigurationTest.java
@@ -83,8 +83,7 @@ public class TableConfigurationTest {
   @Test
   public void testGet_InZK() {
     Property p = Property.INSTANCE_SECRET;
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + p.getKey())).andReturn(
-        "sekrit".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + p.getKey())).andReturn("sekrit".getBytes(UTF_8));
     replay(zc);
     assertEquals("sekrit", c.get(Property.INSTANCE_SECRET));
   }
@@ -109,10 +108,8 @@ public class TableConfigurationTest {
     children.add("foo");
     children.add("ding");
     expect(zc.getChildren(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF)).andReturn(children);
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + "foo"))
-        .andReturn("bar".getBytes(UTF_8));
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + "ding")).andReturn(
-        "dong".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + "foo")).andReturn("bar".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + "ding")).andReturn("dong".getBytes(UTF_8));
     replay(zc);
     c.getProperties(props, filter);
     assertEquals(2, props.size());
@@ -136,8 +133,7 @@ public class TableConfigurationTest {
   public void testInvalidateCache() {
     // need to do a get so the accessor is created
     Property p = Property.INSTANCE_SECRET;
-    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + p.getKey())).andReturn(
-        "sekrit".getBytes(UTF_8));
+    expect(zc.get(ZooUtil.getRoot(iid) + Constants.ZTABLES + "/" + TID + Constants.ZTABLE_CONF + "/" + p.getKey())).andReturn("sekrit".getBytes(UTF_8));
     zc.clear();
     replay(zc);
     c.get(Property.INSTANCE_SECRET);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/conf/ZooConfigurationFactoryTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/conf/ZooConfigurationFactoryTest.java b/server/base/src/test/java/org/apache/accumulo/server/conf/ZooConfigurationFactoryTest.java
index b54f3e7..b069163 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/conf/ZooConfigurationFactoryTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/conf/ZooConfigurationFactoryTest.java
@@ -16,19 +16,20 @@
  */
 package org.apache.accumulo.server.conf;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.expectLastCall;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertNotNull;
+import static org.junit.Assert.assertSame;
+
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.fate.zookeeper.ZooCache;
 import org.apache.accumulo.fate.zookeeper.ZooCacheFactory;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertSame;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.expectLastCall;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
 
 public class ZooConfigurationFactoryTest {
   private Instance instance;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/constraints/MetadataConstraintsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/constraints/MetadataConstraintsTest.java b/server/base/src/test/java/org/apache/accumulo/server/constraints/MetadataConstraintsTest.java
index d0b2e9e..7b6eec2 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/constraints/MetadataConstraintsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/constraints/MetadataConstraintsTest.java
@@ -28,26 +28,25 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection;
 import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection.DataFileColumnFamily;
 import org.apache.accumulo.fate.zookeeper.TransactionWatcher.Arbitrator;
-import org.apache.accumulo.server.constraints.MetadataConstraints;
 import org.apache.hadoop.io.Text;
 import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 import org.junit.Test;
 
 public class MetadataConstraintsTest {
-  
+
   static class TestMetadataConstraints extends MetadataConstraints {
     @Override
     protected Arbitrator getArbitrator() {
       return new Arbitrator() {
-        
+
         @Override
         public boolean transactionAlive(String type, long tid) throws Exception {
           if (tid == 9)
             throw new RuntimeException("txid 9 reserved for future use");
           return tid == 5 || tid == 7;
         }
-        
+
         @Override
         public boolean transactionComplete(String type, long tid) throws Exception {
           return tid != 5 && tid != 7;
@@ -55,89 +54,89 @@ public class MetadataConstraintsTest {
       };
     }
   }
-  
+
   @Test
   public void testCheck() {
     Logger.getLogger(AccumuloConfiguration.class).setLevel(Level.ERROR);
     Mutation m = new Mutation(new Text("0;foo"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("1foo".getBytes()));
-    
+
     MetadataConstraints mc = new MetadataConstraints();
-    
+
     List<Short> violations = mc.check(null, m);
-    
+
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 3), violations.get(0));
-    
+
     m = new Mutation(new Text("0:foo"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("1poo".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 4), violations.get(0));
-    
+
     m = new Mutation(new Text("0;foo"));
     m.put(new Text("bad_column_name"), new Text(""), new Value("e".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 2), violations.get(0));
-    
+
     m = new Mutation(new Text("!!<"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("1poo".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertNotNull(violations);
     assertEquals(2, violations.size());
     assertEquals(Short.valueOf((short) 4), violations.get(0));
     assertEquals(Short.valueOf((short) 5), violations.get(1));
-    
+
     m = new Mutation(new Text("0;foo"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 6), violations.get(0));
-    
+
     m = new Mutation(new Text("0;foo"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("bar".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertEquals(null, violations);
-    
+
     m = new Mutation(new Text("!0<"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("bar".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertEquals(null, violations);
-    
+
     m = new Mutation(new Text("!1<"));
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.put(m, new Value("bar".getBytes()));
-    
+
     violations = mc.check(null, m);
-    
+
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 4), violations.get(0));
-    
+
   }
-  
+
   @Test
   public void testBulkFileCheck() {
     MetadataConstraints mc = new TestMetadataConstraints();
     Mutation m;
     List<Short> violations;
-    
+
     // inactive txid
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("12345".getBytes()));
@@ -146,7 +145,7 @@ public class MetadataConstraintsTest {
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 8), violations.get(0));
-    
+
     // txid that throws exception
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("9".getBytes()));
@@ -155,14 +154,14 @@ public class MetadataConstraintsTest {
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 8), violations.get(0));
-    
+
     // active txid w/ file
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
     m.put(DataFileColumnFamily.NAME, new Text("/someFile"), new Value("1,1".getBytes()));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
     // active txid w/o file
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
@@ -170,7 +169,7 @@ public class MetadataConstraintsTest {
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 8), violations.get(0));
-    
+
     // two active txids w/ files
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
@@ -181,7 +180,7 @@ public class MetadataConstraintsTest {
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 8), violations.get(0));
-    
+
     // two files w/ one active txid
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
@@ -190,7 +189,7 @@ public class MetadataConstraintsTest {
     m.put(DataFileColumnFamily.NAME, new Text("/someFile2"), new Value("1,1".getBytes()));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
     // two loaded w/ one active txid and one file
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
@@ -200,41 +199,41 @@ public class MetadataConstraintsTest {
     assertNotNull(violations);
     assertEquals(1, violations.size());
     assertEquals(Short.valueOf((short) 8), violations.get(0));
-    
+
     // active txid, mutation that looks like split
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(m, new Value("/t1".getBytes()));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
     // inactive txid, mutation that looks like split
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("12345".getBytes()));
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(m, new Value("/t1".getBytes()));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
     // active txid, mutation that looks like a load
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("5".getBytes()));
     m.put(TabletsSection.CurrentLocationColumnFamily.NAME, new Text("789"), new Value("127.0.0.1:9997".getBytes()));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
     // inactive txid, mutation that looks like a load
     m = new Mutation(new Text("0;foo"));
     m.put(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"), new Value("12345".getBytes()));
     m.put(TabletsSection.CurrentLocationColumnFamily.NAME, new Text("789"), new Value("127.0.0.1:9997".getBytes()));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
     // deleting a load flag
     m = new Mutation(new Text("0;foo"));
     m.putDelete(TabletsSection.BulkFileColumnFamily.NAME, new Text("/someFile"));
     violations = mc.check(null, m);
     assertNull(violations);
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/data/ServerMutationTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/data/ServerMutationTest.java b/server/base/src/test/java/org/apache/accumulo/server/data/ServerMutationTest.java
index 0df27f1..52d3164 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/data/ServerMutationTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/data/ServerMutationTest.java
@@ -30,50 +30,49 @@ import org.apache.hadoop.util.ReflectionUtils;
 import org.junit.Test;
 
 public class ServerMutationTest {
-  
+
   @Test
   public void test() throws Exception {
     ServerMutation m = new ServerMutation(new Text("r1"));
     m.put(new Text("cf1"), new Text("cq1"), new Value("v1".getBytes()));
     m.put(new Text("cf2"), new Text("cq2"), 56, new Value("v2".getBytes()));
     m.setSystemTimestamp(42);
-    
+
     List<ColumnUpdate> updates = m.getUpdates();
-    
+
     assertEquals(2, updates.size());
-    
+
     assertEquals("r1", new String(m.getRow()));
     ColumnUpdate cu = updates.get(0);
-    
+
     assertEquals("cf1", new String(cu.getColumnFamily()));
     assertEquals("cq1", new String(cu.getColumnQualifier()));
     assertEquals("", new String(cu.getColumnVisibility()));
     assertFalse(cu.hasTimestamp());
     assertEquals(42l, cu.getTimestamp());
-    
+
     ServerMutation m2 = new ServerMutation();
     ReflectionUtils.copy(CachedConfiguration.getInstance(), m, m2);
-    
+
     updates = m2.getUpdates();
-    
+
     assertEquals(2, updates.size());
     assertEquals("r1", new String(m2.getRow()));
-    
+
     cu = updates.get(0);
     assertEquals("cf1", new String(cu.getColumnFamily()));
     assertEquals("cq1", new String(cu.getColumnQualifier()));
     assertFalse(cu.hasTimestamp());
     assertEquals(42l, cu.getTimestamp());
-    
+
     cu = updates.get(1);
-    
+
     assertEquals("r1", new String(m2.getRow()));
     assertEquals("cf2", new String(cu.getColumnFamily()));
     assertEquals("cq2", new String(cu.getColumnQualifier()));
     assertTrue(cu.hasTimestamp());
     assertEquals(56, cu.getTimestamp());
-    
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/fs/FileRefTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/fs/FileRefTest.java b/server/base/src/test/java/org/apache/accumulo/server/fs/FileRefTest.java
index 4ad0ea3..402f689 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/fs/FileRefTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/fs/FileRefTest.java
@@ -23,9 +23,9 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
-public class FileRefTest{
+public class FileRefTest {
 
   private void testBadTableSuffix(String badPath) {
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/fs/FileTypeTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/fs/FileTypeTest.java b/server/base/src/test/java/org/apache/accumulo/server/fs/FileTypeTest.java
index ad29c19..9ffe7d3 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/fs/FileTypeTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/fs/FileTypeTest.java
@@ -22,7 +22,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class FileTypeTest {
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/fs/ViewFSUtilsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/fs/ViewFSUtilsTest.java b/server/base/src/test/java/org/apache/accumulo/server/fs/ViewFSUtilsTest.java
index 52c70c0..cea35e2 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/fs/ViewFSUtilsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/fs/ViewFSUtilsTest.java
@@ -26,16 +26,16 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ViewFSUtilsTest {
-  
-  private String[] shuffle(String ... inputs){
+
+  private String[] shuffle(String... inputs) {
     // code below will modify array
     Collections.shuffle(Arrays.asList(inputs));
     return inputs;
   }
-  
+
   @Test
   public void testDisjointMountPoints() throws IllegalArgumentException, IOException {
     if (ViewFSUtils.isViewFSSupported()) {
@@ -47,8 +47,7 @@ public class ViewFSUtilsTest {
 
       String[] tablesDirs1 = shuffle("viewfs:///ns1/accumulo/tables", "viewfs:///ns2/accumulo/tables", "viewfs:///ns22/accumulo/tables",
           "viewfs:///ns/accumulo/tables");
-      String[] tablesDirs2 = shuffle("viewfs:/ns1/accumulo/tables", "viewfs:/ns2/accumulo/tables", "viewfs:/ns22/accumulo/tables",
-          "viewfs:/ns/accumulo/tables");
+      String[] tablesDirs2 = shuffle("viewfs:/ns1/accumulo/tables", "viewfs:/ns2/accumulo/tables", "viewfs:/ns22/accumulo/tables", "viewfs:/ns/accumulo/tables");
 
       for (String ns : Arrays.asList("ns1", "ns2", "ns22", "ns")) {
         Path match = ViewFSUtils.matchingFileSystem(new Path("viewfs:/" + ns + "/bulk_import_01"), tablesDirs2, conf);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeManagerImplTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeManagerImplTest.java b/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeManagerImplTest.java
index 23829d1..3bf207a 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeManagerImplTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeManagerImplTest.java
@@ -21,7 +21,6 @@ import java.util.List;
 
 import org.apache.accumulo.core.conf.ConfigurationCopy;
 import org.apache.accumulo.core.conf.Property;
-import org.apache.accumulo.server.fs.VolumeChooserEnvironment;
 import org.apache.accumulo.server.fs.VolumeManager.FileType;
 import org.apache.commons.lang.StringUtils;
 import org.apache.hadoop.fs.Path;
@@ -111,7 +110,7 @@ public class VolumeManagerImplTest {
     List<String> volumes = Arrays.asList("file://one/", "file://two/", "file://three/");
     ConfigurationCopy conf = new ConfigurationCopy();
     conf.set(INSTANCE_DFS_URI, volumes.get(0));
-    conf.set(Property.INSTANCE_VOLUMES, StringUtils.join(volumes,","));
+    conf.set(Property.INSTANCE_VOLUMES, StringUtils.join(volumes, ","));
     conf.set(Property.GENERAL_VOLUME_CHOOSER, WrongVolumeChooser.class.getName());
     VolumeManager vm = VolumeManagerImpl.get(conf);
     String choice = vm.choose(Optional.of("sometable"), volumes.toArray(new String[0]));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeUtilTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeUtilTest.java b/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeUtilTest.java
index 0013d04..9fcd3e0 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeUtilTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/fs/VolumeUtilTest.java
@@ -33,7 +33,7 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
 /**
- * 
+ *
  */
 public class VolumeUtilTest {
 
@@ -79,12 +79,12 @@ public class VolumeUtilTest {
     replacements.add(new Pair<Path,Path>(new Path("hdfs://nn1:9000/accumulo"), new Path("viewfs:/a")));
     replacements.add(new Pair<Path,Path>(new Path("hdfs://nn2/accumulo"), new Path("viewfs:/b")));
 
-    Assert.assertEquals("viewfs:/a/tables/t-00000/C000.rf",
-        VolumeUtil.switchVolume("hdfs://nn1/accumulo/tables/t-00000/C000.rf", FileType.TABLE, replacements));
+    Assert
+        .assertEquals("viewfs:/a/tables/t-00000/C000.rf", VolumeUtil.switchVolume("hdfs://nn1/accumulo/tables/t-00000/C000.rf", FileType.TABLE, replacements));
     Assert.assertEquals("viewfs:/a/tables/t-00000/C000.rf",
         VolumeUtil.switchVolume("hdfs://nn1:9000/accumulo/tables/t-00000/C000.rf", FileType.TABLE, replacements));
-    Assert.assertEquals("viewfs:/b/tables/t-00000/C000.rf",
-        VolumeUtil.switchVolume("hdfs://nn2/accumulo/tables/t-00000/C000.rf", FileType.TABLE, replacements));
+    Assert
+        .assertEquals("viewfs:/b/tables/t-00000/C000.rf", VolumeUtil.switchVolume("hdfs://nn2/accumulo/tables/t-00000/C000.rf", FileType.TABLE, replacements));
     Assert.assertNull(VolumeUtil.switchVolume("viewfs:/a/tables/t-00000/C000.rf", FileType.TABLE, replacements));
     Assert.assertNull(VolumeUtil.switchVolume("file:/nn1/a/accumulo/tables/t-00000/C000.rf", FileType.TABLE, replacements));
 
@@ -205,9 +205,7 @@ public class VolumeUtilTest {
 
     FileType ft = FileType.TABLE;
 
-    Assert.assertEquals("file:/foo/v8/tables/+r/root_tablet",
-        VolumeUtil.switchVolume("file:/foo/v1/tables/+r/root_tablet",
-        ft, replacements));
+    Assert.assertEquals("file:/foo/v8/tables/+r/root_tablet", VolumeUtil.switchVolume("file:/foo/v1/tables/+r/root_tablet", ft, replacements));
   }
 
   private void writeFile(FileSystem fs, Path dir, String filename, String data) throws IOException {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilterTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilterTest.java b/server/base/src/test/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilterTest.java
index 4a45e99..cfdd5e9 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilterTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilterTest.java
@@ -34,13 +34,12 @@ import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection;
 import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection.DataFileColumnFamily;
 import org.apache.accumulo.core.util.ColumnFQ;
 import org.apache.accumulo.fate.zookeeper.TransactionWatcher.Arbitrator;
-import org.apache.accumulo.server.iterators.MetadataBulkLoadFilter;
 import org.apache.hadoop.io.Text;
 import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class MetadataBulkLoadFilterTest {
   static class TestArbitrator implements Arbitrator {
@@ -48,7 +47,7 @@ public class MetadataBulkLoadFilterTest {
     public boolean transactionAlive(String type, long tid) throws Exception {
       return tid == 5;
     }
-    
+
     @Override
     public boolean transactionComplete(String type, long tid) throws Exception {
       if (tid == 9)
@@ -56,19 +55,19 @@ public class MetadataBulkLoadFilterTest {
       return tid != 5 && tid != 7;
     }
   }
-  
+
   static class TestMetadataBulkLoadFilter extends MetadataBulkLoadFilter {
     @Override
     protected Arbitrator getArbitrator() {
       return new TestArbitrator();
     }
   }
-  
+
   private static void put(TreeMap<Key,Value> tm, String row, ColumnFQ cfq, String val) {
     Key k = new Key(new Text(row), cfq.getColumnFamily(), cfq.getColumnQualifier());
     tm.put(k, new Value(val.getBytes()));
   }
-  
+
   private static void put(TreeMap<Key,Value> tm, String row, Text cf, String cq, String val) {
     Key k = new Key(new Text(row), cf, new Text(cq));
     if (val == null) {
@@ -77,12 +76,12 @@ public class MetadataBulkLoadFilterTest {
     } else
       tm.put(k, new Value(val.getBytes()));
   }
-  
+
   @Test
   public void testBasic() throws IOException {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     TreeMap<Key,Value> expected = new TreeMap<Key,Value>();
-    
+
     // following should not be deleted by filter
     put(tm1, "2;m", TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN, "/t1");
     put(tm1, "2;m", DataFileColumnFamily.NAME, "/t1/file1", "1,1");
@@ -95,50 +94,50 @@ public class MetadataBulkLoadFilterTest {
     put(tm1, "2<", TabletsSection.BulkFileColumnFamily.NAME, "/t2/file7", "7");
     put(tm1, "2<", TabletsSection.BulkFileColumnFamily.NAME, "/t2/file8", "9");
     put(tm1, "2<", TabletsSection.BulkFileColumnFamily.NAME, "/t2/fileC", null);
-    
+
     expected.putAll(tm1);
-    
+
     // the following should be deleted by filter
     put(tm1, "2;m", TabletsSection.BulkFileColumnFamily.NAME, "/t1/file5", "8");
     put(tm1, "2<", TabletsSection.BulkFileColumnFamily.NAME, "/t2/file9", "8");
     put(tm1, "2<", TabletsSection.BulkFileColumnFamily.NAME, "/t2/fileA", "2");
-    
+
     TestMetadataBulkLoadFilter iter = new TestMetadataBulkLoadFilter();
     iter.init(new SortedMapIterator(tm1), new HashMap<String,String>(), new IteratorEnvironment() {
-      
+
       @Override
       public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
         return null;
       }
-      
+
       @Override
       public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {}
-      
+
       @Override
       public boolean isFullMajorCompaction() {
         return false;
       }
-      
+
       @Override
       public IteratorScope getIteratorScope() {
         return IteratorScope.majc;
       }
-      
+
       @Override
       public AccumuloConfiguration getConfig() {
         return null;
       }
     });
-    
+
     iter.seek(new Range(), new ArrayList<ByteSequence>(), false);
-    
+
     TreeMap<Key,Value> actual = new TreeMap<Key,Value>();
-    
+
     while (iter.hasTop()) {
       actual.put(iter.getTopKey(), iter.getTopValue());
       iter.next();
     }
-    
+
     Assert.assertEquals(expected, actual);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancerTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancerTest.java b/server/base/src/test/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancerTest.java
index 2dff03d..1b691c7 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancerTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancerTest.java
@@ -33,7 +33,6 @@ import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.accumulo.core.master.thrift.TableInfo;
 import org.apache.accumulo.core.master.thrift.TabletServerStatus;
 import org.apache.accumulo.core.tabletserver.thrift.TabletStats;
-import org.apache.accumulo.server.master.balancer.ChaoticLoadBalancer;
 import org.apache.accumulo.server.master.state.TServerInstance;
 import org.apache.accumulo.server.master.state.TabletMigration;
 import org.apache.hadoop.io.Text;
@@ -43,10 +42,10 @@ import org.junit.Test;
 import com.google.common.net.HostAndPort;
 
 public class ChaoticLoadBalancerTest {
-  
+
   class FakeTServer {
     List<KeyExtent> extents = new ArrayList<KeyExtent>();
-    
+
     TabletServerStatus getStatus(TServerInstance server) {
       TabletServerStatus result = new TabletServerStatus();
       result.tableMap = new HashMap<String,TableInfo>();
@@ -63,11 +62,11 @@ public class ChaoticLoadBalancerTest {
       return result;
     }
   }
-  
+
   Map<TServerInstance,FakeTServer> servers = new HashMap<TServerInstance,FakeTServer>();
-  
+
   class TestChaoticLoadBalancer extends ChaoticLoadBalancer {
-    
+
     @Override
     public List<TabletStats> getOnlineTabletsForTable(TServerInstance tserver, String table) throws ThriftSecurityException, TException {
       List<TabletStats> result = new ArrayList<TabletStats>();
@@ -79,7 +78,7 @@ public class ChaoticLoadBalancerTest {
       return result;
     }
   }
-  
+
   @Test
   public void testAssignMigrations() {
     servers.clear();
@@ -99,20 +98,20 @@ public class ChaoticLoadBalancerTest {
     metadataTable.put(makeExtent(table, "d", "c"), null);
     metadataTable.put(makeExtent(table, "e", "d"), null);
     metadataTable.put(makeExtent(table, null, "e"), null);
-    
+
     TestChaoticLoadBalancer balancer = new TestChaoticLoadBalancer();
-    
+
     SortedMap<TServerInstance,TabletServerStatus> current = new TreeMap<TServerInstance,TabletServerStatus>();
     for (Entry<TServerInstance,FakeTServer> entry : servers.entrySet()) {
       current.put(entry.getKey(), entry.getValue().getStatus(entry.getKey()));
     }
-    
+
     Map<KeyExtent,TServerInstance> assignments = new HashMap<KeyExtent,TServerInstance>();
     balancer.getAssignments(getAssignments(servers), metadataTable, assignments);
-    
+
     assertEquals(assignments.size(), metadataTable.size());
   }
-  
+
   SortedMap<TServerInstance,TabletServerStatus> getAssignments(Map<TServerInstance,FakeTServer> servers) {
     SortedMap<TServerInstance,TabletServerStatus> result = new TreeMap<TServerInstance,TabletServerStatus>();
     for (Entry<TServerInstance,FakeTServer> entry : servers.entrySet()) {
@@ -120,7 +119,7 @@ public class ChaoticLoadBalancerTest {
     }
     return result;
   }
-  
+
   @Test
   public void testUnevenAssignment() {
     servers.clear();
@@ -146,22 +145,22 @@ public class ChaoticLoadBalancerTest {
     first.getValue().extents.add(makeExtent("newTable", "i", null));
     TestChaoticLoadBalancer balancer = new TestChaoticLoadBalancer();
     Set<KeyExtent> migrations = Collections.emptySet();
-    
+
     // Just want to make sure it gets some migrations, randomness prevents guarantee of a defined amount, or even expected amount
     List<TabletMigration> migrationsOut = new ArrayList<TabletMigration>();
     while (migrationsOut.size() != 0) {
       balancer.balance(getAssignments(servers), migrations, migrationsOut);
     }
   }
-  
+
   private static KeyExtent makeExtent(String table, String end, String prev) {
     return new KeyExtent(new Text(table), toText(end), toText(prev));
   }
-  
+
   private static Text toText(String value) {
     if (value != null)
       return new Text(value);
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/master/balancer/DefaultLoadBalancerTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/master/balancer/DefaultLoadBalancerTest.java b/server/base/src/test/java/org/apache/accumulo/server/master/balancer/DefaultLoadBalancerTest.java
index 9f99b1c..aee15bc 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/master/balancer/DefaultLoadBalancerTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/master/balancer/DefaultLoadBalancerTest.java
@@ -36,7 +36,6 @@ import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.accumulo.core.master.thrift.TableInfo;
 import org.apache.accumulo.core.master.thrift.TabletServerStatus;
 import org.apache.accumulo.core.tabletserver.thrift.TabletStats;
-import org.apache.accumulo.server.master.balancer.DefaultLoadBalancer;
 import org.apache.accumulo.server.master.state.TServerInstance;
 import org.apache.accumulo.server.master.state.TabletMigration;
 import org.apache.hadoop.io.Text;
@@ -47,10 +46,10 @@ import org.junit.Test;
 import com.google.common.net.HostAndPort;
 
 public class DefaultLoadBalancerTest {
-  
+
   class FakeTServer {
     List<KeyExtent> extents = new ArrayList<KeyExtent>();
-    
+
     TabletServerStatus getStatus(TServerInstance server) {
       TabletServerStatus result = new TabletServerStatus();
       result.tableMap = new HashMap<String,TableInfo>();
@@ -67,12 +66,12 @@ public class DefaultLoadBalancerTest {
       return result;
     }
   }
-  
+
   Map<TServerInstance,FakeTServer> servers = new HashMap<TServerInstance,FakeTServer>();
-  Map<KeyExtent, TServerInstance> last = new HashMap<KeyExtent, TServerInstance>();
-  
+  Map<KeyExtent,TServerInstance> last = new HashMap<KeyExtent,TServerInstance>();
+
   class TestDefaultLoadBalancer extends DefaultLoadBalancer {
-    
+
     @Override
     public List<TabletStats> getOnlineTabletsForTable(TServerInstance tserver, String table) throws ThriftSecurityException, TException {
       List<TabletStats> result = new ArrayList<TabletStats>();
@@ -84,13 +83,13 @@ public class DefaultLoadBalancerTest {
       return result;
     }
   }
-  
+
   @Before
   public void setUp() {
     last.clear();
     servers.clear();
   }
-  
+
   @Test
   public void testAssignMigrations() {
     servers.put(new TServerInstance(HostAndPort.fromParts("127.0.0.1", 1234), "a"), new FakeTServer());
@@ -110,42 +109,42 @@ public class DefaultLoadBalancerTest {
     metadataTable.add(makeExtent(table, "e", "d"));
     metadataTable.add(makeExtent(table, null, "e"));
     Collections.sort(metadataTable);
-    
+
     TestDefaultLoadBalancer balancer = new TestDefaultLoadBalancer();
-    
+
     SortedMap<TServerInstance,TabletServerStatus> current = new TreeMap<TServerInstance,TabletServerStatus>();
     for (Entry<TServerInstance,FakeTServer> entry : servers.entrySet()) {
       current.put(entry.getKey(), entry.getValue().getStatus(entry.getKey()));
     }
     assignTablets(metadataTable, servers, current, balancer);
-    
+
     // Verify that the counts on the tables are correct
     Map<String,Integer> expectedCounts = new HashMap<String,Integer>();
     expectedCounts.put("t1", 1);
     expectedCounts.put("t2", 1);
     expectedCounts.put("t3", 2);
     checkBalance(metadataTable, servers, expectedCounts);
-    
+
     // Rebalance once
     for (Entry<TServerInstance,FakeTServer> entry : servers.entrySet()) {
       current.put(entry.getKey(), entry.getValue().getStatus(entry.getKey()));
     }
-    
+
     // Nothing should happen, we are balanced
     ArrayList<TabletMigration> out = new ArrayList<TabletMigration>();
     balancer.getMigrations(current, out);
     assertEquals(out.size(), 0);
-    
+
     // Take down a tabletServer
     TServerInstance first = current.keySet().iterator().next();
     current.remove(first);
     FakeTServer remove = servers.remove(first);
-    
+
     // reassign offline extents
     assignTablets(remove.extents, servers, current, balancer);
     checkBalance(metadataTable, servers, null);
   }
-  
+
   private void assignTablets(List<KeyExtent> metadataTable, Map<TServerInstance,FakeTServer> servers, SortedMap<TServerInstance,TabletServerStatus> status,
       TestDefaultLoadBalancer balancer) {
     // Assign tablets
@@ -157,7 +156,7 @@ public class DefaultLoadBalancerTest {
       last.put(extent, assignment);
     }
   }
-  
+
   SortedMap<TServerInstance,TabletServerStatus> getAssignments(Map<TServerInstance,FakeTServer> servers) {
     SortedMap<TServerInstance,TabletServerStatus> result = new TreeMap<TServerInstance,TabletServerStatus>();
     for (Entry<TServerInstance,FakeTServer> entry : servers.entrySet()) {
@@ -165,7 +164,7 @@ public class DefaultLoadBalancerTest {
     }
     return result;
   }
-  
+
   @Test
   public void testUnevenAssignment() {
     for (char c : "abcdefghijklmnopqrstuvwxyz".toCharArray()) {
@@ -205,7 +204,7 @@ public class DefaultLoadBalancerTest {
     }
     assertEquals(8, moved);
   }
-  
+
   @Test
   public void testUnevenAssignment2() {
     // make 26 servers
@@ -230,7 +229,7 @@ public class DefaultLoadBalancerTest {
     for (int i = 0; i < 10; i++) {
       shortServer.getValue().extents.add(makeExtent("s" + i, null, null));
     }
-    
+
     TestDefaultLoadBalancer balancer = new TestDefaultLoadBalancer();
     Set<KeyExtent> migrations = Collections.emptySet();
     int moved = 0;
@@ -251,7 +250,7 @@ public class DefaultLoadBalancerTest {
     // average is 58, with 2 at 59: we need 48 more moved to the short server
     assertEquals(48, moved);
   }
-  
+
   private void checkBalance(List<KeyExtent> metadataTable, Map<TServerInstance,FakeTServer> servers, Map<String,Integer> expectedCounts) {
     // Verify they are spread evenly over the cluster
     int average = metadataTable.size() / servers.size();
@@ -262,7 +261,7 @@ public class DefaultLoadBalancerTest {
       if (diff > 1)
         fail("average number of tablets is " + average + " but a server has " + server.extents.size());
     }
-    
+
     if (expectedCounts != null) {
       for (FakeTServer server : servers.values()) {
         Map<String,Integer> counts = new HashMap<String,Integer>();
@@ -278,15 +277,15 @@ public class DefaultLoadBalancerTest {
       }
     }
   }
-  
+
   private static KeyExtent makeExtent(String table, String end, String prev) {
     return new KeyExtent(new Text(table), toText(end), toText(prev));
   }
-  
+
   private static Text toText(String value) {
     if (value != null)
       return new Text(value);
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/master/balancer/TableLoadBalancerTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/master/balancer/TableLoadBalancerTest.java b/server/base/src/test/java/org/apache/accumulo/server/master/balancer/TableLoadBalancerTest.java
index f107759..8e68d84 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/master/balancer/TableLoadBalancerTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/master/balancer/TableLoadBalancerTest.java
@@ -48,11 +48,11 @@ import org.junit.Test;
 import com.google.common.net.HostAndPort;
 
 public class TableLoadBalancerTest {
-  
+
   static private TServerInstance mkts(String address, String session) throws Exception {
     return new TServerInstance(HostAndPort.fromParts(address, 1234), session);
   }
-  
+
   static private TabletServerStatus status(Object... config) {
     TabletServerStatus result = new TabletServerStatus();
     result.tableMap = new HashMap<String,TableInfo>();
@@ -70,11 +70,11 @@ public class TableLoadBalancerTest {
     }
     return result;
   }
-  
+
   static MockInstance instance = new MockInstance("mockamatic");
-  
+
   static SortedMap<TServerInstance,TabletServerStatus> state;
-  
+
   static List<TabletStats> generateFakeTablets(TServerInstance tserver, String tableId) {
     List<TabletStats> result = new ArrayList<TabletStats>();
     TabletServerStatus tableInfo = state.get(tserver);
@@ -87,39 +87,39 @@ public class TableLoadBalancerTest {
     }
     return result;
   }
-  
+
   static class DefaultLoadBalancer extends org.apache.accumulo.server.master.balancer.DefaultLoadBalancer {
-    
+
     public DefaultLoadBalancer(String table) {
       super(table);
     }
-    
+
     @Override
     public List<TabletStats> getOnlineTabletsForTable(TServerInstance tserver, String tableId) throws ThriftSecurityException, TException {
       return generateFakeTablets(tserver, tableId);
     }
   }
-  
+
   // ugh... so wish I had provided mock objects to the LoadBalancer in the master
   static class TableLoadBalancer extends org.apache.accumulo.server.master.balancer.TableLoadBalancer {
-    
+
     TableLoadBalancer() {
       super();
     }
-    
+
     // use our new classname to test class loading
     @Override
     protected String getLoadBalancerClassNameForTable(String table) {
       return DefaultLoadBalancer.class.getName();
     }
-    
+
     // we don't have real tablet servers to ask: invent some online tablets
     @Override
     public List<TabletStats> getOnlineTabletsForTable(TServerInstance tserver, String tableId) throws ThriftSecurityException, TException {
       return generateFakeTablets(tserver, tableId);
     }
   }
-  
+
   @Test
   public void test() throws Exception {
     Connector c = instance.getConnector("user", new PasswordToken("pass"));
@@ -143,14 +143,14 @@ public class TableLoadBalancerTest {
     state = new TreeMap<TServerInstance,TabletServerStatus>();
     TServerInstance svr = mkts("10.0.0.1", "0x01020304");
     state.put(svr, status(t1Id, 10, t2Id, 10, t3Id, 10));
-    
+
     Set<KeyExtent> migrations = Collections.emptySet();
     List<TabletMigration> migrationsOut = new ArrayList<TabletMigration>();
     TableLoadBalancer tls = new TableLoadBalancer();
     tls.init(confFactory);
     tls.balance(state, migrations, migrationsOut);
     Assert.assertEquals(0, migrationsOut.size());
-    
+
     state.put(mkts("10.0.0.2", "0x02030405"), status());
     tls = new TableLoadBalancer();
     tls.init(confFactory);
@@ -171,5 +171,5 @@ public class TableLoadBalancerTest {
       Assert.assertEquals(5, moved.intValue());
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/master/state/MergeInfoTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/master/state/MergeInfoTest.java b/server/base/src/test/java/org/apache/accumulo/server/master/state/MergeInfoTest.java
index 58484af..7c26f85 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/master/state/MergeInfoTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/master/state/MergeInfoTest.java
@@ -16,19 +16,25 @@
  */
 package org.apache.accumulo.server.master.state;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.expectLastCall;
+import static org.easymock.EasyMock.replay;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertSame;
+import static org.junit.Assert.assertTrue;
+
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
 import java.io.DataInputStream;
 import java.io.DataOutputStream;
+
 import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.hadoop.io.Text;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.expectLastCall;
-import static org.easymock.EasyMock.replay;
 
 public class MergeInfoTest {
   private KeyExtent keyExtent;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/master/state/TabletLocationStateTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/master/state/TabletLocationStateTest.java b/server/base/src/test/java/org/apache/accumulo/server/master/state/TabletLocationStateTest.java
index c1f312d..a7b673b 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/master/state/TabletLocationStateTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/master/state/TabletLocationStateTest.java
@@ -16,17 +16,23 @@
  */
 package org.apache.accumulo.server.master.state;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNotNull;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertSame;
+import static org.junit.Assert.assertTrue;
+
 import java.util.Collection;
 import java.util.Set;
+
 import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.hadoop.io.Text;
 import org.junit.Before;
 import org.junit.BeforeClass;
 import org.junit.Test;
-import static org.junit.Assert.*;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.replay;
 
 public class TabletLocationStateTest {
   private static final Collection<String> innerWalogs = new java.util.HashSet<String>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/problems/ProblemReportTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/problems/ProblemReportTest.java b/server/base/src/test/java/org/apache/accumulo/server/problems/ProblemReportTest.java
index dbad326..1ca3e8d 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/problems/ProblemReportTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/problems/ProblemReportTest.java
@@ -16,8 +16,21 @@
  */
 package org.apache.accumulo.server.problems;
 
+import static org.easymock.EasyMock.aryEq;
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.eq;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertSame;
+import static org.junit.Assert.assertTrue;
+
 import java.io.ByteArrayOutputStream;
 import java.io.DataOutputStream;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.util.Encoding;
@@ -28,13 +41,6 @@ import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.apache.hadoop.io.Text;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
-import static org.easymock.EasyMock.aryEq;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.eq;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
 
 public class ProblemReportTest {
   private static final String TABLE = "table";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/replication/StatusCombinerTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/replication/StatusCombinerTest.java b/server/base/src/test/java/org/apache/accumulo/server/replication/StatusCombinerTest.java
index a9801b0..74e5455 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/replication/StatusCombinerTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/replication/StatusCombinerTest.java
@@ -177,8 +177,8 @@ public class StatusCombinerTest {
 
   @Test
   public void commutativeWithMultipleUpdates() {
-    Status newFile = StatusUtil.fileCreated(100), update1 = StatusUtil.ingestedUntil(100), update2 = StatusUtil.ingestedUntil(200), repl1 = StatusUtil.replicated(50), repl2 = StatusUtil
-        .replicated(150);
+    Status newFile = StatusUtil.fileCreated(100), update1 = StatusUtil.ingestedUntil(100), update2 = StatusUtil.ingestedUntil(200), repl1 = StatusUtil
+        .replicated(50), repl2 = StatusUtil.replicated(150);
 
     Status order1 = combiner.typedReduce(key, Arrays.asList(newFile, update1, repl1, update2, repl2).iterator());
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/replication/ZooKeeperInitializationTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/replication/ZooKeeperInitializationTest.java b/server/base/src/test/java/org/apache/accumulo/server/replication/ZooKeeperInitializationTest.java
index 25e7bc8..5604242 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/replication/ZooKeeperInitializationTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/replication/ZooKeeperInitializationTest.java
@@ -27,7 +27,7 @@ import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ZooKeeperInitializationTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/security/SystemCredentialsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/security/SystemCredentialsTest.java b/server/base/src/test/java/org/apache/accumulo/server/security/SystemCredentialsTest.java
index 4202a7e..01ff9ac 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/security/SystemCredentialsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/security/SystemCredentialsTest.java
@@ -32,10 +32,10 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class SystemCredentialsTest {
-  
+
   private static MockInstance inst;
 
   @BeforeClass
@@ -54,7 +54,7 @@ public class SystemCredentialsTest {
       testInstanceVersion.createNewFile();
     }
   }
-  
+
   /**
    * This is a test to ensure the string literal in {@link ConnectorImpl#ConnectorImpl(org.apache.accumulo.core.client.impl.ClientContext)} is kept up-to-date
    * if we move the {@link SystemToken}<br/>
@@ -65,7 +65,7 @@ public class SystemCredentialsTest {
     assertEquals("org.apache.accumulo.server.security.SystemCredentials$SystemToken", SystemToken.class.getName());
     assertEquals(SystemCredentials.get(inst).getToken().getClass(), SystemToken.class);
   }
-  
+
   @Test
   public void testSystemCredentials() {
     Credentials a = SystemCredentials.get(inst);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/security/handler/ZKAuthenticatorTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/security/handler/ZKAuthenticatorTest.java b/server/base/src/test/java/org/apache/accumulo/server/security/handler/ZKAuthenticatorTest.java
index 2f01e6e..eb5e2f3 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/security/handler/ZKAuthenticatorTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/security/handler/ZKAuthenticatorTest.java
@@ -19,31 +19,31 @@ package org.apache.accumulo.server.security.handler;
 import java.util.Set;
 import java.util.TreeSet;
 
+import junit.framework.TestCase;
+
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
 import org.apache.accumulo.core.util.ByteArraySet;
-
-import junit.framework.TestCase;
 import org.apache.log4j.Logger;
 
 public class ZKAuthenticatorTest extends TestCase {
   private static final Logger log = Logger.getLogger(ZKAuthenticatorTest.class);
-  
+
   public void testPermissionIdConversions() {
     for (SystemPermission s : SystemPermission.values())
       assertTrue(s.equals(SystemPermission.getPermissionById(s.getId())));
-    
+
     for (TablePermission s : TablePermission.values())
       assertTrue(s.equals(TablePermission.getPermissionById(s.getId())));
   }
-  
+
   public void testAuthorizationConversion() {
     ByteArraySet auths = new ByteArraySet();
     for (int i = 0; i < 300; i += 3)
       auths.add(Integer.toString(i).getBytes());
-    
+
     Authorizations converted = new Authorizations(auths);
     byte[] test = ZKSecurityTool.convertAuthorizations(converted);
     Authorizations test2 = ZKSecurityTool.convertAuthorizations(test);
@@ -52,29 +52,29 @@ public class ZKAuthenticatorTest extends TestCase {
       assertTrue(test2.contains(s));
     }
   }
-  
+
   public void testSystemConversion() {
     Set<SystemPermission> perms = new TreeSet<SystemPermission>();
     for (SystemPermission s : SystemPermission.values())
       perms.add(s);
-    
+
     Set<SystemPermission> converted = ZKSecurityTool.convertSystemPermissions(ZKSecurityTool.convertSystemPermissions(perms));
     assertTrue(perms.size() == converted.size());
     for (SystemPermission s : perms)
       assertTrue(converted.contains(s));
   }
-  
+
   public void testTableConversion() {
     Set<TablePermission> perms = new TreeSet<TablePermission>();
     for (TablePermission s : TablePermission.values())
       perms.add(s);
-    
+
     Set<TablePermission> converted = ZKSecurityTool.convertTablePermissions(ZKSecurityTool.convertTablePermissions(perms));
     assertTrue(perms.size() == converted.size());
     for (TablePermission s : perms)
       assertTrue(converted.contains(s));
   }
-  
+
   public void testEncryption() {
     byte[] rawPass = "myPassword".getBytes();
     byte[] storedBytes;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/tablets/MillisTimeTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/tablets/MillisTimeTest.java b/server/base/src/test/java/org/apache/accumulo/server/tablets/MillisTimeTest.java
index 15c8465..3ee220e 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/tablets/MillisTimeTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/tablets/MillisTimeTest.java
@@ -16,18 +16,20 @@
  */
 package org.apache.accumulo.server.tablets;
 
+import static org.easymock.EasyMock.anyLong;
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+
+import java.util.List;
+
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.server.data.ServerMutation;
 import org.apache.accumulo.server.tablets.TabletTime.MillisTime;
-
-import java.util.List;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
-import static org.easymock.EasyMock.anyLong;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
 
 public class MillisTimeTest {
   private static final long TIME = 1234L;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/tablets/TabletTimeTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/tablets/TabletTimeTest.java b/server/base/src/test/java/org/apache/accumulo/server/tablets/TabletTimeTest.java
index 38cca85..46068e3 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/tablets/TabletTimeTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/tablets/TabletTimeTest.java
@@ -16,17 +16,18 @@
  */
 package org.apache.accumulo.server.tablets;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNull;
+
 import org.apache.accumulo.core.client.admin.TimeType;
 import org.apache.accumulo.server.data.ServerMutation;
 import org.apache.accumulo.server.tablets.TabletTime.LogicalTime;
 import org.apache.accumulo.server.tablets.TabletTime.MillisTime;
-
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
 
 public class TabletTimeTest {
   private static final long TIME = 1234L;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/AdminCommandsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/AdminCommandsTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/AdminCommandsTest.java
index f95c0f1..ab799ec 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/AdminCommandsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/AdminCommandsTest.java
@@ -16,8 +16,11 @@
  */
 package org.apache.accumulo.server.util;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNull;
+
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class AdminCommandsTest {
   @Test


[21/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java b/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
index c424f1a..b55cfd7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
@@ -107,12 +107,13 @@ public class Accumulo {
   }
 
   /**
-   * Finds the best log4j configuration file. A generic file is used only if an
-   * application-specific file is not available. An XML file is preferred over
-   * a properties file, if possible.
+   * Finds the best log4j configuration file. A generic file is used only if an application-specific file is not available. An XML file is preferred over a
+   * properties file, if possible.
    *
-   * @param confDir directory where configuration files should reside
-   * @param application application name for configuration file name
+   * @param confDir
+   *          directory where configuration files should reside
+   * @param application
+   *          application name for configuration file name
    * @return configuration file name
    */
   static String locateLogConfig(String confDir, String application) {
@@ -120,13 +121,9 @@ public class Accumulo {
     if (explicitConfigFile != null) {
       return explicitConfigFile;
     }
-    String[] configFiles = {
-      String.format("%s/%s_logger.xml", confDir, application),
-      String.format("%s/%s_logger.properties", confDir, application),
-      String.format("%s/generic_logger.xml", confDir),
-      String.format("%s/generic_logger.properties", confDir)
-    };
-    String defaultConfigFile = configFiles[2];  // generic_logger.xml
+    String[] configFiles = {String.format("%s/%s_logger.xml", confDir, application), String.format("%s/%s_logger.properties", confDir, application),
+        String.format("%s/generic_logger.xml", confDir), String.format("%s/generic_logger.properties", confDir)};
+    String defaultConfigFile = configFiles[2]; // generic_logger.xml
     for (String f : configFiles) {
       if (new File(f).exists()) {
         return f;
@@ -196,7 +193,8 @@ public class Accumulo {
 
     // Encourage users to configure TLS
     final String SSL = "SSL";
-    for (Property sslProtocolProperty : Arrays.asList(Property.RPC_SSL_CLIENT_PROTOCOL, Property.RPC_SSL_ENABLED_PROTOCOLS, Property.MONITOR_SSL_INCLUDE_PROTOCOLS)) {
+    for (Property sslProtocolProperty : Arrays.asList(Property.RPC_SSL_CLIENT_PROTOCOL, Property.RPC_SSL_ENABLED_PROTOCOLS,
+        Property.MONITOR_SSL_INCLUDE_PROTOCOLS)) {
       String value = conf.get(sslProtocolProperty);
       if (value.contains(SSL)) {
         log.warn("It is recommended that " + sslProtocolProperty + " only allow TLS");
@@ -206,7 +204,9 @@ public class Accumulo {
 
   /**
    * Sanity check that the current persistent version is allowed to upgrade to the version of Accumulo running.
-   * @param dataVersion the version that is persisted in the backing Volumes
+   *
+   * @param dataVersion
+   *          the version that is persisted in the backing Volumes
    */
   public static boolean canUpgradeFromDataVersion(final int dataVersion) {
     return ServerConstants.CAN_UPGRADE.get(dataVersion);
@@ -280,7 +280,7 @@ public class Accumulo {
           if (unknownHostTries > 0) {
             log.warn("Unable to connect to HDFS, will retry. cause: " + exception.getCause());
             /* We need to make sure our sleep period is long enough to avoid getting a cached failure of the host lookup. */
-            sleep = Math.max(sleep, (AddressUtil.getAddressCacheNegativeTtl((UnknownHostException)(exception.getCause()))+1)*1000);
+            sleep = Math.max(sleep, (AddressUtil.getAddressCacheNegativeTtl((UnknownHostException) (exception.getCause())) + 1) * 1000);
           } else {
             log.error("Unable to connect to HDFS and have exceeded max number of retries.", exception);
             throw exception;
@@ -299,10 +299,9 @@ public class Accumulo {
   }
 
   /**
-   * Exit loudly if there are outstanding Fate operations.
-   * Since Fate serializes class names, we need to make sure there are no queued
-   * transactions from a previous version before continuing an upgrade. The status of the operations is
-   * irrelevant; those in SUCCESSFUL status cause the same problem as those just queued.
+   * Exit loudly if there are outstanding Fate operations. Since Fate serializes class names, we need to make sure there are no queued transactions from a
+   * previous version before continuing an upgrade. The status of the operations is irrelevant; those in SUCCESSFUL status cause the same problem as those just
+   * queued.
    *
    * Note that the Master should not allow write access to Fate until after all upgrade steps are complete.
    *
@@ -312,10 +311,11 @@ public class Accumulo {
    */
   public static void abortIfFateTransactions() {
     try {
-      final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE,
-          ZooReaderWriter.getInstance()));
+      final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(
+          ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE, ZooReaderWriter.getInstance()));
       if (!(fate.list().isEmpty())) {
-        throw new AccumuloException("Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
+        throw new AccumuloException(
+            "Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
       }
     } catch (Exception exception) {
       log.fatal("Problem verifying Fate readiness", exception);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/ServerOpts.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/ServerOpts.java b/server/base/src/main/java/org/apache/accumulo/server/ServerOpts.java
index 95fee8f..bbe0dd2 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/ServerOpts.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/ServerOpts.java
@@ -21,12 +21,12 @@ import org.apache.accumulo.core.cli.Help;
 import com.beust.jcommander.Parameter;
 
 public class ServerOpts extends Help {
-  @Parameter(names={"-a", "--address"}, description = "address to bind to")
+  @Parameter(names = {"-a", "--address"}, description = "address to bind to")
   String address = null;
-  
+
   public String getAddress() {
     if (address != null)
       return address;
     return "0.0.0.0";
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnDefaultTable.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnDefaultTable.java b/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnDefaultTable.java
index 588c35c..c347994 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnDefaultTable.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnDefaultTable.java
@@ -30,7 +30,7 @@ public class ClientOnDefaultTable extends org.apache.accumulo.core.cli.ClientOnD
   synchronized public Instance getInstance() {
     if (cachedInstance != null)
       return cachedInstance;
-    
+
     if (mock)
       return cachedInstance = new MockInstance(instance);
     if (instance == null) {
@@ -38,6 +38,7 @@ public class ClientOnDefaultTable extends org.apache.accumulo.core.cli.ClientOnD
     }
     return cachedInstance = new ZooKeeperInstance(this.getClientConfiguration());
   }
+
   public ClientOnDefaultTable(String table) {
     super(table);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnRequiredTable.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnRequiredTable.java b/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnRequiredTable.java
index f2e04e4..38926be 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnRequiredTable.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOnRequiredTable.java
@@ -25,12 +25,12 @@ public class ClientOnRequiredTable extends org.apache.accumulo.core.cli.ClientOn
   {
     principal = "root";
   }
-  
+
   @Override
   synchronized public Instance getInstance() {
     if (cachedInstance != null)
       return cachedInstance;
-    
+
     if (mock)
       return cachedInstance = new MockInstance(instance);
     if (instance == null) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOpts.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOpts.java b/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOpts.java
index c19b7b0..0a7714d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOpts.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/cli/ClientOpts.java
@@ -22,7 +22,7 @@ import org.apache.accumulo.core.client.mock.MockInstance;
 import org.apache.accumulo.server.client.HdfsZooInstance;
 
 public class ClientOpts extends org.apache.accumulo.core.cli.ClientOpts {
-  
+
   {
     principal = "root";
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/client/BulkImporter.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/client/BulkImporter.java b/server/base/src/main/java/org/apache/accumulo/server/client/BulkImporter.java
index 8171555..01d03ed 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/client/BulkImporter.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/client/BulkImporter.java
@@ -87,34 +87,34 @@ public class BulkImporter {
     }
     return result;
   }
-  
+
   private StopWatch<Timers> timer;
-  
+
   private static enum Timers {
     EXAMINE_MAP_FILES, QUERY_METADATA, IMPORT_MAP_FILES, SLEEP, TOTAL
   }
-  
+
   private final ClientContext context;
   private String tableId;
   private long tid;
   private boolean setTime;
-  
+
   public BulkImporter(ClientContext context, long tid, String tableId, boolean setTime) {
     this.context = context;
     this.tid = tid;
     this.tableId = tableId;
     this.setTime = setTime;
   }
-  
+
   public AssignmentStats importFiles(List<String> files, Path failureDir) throws IOException, AccumuloException, AccumuloSecurityException,
       ThriftTableOperationException {
-    
+
     int numThreads = context.getConfiguration().getCount(Property.TSERV_BULK_PROCESS_THREADS);
     int numAssignThreads = context.getConfiguration().getCount(Property.TSERV_BULK_ASSIGNMENT_THREADS);
-    
+
     timer = new StopWatch<Timers>(Timers.class);
     timer.start(Timers.TOTAL);
-    
+
     Configuration conf = CachedConfiguration.getInstance();
     VolumeManagerImpl.get(context.getConfiguration());
     final VolumeManager fs = VolumeManagerImpl.get(context.getConfiguration());
@@ -124,18 +124,18 @@ public class BulkImporter {
       paths.add(new Path(file));
     }
     AssignmentStats assignmentStats = new AssignmentStats(paths.size());
-    
+
     final Map<Path,List<KeyExtent>> completeFailures = Collections.synchronizedSortedMap(new TreeMap<Path,List<KeyExtent>>());
-    
+
     ClientService.Client client = null;
     final TabletLocator locator = TabletLocator.getLocator(context, new Text(tableId));
-    
+
     try {
       final Map<Path,List<TabletLocation>> assignments = Collections.synchronizedSortedMap(new TreeMap<Path,List<TabletLocation>>());
-      
+
       timer.start(Timers.EXAMINE_MAP_FILES);
       ExecutorService threadPool = Executors.newFixedThreadPool(numThreads, new NamingThreadFactory("findOverlapping"));
-      
+
       for (Path path : paths) {
         final Path mapFile = path;
         Runnable getAssignments = new Runnable() {
@@ -166,16 +166,16 @@ public class BulkImporter {
         }
       }
       timer.stop(Timers.EXAMINE_MAP_FILES);
-      
+
       assignmentStats.attemptingAssignments(assignments);
       Map<Path,List<KeyExtent>> assignmentFailures = assignMapFiles(context, conf, fs, tableId, assignments, paths, numAssignThreads, numThreads);
       assignmentStats.assignmentsFailed(assignmentFailures);
-      
+
       Map<Path,Integer> failureCount = new TreeMap<Path,Integer>();
-      
+
       for (Entry<Path,List<KeyExtent>> entry : assignmentFailures.entrySet())
         failureCount.put(entry.getKey(), 1);
-      
+
       long sleepTime = 2 * 1000;
       while (assignmentFailures.size() > 0) {
         sleepTime = Math.min(sleepTime * 2, 60 * 1000);
@@ -185,24 +185,24 @@ public class BulkImporter {
         //
         // for splits we need to find children key extents that cover the
         // same key range and are contiguous (no holes, no overlap)
-        
+
         timer.start(Timers.SLEEP);
         UtilWaitThread.sleep(sleepTime);
         timer.stop(Timers.SLEEP);
-        
+
         log.debug("Trying to assign " + assignmentFailures.size() + " map files that previously failed on some key extents");
         assignments.clear();
-        
+
         // for failed key extents, try to find children key extents to
         // assign to
         for (Entry<Path,List<KeyExtent>> entry : assignmentFailures.entrySet()) {
           Iterator<KeyExtent> keListIter = entry.getValue().iterator();
-          
+
           List<TabletLocation> tabletsToAssignMapFileTo = new ArrayList<TabletLocation>();
-          
+
           while (keListIter.hasNext()) {
             KeyExtent ke = keListIter.next();
-            
+
             try {
               timer.start(Timers.QUERY_METADATA);
               tabletsToAssignMapFileTo.addAll(findOverlappingTablets(context, fs, locator, entry.getKey(), ke));
@@ -212,26 +212,26 @@ public class BulkImporter {
               log.warn("Exception finding overlapping tablets, will retry tablet " + ke, ex);
             }
           }
-          
+
           if (tabletsToAssignMapFileTo.size() > 0)
             assignments.put(entry.getKey(), tabletsToAssignMapFileTo);
         }
-        
+
         assignmentStats.attemptingAssignments(assignments);
         Map<Path,List<KeyExtent>> assignmentFailures2 = assignMapFiles(context, conf, fs, tableId, assignments, paths, numAssignThreads, numThreads);
         assignmentStats.assignmentsFailed(assignmentFailures2);
-        
+
         // merge assignmentFailures2 into assignmentFailures
         for (Entry<Path,List<KeyExtent>> entry : assignmentFailures2.entrySet()) {
           assignmentFailures.get(entry.getKey()).addAll(entry.getValue());
-          
+
           Integer fc = failureCount.get(entry.getKey());
           if (fc == null)
             fc = 0;
-          
+
           failureCount.put(entry.getKey(), fc + 1);
         }
-        
+
         // remove map files that have no more key extents to assign
         Iterator<Entry<Path,List<KeyExtent>>> afIter = assignmentFailures.entrySet().iterator();
         while (afIter.hasNext()) {
@@ -239,7 +239,7 @@ public class BulkImporter {
           if (entry.getValue().size() == 0)
             afIter.remove();
         }
-        
+
         Set<Entry<Path,Integer>> failureIter = failureCount.entrySet();
         for (Entry<Path,Integer> entry : failureIter) {
           int retries = context.getConfiguration().getCount(Property.TSERV_BULK_RETRY);
@@ -253,7 +253,7 @@ public class BulkImporter {
       assignmentStats.assignmentsAbandoned(completeFailures);
       Set<Path> failedFailures = processFailures(completeFailures);
       assignmentStats.unrecoveredMapFiles(failedFailures);
-      
+
       timer.stop(Timers.TOTAL);
       printReport(paths);
       return assignmentStats;
@@ -263,13 +263,13 @@ public class BulkImporter {
       locator.invalidateCache();
     }
   }
-  
+
   private void printReport(Set<Path> paths) {
     long totalTime = 0;
     for (Timers t : Timers.values()) {
       if (t == Timers.TOTAL)
         continue;
-      
+
       totalTime += timer.get(t);
     }
     List<String> files = new ArrayList<String>();
@@ -277,7 +277,7 @@ public class BulkImporter {
       files.add(path.getName());
     }
     Collections.sort(files);
-    
+
     log.debug("BULK IMPORT TIMING STATISTICS");
     log.debug("Files: " + files);
     log.debug(String.format("Examine map files    : %,10.2f secs %6.2f%s", timer.getSecs(Timers.EXAMINE_MAP_FILES), 100.0 * timer.get(Timers.EXAMINE_MAP_FILES)
@@ -292,51 +292,51 @@ public class BulkImporter {
         * (timer.get(Timers.TOTAL) - totalTime) / timer.get(Timers.TOTAL), "%"));
     log.debug(String.format("Total                : %,10.2f secs", timer.getSecs(Timers.TOTAL)));
   }
-  
+
   private Set<Path> processFailures(Map<Path,List<KeyExtent>> completeFailures) {
     // we should check if map file was not assigned to any tablets, then we
     // should just move it; not currently being done?
-    
+
     Set<Entry<Path,List<KeyExtent>>> es = completeFailures.entrySet();
-    
+
     if (completeFailures.size() == 0)
       return Collections.emptySet();
-    
+
     log.debug("The following map files failed ");
-    
+
     for (Entry<Path,List<KeyExtent>> entry : es) {
       List<KeyExtent> extents = entry.getValue();
-      
+
       for (KeyExtent keyExtent : extents)
         log.debug("\t" + entry.getKey() + " -> " + keyExtent);
     }
-    
+
     return Collections.emptySet();
   }
-  
+
   private class AssignmentInfo {
     public AssignmentInfo(KeyExtent keyExtent, Long estSize) {
       this.ke = keyExtent;
       this.estSize = estSize;
     }
-    
+
     KeyExtent ke;
     long estSize;
   }
-  
+
   private static List<KeyExtent> extentsOf(List<TabletLocation> locations) {
     List<KeyExtent> result = new ArrayList<KeyExtent>(locations.size());
     for (TabletLocation tl : locations)
       result.add(tl.tablet_extent);
     return result;
   }
-  
+
   private Map<Path,List<AssignmentInfo>> estimateSizes(final AccumuloConfiguration acuConf, final Configuration conf, final VolumeManager vm,
       Map<Path,List<TabletLocation>> assignments, Collection<Path> paths, int numThreads) {
-    
+
     long t1 = System.currentTimeMillis();
     final Map<Path,Long> mapFileSizes = new TreeMap<Path,Long>();
-    
+
     try {
       for (Path path : paths) {
         FileSystem fs = vm.getVolumeByPath(path).getFileSystem();
@@ -346,33 +346,33 @@ public class BulkImporter {
       log.error("Failed to get map files in for " + paths + ": " + e.getMessage(), e);
       throw new RuntimeException(e);
     }
-    
+
     final Map<Path,List<AssignmentInfo>> ais = Collections.synchronizedMap(new TreeMap<Path,List<AssignmentInfo>>());
-    
+
     ExecutorService threadPool = Executors.newFixedThreadPool(numThreads, new NamingThreadFactory("estimateSizes"));
-    
+
     for (final Entry<Path,List<TabletLocation>> entry : assignments.entrySet()) {
       if (entry.getValue().size() == 1) {
         TabletLocation tabletLocation = entry.getValue().get(0);
-        
+
         // if the tablet completely contains the map file, there is no
         // need to estimate its
         // size
         ais.put(entry.getKey(), Collections.singletonList(new AssignmentInfo(tabletLocation.tablet_extent, mapFileSizes.get(entry.getKey()))));
         continue;
       }
-      
+
       Runnable estimationTask = new Runnable() {
         @Override
         public void run() {
           Map<KeyExtent,Long> estimatedSizes = null;
-          
+
           try {
             estimatedSizes = FileUtil.estimateSizes(acuConf, entry.getKey(), mapFileSizes.get(entry.getKey()), extentsOf(entry.getValue()), conf, vm);
           } catch (IOException e) {
             log.warn("Failed to estimate map file sizes " + e.getMessage());
           }
-          
+
           if (estimatedSizes == null) {
             // estimation failed, do a simple estimation
             estimatedSizes = new TreeMap<KeyExtent,Long>();
@@ -380,21 +380,21 @@ public class BulkImporter {
             for (TabletLocation tl : entry.getValue())
               estimatedSizes.put(tl.tablet_extent, estSize);
           }
-          
+
           List<AssignmentInfo> assignmentInfoList = new ArrayList<AssignmentInfo>(estimatedSizes.size());
-          
+
           for (Entry<KeyExtent,Long> entry2 : estimatedSizes.entrySet())
             assignmentInfoList.add(new AssignmentInfo(entry2.getKey(), entry2.getValue()));
-          
+
           ais.put(entry.getKey(), assignmentInfoList);
         }
       };
-      
+
       threadPool.submit(new TraceRunnable(new LoggingRunnable(log, estimationTask)));
     }
-    
+
     threadPool.shutdown();
-    
+
     while (!threadPool.isTerminated()) {
       try {
         threadPool.awaitTermination(60, TimeUnit.SECONDS);
@@ -403,14 +403,14 @@ public class BulkImporter {
         throw new RuntimeException(e);
       }
     }
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     log.debug(String.format("Estimated map files sizes in %6.2f secs", (t2 - t1) / 1000.0));
-    
+
     return ais;
   }
-  
+
   private static Map<KeyExtent,String> locationsOf(Map<Path,List<TabletLocation>> assignments) {
     Map<KeyExtent,String> result = new HashMap<KeyExtent,String>();
     for (List<TabletLocation> entry : assignments.values()) {
@@ -420,33 +420,33 @@ public class BulkImporter {
     }
     return result;
   }
-  
+
   private Map<Path,List<KeyExtent>> assignMapFiles(ClientContext context, Configuration conf, VolumeManager fs, String tableId,
       Map<Path,List<TabletLocation>> assignments, Collection<Path> paths, int numThreads, int numMapThreads) {
     timer.start(Timers.EXAMINE_MAP_FILES);
     Map<Path,List<AssignmentInfo>> assignInfo = estimateSizes(context.getConfiguration(), conf, fs, assignments, paths, numMapThreads);
     timer.stop(Timers.EXAMINE_MAP_FILES);
-    
+
     Map<Path,List<KeyExtent>> ret;
-    
+
     timer.start(Timers.IMPORT_MAP_FILES);
     ret = assignMapFiles(tableId, assignInfo, locationsOf(assignments), numThreads);
     timer.stop(Timers.IMPORT_MAP_FILES);
-    
+
     return ret;
   }
-  
+
   private class AssignmentTask implements Runnable {
     final Map<Path,List<KeyExtent>> assignmentFailures;
     HostAndPort location;
     private Map<KeyExtent,List<PathSize>> assignmentsPerTablet;
-    
+
     public AssignmentTask(Map<Path,List<KeyExtent>> assignmentFailures, String tableName, String location, Map<KeyExtent,List<PathSize>> assignmentsPerTablet) {
       this.assignmentFailures = assignmentFailures;
       this.location = HostAndPort.fromString(location);
       this.assignmentsPerTablet = assignmentsPerTablet;
     }
-    
+
     private void handleFailures(Collection<KeyExtent> failures, String message) {
       for (KeyExtent ke : failures) {
         List<PathSize> mapFiles = assignmentsPerTablet.get(ke);
@@ -457,24 +457,24 @@ public class BulkImporter {
               existingFailures = new ArrayList<KeyExtent>();
               assignmentFailures.put(pathSize.path, existingFailures);
             }
-            
+
             existingFailures.add(ke);
           }
         }
-        
+
         log.info("Could not assign " + mapFiles.size() + " map files to tablet " + ke + " because : " + message + ".  Will retry ...");
       }
     }
-    
+
     @Override
     public void run() {
       HashSet<Path> uniqMapFiles = new HashSet<Path>();
       for (List<PathSize> mapFiles : assignmentsPerTablet.values())
         for (PathSize ps : mapFiles)
           uniqMapFiles.add(ps.path);
-      
+
       log.debug("Assigning " + uniqMapFiles.size() + " map files to " + assignmentsPerTablet.size() + " tablets at " + location);
-      
+
       try {
         List<KeyExtent> failures = assignMapFiles(context, location, assignmentsPerTablet);
         handleFailures(failures, "Not Serving Tablet");
@@ -484,53 +484,53 @@ public class BulkImporter {
         handleFailures(assignmentsPerTablet.keySet(), e.getMessage());
       }
     }
-    
+
   }
-  
+
   private class PathSize {
     public PathSize(Path mapFile, long estSize) {
       this.path = mapFile;
       this.estSize = estSize;
     }
-    
+
     Path path;
     long estSize;
-    
+
     @Override
     public String toString() {
       return path + " " + estSize;
     }
   }
-  
+
   private Map<Path,List<KeyExtent>> assignMapFiles(String tableName, Map<Path,List<AssignmentInfo>> assignments, Map<KeyExtent,String> locations, int numThreads) {
-    
+
     // group assignments by tablet
     Map<KeyExtent,List<PathSize>> assignmentsPerTablet = new TreeMap<KeyExtent,List<PathSize>>();
     for (Entry<Path,List<AssignmentInfo>> entry : assignments.entrySet()) {
       Path mapFile = entry.getKey();
       List<AssignmentInfo> tabletsToAssignMapFileTo = entry.getValue();
-      
+
       for (AssignmentInfo ai : tabletsToAssignMapFileTo) {
         List<PathSize> mapFiles = assignmentsPerTablet.get(ai.ke);
         if (mapFiles == null) {
           mapFiles = new ArrayList<PathSize>();
           assignmentsPerTablet.put(ai.ke, mapFiles);
         }
-        
+
         mapFiles.add(new PathSize(mapFile, ai.estSize));
       }
     }
-    
+
     // group assignments by tabletserver
-    
+
     Map<Path,List<KeyExtent>> assignmentFailures = Collections.synchronizedMap(new TreeMap<Path,List<KeyExtent>>());
-    
+
     TreeMap<String,Map<KeyExtent,List<PathSize>>> assignmentsPerTabletServer = new TreeMap<String,Map<KeyExtent,List<PathSize>>>();
-    
+
     for (Entry<KeyExtent,List<PathSize>> entry : assignmentsPerTablet.entrySet()) {
       KeyExtent ke = entry.getKey();
       String location = locations.get(ke);
-      
+
       if (location == null) {
         for (PathSize pathSize : entry.getValue()) {
           synchronized (assignmentFailures) {
@@ -539,34 +539,34 @@ public class BulkImporter {
               failures = new ArrayList<KeyExtent>();
               assignmentFailures.put(pathSize.path, failures);
             }
-            
+
             failures.add(ke);
           }
         }
-        
+
         log.warn("Could not assign " + entry.getValue().size() + " map files to tablet " + ke + " because it had no location, will retry ...");
-        
+
         continue;
       }
-      
+
       Map<KeyExtent,List<PathSize>> apt = assignmentsPerTabletServer.get(location);
       if (apt == null) {
         apt = new TreeMap<KeyExtent,List<PathSize>>();
         assignmentsPerTabletServer.put(location, apt);
       }
-      
+
       apt.put(entry.getKey(), entry.getValue());
     }
-    
+
     ExecutorService threadPool = Executors.newFixedThreadPool(numThreads, new NamingThreadFactory("submit"));
-    
+
     for (Entry<String,Map<KeyExtent,List<PathSize>>> entry : assignmentsPerTabletServer.entrySet()) {
       String location = entry.getKey();
       threadPool.submit(new AssignmentTask(assignmentFailures, tableName, location, entry.getValue()));
     }
-    
+
     threadPool.shutdown();
-    
+
     while (!threadPool.isTerminated()) {
       try {
         threadPool.awaitTermination(60, TimeUnit.SECONDS);
@@ -575,7 +575,7 @@ public class BulkImporter {
         throw new RuntimeException(e);
       }
     }
-    
+
     return assignmentFailures;
   }
 
@@ -589,16 +589,16 @@ public class BulkImporter {
         for (Entry<KeyExtent,List<PathSize>> entry : assignmentsPerTablet.entrySet()) {
           HashMap<String,org.apache.accumulo.core.data.thrift.MapFileInfo> tabletFiles = new HashMap<String,org.apache.accumulo.core.data.thrift.MapFileInfo>();
           files.put(entry.getKey(), tabletFiles);
-          
+
           for (PathSize pathSize : entry.getValue()) {
             org.apache.accumulo.core.data.thrift.MapFileInfo mfi = new org.apache.accumulo.core.data.thrift.MapFileInfo(pathSize.estSize);
             tabletFiles.put(pathSize.path.toString(), mfi);
           }
         }
-        
+
         log.debug("Asking " + location + " to bulk load " + files);
         List<TKeyExtent> failures = client.bulkImport(Tracer.traceInfo(), context.rpcCreds(), tid, Translator.translate(files, Translators.KET), setTime);
-        
+
         return Translator.translate(failures, Translators.TKET);
       } finally {
         ThriftUtil.returnClient((TServiceClient) client);
@@ -610,11 +610,11 @@ public class BulkImporter {
       throw new AccumuloException(t);
     }
   }
-  
+
   public static List<TabletLocation> findOverlappingTablets(ClientContext context, VolumeManager fs, TabletLocator locator, Path file) throws Exception {
     return findOverlappingTablets(context, fs, locator, file, null, null);
   }
-  
+
   public static List<TabletLocation> findOverlappingTablets(ClientContext context, VolumeManager fs, TabletLocator locator, Path file, KeyExtent failed)
       throws Exception {
     locator.invalidateCache(failed);
@@ -623,9 +623,9 @@ public class BulkImporter {
       start = Range.followingPrefix(start);
     return findOverlappingTablets(context, fs, locator, file, start, failed.getEndRow());
   }
-  
+
   final static byte[] byte0 = {0};
-  
+
   public static List<TabletLocation> findOverlappingTablets(ClientContext context, VolumeManager vm, TabletLocator locator, Path file, Text startRow,
       Text endRow) throws Exception {
     List<TabletLocation> result = new ArrayList<TabletLocation>();
@@ -662,98 +662,98 @@ public class BulkImporter {
     // log.debug(filename + " to be sent to " + result);
     return result;
   }
-  
+
   public static class AssignmentStats {
     private Map<KeyExtent,Integer> counts;
     private int numUniqueMapFiles;
     private Map<Path,List<KeyExtent>> completeFailures = null;
     private Set<Path> failedFailures = null;
-    
+
     AssignmentStats(int fileCount) {
       counts = new HashMap<KeyExtent,Integer>();
       numUniqueMapFiles = fileCount;
     }
-    
+
     void attemptingAssignments(Map<Path,List<TabletLocation>> assignments) {
       for (Entry<Path,List<TabletLocation>> entry : assignments.entrySet()) {
         for (TabletLocation tl : entry.getValue()) {
-          
+
           Integer count = getCount(tl.tablet_extent);
-          
+
           counts.put(tl.tablet_extent, count + 1);
         }
       }
     }
-    
+
     void assignmentsFailed(Map<Path,List<KeyExtent>> assignmentFailures) {
       for (Entry<Path,List<KeyExtent>> entry : assignmentFailures.entrySet()) {
         for (KeyExtent ke : entry.getValue()) {
-          
+
           Integer count = getCount(ke);
-          
+
           counts.put(ke, count - 1);
         }
       }
     }
-    
+
     void assignmentsAbandoned(Map<Path,List<KeyExtent>> completeFailures) {
       this.completeFailures = completeFailures;
     }
-    
+
     void tabletSplit(KeyExtent parent, Collection<KeyExtent> children) {
       Integer count = getCount(parent);
-      
+
       counts.remove(parent);
-      
+
       for (KeyExtent keyExtent : children)
         counts.put(keyExtent, count);
     }
-    
+
     private Integer getCount(KeyExtent parent) {
       Integer count = counts.get(parent);
-      
+
       if (count == null) {
         count = 0;
       }
       return count;
     }
-    
+
     void unrecoveredMapFiles(Set<Path> failedFailures) {
       this.failedFailures = failedFailures;
     }
-    
+
     @Override
     public String toString() {
       StringBuilder sb = new StringBuilder();
       int totalAssignments = 0;
       int tabletsImportedTo = 0;
-      
+
       int min = Integer.MAX_VALUE, max = Integer.MIN_VALUE;
-      
+
       for (Entry<KeyExtent,Integer> entry : counts.entrySet()) {
         totalAssignments += entry.getValue();
         if (entry.getValue() > 0)
           tabletsImportedTo++;
-        
+
         if (entry.getValue() < min)
           min = entry.getValue();
-        
+
         if (entry.getValue() > max)
           max = entry.getValue();
       }
-      
+
       double stddev = 0;
-      
+
       for (Entry<KeyExtent,Integer> entry : counts.entrySet())
         stddev += Math.pow(entry.getValue() - totalAssignments / (double) counts.size(), 2);
-      
+
       stddev = stddev / counts.size();
       stddev = Math.sqrt(stddev);
-      
+
       Set<KeyExtent> failedTablets = new HashSet<KeyExtent>();
       for (List<KeyExtent> ft : completeFailures.values())
         failedTablets.addAll(ft);
-      
+
       sb.append("BULK IMPORT ASSIGNMENT STATISTICS\n");
       sb.append(String.format("# of map files            : %,10d%n", numUniqueMapFiles));
       sb.append(String.format("# map files with failures : %,10d %6.2f%s%n", completeFailures.size(), completeFailures.size() * 100.0 / numUniqueMapFiles, "%"));
@@ -767,5 +767,5 @@ public class BulkImporter {
       return sb.toString();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/client/HdfsZooInstance.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/client/HdfsZooInstance.java b/server/base/src/main/java/org/apache/accumulo/server/client/HdfsZooInstance.java
index 4ab9f90..3175fff 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/client/HdfsZooInstance.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/client/HdfsZooInstance.java
@@ -58,7 +58,7 @@ import com.google.common.base.Joiner;
 
 /**
  * An implementation of Instance that looks in HDFS and ZooKeeper to find the master and root tablet location.
- * 
+ *
  */
 public class HdfsZooInstance implements Instance {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/conf/ConfigSanityCheck.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/conf/ConfigSanityCheck.java b/server/base/src/main/java/org/apache/accumulo/server/conf/ConfigSanityCheck.java
index b90051f..658d249 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/conf/ConfigSanityCheck.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/conf/ConfigSanityCheck.java
@@ -19,9 +19,9 @@ package org.apache.accumulo.server.conf;
 import org.apache.accumulo.server.client.HdfsZooInstance;
 
 public class ConfigSanityCheck {
-  
+
   public static void main(String[] args) {
     new ServerConfigurationFactory(HdfsZooInstance.getInstance()).getConfiguration();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/conf/NamespaceConfWatcher.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/conf/NamespaceConfWatcher.java b/server/base/src/main/java/org/apache/accumulo/server/conf/NamespaceConfWatcher.java
index 65d9388..945e904 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/conf/NamespaceConfWatcher.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/conf/NamespaceConfWatcher.java
@@ -48,9 +48,8 @@ class NamespaceConfWatcher implements Watcher {
   }
 
   static String toString(WatchedEvent event) {
-    return new StringBuilder("{path=").append(event.getPath()).append(",state=")
-      .append(event.getState()).append(",type=").append(event.getType())
-      .append("}").toString();
+    return new StringBuilder("{path=").append(event.getPath()).append(",state=").append(event.getState()).append(",type=").append(event.getType()).append("}")
+        .toString();
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfiguration.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfiguration.java b/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfiguration.java
index 342aebe..f2b2042 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfiguration.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfiguration.java
@@ -21,7 +21,7 @@ import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.core.data.KeyExtent;
 
 public abstract class ServerConfiguration {
-  
+
   abstract public TableConfiguration getTableConfiguration(String tableId);
 
   abstract public TableConfiguration getTableConfiguration(KeyExtent extent);
@@ -31,5 +31,5 @@ public abstract class ServerConfiguration {
   abstract public AccumuloConfiguration getConfiguration();
 
   abstract public Instance getInstance();
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfigurationFactory.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfigurationFactory.java b/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfigurationFactory.java
index 128f74e..2ec9ba1 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfigurationFactory.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/conf/ServerConfigurationFactory.java
@@ -154,14 +154,14 @@ public class ServerConfigurationFactory extends ServerConfiguration {
     synchronized (tableConfigs) {
       conf = tableConfigs.get(instanceID).get(tableId);
     }
-    // can't hold the lock during the construction and validation of the config, 
+    // can't hold the lock during the construction and validation of the config,
     // which may result in creating multiple objects for the same id, but that's ok.
     if (conf == null && Tables.exists(instance, tableId)) {
-        conf = new TableConfiguration(instance, tableId, getNamespaceConfigurationForTable(tableId));
-        ConfigSanityCheck.validate(conf);
-        synchronized (tableConfigs) {
-          tableConfigs.get(instanceID).put(tableId, conf);
-        }
+      conf = new TableConfiguration(instance, tableId, getNamespaceConfigurationForTable(tableId));
+      ConfigSanityCheck.validate(conf);
+      synchronized (tableConfigs) {
+        tableConfigs.get(instanceID).put(tableId, conf);
+      }
     }
     return conf;
   }
@@ -177,7 +177,7 @@ public class ServerConfigurationFactory extends ServerConfiguration {
     synchronized (tableParentConfigs) {
       conf = tableParentConfigs.get(instanceID).get(tableId);
     }
-    // can't hold the lock during the construction and validation of the config, 
+    // can't hold the lock during the construction and validation of the config,
     // which may result in creating multiple objects for the same id, but that's ok.
     if (conf == null) {
       // changed - include instance in constructor call
@@ -194,7 +194,7 @@ public class ServerConfigurationFactory extends ServerConfiguration {
   public NamespaceConfiguration getNamespaceConfiguration(String namespaceId) {
     checkPermissions();
     NamespaceConfiguration conf;
-    // can't hold the lock during the construction and validation of the config, 
+    // can't hold the lock during the construction and validation of the config,
     // which may result in creating multiple objects for the same id, but that's ok.
     synchronized (namespaceConfigs) {
       conf = namespaceConfigs.get(instanceID).get(namespaceId);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/conf/TableConfWatcher.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/conf/TableConfWatcher.java b/server/base/src/main/java/org/apache/accumulo/server/conf/TableConfWatcher.java
index b657056..3c8d45d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/conf/TableConfWatcher.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/conf/TableConfWatcher.java
@@ -29,7 +29,7 @@ class TableConfWatcher implements Watcher {
     Logger.getLogger("org.apache.zookeeper").setLevel(Level.WARN);
     Logger.getLogger("org.apache.hadoop.io.compress").setLevel(Level.WARN);
   }
-  
+
   private static final Logger log = Logger.getLogger(TableConfWatcher.class);
   private final Instance instance;
   private final String tablesPrefix;
@@ -46,9 +46,8 @@ class TableConfWatcher implements Watcher {
   }
 
   static String toString(WatchedEvent event) {
-    return new StringBuilder("{path=").append(event.getPath()).append(",state=")
-      .append(event.getState()).append(",type=").append(event.getType())
-      .append("}").toString();
+    return new StringBuilder("{path=").append(event.getPath()).append(",state=").append(event.getState()).append(",type=").append(event.getType()).append("}")
+        .toString();
   }
 
   @Override
@@ -56,10 +55,10 @@ class TableConfWatcher implements Watcher {
     String path = event.getPath();
     if (log.isTraceEnabled())
       log.trace("WatchedEvent : " + toString(event));
-    
+
     String tableId = null;
     String key = null;
-    
+
     if (path != null) {
       if (path.startsWith(tablesPrefix)) {
         tableId = path.substring(tablesPrefix.length());
@@ -69,13 +68,13 @@ class TableConfWatcher implements Watcher {
             key = path.substring((tablesPrefix + tableId + Constants.ZTABLE_CONF + "/").length());
         }
       }
-      
+
       if (tableId == null) {
         log.warn("Zookeeper told me about a path I was not watching: " + path + ", event " + toString(event));
         return;
       }
     }
-    
+
     switch (event.getType()) {
       case NodeDataChanged:
         if (log.isTraceEnabled())

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/conf/TableParentConfiguration.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/conf/TableParentConfiguration.java b/server/base/src/main/java/org/apache/accumulo/server/conf/TableParentConfiguration.java
index 2a2bfce..bd2e5ab 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/conf/TableParentConfiguration.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/conf/TableParentConfiguration.java
@@ -32,6 +32,7 @@ public class TableParentConfiguration extends NamespaceConfiguration {
     this.tableId = tableId;
     this.namespaceId = getNamespaceId();
   }
+
   public TableParentConfiguration(String tableId, Instance inst, AccumuloConfiguration parent) {
     super(null, inst, parent);
     this.tableId = tableId;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/data/ServerColumnUpdate.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/data/ServerColumnUpdate.java b/server/base/src/main/java/org/apache/accumulo/server/data/ServerColumnUpdate.java
index af992a6..f891065 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/data/ServerColumnUpdate.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/data/ServerColumnUpdate.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.server.data;
 import org.apache.accumulo.core.data.ColumnUpdate;
 
 public class ServerColumnUpdate extends ColumnUpdate {
-  
+
   ServerMutation parent;
 
   public ServerColumnUpdate(byte[] cf, byte[] cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val, ServerMutation serverMutation) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/data/ServerMutation.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/data/ServerMutation.java b/server/base/src/main/java/org/apache/accumulo/server/data/ServerMutation.java
index 389cc33..cb4fa97 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/data/ServerMutation.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/data/ServerMutation.java
@@ -31,7 +31,7 @@ import org.apache.hadoop.io.WritableUtils;
  */
 public class ServerMutation extends Mutation {
   private long systemTime = 0l;
-  
+
   public ServerMutation(TMutation tmutation) {
     super(tmutation);
   }
@@ -40,8 +40,7 @@ public class ServerMutation extends Mutation {
     super(key);
   }
 
-  public ServerMutation() {
-  }
+  public ServerMutation() {}
 
   protected void droppingOldTimestamp(long ts) {
     this.systemTime = ts;
@@ -54,7 +53,7 @@ public class ServerMutation extends Mutation {
     if (getSerializedFormat() == SERIALIZED_FORMAT.VERSION2)
       systemTime = WritableUtils.readVLong(in);
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     super.write(out);
@@ -64,7 +63,7 @@ public class ServerMutation extends Mutation {
   public void setSystemTimestamp(long v) {
     this.systemTime = v;
   }
-  
+
   public long getSystemTimestamp() {
     return this.systemTime;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/FileRef.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/FileRef.java b/server/base/src/main/java/org/apache/accumulo/server/fs/FileRef.java
index c0bb275..eb42a11 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/FileRef.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/FileRef.java
@@ -21,40 +21,37 @@ import org.apache.accumulo.server.fs.VolumeManager.FileType;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 
-
 /**
- * This is a glue object, to convert short file references to long references.
- * The metadata may contain old relative file references.  This class keeps 
- * track of the short file reference, so it can be removed properly from the
- * metadata tables.
+ * This is a glue object, to convert short file references to long references. The metadata may contain old relative file references. This class keeps track of
+ * the short file reference, so it can be removed properly from the metadata tables.
  */
 public class FileRef implements Comparable<FileRef> {
   private String metaReference; // something like ../2/d-00000/A00001.rf
   private Path fullReference; // something like hdfs://nn:9001/accumulo/tables/2/d-00000/A00001.rf
   private Path suffix;
-  
+
   public FileRef(VolumeManager fs, Key key) {
     this(key.getColumnQualifier().toString(), fs.getFullPath(key));
   }
-  
+
   public FileRef(String metaReference, Path fullReference) {
     this.metaReference = metaReference;
     this.fullReference = fullReference;
     this.suffix = extractSuffix(fullReference);
   }
-  
+
   public FileRef(String path) {
     this(path, new Path(path));
   }
-  
+
   public String toString() {
     return fullReference.toString();
   }
-  
+
   public Path path() {
     return fullReference;
   }
-  
+
   public Text meta() {
     return new Text(metaReference);
   }
@@ -89,10 +86,9 @@ public class FileRef implements Comparable<FileRef> {
   @Override
   public boolean equals(Object obj) {
     if (obj instanceof FileRef) {
-      return compareTo((FileRef)obj) == 0;
+      return compareTo((FileRef) obj) == 0;
     }
     return false;
   }
-  
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/PerTableVolumeChooser.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/PerTableVolumeChooser.java b/server/base/src/main/java/org/apache/accumulo/server/fs/PerTableVolumeChooser.java
index a579cc8..e51df03 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/PerTableVolumeChooser.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/PerTableVolumeChooser.java
@@ -24,15 +24,15 @@ import org.apache.accumulo.server.conf.ServerConfigurationFactory;
 import org.apache.accumulo.server.conf.TableConfiguration;
 
 /**
- * A {@link VolumeChooser} that delegates to another volume chooser based on the presence of an experimental table
- * property, {@link Property#TABLE_VOLUME_CHOOSER}. If it isn't found, defaults back to {@link RandomVolumeChooser}.
+ * A {@link VolumeChooser} that delegates to another volume chooser based on the presence of an experimental table property,
+ * {@link Property#TABLE_VOLUME_CHOOSER}. If it isn't found, defaults back to {@link RandomVolumeChooser}.
  */
 public class PerTableVolumeChooser implements VolumeChooser {
 
   private final VolumeChooser fallbackVolumeChooser = new RandomVolumeChooser();
   // TODO Add hint of expected size to construction, see ACCUMULO-3410
   /* Track VolumeChooser instances so they can keep state. */
-  private final ConcurrentHashMap<String, VolumeChooser> tableSpecificChooser = new ConcurrentHashMap<String, VolumeChooser>();
+  private final ConcurrentHashMap<String,VolumeChooser> tableSpecificChooser = new ConcurrentHashMap<String,VolumeChooser>();
   // TODO has to be lazily initialized currently because of the reliance on HdfsZooInstance. see ACCUMULO-3411
   private volatile ServerConfigurationFactory serverConfs;
 
@@ -60,7 +60,7 @@ public class PerTableVolumeChooser implements VolumeChooser {
         // the configuration for this table's chooser has been updated. In the case of failure to instantiate we'll repeat here next call.
         // TODO stricter definition of when the updated property is used, ref ACCUMULO-3412
         VolumeChooser temp = Property.createTableInstanceFromPropertyName(tableConf, Property.TABLE_VOLUME_CHOOSER, VolumeChooser.class, fallbackVolumeChooser);
-        VolumeChooser last  = tableSpecificChooser.replace(env.getTableId(), temp);
+        VolumeChooser last = tableSpecificChooser.replace(env.getTableId(), temp);
         if (chooser.equals(last)) {
           chooser = temp;
         } else {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/PreferredVolumeChooser.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/PreferredVolumeChooser.java b/server/base/src/main/java/org/apache/accumulo/server/fs/PreferredVolumeChooser.java
index 4ddf9bb..68621fb 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/PreferredVolumeChooser.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/PreferredVolumeChooser.java
@@ -37,10 +37,9 @@ import org.apache.commons.lang.StringUtils;
 import org.apache.log4j.Logger;
 
 /**
- * A {@link RandomVolumeChooser} that limits its choices from a given set of options to the subset of those options preferred for a
- * particular table. Defaults to selecting from all of the options presented. Can be customized via the table property
- * {@value #PREFERRED_VOLUMES_CUSTOM_KEY}, which should contain a comma separated list of {@link Volume} URIs. Note that both the property
- * name and the format of its value are specific to this particular implementation.
+ * A {@link RandomVolumeChooser} that limits its choices from a given set of options to the subset of those options preferred for a particular table. Defaults
+ * to selecting from all of the options presented. Can be customized via the table property {@value #PREFERRED_VOLUMES_CUSTOM_KEY}, which should contain a comma
+ * separated list of {@link Volume} URIs. Note that both the property name and the format of its value are specific to this particular implementation.
  */
 public class PreferredVolumeChooser extends RandomVolumeChooser implements VolumeChooser {
   private static final Logger log = Logger.getLogger(PreferredVolumeChooser.class);
@@ -55,7 +54,7 @@ public class PreferredVolumeChooser extends RandomVolumeChooser implements Volum
   };
 
   @SuppressWarnings("unchecked")
-  private final Map<String, Set<String>> parsedPreferredVolumes = Collections.synchronizedMap(new LRUMap(1000));
+  private final Map<String,Set<String>> parsedPreferredVolumes = Collections.synchronizedMap(new LRUMap(1000));
   // TODO has to be lazily initialized currently because of the reliance on HdfsZooInstance. see ACCUMULO-3411
   private volatile ServerConfigurationFactory serverConfs;
 
@@ -73,7 +72,7 @@ public class PreferredVolumeChooser extends RandomVolumeChooser implements Volum
       serverConfs = localConf;
     }
     TableConfiguration tableConf = localConf.getTableConfiguration(env.getTableId());
-    final Map<String,String> props = new HashMap<String, String>();
+    final Map<String,String> props = new HashMap<String,String>();
     tableConf.getProperties(props, PREFERRED_VOLUMES_FILTER);
     if (props.isEmpty()) {
       log.warn("No preferred volumes specified. Defaulting to randomly choosing from instance volumes");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/ViewFSUtils.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/ViewFSUtils.java b/server/base/src/main/java/org/apache/accumulo/server/fs/ViewFSUtils.java
index 34912f3..73535d9 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/ViewFSUtils.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/ViewFSUtils.java
@@ -25,7 +25,7 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
 /**
- * 
+ *
  */
 public class ViewFSUtils {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeChooser.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeChooser.java b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeChooser.java
index 9865512..8b70721 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeChooser.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeChooser.java
@@ -19,8 +19,8 @@ package org.apache.accumulo.server.fs;
 import org.apache.accumulo.core.volume.Volume;
 
 /**
- * Helper used by {@link VolumeManager}s to select from a set of {@link Volume} URIs.
- * N.B. implemenations must be threadsafe. VolumeChooser.equals will be used for internal caching.
+ * Helper used by {@link VolumeManager}s to select from a set of {@link Volume} URIs. N.B. implemenations must be threadsafe. VolumeChooser.equals will be used
+ * for internal caching.
  */
 public interface VolumeChooser {
   String choose(VolumeChooserEnvironment env, String[] options);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManager.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManager.java b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManager.java
index e2353d4..e761e4f 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManager.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManager.java
@@ -32,8 +32,7 @@ import com.google.common.base.Optional;
 
 /**
  * A wrapper around multiple hadoop FileSystem objects, which are assumed to be different volumes. This also concentrates a bunch of meta-operations like
- * waiting for SAFE_MODE, and closing WALs.
- * N.B. implementations must be thread safe.
+ * waiting for SAFE_MODE, and closing WALs. N.B. implementations must be thread safe.
  */
 public interface VolumeManager {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManagerImpl.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManagerImpl.java b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManagerImpl.java
index 8202d27..4423495 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManagerImpl.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeManagerImpl.java
@@ -582,8 +582,8 @@ public class VolumeManagerImpl implements VolumeManager {
     final VolumeChooserEnvironment env = new VolumeChooserEnvironment(tableId);
     final String choice = chooser.choose(env, options);
     if (!(ArrayUtils.contains(options, choice))) {
-      log.error("The configured volume chooser, '" +  chooser.getClass() + "', or one of its delegates returned a volume not in the set of options provided; " +
-          "will continue by relying on a RandomVolumeChooser. You should investigate and correct the named chooser.");
+      log.error("The configured volume chooser, '" + chooser.getClass() + "', or one of its delegates returned a volume not in the set of options provided; "
+          + "will continue by relying on a RandomVolumeChooser. You should investigate and correct the named chooser.");
       return failsafeChooser.choose(env, options);
     }
     return choice;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeUtil.java b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeUtil.java
index e229209..d40106d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/fs/VolumeUtil.java
@@ -265,8 +265,8 @@ public class VolumeUtil {
       throw new IllegalArgumentException("Unexpected table dir " + dir);
     }
 
-    Path newDir = new Path(vm.choose(Optional.of(extent.getTableId().toString()), ServerConstants.getBaseUris()) + Path.SEPARATOR + ServerConstants.TABLE_DIR +
-        Path.SEPARATOR + dir.getParent().getName() + Path.SEPARATOR + dir.getName());
+    Path newDir = new Path(vm.choose(Optional.of(extent.getTableId().toString()), ServerConstants.getBaseUris()) + Path.SEPARATOR + ServerConstants.TABLE_DIR
+        + Path.SEPARATOR + dir.getParent().getName() + Path.SEPARATOR + dir.getName());
 
     log.info("Updating directory for " + extent + " from " + dir + " to " + newDir);
     if (extent.isRootTablet()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java b/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
index 046cfb5..e14ef72 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
@@ -89,11 +89,11 @@ import org.apache.accumulo.server.fs.VolumeManagerImpl;
 import org.apache.accumulo.server.iterators.MetadataBulkLoadFilter;
 import org.apache.accumulo.server.replication.StatusCombiner;
 import org.apache.accumulo.server.security.AuditedSecurityOperation;
+import org.apache.accumulo.server.security.SecurityUtil;
 import org.apache.accumulo.server.tables.TableManager;
 import org.apache.accumulo.server.tablets.TabletTime;
 import org.apache.accumulo.server.util.ReplicationTableUtil;
 import org.apache.accumulo.server.util.TablePropUtil;
-import org.apache.accumulo.server.security.SecurityUtil;
 import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileStatus;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilter.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilter.java b/server/base/src/main/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilter.java
index 536c617..c689fd3 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilter.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/iterators/MetadataBulkLoadFilter.java
@@ -34,23 +34,23 @@ import org.apache.log4j.Logger;
 
 /**
  * A special iterator for the metadata table that removes inactive bulk load flags
- * 
+ *
  */
 public class MetadataBulkLoadFilter extends Filter {
   private static final Logger log = Logger.getLogger(MetadataBulkLoadFilter.class);
-  
+
   enum Status {
     ACTIVE, INACTIVE
   }
-  
+
   Map<Long,Status> bulkTxStatusCache;
   Arbitrator arbitrator;
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     if (!k.isDeleted() && k.compareColumnFamily(TabletsSection.BulkFileColumnFamily.NAME) == 0) {
       long txid = Long.valueOf(v.toString());
-      
+
       Status status = bulkTxStatusCache.get(txid);
       if (status == null) {
         try {
@@ -63,28 +63,28 @@ public class MetadataBulkLoadFilter extends Filter {
           status = Status.ACTIVE;
           log.error(e, e);
         }
-        
+
         bulkTxStatusCache.put(txid, status);
       }
-      
+
       return status == Status.ACTIVE;
     }
-    
+
     return true;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
-    
+
     if (env.getIteratorScope() == IteratorScope.scan) {
       throw new IOException("This iterator not intended for use at scan time");
     }
-    
+
     bulkTxStatusCache = new HashMap<Long,MetadataBulkLoadFilter.Status>();
     arbitrator = getArbitrator();
   }
-  
+
   protected Arbitrator getArbitrator() {
     return new ZooArbitrator();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/log/SortedLogState.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/log/SortedLogState.java b/server/base/src/main/java/org/apache/accumulo/server/log/SortedLogState.java
index 49d744e..c0580ac 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/log/SortedLogState.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/log/SortedLogState.java
@@ -19,8 +19,8 @@ package org.apache.accumulo.server.log;
 import org.apache.hadoop.fs.Path;
 
 /**
- * A file is written in the destination directory for the sorting of write-ahead logs that need recovering. The value of {@link #getMarker()} is the name of the file
- * that will exist in the sorted output directory.
+ * A file is written in the destination directory for the sorting of write-ahead logs that need recovering. The value of {@link #getMarker()} is the name of the
+ * file that will exist in the sorted output directory.
  */
 public enum SortedLogState {
   FINISHED("finished"), FAILED("failed");
@@ -54,7 +54,7 @@ public enum SortedLogState {
   public static Path getFailedMarkerPath(String rootPath) {
     return new Path(rootPath, FAILED.getMarker());
   }
-  
+
   public static Path getFailedMarkerPath(Path rootPath) {
     return new Path(rootPath, FAILED.getMarker());
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancer.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancer.java b/server/base/src/main/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancer.java
index 1820d8e..7d11066 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancer.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/balancer/ChaoticLoadBalancer.java
@@ -45,9 +45,9 @@ import org.apache.thrift.TException;
  */
 public class ChaoticLoadBalancer extends TabletBalancer {
   private static final Logger log = Logger.getLogger(ChaoticLoadBalancer.class);
-  
+
   Random r = new Random();
-  
+
   @Override
   public void getAssignments(SortedMap<TServerInstance,TabletServerStatus> current, Map<KeyExtent,TServerInstance> unassigned,
       Map<KeyExtent,TServerInstance> assignments) {
@@ -65,7 +65,7 @@ public class ChaoticLoadBalancer extends TabletBalancer {
         toAssign.put(e.getKey(), avg - numTablets);
       }
     }
-    
+
     for (KeyExtent ke : unassigned.keySet()) {
       int index = r.nextInt(tServerArray.size());
       TServerInstance dest = tServerArray.get(index);
@@ -79,7 +79,7 @@ public class ChaoticLoadBalancer extends TabletBalancer {
       }
     }
   }
-  
+
   protected final OutstandingMigrations outstandingMigrations = new OutstandingMigrations(log);
 
   /**
@@ -111,7 +111,7 @@ public class ChaoticLoadBalancer extends TabletBalancer {
     // totalTablets is fuzzy due to asynchronicity of the stats
     // *1.2 to handle fuzziness, and prevent locking for 'perfect' balancing scenarios
     long avg = (long) Math.ceil(((double) totalTablets) / current.size() * 1.2);
-    
+
     for (Entry<TServerInstance,TabletServerStatus> e : current.entrySet()) {
       for (String table : e.getValue().getTableMap().keySet()) {
         if (!moveMetadata && MetadataTable.NAME.equals(table))
@@ -128,7 +128,7 @@ public class ChaoticLoadBalancer extends TabletBalancer {
               underCapacityTServer.remove(index);
             if (numTablets.put(e.getKey(), numTablets.get(e.getKey()) - 1) <= avg && !underCapacityTServer.contains(e.getKey()))
               underCapacityTServer.add(e.getKey());
-            
+
             // We can get some craziness with only 1 tserver, so lets make sure there's always an option!
             if (underCapacityTServer.isEmpty())
               underCapacityTServer.addAll(numTablets.keySet());
@@ -142,17 +142,16 @@ public class ChaoticLoadBalancer extends TabletBalancer {
         }
       }
     }
-    
+
     return 100;
   }
-  
+
   @Override
   public void init(ServerConfiguration conf) {
     throw new NotImplementedException();
   }
 
   @Override
-  public void init(ServerConfigurationFactory conf) {
-  }
-  
+  public void init(ServerConfigurationFactory conf) {}
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/balancer/TabletBalancer.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/balancer/TabletBalancer.java b/server/base/src/main/java/org/apache/accumulo/server/master/balancer/TabletBalancer.java
index 9822d0f..ecf59b3 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/balancer/TabletBalancer.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/balancer/TabletBalancer.java
@@ -43,11 +43,11 @@ import org.apache.thrift.transport.TTransportException;
 import com.google.common.collect.Iterables;
 
 public abstract class TabletBalancer {
-  
+
   private static final Logger log = Logger.getLogger(TabletBalancer.class);
-  
+
   protected ServerConfigurationFactory configuration;
-  
+
   protected AccumuloServerContext context;
 
   /**
@@ -64,7 +64,7 @@ public abstract class TabletBalancer {
 
   /**
    * Assign tablets to tablet servers. This method is called whenever the master finds tablets that are unassigned.
-   * 
+   *
    * @param current
    *          The current table-summary state of all the online tablet servers. Read-only. The TabletServerStatus for each server may be null if the tablet
    *          server has not yet responded to a recent request for status.
@@ -75,14 +75,14 @@ public abstract class TabletBalancer {
    */
   abstract public void getAssignments(SortedMap<TServerInstance,TabletServerStatus> current, Map<KeyExtent,TServerInstance> unassigned,
       Map<KeyExtent,TServerInstance> assignments);
-  
+
   /**
    * Ask the balancer if any migrations are necessary.
-   * 
+   *
    * If the balancer is going to self-abort due to some environmental constraint (e.g. it requires some minimum number of tservers, or a maximum number of
    * outstanding migrations), it should issue a log message to alert operators. The message should be at WARN normally and at ERROR if the balancer knows that
    * the problem can not self correct. It should not issue these messages more than once a minute.
-   * 
+   *
    * @param current
    *          The current table-summary state of all the online tablet servers. Read-only.
    * @param migrations
@@ -90,7 +90,7 @@ public abstract class TabletBalancer {
    * @param migrationsOut
    *          new migrations to perform; should not contain tablets in the current set of migrations. Write-only.
    * @return the time, in milliseconds, to wait before re-balancing.
-   * 
+   *
    *         This method will not be called when there are unassigned tablets.
    */
   public abstract long balance(SortedMap<TServerInstance,TabletServerStatus> current, Set<KeyExtent> migrations, List<TabletMigration> migrationsOut);
@@ -102,25 +102,24 @@ public abstract class TabletBalancer {
   protected static final long TIME_BETWEEN_BALANCER_WARNINGS = 60 * ONE_SECOND;
 
   /**
-   * A deferred call descendent TabletBalancers use to log why they can't continue.
-   * The call is deferred so that TabletBalancer can limit how often messages happen.
+   * A deferred call descendent TabletBalancers use to log why they can't continue. The call is deferred so that TabletBalancer can limit how often messages
+   * happen.
    *
    * Implementations should be reused as much as possible.
    *
-   * Be sure to pass in a properly scoped Logger instance so that messages indicate
-   * what part of the system is having trouble.
+   * Be sure to pass in a properly scoped Logger instance so that messages indicate what part of the system is having trouble.
    */
   protected static abstract class BalancerProblem implements Runnable {
     protected final Logger balancerLog;
+
     public BalancerProblem(Logger logger) {
       balancerLog = logger;
     }
   }
 
   /**
-   * If a TabletBalancer requires active tservers, it should use this problem to indicate when there are none.
-   * NoTservers is safe to share with anyone who uses the same Logger. TabletBalancers should have a single
-   * static instance.
+   * If a TabletBalancer requires active tservers, it should use this problem to indicate when there are none. NoTservers is safe to share with anyone who uses
+   * the same Logger. TabletBalancers should have a single static instance.
    */
   protected static class NoTservers extends BalancerProblem {
     public NoTservers(Logger logger) {
@@ -134,14 +133,12 @@ public abstract class TabletBalancer {
   }
 
   /**
-   * If a TabletBalancer only balances when there are no outstanding migrations, it should use this problem
-   * to indicate when they exist.
+   * If a TabletBalancer only balances when there are no outstanding migrations, it should use this problem to indicate when they exist.
    *
-   * Iff a TabletBalancer makes use of the migrations member to provide samples, then OutstandingMigrations
-   * is not thread safe.
+   * Iff a TabletBalancer makes use of the migrations member to provide samples, then OutstandingMigrations is not thread safe.
    */
   protected static class OutstandingMigrations extends BalancerProblem {
-    public Set<KeyExtent> migrations = Collections.<KeyExtent>emptySet();
+    public Set<KeyExtent> migrations = Collections.<KeyExtent> emptySet();
 
     public OutstandingMigrations(Logger logger) {
       super(logger);
@@ -156,8 +153,8 @@ public abstract class TabletBalancer {
   }
 
   /**
-   * Warn that a Balancer can't work because of some external restriction.
-   * Will not call the provided logging handler  more often than TIME_BETWEEN_BALANCER_WARNINGS
+   * Warn that a Balancer can't work because of some external restriction. Will not call the provided logging handler more often than
+   * TIME_BETWEEN_BALANCER_WARNINGS
    */
   protected void constraintNotMet(BalancerProblem cause) {
     if (!stuck) {
@@ -177,11 +174,11 @@ public abstract class TabletBalancer {
   protected void resetBalancerErrors() {
     stuck = false;
   }
-  
+
   /**
    * Fetch the tablets for the given table by asking the tablet server. Useful if your balance strategy needs details at the tablet level to decide what tablets
    * to move.
-   * 
+   *
    * @param tserver
    *          The tablet server to ask.
    * @param tableId
@@ -204,14 +201,14 @@ public abstract class TabletBalancer {
     }
     return null;
   }
-  
+
   /**
    * Utility to ensure that the migrations from balance() are consistent:
    * <ul>
    * <li>Tablet objects are not null
    * <li>Source and destination tablet servers are not null and current
    * </ul>
-   * 
+   *
    * @return A list of TabletMigration object that passed sanity checks.
    */
   public static List<TabletMigration> checkMigrationSanity(Set<TServerInstance> current, List<TabletMigration> migrations) {
@@ -241,5 +238,5 @@ public abstract class TabletBalancer {
     }
     return result;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/recovery/RecoveryPath.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/recovery/RecoveryPath.java b/server/base/src/main/java/org/apache/accumulo/server/master/recovery/RecoveryPath.java
index 4a6638a..cc9ac3e 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/recovery/RecoveryPath.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/recovery/RecoveryPath.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.server.fs.VolumeManager.FileType;
 import org.apache.hadoop.fs.Path;
 
 /**
- * 
+ *
  */
 public class RecoveryPath {
 
@@ -39,21 +39,21 @@ public class RecoveryPath {
         // drop server
         walPath = walPath.getParent();
       }
-  
+
       if (!walPath.getName().equals(FileType.WAL.getDirectory()))
         throw new IllegalArgumentException("Bad path " + walPath);
-  
+
       // drop wal
       walPath = walPath.getParent();
-  
+
       walPath = new Path(walPath, FileType.RECOVERY.getDirectory());
       walPath = new Path(walPath, uuid);
 
       return walPath;
     }
-  
+
     throw new IllegalArgumentException("Bad path " + walPath);
-  
+
   }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/Assignment.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/Assignment.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/Assignment.java
index 40b7a93..f1a9b3f 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/Assignment.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/Assignment.java
@@ -21,7 +21,7 @@ import org.apache.accumulo.core.data.KeyExtent;
 public class Assignment {
   public KeyExtent tablet;
   public TServerInstance server;
-  
+
   public Assignment(KeyExtent tablet, TServerInstance server) {
     this.tablet = tablet;
     this.server = server;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/ClosableIterator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/ClosableIterator.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/ClosableIterator.java
index 18af9ed..3e77d93 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/ClosableIterator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/ClosableIterator.java
@@ -19,5 +19,4 @@ package org.apache.accumulo.server.master.state;
 import java.io.Closeable;
 import java.util.Iterator;
 
-public interface ClosableIterator<T> extends Iterator<T>, Closeable {
-}
+public interface ClosableIterator<T> extends Iterator<T>, Closeable {}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/CurrentState.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/CurrentState.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/CurrentState.java
index f4d98bf..501d66a 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/CurrentState.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/CurrentState.java
@@ -20,11 +20,11 @@ import java.util.Collection;
 import java.util.Set;
 
 public interface CurrentState {
-  
+
   Set<String> onlineTables();
-  
+
   Set<TServerInstance> onlineTabletServers();
-  
+
   Collection<MergeInfo> merges();
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/DeadServerList.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/DeadServerList.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/DeadServerList.java
index 3f7f167..9a2441b 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/DeadServerList.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/DeadServerList.java
@@ -27,13 +27,13 @@ import org.apache.accumulo.fate.zookeeper.ZooUtil.NodeExistsPolicy;
 import org.apache.accumulo.fate.zookeeper.ZooUtil.NodeMissingPolicy;
 import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.apache.log4j.Logger;
-import org.apache.zookeeper.data.Stat;
 import org.apache.zookeeper.KeeperException.NoNodeException;
+import org.apache.zookeeper.data.Stat;
 
 public class DeadServerList {
   private static final Logger log = Logger.getLogger(DeadServerList.class);
   private final String path;
-  
+
   public DeadServerList(String path) {
     this.path = path;
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
@@ -43,7 +43,7 @@ public class DeadServerList {
       log.error("Unable to make parent directories of " + path, ex);
     }
   }
-  
+
   public List<DeadServer> getList() {
     List<DeadServer> result = new ArrayList<DeadServer>();
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
@@ -70,7 +70,7 @@ public class DeadServerList {
     }
     return result;
   }
-  
+
   public void delete(String server) {
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
     try {
@@ -79,7 +79,7 @@ public class DeadServerList {
       log.error(ex, ex);
     }
   }
-  
+
   public void post(String server, String cause) {
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStore.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStore.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStore.java
index 10a1311..3276945 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStore.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStore.java
@@ -22,13 +22,13 @@ import java.util.List;
  * An abstract version of ZooKeeper that we can write tests against.
  */
 public interface DistributedStore {
-  
+
   List<String> getChildren(String path) throws DistributedStoreException;
-  
+
   byte[] get(String path) throws DistributedStoreException;
-  
+
   void put(String path, byte[] bs) throws DistributedStoreException;
-  
+
   void remove(String path) throws DistributedStoreException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStoreException.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStoreException.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStoreException.java
index 3d3a725..3290075 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStoreException.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/DistributedStoreException.java
@@ -17,17 +17,17 @@
 package org.apache.accumulo.server.master.state;
 
 public class DistributedStoreException extends Exception {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   public DistributedStoreException(String why) {
     super(why);
   }
-  
+
   public DistributedStoreException(Exception cause) {
     super(cause);
   }
-  
+
   public DistributedStoreException(String why, Exception cause) {
     super(why, cause);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeInfo.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeInfo.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeInfo.java
index 708b1b7..388da05 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeInfo.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeInfo.java
@@ -25,21 +25,21 @@ import org.apache.hadoop.io.Writable;
 
 /**
  * Information about the current merge/rangeDelete.
- * 
+ *
  * Writable to serialize for zookeeper and the Tablet
  */
 public class MergeInfo implements Writable {
-  
+
   public enum Operation {
     MERGE, DELETE,
   }
-  
+
   MergeState state = MergeState.NONE;
   KeyExtent extent;
   Operation operation = Operation.MERGE;
-  
+
   public MergeInfo() {}
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     extent = new KeyExtent();
@@ -47,39 +47,39 @@ public class MergeInfo implements Writable {
     state = MergeState.values()[in.readInt()];
     operation = Operation.values()[in.readInt()];
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     extent.write(out);
     out.writeInt(state.ordinal());
     out.writeInt(operation.ordinal());
   }
-  
+
   public MergeInfo(KeyExtent range, Operation op) {
     this.extent = range;
     this.operation = op;
   }
-  
+
   public MergeState getState() {
     return state;
   }
-  
+
   public KeyExtent getExtent() {
     return extent;
   }
-  
+
   public Operation getOperation() {
     return operation;
   }
-  
+
   public void setState(MergeState state) {
     this.state = state;
   }
-  
+
   public boolean isDelete() {
     return this.operation.equals(Operation.DELETE);
   }
-  
+
   public boolean needsToBeChopped(KeyExtent otherExtent) {
     // During a delete, the block after the merge will be stretched to cover the deleted area.
     // Therefore, it needs to be chopped
@@ -90,14 +90,14 @@ public class MergeInfo implements Writable {
     else
       return this.extent.overlaps(otherExtent);
   }
-  
+
   public boolean overlaps(KeyExtent otherExtent) {
     boolean result = this.extent.overlaps(otherExtent);
     if (!result && needsToBeChopped(otherExtent))
       return true;
     return result;
   }
-  
+
   @Override
   public String toString() {
     if (!state.equals(MergeState.NONE))

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeState.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeState.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeState.java
index 29b6ae3..47bfd95 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeState.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/MergeState.java
@@ -45,5 +45,5 @@ public enum MergeState {
    * merge is complete, the resulting tablet can be brought online, remove the marker in zookeeper
    */
   COMPLETE;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataStateStore.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataStateStore.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataStateStore.java
index d57a3ef..bf56a7a 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataStateStore.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataStateStore.java
@@ -31,38 +31,38 @@ import org.apache.accumulo.server.AccumuloServerContext;
 
 public class MetaDataStateStore extends TabletStateStore {
   // private static final Logger log = Logger.getLogger(MetaDataStateStore.class);
-  
+
   private static final int THREADS = 4;
   private static final int LATENCY = 1000;
   private static final int MAX_MEMORY = 200 * 1024 * 1024;
-  
+
   final protected ClientContext context;
   final protected CurrentState state;
   final private String targetTableName;
-  
+
   protected MetaDataStateStore(ClientContext context, CurrentState state, String targetTableName) {
     this.context = context;
     this.state = state;
     this.targetTableName = targetTableName;
   }
-  
+
   public MetaDataStateStore(ClientContext context, CurrentState state) {
     this(context, state, MetadataTable.NAME);
   }
-  
+
   protected MetaDataStateStore(AccumuloServerContext context, String tableName) {
     this(context, null, tableName);
   }
-  
+
   public MetaDataStateStore(AccumuloServerContext context) {
     this(context, MetadataTable.NAME);
   }
-  
+
   @Override
   public ClosableIterator<TabletLocationState> iterator() {
     return new MetaDataTableScanner(context, MetadataSchema.TabletsSection.getRange(), state);
   }
-  
+
   @Override
   public void setLocations(Collection<Assignment> assignments) throws DistributedStoreException {
     BatchWriter writer = createBatchWriter();
@@ -83,7 +83,7 @@ public class MetaDataStateStore extends TabletStateStore {
       }
     }
   }
-  
+
   BatchWriter createBatchWriter() {
     try {
       return context.getConnector().createBatchWriter(targetTableName,
@@ -95,7 +95,7 @@ public class MetaDataStateStore extends TabletStateStore {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public void setFutureLocations(Collection<Assignment> assignments) throws DistributedStoreException {
     BatchWriter writer = createBatchWriter();
@@ -115,10 +115,10 @@ public class MetaDataStateStore extends TabletStateStore {
       }
     }
   }
-  
+
   @Override
   public void unassign(Collection<TabletLocationState> tablets) throws DistributedStoreException {
-    
+
     BatchWriter writer = createBatchWriter();
     try {
       for (TabletLocationState tls : tablets) {
@@ -141,7 +141,7 @@ public class MetaDataStateStore extends TabletStateStore {
       }
     }
   }
-  
+
   @Override
   public String name() {
     return "Normal Tablets";


[11/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/ShellCommandException.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/ShellCommandException.java b/shell/src/main/java/org/apache/accumulo/shell/ShellCommandException.java
index b25e9da..49e7163 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/ShellCommandException.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/ShellCommandException.java
@@ -18,41 +18,41 @@ package org.apache.accumulo.shell;
 
 public class ShellCommandException extends Exception {
   private static final long serialVersionUID = 1L;
-  
+
   public enum ErrorCode {
     UNKNOWN_ERROR("Unknown error"),
     UNSUPPORTED_LANGUAGE("Programming language used is not supported"),
     UNRECOGNIZED_COMMAND("Command is not supported"),
     INITIALIZATION_FAILURE("Command could not be initialized"),
     XML_PARSING_ERROR("Failed to parse the XML file");
-    
+
     private String description;
-    
+
     private ErrorCode(String description) {
       this.description = description;
     }
-    
+
     public String getDescription() {
       return this.description;
     }
-    
+
     public String toString() {
       return getDescription();
     }
   }
-  
+
   private ErrorCode code;
   private String command;
-  
+
   public ShellCommandException(ErrorCode code) {
     this(code, null);
   }
-  
+
   public ShellCommandException(ErrorCode code, String command) {
     this.code = code;
     this.command = command;
   }
-  
+
   public String getMessage() {
     return code + (command != null ? " (" + command + ")" : "");
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/ShellExtension.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/ShellExtension.java b/shell/src/main/java/org/apache/accumulo/shell/ShellExtension.java
index d29d276..f8f882a 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/ShellExtension.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/ShellExtension.java
@@ -19,9 +19,9 @@ package org.apache.accumulo.shell;
 import org.apache.accumulo.shell.Shell.Command;
 
 public abstract class ShellExtension {
-    
-    public abstract String getExtensionName();
 
-    public abstract Command[] getCommands();
-    
+  public abstract String getExtensionName();
+
+  public abstract Command[] getCommands();
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java b/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
index 8167ef8..67c8c40 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
@@ -24,11 +24,6 @@ import java.util.Map;
 import java.util.Scanner;
 import java.util.TreeMap;
 
-import com.beust.jcommander.DynamicParameter;
-import com.beust.jcommander.IStringConverter;
-import com.beust.jcommander.Parameter;
-import com.beust.jcommander.ParameterException;
-import com.beust.jcommander.converters.FileConverter;
 import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken;
@@ -37,6 +32,12 @@ import org.apache.commons.configuration.PropertiesConfiguration;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
+import com.beust.jcommander.DynamicParameter;
+import com.beust.jcommander.IStringConverter;
+import com.beust.jcommander.Parameter;
+import com.beust.jcommander.ParameterException;
+import com.beust.jcommander.converters.FileConverter;
+
 public class ShellOptionsJC {
   private static final Logger log = LoggerFactory.getLogger(Shell.class);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/ShellUtil.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/ShellUtil.java b/shell/src/main/java/org/apache/accumulo/shell/ShellUtil.java
index 966fb00..8505756 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/ShellUtil.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/ShellUtil.java
@@ -33,7 +33,7 @@ public class ShellUtil {
   /**
    * Scans the given file line-by-line (ignoring empty lines) and returns a list containing those lines. If decode is set to true, every line is decoded using
    * {@link Base64#decodeBase64(byte[])} from the UTF-8 bytes of that line before inserting in the list.
-   * 
+   *
    * @param filename
    *          Path to the file that needs to be scanned
    * @param decode

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/Token.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/Token.java b/shell/src/main/java/org/apache/accumulo/shell/Token.java
index 5083457..cd25ada 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/Token.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/Token.java
@@ -24,7 +24,7 @@ import java.util.Set;
 /*
  * A token is a word in a command in the shell.  The tree that this builds is used for
  * tab-completion of tables, users, commands and certain other parts of the shell that
- * can be realistically and quickly gathered. Tokens can have multiple commands grouped 
+ * can be realistically and quickly gathered. Tokens can have multiple commands grouped
  * together and many possible subcommands, although they are stored in a set so duplicates
  * aren't allowed.
  */
@@ -33,41 +33,41 @@ public class Token {
   private Set<String> command = new HashSet<String>();
   private Set<Token> subcommands = new HashSet<Token>();
   private boolean caseSensitive = false;
-  
+
   public Token() {}
-  
+
   public Token(String commandName) {
     this();
     command.add(commandName);
   }
-  
+
   public Token(Collection<String> commandNames) {
     this();
     command.addAll(commandNames);
   }
-  
+
   public Token(Set<String> commandNames, Set<Token> subCommandNames) {
     this();
     command.addAll(commandNames);
     subcommands.addAll(subCommandNames);
   }
-  
+
   public void setCaseSensitive(boolean cs) {
     caseSensitive = cs;
   }
-  
+
   public boolean getCaseSensitive() {
     return caseSensitive;
   }
-  
+
   public Set<String> getCommandNames() {
     return command;
   }
-  
+
   public Set<Token> getSubcommandList() {
     return subcommands;
   }
-  
+
   public Token getSubcommand(String name) {
     Iterator<Token> iter = subcommands.iterator();
     while (iter.hasNext()) {
@@ -77,14 +77,14 @@ public class Token {
     }
     return null;
   }
-  
+
   public Set<String> getSubcommandNames() {
     HashSet<String> set = new HashSet<String>();
     for (Token t : subcommands)
       set.addAll(t.getCommandNames());
     return set;
   }
-  
+
   public Set<String> getSubcommandNames(String startsWith) {
     Iterator<Token> iter = subcommands.iterator();
     HashSet<String> set = new HashSet<String>();
@@ -105,7 +105,7 @@ public class Token {
     }
     return set;
   }
-  
+
   public boolean containsCommand(String match) {
     Iterator<String> iter = command.iterator();
     while (iter.hasNext()) {
@@ -120,17 +120,17 @@ public class Token {
     }
     return false;
   }
-  
+
   public void addSubcommand(Token t) {
     subcommands.add(t);
   }
-  
+
   public void addSubcommand(Collection<String> t) {
     for (String a : t) {
       addSubcommand(new Token(a));
     }
   }
-  
+
   public String toString() {
     return this.command.toString();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/AboutCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/AboutCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/AboutCommand.java
index 9ba8460..6a9da5c 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/AboutCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/AboutCommand.java
@@ -26,12 +26,12 @@ import org.apache.commons.cli.Options;
 
 public class AboutCommand extends Command {
   private Option verboseOption;
-  
+
   @Override
   public String description() {
     return "displays information about this program";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException {
     shellState.printInfo();
@@ -40,12 +40,12 @@ public class AboutCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveCompactionIterator.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveCompactionIterator.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveCompactionIterator.java
index 159a2a6..c372246 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveCompactionIterator.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveCompactionIterator.java
@@ -30,11 +30,11 @@ import org.apache.accumulo.core.client.admin.InstanceOperations;
 import org.apache.accumulo.core.util.Duration;
 
 class ActiveCompactionIterator implements Iterator<String> {
-  
+
   private InstanceOperations instanceOps;
   private Iterator<String> tsIter;
   private Iterator<String> compactionIter;
-  
+
   private static String maxDecimal(double count) {
     if (count < 9.995)
       return String.format("%.2f", count);
@@ -55,15 +55,15 @@ class ActiveCompactionIterator implements Iterator<String> {
 
   private void readNext() {
     final List<String> compactions = new ArrayList<String>();
-    
+
     while (tsIter.hasNext()) {
-      
+
       final String tserver = tsIter.next();
       try {
         List<ActiveCompaction> acl = instanceOps.getActiveCompactions(tserver);
-        
+
         acl = new ArrayList<ActiveCompaction>(acl);
-        
+
         Collections.sort(acl, new Comparator<ActiveCompaction>() {
           @Override
           public int compare(ActiveCompaction o1, ActiveCompaction o2) {
@@ -77,9 +77,9 @@ class ActiveCompactionIterator implements Iterator<String> {
           if (index > 0) {
             output = output.substring(index + 6);
           }
-          
+
           ac.getIterators();
-          
+
           List<String> iterList = new ArrayList<String>();
           Map<String,Map<String,String>> iterOpts = new HashMap<String,Map<String,String>>();
           for (IteratorSetting is : ac.getIterators()) {
@@ -94,43 +94,43 @@ class ActiveCompactionIterator implements Iterator<String> {
       } catch (Exception e) {
         compactions.add(tserver + " ERROR " + e.getMessage());
       }
-      
+
       if (compactions.size() > 0) {
         break;
       }
     }
-    
+
     compactionIter = compactions.iterator();
   }
-  
+
   ActiveCompactionIterator(List<String> tservers, InstanceOperations instanceOps) {
     this.instanceOps = instanceOps;
     this.tsIter = tservers.iterator();
-    
+
     final String header = String.format(" %-21s| %-9s | %-5s | %-6s | %-5s | %-5s | %-15s | %-40s | %-5s | %-35s | %-9s | %s", "TABLET SERVER", "AGE", "TYPE",
         "REASON", "READ", "WROTE", "TABLE", "TABLET", "INPUT", "OUTPUT", "ITERATORS", "ITERATOR OPTIONS");
-    
+
     compactionIter = Collections.singletonList(header).iterator();
   }
-  
+
   @Override
   public boolean hasNext() {
     return compactionIter.hasNext();
   }
-  
+
   @Override
   public String next() {
     final String next = compactionIter.next();
-    
+
     if (!compactionIter.hasNext())
       readNext();
-    
+
     return next;
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveScanIterator.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveScanIterator.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveScanIterator.java
index 0412317..498eb2f 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveScanIterator.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ActiveScanIterator.java
@@ -27,65 +27,67 @@ import org.apache.accumulo.core.client.admin.ScanType;
 import org.apache.accumulo.core.util.Duration;
 
 class ActiveScanIterator implements Iterator<String> {
-  
+
   private InstanceOperations instanceOps;
   private Iterator<String> tsIter;
   private Iterator<String> scansIter;
-  
+
   private void readNext() {
     final List<String> scans = new ArrayList<String>();
-    
+
     while (tsIter.hasNext()) {
-      
+
       final String tserver = tsIter.next();
       try {
         final List<ActiveScan> asl = instanceOps.getActiveScans(tserver);
-        
+
         for (ActiveScan as : asl) {
-          scans.add(String.format("%21s |%21s |%9s |%9s |%7s |%6s |%8s |%8s |%10s |%20s |%10s |%10s | %s", tserver, as.getClient(),
-              Duration.format(as.getAge(), "", "-"), Duration.format(as.getLastContactTime(), "", "-"), as.getState(), as.getType(), as.getUser(),
-              as.getTable(), as.getColumns(), as.getAuthorizations(), (as.getType() == ScanType.SINGLE ? as.getExtent() : "N/A"), as.getSsiList(), as.getSsio()));
+          scans
+              .add(String.format("%21s |%21s |%9s |%9s |%7s |%6s |%8s |%8s |%10s |%20s |%10s |%10s | %s", tserver, as.getClient(),
+                  Duration.format(as.getAge(), "", "-"), Duration.format(as.getLastContactTime(), "", "-"), as.getState(), as.getType(), as.getUser(),
+                  as.getTable(), as.getColumns(), as.getAuthorizations(), (as.getType() == ScanType.SINGLE ? as.getExtent() : "N/A"), as.getSsiList(),
+                  as.getSsio()));
         }
       } catch (Exception e) {
         scans.add(tserver + " ERROR " + e.getMessage());
       }
-      
+
       if (scans.size() > 0) {
         break;
       }
     }
-    
+
     scansIter = scans.iterator();
   }
-  
+
   ActiveScanIterator(List<String> tservers, InstanceOperations instanceOps) {
     this.instanceOps = instanceOps;
     this.tsIter = tservers.iterator();
-    
+
     final String header = String.format(" %-21s| %-21s| %-9s| %-9s| %-7s| %-6s| %-8s| %-8s| %-10s| %-20s| %-10s| %-10s | %s", "TABLET SERVER", "CLIENT", "AGE",
         "LAST", "STATE", "TYPE", "USER", "TABLE", "COLUMNS", "AUTHORIZATIONS", "TABLET", "ITERATORS", "ITERATOR OPTIONS");
-    
+
     scansIter = Collections.singletonList(header).iterator();
   }
-  
+
   @Override
   public boolean hasNext() {
     return scansIter.hasNext();
   }
-  
+
   @Override
   public String next() {
     final String next = scansIter.next();
-    
+
     if (!scansIter.hasNext())
       readNext();
-    
+
     return next;
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/AddAuthsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/AddAuthsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/AddAuthsCommand.java
index fd57648..ef414e8 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/AddAuthsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/AddAuthsCommand.java
@@ -23,9 +23,9 @@ import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;
@@ -34,7 +34,7 @@ import org.apache.commons.cli.Options;
 public class AddAuthsCommand extends Command {
   private Option userOpt;
   private Option scanOptAuths;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException {
     final String user = cl.getOptionValue(userOpt.getOpt(), shellState.getConnector().whoami());
@@ -50,17 +50,17 @@ public class AddAuthsCommand extends Command {
     Shell.log.debug("Changed record-level authorizations for user " + user);
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "adds authorizations to the maximum scan authorizations for a user";
   }
-  
+
   @Override
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> completionSet) {
     registerCompletionForUsers(root, completionSet);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -75,7 +75,7 @@ public class AddAuthsCommand extends Command {
     o.addOption(userOpt);
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/AddSplitsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/AddSplitsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/AddSplitsCommand.java
index b695d4d..37cf644 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/AddSplitsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/AddSplitsCommand.java
@@ -20,8 +20,8 @@ import java.util.TreeSet;
 
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.ShellUtil;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.ShellUtil;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.MissingArgumentException;
 import org.apache.commons.cli.Option;
@@ -30,13 +30,13 @@ import org.apache.hadoop.io.Text;
 
 public class AddSplitsCommand extends Command {
   private Option optSplitsFile, base64Opt;
-  
+
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
     final boolean decode = cl.hasOption(base64Opt.getOpt());
-    
+
     final TreeSet<Text> splits = new TreeSet<Text>();
-    
+
     if (cl.hasOption(optSplitsFile.getOpt())) {
       splits.addAll(ShellUtil.scanFile(cl.getOptionValue(optSplitsFile.getOpt()), decode));
     } else {
@@ -47,40 +47,40 @@ public class AddSplitsCommand extends Command {
         splits.add(new Text(s.getBytes(Shell.CHARSET)));
       }
     }
-    
+
     if (!shellState.getConnector().tableOperations().exists(tableName)) {
       throw new TableNotFoundException(null, tableName, null);
     }
     shellState.getConnector().tableOperations().addSplits(tableName, splits);
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "adds split points to an existing table";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     optSplitsFile = new Option("sf", "splits-file", true, "file with a newline-separated list of rows to split the table with");
     optSplitsFile.setArgName("filename");
-    
+
     base64Opt = new Option("b64", "base64encoded", false, "decode encoded split points (splits file only)");
-    
+
     o.addOption(OptUtil.tableOpt("name of the table to add split points to"));
     o.addOption(optSplitsFile);
     o.addOption(base64Opt);
     return o;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " [<split>{ <split>} ]";
   }
-  
+
   @Override
   public int numArgs() {
     return Shell.NO_FIXED_ARG_LENGTH_CHECK;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/AuthenticateCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/AuthenticateCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/AuthenticateCommand.java
index ef35f52..dd1cedd 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/AuthenticateCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/AuthenticateCommand.java
@@ -26,8 +26,8 @@ import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 
 public class AuthenticateCommand extends Command {
@@ -44,22 +44,22 @@ public class AuthenticateCommand extends Command {
     shellState.getReader().println((valid ? "V" : "Not v") + "alid");
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "verifies a user's credentials";
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <username>";
   }
-  
+
   @Override
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> completionSet) {
     registerCompletionForUsers(root, completionSet);
   }
-  
+
   @Override
   public int numArgs() {
     return 1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ClasspathCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ClasspathCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ClasspathCommand.java
index c056581..071571c 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ClasspathCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ClasspathCommand.java
@@ -42,12 +42,12 @@ public class ClasspathCommand extends Command {
     });
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "lists the current files on the classpath";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ClearCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ClearCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ClearCommand.java
index 9340d4e..77889a1 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ClearCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ClearCommand.java
@@ -27,7 +27,7 @@ public class ClearCommand extends Command {
   public String description() {
     return "clears the screen";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException {
     // custom clear screen, so I don't have to redraw the prompt twice
@@ -37,14 +37,14 @@ public class ClearCommand extends Command {
     // send the ANSI code to clear the screen
     shellState.getReader().print(((char) 27) + "[2J");
     shellState.getReader().flush();
-    
+
     // then send the ANSI code to go to position 1,1
     shellState.getReader().print(((char) 27) + "[1;1H");
     shellState.getReader().flush();
-    
+
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/CloneTableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/CloneTableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/CloneTableCommand.java
index 37de39d..bc58a92 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/CloneTableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/CloneTableCommand.java
@@ -26,26 +26,26 @@ import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class CloneTableCommand extends Command {
-  
+
   private Option setPropsOption;
   private Option excludePropsOption;
   private Option noFlushOption;
-  
+
   @Override
-  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, TableNotFoundException,
-      TableExistsException {
-    
+  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException, TableExistsException {
+
     final HashMap<String,String> props = new HashMap<String,String>();
     final HashSet<String> exclude = new HashSet<String>();
     boolean flush = true;
-    
+
     if (cl.hasOption(setPropsOption.getOpt())) {
       String[] keyVals = cl.getOptionValue(setPropsOption.getOpt()).split(",");
       for (String keyVal : keyVals) {
@@ -53,36 +53,36 @@ public class CloneTableCommand extends Command {
         props.put(sa[0], sa[1]);
       }
     }
-    
+
     if (cl.hasOption(excludePropsOption.getOpt())) {
       String[] keys = cl.getOptionValue(excludePropsOption.getOpt()).split(",");
       for (String key : keys) {
         exclude.add(key);
       }
     }
-    
+
     if (cl.hasOption(noFlushOption.getOpt())) {
       flush = false;
     }
-    
+
     shellState.getConnector().tableOperations().clone(cl.getArgs()[0], cl.getArgs()[1], flush, props, exclude);
     return 0;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <current table name> <new table name>";
   }
-  
+
   @Override
   public String description() {
     return "clones a table";
   }
-  
+
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> completionSet) {
     registerCompletionForTables(root, completionSet);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -94,7 +94,7 @@ public class CloneTableCommand extends Command {
     o.addOption(noFlushOption);
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 2;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/CompactCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/CompactCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/CompactCommand.java
index 9e599ae..131534f 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/CompactCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/CompactCommand.java
@@ -41,9 +41,9 @@ public class CompactCommand extends TableOperation {
       outCompressionOpt, outReplication;
 
   private CompactionConfig compactionConfig = null;
-  
+
   boolean override = false;
-  
+
   private boolean cancel = false;
 
   @Override
@@ -52,10 +52,10 @@ public class CompactCommand extends TableOperation {
         + "all files will be compacted.  Options that configure output settings are only applied to this compaction and not later compactions.  If multiple "
         + "concurrent user initiated compactions specify iterators or a compaction strategy, then all but one will fail to start.";
   }
-  
+
   protected void doTableOp(final Shell shellState, final String tableName) throws AccumuloException, AccumuloSecurityException {
     // compact the tables
-    
+
     if (cancel) {
       try {
         shellState.getConnector().tableOperations().cancelCompaction(tableName);
@@ -68,9 +68,9 @@ public class CompactCommand extends TableOperation {
         if (compactionConfig.getWait()) {
           Shell.log.info("Compacting table ...");
         }
-        
+
         shellState.getConnector().tableOperations().compact(tableName, compactionConfig);
-        
+
         Shell.log.info("Compaction of table " + tableName + " " + (compactionConfig.getWait() ? "completed" : "started") + " for given range");
       } catch (Exception ex) {
         throw new AccumuloException(ex);
@@ -102,10 +102,10 @@ public class CompactCommand extends TableOperation {
 
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
-    
+
     if (cl.hasOption(cancelOpt.getLongOpt())) {
       cancel = true;
-      
+
       if (cl.getOptions().length > 2) {
         throw new IllegalArgumentException("Can not specify other options with cancel");
       }
@@ -119,14 +119,14 @@ public class CompactCommand extends TableOperation {
     compactionConfig.setWait(cl.hasOption(waitOpt.getOpt()));
     compactionConfig.setStartRow(OptUtil.getStartRow(cl));
     compactionConfig.setEndRow(OptUtil.getEndRow(cl));
-    
+
     if (cl.hasOption(profileOpt.getOpt())) {
       List<IteratorSetting> iterators = shellState.iteratorProfiles.get(cl.getOptionValue(profileOpt.getOpt()));
       if (iterators == null) {
         Shell.log.error("Profile " + cl.getOptionValue(profileOpt.getOpt()) + " does not exist");
         return -1;
       }
-      
+
       compactionConfig.setIterators(new ArrayList<>(iterators));
     }
 
@@ -159,7 +159,7 @@ public class CompactCommand extends TableOperation {
 
     return super.execute(fullCommand, cl, shellState);
   }
-  
+
   private Option newLAO(String lopt, String desc) {
     return new Option(null, lopt, true, desc);
   }
@@ -167,14 +167,14 @@ public class CompactCommand extends TableOperation {
   @Override
   public Options getOptions() {
     final Options opts = super.getOptions();
-    
+
     opts.addOption(OptUtil.startRowOpt());
     opts.addOption(OptUtil.endRowOpt());
     noFlushOption = new Option("nf", "noFlush", false, "do not flush table data in memory before compacting.");
     opts.addOption(noFlushOption);
     waitOpt = new Option("w", "wait", false, "wait for compact to finish");
     opts.addOption(waitOpt);
-    
+
     profileOpt = new Option("pn", "profile", true, "Iterator profile name.");
     profileOpt.setArgName("profile");
     opts.addOption(profileOpt);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ConfigCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ConfigCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ConfigCommand.java
index e720e16..ec3f276 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ConfigCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ConfigCommand.java
@@ -37,10 +37,10 @@ import org.apache.accumulo.core.conf.Property;
 import org.apache.accumulo.core.security.ColumnVisibility;
 import org.apache.accumulo.core.util.BadArgumentException;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.ShellOptions;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.Shell.PrintFile;
+import org.apache.accumulo.shell.ShellOptions;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ConstraintCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ConstraintCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ConstraintCommand.java
index c6c13e8..9ffb349 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ConstraintCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ConstraintCommand.java
@@ -20,8 +20,8 @@ import java.util.Map.Entry;
 
 import org.apache.accumulo.core.constraints.Constraint;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.ShellCommandException;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.ShellCommandException;
 import org.apache.accumulo.shell.ShellCommandException.ErrorCode;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.commons.cli.CommandLine;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/CreateNamespaceCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/CreateNamespaceCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/CreateNamespaceCommand.java
index e278105..67da2b3 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/CreateNamespaceCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/CreateNamespaceCommand.java
@@ -55,7 +55,7 @@ public class CreateNamespaceCommand extends Command {
       if (shellState.getConnector().namespaceOperations().exists(namespace)) {
         configuration = shellState.getConnector().namespaceOperations().getProperties(copy);
       }
-    } 
+    }
     if (configuration != null) {
       for (Entry<String,String> entry : configuration) {
         if (Property.isValidTablePropertyKey(entry.getKey())) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/CreateTableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/CreateTableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/CreateTableCommand.java
index bcf3812..7a4e551 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/CreateTableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/CreateTableCommand.java
@@ -17,9 +17,9 @@
 package org.apache.accumulo.shell.commands;
 
 import java.io.IOException;
+import java.util.HashMap;
 import java.util.Map;
 import java.util.Map.Entry;
-import java.util.HashMap;
 import java.util.Set;
 import java.util.SortedSet;
 import java.util.TreeSet;
@@ -35,9 +35,9 @@ import org.apache.accumulo.core.conf.Property;
 import org.apache.accumulo.core.iterators.IteratorUtil;
 import org.apache.accumulo.core.security.VisibilityConstraint;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellUtil;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/CreateUserCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/CreateUserCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/CreateUserCommand.java
index 2a7359f..161486c 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/CreateUserCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/CreateUserCommand.java
@@ -33,7 +33,7 @@ public class CreateUserCommand extends Command {
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, TableNotFoundException,
       AccumuloSecurityException, TableExistsException, IOException {
     final String user = cl.getArgs()[0];
-    
+
     final String password = shellState.readMaskedLine("Enter new password for '" + user + "': ", '*');
     if (password == null) {
       shellState.getReader().println();
@@ -44,7 +44,7 @@ public class CreateUserCommand extends Command {
       shellState.getReader().println();
       return 0;
     } // user canceled
-    
+
     if (!password.equals(passwordConfirm)) {
       throw new IllegalArgumentException("Passwords do not match");
     }
@@ -52,23 +52,23 @@ public class CreateUserCommand extends Command {
     Shell.log.debug("Created user " + user);
     return 0;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <username>";
   }
-  
+
   @Override
   public String description() {
     return "creates a new user";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DebugCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DebugCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DebugCommand.java
index ab40adc..6abcdf0 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DebugCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DebugCommand.java
@@ -23,8 +23,8 @@ import java.util.Set;
 
 import org.apache.accumulo.core.util.BadArgumentException;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 
 public class DebugCommand extends Command {
@@ -46,24 +46,24 @@ public class DebugCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "turns debug logging on or off";
   }
-  
+
   @Override
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> special) {
     final Token debug_command = new Token(getName());
     debug_command.addSubcommand(Arrays.asList(new String[] {"on", "off"}));
     root.addSubcommand(debug_command);
   }
-  
+
   @Override
   public String usage() {
     return getName() + " [ on | off ]";
   }
-  
+
   @Override
   public int numArgs() {
     return Shell.NO_FIXED_ARG_LENGTH_CHECK;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteCommand.java
index 2601f58..915e3f8 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteCommand.java
@@ -38,24 +38,24 @@ import org.apache.hadoop.io.Text;
 public class DeleteCommand extends Command {
   private Option deleteOptAuths, timestampOpt;
   private Option timeoutOption;
-  
+
   protected long getTimeout(final CommandLine cl) {
     if (cl.hasOption(timeoutOption.getLongOpt())) {
       return AccumuloConfiguration.getTimeInMillis(cl.getOptionValue(timeoutOption.getLongOpt()));
     }
-    
+
     return Long.MAX_VALUE;
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
       TableNotFoundException, IOException, ConstraintViolationException {
     shellState.checkTableState();
-    
+
     final Mutation m = new Mutation(new Text(cl.getArgs()[0].getBytes(Shell.CHARSET)));
     final Text colf = new Text(cl.getArgs()[1].getBytes(Shell.CHARSET));
     final Text colq = new Text(cl.getArgs()[2].getBytes(Shell.CHARSET));
-    
+
     if (cl.hasOption(deleteOptAuths.getOpt())) {
       final ColumnVisibility le = new ColumnVisibility(cl.getOptionValue(deleteOptAuths.getOpt()));
       if (cl.hasOption(timestampOpt.getOpt())) {
@@ -74,37 +74,37 @@ public class DeleteCommand extends Command {
     bw.close();
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "deletes a record from a table";
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <row> <colfamily> <colqualifier>";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     deleteOptAuths = new Option("l", "visibility-label", true, "formatted visibility");
     deleteOptAuths.setArgName("expression");
     o.addOption(deleteOptAuths);
-    
+
     timestampOpt = new Option("ts", "timestamp", true, "timestamp to use for deletion");
     timestampOpt.setArgName("timestamp");
     o.addOption(timestampOpt);
-    
+
     timeoutOption = new Option(null, "timeout", true,
         "time before insert should fail if no data is written. If no unit is given assumes seconds.  Units d,h,m,s,and ms are supported.  e.g. 30s or 100ms");
     timeoutOption.setArgName("timeout");
     o.addOption(timeoutOption);
-    
+
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 3;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteManyCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteManyCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteManyCommand.java
index c29a92c..c8ab6d8 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteManyCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteManyCommand.java
@@ -33,43 +33,43 @@ import org.apache.commons.cli.Options;
 
 public class DeleteManyCommand extends ScanCommand {
   private Option forceOpt;
-  
+
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     final ScanInterpreter interpeter = getInterpreter(cl, tableName, shellState);
 
     // handle first argument, if present, the authorizations list to
     // scan with
     final Authorizations auths = getAuths(cl, shellState);
     final Scanner scanner = shellState.getConnector().createScanner(tableName, auths);
-    
+
     scanner.addScanIterator(new IteratorSetting(Integer.MAX_VALUE, "NOVALUE", SortedKeyIterator.class));
-    
+
     // handle session-specific scan iterators
     addScanIterators(shellState, cl, scanner, tableName);
-    
+
     // handle remaining optional arguments
     scanner.setRange(getRange(cl, interpeter));
-    
+
     scanner.setTimeout(getTimeout(cl), TimeUnit.MILLISECONDS);
 
     // handle columns
     fetchColumns(cl, scanner, interpeter);
-    
+
     // output / delete the records
     final BatchWriter writer = shellState.getConnector()
         .createBatchWriter(tableName, new BatchWriterConfig().setTimeout(getTimeout(cl), TimeUnit.MILLISECONDS));
     shellState.printLines(new DeleterFormatter(writer, scanner, cl.hasOption(timestampOpt.getOpt()), shellState, cl.hasOption(forceOpt.getOpt())), false);
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "scans a table and deletes the resulting records";
   }
-  
+
   @Override
   public Options getOptions() {
     forceOpt = new Option("f", "force", false, "force deletion without prompting");
@@ -78,5 +78,5 @@ public class DeleteManyCommand extends ScanCommand {
     opts.addOption(OptUtil.tableOpt("table to delete entries from"));
     return opts;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteNamespaceCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteNamespaceCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteNamespaceCommand.java
index faf1147..b749f06 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteNamespaceCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteNamespaceCommand.java
@@ -23,8 +23,8 @@ import java.util.Set;
 import org.apache.accumulo.core.client.NamespaceNotFoundException;
 import org.apache.accumulo.core.client.impl.Namespaces;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteRowsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteRowsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteRowsCommand.java
index 09f2938..af7eb9a 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteRowsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteRowsCommand.java
@@ -26,7 +26,7 @@ import org.apache.hadoop.io.Text;
 public class DeleteRowsCommand extends Command {
   private Option forceOpt;
   private Option startRowOptExclusive;
- 
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
@@ -39,17 +39,17 @@ public class DeleteRowsCommand extends Command {
     shellState.getConnector().tableOperations().deleteRows(tableName, startRow, endRow);
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "deletes a range of rows in a table.  Note that rows matching the start row ARE NOT deleted, but rows matching the end row ARE deleted.";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteScanIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteScanIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteScanIterCommand.java
index f3c9823..3ee6ca7 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteScanIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteScanIterCommand.java
@@ -29,11 +29,11 @@ import org.apache.commons.cli.Options;
 
 public class DeleteScanIterCommand extends Command {
   private Option nameOpt, allOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     if (cl.hasOption(allOpt.getOpt())) {
       final List<IteratorSetting> tableScanIterators = shellState.scanIteratorOptions.remove(tableName);
       if (tableScanIterators == null) {
@@ -65,36 +65,36 @@ public class DeleteScanIterCommand extends Command {
         Shell.log.info("No iterator named " + name + " found for table " + tableName);
       }
     }
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "deletes a table-specific scan iterator so it is no longer used during this shell session";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     OptionGroup nameGroup = new OptionGroup();
-    
+
     nameOpt = new Option("n", "name", true, "iterator to delete");
     nameOpt.setArgName("itername");
-    
+
     allOpt = new Option("a", "all", false, "delete all scan iterators");
     allOpt.setArgName("all");
-    
+
     nameGroup.addOption(nameOpt);
     nameGroup.addOption(allOpt);
     nameGroup.setRequired(true);
     o.addOptionGroup(nameGroup);
     o.addOption(OptUtil.tableOpt("table to delete scan iterators from"));
-    
+
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteShellIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteShellIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteShellIterCommand.java
index 3010f0b..2184331 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteShellIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DeleteShellIterCommand.java
@@ -29,7 +29,7 @@ import org.apache.commons.cli.Options;
 
 public class DeleteShellIterCommand extends Command {
   private Option nameOpt, allOpt, profileOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
 
@@ -55,36 +55,36 @@ public class DeleteShellIterCommand extends Command {
           Shell.log.info("Removed iterator " + name + " from profile " + profile + " (" + iterSettings.size() + " left)");
         }
       }
-      
+
     } else {
       Shell.log.info("No profile named " + profile);
     }
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "deletes iterators profiles configured in this shell session";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     OptionGroup nameGroup = new OptionGroup();
-    
+
     nameOpt = new Option("n", "name", true, "iterator to delete");
     nameOpt.setArgName("itername");
-    
+
     allOpt = new Option("a", "all", false, "delete all scan iterators");
     allOpt.setArgName("all");
-    
+
     nameGroup.addOption(nameOpt);
     nameGroup.addOption(allOpt);
     nameGroup.setRequired(true);
     o.addOptionGroup(nameGroup);
-    
+
     profileOpt = new Option("pn", "profile", true, "iterator profile name");
     profileOpt.setRequired(true);
     profileOpt.setArgName("profile");
@@ -92,7 +92,7 @@ public class DeleteShellIterCommand extends Command {
 
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
index 958f5eb..eeac50c 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
@@ -26,9 +26,9 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class EGrepCommand extends GrepCommand {
-  
+
   private Option matchSubstringOption;
-  
+
   @Override
   protected void setUpIterator(final int prio, final String name, final String term, final BatchScanner scanner, CommandLine cl) throws IOException {
     if (prio < 0) {
@@ -38,17 +38,17 @@ public class EGrepCommand extends GrepCommand {
     RegExFilter.setRegexs(si, term, term, term, term, true, cl.hasOption(matchSubstringOption.getOpt()));
     scanner.addScanIterator(si);
   }
-  
+
   @Override
   public String description() {
     return "searches each row, column family, column qualifier and value, in parallel, on the server side (using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <regex>{ <regex>}";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = super.getOptions();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ExecfileCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ExecfileCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ExecfileCommand.java
index fb8af7f..9d6278e 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ExecfileCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ExecfileCommand.java
@@ -29,12 +29,12 @@ import org.apache.commons.cli.Options;
 
 public class ExecfileCommand extends Command {
   private Option verboseOption;
-  
+
   @Override
   public String description() {
     return "specifies a file containing accumulo commands to execute";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     Scanner scanner = new Scanner(new File(cl.getArgs()[0]), UTF_8.name());
@@ -47,17 +47,17 @@ public class ExecfileCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <fileName>";
   }
-  
+
   @Override
   public int numArgs() {
     return 1;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ExitCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ExitCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ExitCommand.java
index 1d44409..2a424ad 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ExitCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ExitCommand.java
@@ -26,12 +26,12 @@ public class ExitCommand extends Command {
     shellState.setExit(true);
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "exits the shell";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ExportTableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ExportTableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ExportTableCommand.java
index b8fef1f..7bcc67d 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ExportTableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ExportTableCommand.java
@@ -24,54 +24,54 @@ import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class ExportTableCommand extends Command {
-  
+
   private Option tableOpt;
 
   @Override
-  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, TableNotFoundException,
-      TableExistsException {
-    
+  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException, TableExistsException {
+
     final String tableName = OptUtil.getTableOpt(cl, shellState);
 
     shellState.getConnector().tableOperations().exportTable(tableName, cl.getArgs()[0]);
     return 0;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <export dir>";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     tableOpt = new Option(ShellOptions.tableOption, "table", true, "table to export");
-    
+
     tableOpt.setArgName("table");
-    
+
     o.addOption(tableOpt);
 
     return o;
   }
-  
+
   @Override
   public String description() {
     return "exports a table";
   }
-  
+
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> completionSet) {
     registerCompletionForTables(root, completionSet);
   }
-  
+
   @Override
   public int numArgs() {
     return 1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ExtensionCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ExtensionCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ExtensionCommand.java
index fbc1833..7dc457b 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ExtensionCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ExtensionCommand.java
@@ -22,32 +22,32 @@ import java.util.Set;
 import java.util.TreeSet;
 
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.ShellExtension;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.ShellExtension;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class ExtensionCommand extends Command {
-  
+
   protected Option enable, disable, list;
-  
+
   private static ServiceLoader<ShellExtension> extensions = null;
-  
+
   private Set<String> loadedHeaders = new HashSet<String>();
   private Set<String> loadedCommands = new HashSet<String>();
   private Set<String> loadedExtensions = new TreeSet<String>();
-  
+
   public int execute(String fullCommand, CommandLine cl, Shell shellState) throws Exception {
     if (cl.hasOption(enable.getOpt())) {
       extensions = ServiceLoader.load(ShellExtension.class);
       for (ShellExtension se : extensions) {
-        
+
         loadedExtensions.add(se.getExtensionName());
         String header = "-- " + se.getExtensionName() + " Extension Commands ---------";
         loadedHeaders.add(header);
         shellState.commandGrouping.put(header, se.getCommands());
-        
+
         for (Command cmd : se.getCommands()) {
           String name = se.getExtensionName() + "::" + cmd.getName();
           loadedCommands.add(name);
@@ -55,15 +55,15 @@ public class ExtensionCommand extends Command {
         }
       }
     } else if (cl.hasOption(disable.getOpt())) {
-      //Remove the headers
+      // Remove the headers
       for (String header : loadedHeaders) {
         shellState.commandGrouping.remove(header);
       }
-      //remove the commands
+      // remove the commands
       for (String name : loadedCommands) {
         shellState.commandFactory.remove(name);
       }
-      //Reset state
+      // Reset state
       loadedExtensions.clear();
       extensions.reload();
     } else if (cl.hasOption(list.getOpt())) {
@@ -73,15 +73,15 @@ public class ExtensionCommand extends Command {
     }
     return 0;
   }
-  
+
   public String description() {
     return "Enable, disable, or list shell extensions";
   }
-  
+
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public String getName() {
     return "extensions";
@@ -98,5 +98,5 @@ public class ExtensionCommand extends Command {
     o.addOption(list);
     return o;
   }
-    
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/FateCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/FateCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/FateCommand.java
index 85dc6b5..cef9a6d 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/FateCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/FateCommand.java
@@ -27,8 +27,8 @@ import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.core.conf.DefaultConfiguration;
-import org.apache.accumulo.core.conf.SiteConfiguration;
 import org.apache.accumulo.core.conf.Property;
+import org.apache.accumulo.core.conf.SiteConfiguration;
 import org.apache.accumulo.core.zookeeper.ZooUtil;
 import org.apache.accumulo.fate.AdminUtil;
 import org.apache.accumulo.fate.ReadOnlyTStore.TStatus;
@@ -45,18 +45,18 @@ import org.apache.zookeeper.KeeperException;
 
 /**
  * Manage FATE transactions
- * 
+ *
  */
 public class FateCommand extends Command {
-  
+
   private static final String SCHEME = "digest";
-  
+
   private static final String USER = "accumulo";
-  
+
   private Option secretOption;
   private Option statusOption;
   private Option disablePaginationOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws ParseException, KeeperException, InterruptedException,
       IOException {
@@ -67,14 +67,14 @@ public class FateCommand extends Command {
     }
     String cmd = args[0];
     boolean failedCommand = false;
-    
+
     AdminUtil<FateCommand> admin = new AdminUtil<FateCommand>(false);
-    
+
     String path = ZooUtil.getRoot(instance) + Constants.ZFATE;
     String masterPath = ZooUtil.getRoot(instance) + Constants.ZMASTER_LOCK;
     IZooReaderWriter zk = getZooReaderWriter(shellState.getInstance(), cl.getOptionValue(secretOption.getOpt()));
     ZooStore<FateCommand> zs = new ZooStore<FateCommand>(path, zk);
-    
+
     if ("fail".equals(cmd)) {
       if (args.length <= 1) {
         throw new ParseException("Must provide transaction ID");
@@ -113,7 +113,7 @@ public class FateCommand extends Command {
           }
         }
       }
-      
+
       // Parse TStatus filters for print display
       EnumSet<TStatus> filterStatus = null;
       if (cl.hasOption(statusOption.getOpt())) {
@@ -128,7 +128,7 @@ public class FateCommand extends Command {
           }
         }
       }
-      
+
       StringBuilder buf = new StringBuilder(8096);
       Formatter fmt = new Formatter(buf);
       admin.print(zs, zk, ZooUtil.getRoot(instance) + Constants.ZTABLE_LOCKS, fmt, filterTxid, filterStatus);
@@ -136,30 +136,30 @@ public class FateCommand extends Command {
     } else {
       throw new ParseException("Invalid command option");
     }
-    
+
     return failedCommand ? 1 : 0;
   }
-  
+
   protected synchronized IZooReaderWriter getZooReaderWriter(Instance instance, String secret) {
-    
+
     if (secret == null) {
       AccumuloConfiguration conf = SiteConfiguration.getInstance(DefaultConfiguration.getInstance());
       secret = conf.get(Property.INSTANCE_SECRET);
     }
-    
+
     return new ZooReaderWriter(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut(), SCHEME, (USER + ":" + secret).getBytes());
   }
-  
+
   @Override
   public String description() {
     return "manage FATE transactions";
   }
-  
+
   @Override
   public String usage() {
     return getName() + " fail <txid>... | delete <txid>... | print [<txid>...]";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -175,7 +175,7 @@ public class FateCommand extends Command {
     o.addOption(disablePaginationOpt);
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     // Arg length varies between 1 to n

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/FlushCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/FlushCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/FlushCommand.java
index 34e22a5..88a023b 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/FlushCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/FlushCommand.java
@@ -28,20 +28,20 @@ import org.apache.hadoop.io.Text;
 public class FlushCommand extends TableOperation {
   private Text startRow;
   private Text endRow;
-  
+
   private boolean wait;
   private Option waitOpt;
-  
+
   @Override
   public String description() {
     return "flushes a tables data that is currently in memory to disk";
   }
-  
+
   protected void doTableOp(final Shell shellState, final String tableName) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     shellState.getConnector().tableOperations().flush(tableName, startRow, endRow, wait);
     Shell.log.info("Flush of table " + tableName + (wait ? " completed." : " initiated..."));
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     wait = cl.hasOption(waitOpt.getLongOpt());
@@ -49,7 +49,7 @@ public class FlushCommand extends TableOperation {
     endRow = OptUtil.getEndRow(cl);
     return super.execute(fullCommand, cl, shellState);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = super.getOptions();
@@ -57,7 +57,7 @@ public class FlushCommand extends TableOperation {
     opts.addOption(waitOpt);
     opts.addOption(OptUtil.startRowOpt());
     opts.addOption(OptUtil.endRowOpt());
-    
+
     return opts;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/FormatterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/FormatterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/FormatterCommand.java
index e416699..931383b 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/FormatterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/FormatterCommand.java
@@ -26,7 +26,7 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class FormatterCommand extends ShellPluginConfigurationCommand {
-  
+
   private Option interpeterOption;
 
   public FormatterCommand() {
@@ -41,25 +41,26 @@ public class FormatterCommand extends ShellPluginConfigurationCommand {
   public static Class<? extends Formatter> getCurrentFormatter(final String tableName, final Shell shellState) {
     return ShellPluginConfigurationCommand.getPluginClass(tableName, shellState, Formatter.class, Property.TABLE_FORMATTER_CLASS);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options options = super.getOptions();
-    
+
     interpeterOption = new Option("i", "interpeter", false, "configure class as interpreter also");
-    
+
     options.addOption(interpeterOption);
-    
+
     return options;
   }
-  
-  protected void setPlugin(final CommandLine cl, final Shell shellState, final String tableName, final String className) throws AccumuloException, AccumuloSecurityException {
+
+  protected void setPlugin(final CommandLine cl, final Shell shellState, final String tableName, final String className) throws AccumuloException,
+      AccumuloSecurityException {
     super.setPlugin(cl, shellState, tableName, className);
     if (cl.hasOption(interpeterOption.getOpt())) {
       shellState.getConnector().tableOperations().setProperty(tableName, Property.TABLE_INTERPRETER_CLASS.toString(), className);
     }
   }
-  
+
   protected void removePlugin(final CommandLine cl, final Shell shellState, final String tableName) throws AccumuloException, AccumuloSecurityException {
     super.removePlugin(cl, shellState, tableName);
     if (cl.hasOption(interpeterOption.getOpt())) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/GetAuthsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/GetAuthsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/GetAuthsCommand.java
index 48ac92b..c4c1b67 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/GetAuthsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/GetAuthsCommand.java
@@ -33,7 +33,7 @@ import org.apache.commons.lang.StringUtils;
 
 public class GetAuthsCommand extends Command {
   private Option userOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, IOException {
     final String user = cl.getOptionValue(userOpt.getOpt(), shellState.getConnector().whoami());
@@ -46,12 +46,12 @@ public class GetAuthsCommand extends Command {
     shellState.getReader().println(StringUtils.join(set, ','));
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "displays the maximum scan authorizations for a user";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -60,7 +60,7 @@ public class GetAuthsCommand extends Command {
     o.addOption(userOpt);
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/GetGroupsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/GetGroupsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/GetGroupsCommand.java
index 6cae7e0..b2bba28 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/GetGroupsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/GetGroupsCommand.java
@@ -28,29 +28,29 @@ import org.apache.commons.cli.Options;
 import org.apache.hadoop.io.Text;
 
 public class GetGroupsCommand extends Command {
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     final Map<String,Set<Text>> groups = shellState.getConnector().tableOperations().getLocalityGroups(tableName);
-    
+
     for (Entry<String,Set<Text>> entry : groups.entrySet()) {
       shellState.getReader().println(entry.getKey() + "=" + LocalityGroupUtil.encodeColumnFamilies(entry.getValue()));
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "gets the locality groups for a given table";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/GetSplitsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/GetSplitsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/GetSplitsCommand.java
index 26493fd..d780aee 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/GetSplitsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/GetSplitsCommand.java
@@ -48,22 +48,22 @@ import org.apache.commons.cli.Options;
 import org.apache.hadoop.io.Text;
 
 public class GetSplitsCommand extends Command {
-  
+
   private Option outputFileOpt, maxSplitsOpt, base64Opt, verboseOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException, AccumuloException, AccumuloSecurityException,
       TableNotFoundException {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     final String outputFile = cl.getOptionValue(outputFileOpt.getOpt());
     final String m = cl.getOptionValue(maxSplitsOpt.getOpt());
     final int maxSplits = m == null ? 0 : Integer.parseInt(m);
     final boolean encode = cl.hasOption(base64Opt.getOpt());
     final boolean verbose = cl.hasOption(verboseOpt.getOpt());
-    
+
     final PrintLine p = outputFile == null ? new PrintShell(shellState.getReader()) : new PrintFile(outputFile);
-    
+
     try {
       if (!verbose) {
         for (Text row : maxSplits > 0 ? shellState.getConnector().tableOperations().listSplits(tableName, maxSplits) : shellState.getConnector()
@@ -90,14 +90,14 @@ public class GetSplitsCommand extends Command {
           }
         }
       }
-      
+
     } finally {
       p.close();
     }
-    
+
     return 0;
   }
-  
+
   private static String encode(final boolean encode, final Text text) {
     if (text == null) {
       return null;
@@ -105,7 +105,7 @@ public class GetSplitsCommand extends Command {
     BinaryFormatter.getlength(text.getLength());
     return encode ? Base64.encodeBase64String(TextUtil.getBytes(text)) : BinaryFormatter.appendText(new StringBuilder(), text).toString();
   }
-  
+
   private static String obscuredTabletName(final KeyExtent extent) {
     MessageDigest digester;
     try {
@@ -118,37 +118,37 @@ public class GetSplitsCommand extends Command {
     }
     return Base64.encodeBase64String(digester.digest());
   }
-  
+
   @Override
   public String description() {
     return "retrieves the current split points for tablets in the current table";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();
-    
+
     outputFileOpt = new Option("o", "output", true, "local file to write the splits to");
     outputFileOpt.setArgName("file");
-    
+
     maxSplitsOpt = new Option("m", "max", true, "maximum number of splits to return (evenly spaced)");
     maxSplitsOpt.setArgName("num");
-    
+
     base64Opt = new Option("b64", "base64encoded", false, "encode the split points");
-    
+
     verboseOpt = new Option("v", "verbose", false, "print out the tablet information with start/end rows");
-    
+
     opts.addOption(outputFileOpt);
     opts.addOption(maxSplitsOpt);
     opts.addOption(base64Opt);
     opts.addOption(verboseOpt);
     opts.addOption(OptUtil.tableOpt("table to get splits for"));
-    
+
     return opts;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/GrantCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/GrantCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/GrantCommand.java
index 84952e3..42d5f12 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/GrantCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/GrantCommand.java
@@ -24,9 +24,9 @@ import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
 import org.apache.accumulo.core.util.BadArgumentException;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/GrepCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/GrepCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/GrepCommand.java
index 1a3c484..97bddc9 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/GrepCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/GrepCommand.java
@@ -34,21 +34,21 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class GrepCommand extends ScanCommand {
-  
+
   private Option numThreadsOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final PrintFile printFile = getOutputFile(cl);
-    
+
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     if (cl.getArgList().isEmpty()) {
       throw new MissingArgumentException("No terms specified");
     }
     final Class<? extends Formatter> formatter = getFormatter(cl, tableName, shellState);
     final ScanInterpreter interpeter = getInterpreter(cl, tableName, shellState);
-    
+
     // handle first argument, if present, the authorizations list to
     // scan with
     int numThreads = 20;
@@ -58,25 +58,25 @@ public class GrepCommand extends ScanCommand {
     final Authorizations auths = getAuths(cl, shellState);
     final BatchScanner scanner = shellState.getConnector().createBatchScanner(tableName, auths, numThreads);
     scanner.setRanges(Collections.singletonList(getRange(cl, interpeter)));
-    
+
     scanner.setTimeout(getTimeout(cl), TimeUnit.MILLISECONDS);
-    
+
     for (int i = 0; i < cl.getArgs().length; i++) {
       setUpIterator(Integer.MAX_VALUE - cl.getArgs().length + i, "grep" + i, cl.getArgs()[i], scanner, cl);
     }
     try {
       // handle columns
       fetchColumns(cl, scanner, interpeter);
-      
+
       // output the records
       printRecords(cl, shellState, scanner, formatter, printFile);
     } finally {
       scanner.close();
     }
-    
+
     return 0;
   }
-  
+
   protected void setUpIterator(final int prio, final String name, final String term, final BatchScanner scanner, CommandLine cl) throws IOException {
     if (prio < 0) {
       throw new IllegalArgumentException("Priority < 0 " + prio);
@@ -85,12 +85,12 @@ public class GrepCommand extends ScanCommand {
     GrepIterator.setTerm(grep, term);
     scanner.addScanIterator(grep);
   }
-  
+
   @Override
   public String description() {
     return "searches each row, column family, column qualifier and value in a table for a substring (not a regular expression), in parallel, on the server side";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = super.getOptions();
@@ -98,12 +98,12 @@ public class GrepCommand extends ScanCommand {
     opts.addOption(numThreadsOpt);
     return opts;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <term>{ <term>}";
   }
-  
+
   @Override
   public int numArgs() {
     return Shell.NO_FIXED_ARG_LENGTH_CHECK;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/HelpCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/HelpCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/HelpCommand.java
index 90e8c4f..6d0426a 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/HelpCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/HelpCommand.java
@@ -23,9 +23,9 @@ import java.util.Map.Entry;
 import java.util.Set;
 
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellCommandException;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
@@ -33,7 +33,7 @@ import org.apache.commons.cli.Options;
 public class HelpCommand extends Command {
   private Option disablePaginationOpt;
   private Option noWrapOpt;
-  
+
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws ShellCommandException, IOException {
     int numColumns = shellState.getReader().getTerminal().getWidth();
     if (cl.hasOption(noWrapOpt.getOpt())) {
@@ -86,7 +86,7 @@ public class HelpCommand extends Command {
       }
       shellState.printLines(output.iterator(), !cl.hasOption(disablePaginationOpt.getOpt()));
     }
-    
+
     // print help for every command on command line
     for (String cmd : cl.getArgs()) {
       final Command c = shellState.commandFactory.get(cmd);
@@ -98,15 +98,15 @@ public class HelpCommand extends Command {
     }
     return 0;
   }
-  
+
   public String description() {
     return "provides information about the available commands";
   }
-  
+
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> special) {
     registerCompletionForCommands(root, special);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -116,12 +116,12 @@ public class HelpCommand extends Command {
     o.addOption(noWrapOpt);
     return o;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " [ <command>{ <command>} ]";
   }
-  
+
   @Override
   public int numArgs() {
     return Shell.NO_FIXED_ARG_LENGTH_CHECK;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/HiddenCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/HiddenCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/HiddenCommand.java
index f21e33a..14d6c2c 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/HiddenCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/HiddenCommand.java
@@ -23,19 +23,19 @@ import java.util.Random;
 
 import org.apache.accumulo.core.util.Base64;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.ShellCommandException;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.ShellCommandException;
 import org.apache.accumulo.shell.ShellCommandException.ErrorCode;
 import org.apache.commons.cli.CommandLine;
 
 public class HiddenCommand extends Command {
   private static Random rand = new SecureRandom();
-  
+
   @Override
   public String description() {
     return "The first rule of Accumulo is: \"You don't talk about Accumulo.\"";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     if (rand.nextInt(10) == 0) {
@@ -50,12 +50,12 @@ public class HiddenCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return Shell.NO_FIXED_ARG_LENGTH_CHECK;
   }
-  
+
   @Override
   public String getName() {
     return "accvmvlo";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/HistoryCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/HistoryCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/HistoryCommand.java
index 74817a3..785b49e 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/HistoryCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/HistoryCommand.java
@@ -33,7 +33,7 @@ import com.google.common.collect.Iterators;
 public class HistoryCommand extends Command {
   private Option clearHist;
   private Option disablePaginationOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException {
     if (cl.hasOption(clearHist.getOpt())) {
@@ -49,20 +49,20 @@ public class HistoryCommand extends Command {
 
       shellState.printLines(historyIterator, !cl.hasOption(disablePaginationOpt.getOpt()));
     }
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return ("generates a list of commands previously executed");
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ImportDirectoryCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ImportDirectoryCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ImportDirectoryCommand.java
index 56c27c2..7f75830 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ImportDirectoryCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ImportDirectoryCommand.java
@@ -26,17 +26,17 @@ import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 
 public class ImportDirectoryCommand extends Command {
-  
+
   @Override
   public String description() {
     return "bulk imports an entire directory of data files to the current table.  The boolean argument determines if accumulo sets the time.";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException, AccumuloException, AccumuloSecurityException,
       TableNotFoundException {
     shellState.checkTableState();
-    
+
     String dir = cl.getArgs()[0];
     String failureDir = cl.getArgs()[1];
     final boolean setTime = Boolean.parseBoolean(cl.getArgs()[2]);
@@ -44,15 +44,15 @@ public class ImportDirectoryCommand extends Command {
     shellState.getConnector().tableOperations().importDirectory(shellState.getTableName(), dir, failureDir, setTime);
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return 3;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <directory> <failureDirectory> true|false";
   }
-  
+
 }


[34/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/RegExFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/RegExFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/RegExFilter.java
index 8d56889..5d1a0c7 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/RegExFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/RegExFilter.java
@@ -37,9 +37,9 @@ import org.apache.log4j.Logger;
  * A Filter that matches entries based on Java regular expressions.
  */
 public class RegExFilter extends Filter {
-  
+
   private static final Logger log = Logger.getLogger(RegExFilter.class);
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     RegExFilter result = (RegExFilter) super.deepCopy(env);
@@ -50,7 +50,7 @@ public class RegExFilter extends Filter {
     result.orFields = orFields;
     return result;
   }
-  
+
   public static final String ROW_REGEX = "rowRegex";
   public static final String COLF_REGEX = "colfRegex";
   public static final String COLQ_REGEX = "colqRegex";
@@ -58,25 +58,25 @@ public class RegExFilter extends Filter {
   public static final String OR_FIELDS = "orFields";
   public static final String ENCODING = "encoding";
   public static final String MATCH_SUBSTRING = "matchSubstring";
-  
+
   public static final String ENCODING_DEFAULT = UTF_8.name();
-  
+
   private Matcher rowMatcher;
   private Matcher colfMatcher;
   private Matcher colqMatcher;
   private Matcher valueMatcher;
   private boolean orFields = false;
   private boolean matchSubstring = false;
-  
+
   private String encoding = ENCODING_DEFAULT;
-  
+
   private Matcher copyMatcher(Matcher m) {
     if (m == null)
       return m;
     else
       return m.pattern().matcher("");
   }
-  
+
   private boolean matches(Matcher matcher, ByteSequence bs) {
     if (matcher != null) {
       try {
@@ -88,7 +88,7 @@ public class RegExFilter extends Filter {
     }
     return !orFields;
   }
-  
+
   private boolean matches(Matcher matcher, byte data[], int offset, int len) {
     if (matcher != null) {
       try {
@@ -100,7 +100,7 @@ public class RegExFilter extends Filter {
     }
     return !orFields;
   }
-  
+
   @Override
   public boolean accept(Key key, Value value) {
     if (orFields)
@@ -111,7 +111,7 @@ public class RegExFilter extends Filter {
         && (matches(colfMatcher, colfMatcher == null ? null : key.getColumnFamilyData()))
         && (matches(colqMatcher, colqMatcher == null ? null : key.getColumnQualifierData())) && (matches(valueMatcher, value.get(), 0, value.get().length)));
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
@@ -120,42 +120,42 @@ public class RegExFilter extends Filter {
     } else {
       rowMatcher = null;
     }
-    
+
     if (options.containsKey(COLF_REGEX)) {
       colfMatcher = Pattern.compile(options.get(COLF_REGEX)).matcher("");
     } else {
       colfMatcher = null;
     }
-    
+
     if (options.containsKey(COLQ_REGEX)) {
       colqMatcher = Pattern.compile(options.get(COLQ_REGEX)).matcher("");
     } else {
       colqMatcher = null;
     }
-    
+
     if (options.containsKey(VALUE_REGEX)) {
       valueMatcher = Pattern.compile(options.get(VALUE_REGEX)).matcher("");
     } else {
       valueMatcher = null;
     }
-    
+
     if (options.containsKey(OR_FIELDS)) {
       orFields = Boolean.parseBoolean(options.get(OR_FIELDS));
     } else {
       orFields = false;
     }
-    
+
     if (options.containsKey(MATCH_SUBSTRING)) {
       matchSubstring = Boolean.parseBoolean(options.get(MATCH_SUBSTRING));
     } else {
       matchSubstring = false;
     }
-    
+
     if (options.containsKey(ENCODING)) {
       encoding = options.get(ENCODING);
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -170,28 +170,28 @@ public class RegExFilter extends Filter {
     io.addNamedOption(RegExFilter.ENCODING, "character encoding of byte array value (default is " + ENCODING_DEFAULT + ")");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
       return false;
-    
+
     try {
       if (options.containsKey(ROW_REGEX))
         Pattern.compile(options.get(ROW_REGEX)).matcher("");
-      
+
       if (options.containsKey(COLF_REGEX))
         Pattern.compile(options.get(COLF_REGEX)).matcher("");
-      
+
       if (options.containsKey(COLQ_REGEX))
         Pattern.compile(options.get(COLQ_REGEX)).matcher("");
-      
+
       if (options.containsKey(VALUE_REGEX))
         Pattern.compile(options.get(VALUE_REGEX)).matcher("");
     } catch (Exception e) {
       throw new IllegalArgumentException("bad regex", e);
     }
-    
+
     if (options.containsKey(ENCODING)) {
       try {
         this.encoding = options.get(ENCODING);
@@ -202,14 +202,14 @@ public class RegExFilter extends Filter {
         throw new IllegalArgumentException("invalid encoding " + ENCODING + ":" + this.encoding, e);
       }
     }
-    
+
     return true;
   }
-  
+
   /**
    * Encode the terms to match against in the iterator. Same as calling {@link #setRegexs(IteratorSetting, String, String, String, String, boolean, boolean)}
    * with matchSubstring set to false
-   * 
+   *
    * @param si
    *          ScanIterator config to be updated
    * @param rowTerm
@@ -226,10 +226,10 @@ public class RegExFilter extends Filter {
   public static void setRegexs(IteratorSetting si, String rowTerm, String cfTerm, String cqTerm, String valueTerm, boolean orFields) {
     setRegexs(si, rowTerm, cfTerm, cqTerm, valueTerm, orFields, false);
   }
-  
+
   /**
    * Encode the terms to match against in the iterator
-   * 
+   *
    * @param si
    *          ScanIterator config to be updated
    * @param rowTerm
@@ -244,7 +244,7 @@ public class RegExFilter extends Filter {
    *          if true then search expressions will match on partial strings
    */
   public static void setRegexs(IteratorSetting si, String rowTerm, String cfTerm, String cqTerm, String valueTerm, boolean orFields, boolean matchSubstring) {
-    
+
     if (rowTerm != null)
       si.addOption(RegExFilter.ROW_REGEX, rowTerm);
     if (cfTerm != null)
@@ -255,17 +255,17 @@ public class RegExFilter extends Filter {
       si.addOption(RegExFilter.VALUE_REGEX, valueTerm);
     si.addOption(RegExFilter.OR_FIELDS, String.valueOf(orFields));
     si.addOption(RegExFilter.MATCH_SUBSTRING, String.valueOf(matchSubstring));
-    
+
   }
-  
+
   /**
    * Set the encoding string to use when interpreting characters
-   * 
+   *
    * @param si
    *          ScanIterator config to be updated
    * @param encoding
    *          the encoding string to use for character interpretation.
-   * 
+   *
    */
   public static void setEncoding(IteratorSetting si, String encoding) {
     if (!encoding.isEmpty()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/ReqVisFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/ReqVisFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/ReqVisFilter.java
index 754b2c3..d7f85f5 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/ReqVisFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/ReqVisFilter.java
@@ -25,13 +25,13 @@ import org.apache.accumulo.core.security.ColumnVisibility;
  * A Filter that matches entries with a non-empty ColumnVisibility.
  */
 public class ReqVisFilter extends Filter {
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     ColumnVisibility vis = new ColumnVisibility(k.getColumnVisibility());
     return vis.getExpression().length > 0;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/RowDeletingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/RowDeletingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/RowDeletingIterator.java
index 3c24f0e..60870d8 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/RowDeletingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/RowDeletingIterator.java
@@ -34,85 +34,85 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 /**
  * An iterator for deleting whole rows.
- * 
+ *
  * After setting this iterator up for your table, to delete a row insert a row with empty column family, empty column qualifier, empty column visibility, and a
  * value of DEL_ROW. Do not use empty columns for anything else when using this iterator.
- * 
+ *
  * When using this iterator the locality group containing the row deletes will always be read. The locality group containing the empty column family will
  * contain row deletes. Always reading this locality group can have an impact on performance.
- * 
+ *
  * For example assume there are two locality groups, one containing large images and one containing small metadata about the images. If row deletes are in the
  * same locality group as the images, then this will significantly slow down scans and major compactions that are only reading the metadata locality group.
  * Therefore, you would want to put the empty column family in the locality group that contains the metadata. Another option is to put the empty column in its
  * own locality group. Which is best depends on your data.
- * 
+ *
  */
 
 public class RowDeletingIterator implements SortedKeyValueIterator<Key,Value> {
-  
+
   public static final Value DELETE_ROW_VALUE = new Value("DEL_ROW".getBytes(UTF_8));
   private SortedKeyValueIterator<Key,Value> source;
   private boolean propogateDeletes;
   private ByteSequence currentRow;
   private boolean currentRowDeleted;
   private long deleteTS;
-  
+
   private boolean dropEmptyColFams;
-  
+
   private static final ByteSequence EMPTY = new ArrayByteSequence(new byte[] {});
-  
+
   private RowDeletingIterator(SortedKeyValueIterator<Key,Value> source, boolean propogateDeletes2) {
     this.source = source;
     this.propogateDeletes = propogateDeletes2;
   }
-  
+
   public RowDeletingIterator() {}
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new RowDeletingIterator(source.deepCopy(env), propogateDeletes);
   }
-  
+
   @Override
   public Key getTopKey() {
     return source.getTopKey();
   }
-  
+
   @Override
   public Value getTopValue() {
     return source.getTopValue();
   }
-  
+
   @Override
   public boolean hasTop() {
     return source.hasTop();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     this.source = source;
     this.propogateDeletes = (env.getIteratorScope() == IteratorScope.majc && !env.isFullMajorCompaction()) || env.getIteratorScope() == IteratorScope.minc;
   }
-  
+
   @Override
   public void next() throws IOException {
     source.next();
     consumeDeleted();
     consumeEmptyColFams();
   }
-  
+
   private void consumeEmptyColFams() throws IOException {
     while (dropEmptyColFams && source.hasTop() && source.getTopKey().getColumnFamilyData().length() == 0) {
       source.next();
       consumeDeleted();
     }
   }
-  
+
   private boolean isDeleteMarker(Key key, Value val) {
     return key.getColumnFamilyData().length() == 0 && key.getColumnQualifierData().length() == 0 && key.getColumnVisibilityData().length() == 0
         && val.equals(DELETE_ROW_VALUE);
   }
-  
+
   private void consumeDeleted() throws IOException {
     // this method tries to do as little work as possible when nothing is deleted
     while (source.hasTop()) {
@@ -120,29 +120,29 @@ public class RowDeletingIterator implements SortedKeyValueIterator<Key,Value> {
         while (source.hasTop() && currentRow.equals(source.getTopKey().getRowData()) && source.getTopKey().getTimestamp() <= deleteTS) {
           source.next();
         }
-        
+
         if (source.hasTop() && !currentRow.equals(source.getTopKey().getRowData())) {
           currentRowDeleted = false;
         }
       }
-      
+
       if (!currentRowDeleted && source.hasTop() && isDeleteMarker(source.getTopKey(), source.getTopValue())) {
         currentRow = source.getTopKey().getRowData();
         currentRowDeleted = true;
         deleteTS = source.getTopKey().getTimestamp();
-        
+
         if (propogateDeletes)
           break;
       } else {
         break;
       }
     }
-    
+
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-    
+
     if (inclusive && !columnFamilies.contains(EMPTY)) {
       columnFamilies = new HashSet<ByteSequence>(columnFamilies);
       columnFamilies.add(EMPTY);
@@ -154,16 +154,16 @@ public class RowDeletingIterator implements SortedKeyValueIterator<Key,Value> {
     } else {
       dropEmptyColFams = false;
     }
-    
+
     currentRowDeleted = false;
-    
+
     if (range.getStartKey() != null) {
       // seek to beginning of row
       Range newRange = new Range(new Key(range.getStartKey().getRow()), true, range.getEndKey(), range.isEndKeyInclusive());
       source.seek(newRange, columnFamilies, inclusive);
       consumeDeleted();
       consumeEmptyColFams();
-      
+
       if (source.hasTop() && range.beforeStartKey(source.getTopKey())) {
         source.seek(range, columnFamilies, inclusive);
         consumeDeleted();
@@ -174,7 +174,7 @@ public class RowDeletingIterator implements SortedKeyValueIterator<Key,Value> {
       consumeDeleted();
       consumeEmptyColFams();
     }
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/RowEncodingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/RowEncodingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/RowEncodingIterator.java
index ce38b08..ef4003c 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/RowEncodingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/RowEncodingIterator.java
@@ -33,23 +33,23 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  * The RowEncodingIterator is designed to provide row-isolation so that queries see mutations as atomic. It does so by encapsulating an entire row of key/value
  * pairs into a single key/value pair, which is returned through the client as an atomic operation. This is an abstract class, allowing the user to implement
  * rowEncoder and rowDecoder such that the columns and values of a given row may be encoded in a format best suited to the client.
- * 
+ *
  * <p>
  * For an example implementation, see {@link WholeRowIterator}.
- * 
+ *
  * <p>
  * One caveat is that when seeking in the WholeRowIterator using a range that starts at a non-inclusive first key in a row, (e.g. seek(new Range(new Key(new
  * Text("row")),false,...),...)) this iterator will skip to the next row. This is done in order to prevent repeated scanning of the same row when system
  * automatically creates ranges of that form, which happens in the case of the client calling continueScan, or in the case of the tablet server continuing a
  * scan after swapping out sources.
- * 
+ *
  * <p>
  * To regain the original key/value pairs of the row, call the rowDecoder function on the key/value pair that this iterator returned.
- * 
+ *
  * @see RowFilter
  */
 public abstract class RowEncodingIterator implements SortedKeyValueIterator<Key,Value> {
@@ -101,7 +101,7 @@ public abstract class RowEncodingIterator implements SortedKeyValueIterator<Key,
   }
 
   /**
-   * 
+   *
    * @param currentRow
    *          All keys have this in their row portion (do not modify!).
    * @param keys

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/RowFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/RowFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/RowFilter.java
index 9c4edc2..1287b81 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/RowFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/RowFilter.java
@@ -34,21 +34,21 @@ import org.apache.hadoop.io.Text;
 /**
  * This iterator makes it easy to select rows that meet a given criteria. Its an alternative to the {@link WholeRowIterator}. There are a few things to consider
  * when deciding which one to use.
- * 
+ *
  * First the WholeRowIterator requires that the row fit in memory and that the entire row is read before a decision is made. This iterator has neither
  * requirement, it allows seeking within a row to avoid reading the entire row to make a decision. So even if your rows fit into memory, this extending this
  * iterator may be better choice because you can seek.
- * 
+ *
  * Second the WholeRowIterator is currently the only way to achieve row isolation with the {@link BatchScanner}. With the normal {@link Scanner} row isolation
  * can be enabled and this Iterator may be used.
- * 
+ *
  * Third the row acceptance test will be executed every time this Iterator is seeked. If the row is large, then the row will fetched in batches of key/values.
  * As each batch is fetched the test may be re-executed because the iterator stack is reseeked for each batch. The batch size may be increased to reduce the
  * number of times the test is executed. With the normal Scanner, if isolation is enabled then it will read an entire row w/o seeking this iterator.
- * 
+ *
  */
 public abstract class RowFilter extends WrappingIterator {
-  
+
   private RowIterator decisionIterator;
   private Collection<ByteSequence> columnFamilies;
   Text currentRow;
@@ -59,23 +59,23 @@ public abstract class RowFilter extends WrappingIterator {
   private static class RowIterator extends WrappingIterator {
     private Range rowRange;
     private boolean hasTop;
-    
+
     RowIterator(SortedKeyValueIterator<Key,Value> source) {
       super.setSource(source);
     }
-    
+
     void setRow(Range row) {
       this.rowRange = row;
     }
-    
+
     @Override
     public boolean hasTop() {
       return hasTop && super.hasTop();
     }
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-      
+
       range = rowRange.clip(range, true);
       if (range == null) {
         hasTop = false;
@@ -90,14 +90,14 @@ public abstract class RowFilter extends WrappingIterator {
     SortedKeyValueIterator<Key,Value> source = getSource();
     while (source.hasTop()) {
       Text row = source.getTopKey().getRow();
-      
+
       if (currentRow != null && currentRow.equals(row))
         break;
-      
+
       Range rowRange = new Range(row);
       decisionIterator.setRow(rowRange);
       decisionIterator.seek(rowRange, columnFamilies, inclusive);
-      
+
       if (acceptRow(decisionIterator)) {
         currentRow = row;
         break;
@@ -108,7 +108,7 @@ public abstract class RowFilter extends WrappingIterator {
           count++;
           source.next();
         }
-        
+
         if (source.hasTop() && source.getTopKey().getRow().equals(row)) {
           Range nextRow = new Range(row, false, null, false);
           nextRow = range.clip(nextRow, true);
@@ -120,11 +120,11 @@ public abstract class RowFilter extends WrappingIterator {
       }
     }
   }
-  
+
   /**
    * Implementation should return false to suppress a row.
-   * 
-   * 
+   *
+   *
    * @param rowIterator
    *          - An iterator over the row. This iterator is confined to the row. Seeking past the end of the row will return no data. Seeking before the row will
    *          always set top to the first column in the current row. By default this iterator will only see the columns the parent was seeked with. To see more
@@ -156,13 +156,13 @@ public abstract class RowFilter extends WrappingIterator {
   public boolean hasTop() {
     return hasTop && super.hasTop();
   }
-  
+
   @Override
   public void next() throws IOException {
     super.next();
     skipRows();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     super.seek(range, columnFamilies, inclusive);
@@ -172,6 +172,6 @@ public abstract class RowFilter extends WrappingIterator {
     currentRow = null;
     hasTop = true;
     skipRows();
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
index 9bdc883..3c9bdd5 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
@@ -46,10 +46,10 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
   public static final Encoder<List<Long>> FIXED_LONG_ARRAY_ENCODER = new FixedLongArrayEncoder();
   public static final Encoder<List<Long>> VAR_LONG_ARRAY_ENCODER = new VarLongArrayEncoder();
   public static final Encoder<List<Long>> STRING_ARRAY_ENCODER = new StringArrayEncoder();
-  
+
   private static final String TYPE = "type";
   private static final String CLASS_PREFIX = "class:";
-  
+
   public static enum Type {
     /**
      * indicates a variable-length encoding of a list of Longs using {@link SummingArrayCombiner.VarLongArrayEncoder}
@@ -64,7 +64,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
      */
     STRING
   }
-  
+
   @Override
   public List<Long> typedReduce(Key key, Iterator<List<Long>> iter) {
     List<Long> sum = new ArrayList<Long>();
@@ -73,7 +73,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
     }
     return sum;
   }
-  
+
   public static List<Long> arrayAdd(List<Long> la, List<Long> lb) {
     if (la.size() > lb.size()) {
       for (int i = 0; i < lb.size(); i++) {
@@ -87,13 +87,13 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
       return lb;
     }
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     setEncoder(options);
   }
-  
+
   private void setEncoder(Map<String,String> options) {
     String type = options.get(TYPE);
     if (type == null)
@@ -117,7 +117,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
       }
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -126,7 +126,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
     io.addNamedOption(TYPE, "<VARLEN|FIXEDLEN|STRING|fullClassName>");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
@@ -138,12 +138,12 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
     }
     return true;
   }
-  
+
   public abstract static class DOSArrayEncoder<V> implements Encoder<List<V>> {
     public abstract void write(DataOutputStream dos, V v) throws IOException;
-    
+
     public abstract V read(DataInputStream dis) throws IOException;
-    
+
     @Override
     public byte[] encode(List<V> vl) {
       ByteArrayOutputStream baos = new ByteArrayOutputStream();
@@ -158,7 +158,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
       }
       return baos.toByteArray();
     }
-    
+
     @Override
     public List<V> decode(byte[] b) {
       DataInputStream dis = new DataInputStream(new ByteArrayInputStream(b));
@@ -174,31 +174,31 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
       }
     }
   }
-  
+
   public static class VarLongArrayEncoder extends DOSArrayEncoder<Long> {
     @Override
     public void write(DataOutputStream dos, Long v) throws IOException {
       WritableUtils.writeVLong(dos, v);
     }
-    
+
     @Override
     public Long read(DataInputStream dis) throws IOException {
       return WritableUtils.readVLong(dis);
     }
   }
-  
+
   public static class FixedLongArrayEncoder extends DOSArrayEncoder<Long> {
     @Override
     public void write(DataOutputStream dos, Long v) throws IOException {
       dos.writeLong(v);
     }
-    
+
     @Override
     public Long read(DataInputStream dis) throws IOException {
       return dis.readLong();
     }
   }
-  
+
   public static class StringArrayEncoder implements Encoder<List<Long>> {
     @Override
     public byte[] encode(List<Long> la) {
@@ -211,7 +211,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
       }
       return sb.toString().getBytes(UTF_8);
     }
-    
+
     @Override
     public List<Long> decode(byte[] b) {
       String[] longstrs = new String(b, UTF_8).split(",");
@@ -229,10 +229,10 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
       return la;
     }
   }
-  
+
   /**
    * A convenience method for setting the encoding type.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param type
@@ -241,10 +241,10 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
   public static void setEncodingType(IteratorSetting is, Type type) {
     is.addOption(TYPE, type.toString());
   }
-  
+
   /**
    * A convenience method for setting the encoding type.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
@@ -253,10 +253,10 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<List<Long>>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());
   }
-  
+
   /**
    * A convenience method for setting the encoding type.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClassName

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingCombiner.java
index e2c5f65..68550b5 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingCombiner.java
@@ -34,7 +34,7 @@ public class SummingCombiner extends LongCombiner {
     }
     return sum;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/TimestampFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/TimestampFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/TimestampFilter.java
index 8747aa6..3cfbf5b 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/TimestampFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/TimestampFilter.java
@@ -35,13 +35,13 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 public class TimestampFilter extends Filter {
   private static final String LONG_PREFIX = "LONG";
   private final SimpleDateFormat dateParser = initDateParser();
-  
+
   private static SimpleDateFormat initDateParser() {
     SimpleDateFormat dateParser = new SimpleDateFormat("yyyyMMddHHmmssz");
     dateParser.setTimeZone(TimeZone.getTimeZone("GMT"));
     return dateParser;
   }
-  
+
   public static final String START = "start";
   public static final String START_INCL = "startInclusive";
   public static final String END = "end";
@@ -52,9 +52,9 @@ public class TimestampFilter extends Filter {
   private boolean endInclusive;
   private boolean hasStart;
   private boolean hasEnd;
-  
+
   public TimestampFilter() {}
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     long ts = k.getTimestamp();
@@ -66,26 +66,26 @@ public class TimestampFilter extends Filter {
       return false;
     return true;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     if (options == null)
       throw new IllegalArgumentException("start and/or end must be set for " + TimestampFilter.class.getName());
-    
+
     super.init(source, options, env);
-    
+
     hasStart = false;
     hasEnd = false;
     startInclusive = true;
     endInclusive = true;
-    
+
     if (options.containsKey(START))
       hasStart = true;
     if (options.containsKey(END))
       hasEnd = true;
     if (!hasStart && !hasEnd)
       throw new IllegalArgumentException("must have either start or end for " + TimestampFilter.class.getName());
-    
+
     try {
       if (hasStart) {
         String s = options.get(START);
@@ -109,7 +109,7 @@ public class TimestampFilter extends Filter {
     if (options.get(END_INCL) != null)
       endInclusive = Boolean.parseBoolean(options.get(END_INCL));
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     TimestampFilter copy = (TimestampFilter) super.deepCopy(env);
@@ -121,7 +121,7 @@ public class TimestampFilter extends Filter {
     copy.endInclusive = endInclusive;
     return copy;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -133,7 +133,7 @@ public class TimestampFilter extends Filter {
     io.addNamedOption("endInclusive", "true or false");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
@@ -168,10 +168,10 @@ public class TimestampFilter extends Filter {
     }
     return true;
   }
-  
+
   /**
    * A convenience method for setting the range of timestamps accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param start
@@ -182,10 +182,10 @@ public class TimestampFilter extends Filter {
   public static void setRange(IteratorSetting is, String start, String end) {
     setRange(is, start, true, end, true);
   }
-  
+
   /**
    * A convenience method for setting the range of timestamps accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param start
@@ -201,10 +201,10 @@ public class TimestampFilter extends Filter {
     setStart(is, start, startInclusive);
     setEnd(is, end, endInclusive);
   }
-  
+
   /**
    * A convenience method for setting the start timestamp accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param start
@@ -221,10 +221,10 @@ public class TimestampFilter extends Filter {
       throw new IllegalArgumentException("couldn't parse " + start);
     }
   }
-  
+
   /**
    * A convenience method for setting the end timestamp accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param end
@@ -241,10 +241,10 @@ public class TimestampFilter extends Filter {
       throw new IllegalArgumentException("couldn't parse " + end);
     }
   }
-  
+
   /**
    * A convenience method for setting the range of timestamps accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param start
@@ -255,10 +255,10 @@ public class TimestampFilter extends Filter {
   public static void setRange(IteratorSetting is, long start, long end) {
     setRange(is, start, true, end, true);
   }
-  
+
   /**
    * A convenience method for setting the range of timestamps accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param start
@@ -274,10 +274,10 @@ public class TimestampFilter extends Filter {
     setStart(is, start, startInclusive);
     setEnd(is, end, endInclusive);
   }
-  
+
   /**
    * A convenience method for setting the start timestamp accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param start
@@ -289,10 +289,10 @@ public class TimestampFilter extends Filter {
     is.addOption(START, LONG_PREFIX + Long.toString(start));
     is.addOption(START_INCL, Boolean.toString(startInclusive));
   }
-  
+
   /**
    * A convenience method for setting the end timestamp accepted by the timestamp filter.
-   * 
+   *
    * @param is
    *          the iterator setting object to configure
    * @param end

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/TransformingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/TransformingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/TransformingIterator.java
index 6b51ac5..9d6c290 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/TransformingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/TransformingIterator.java
@@ -629,7 +629,7 @@ abstract public class TransformingIterator extends WrappingIterator implements O
    * @return the part of the key this iterator is not transforming
    */
   abstract protected PartialKey getKeyPrefix();
-  
+
   public interface KVBuffer {
     void append(Key key, Value val);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/VersioningIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/VersioningIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/VersioningIterator.java
index 2fc3a27..88ba20d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/VersioningIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/VersioningIterator.java
@@ -35,15 +35,15 @@ import org.apache.accumulo.core.iterators.WrappingIterator;
 
 public class VersioningIterator extends WrappingIterator implements OptionDescriber {
   private final int maxCount = 10;
-  
+
   private Key currentKey = new Key();
   private int numVersions;
   protected int maxVersions;
-  
+
   private Range range;
   private Collection<ByteSequence> columnFamilies;
   private boolean inclusive;
-  
+
   @Override
   public VersioningIterator deepCopy(IteratorEnvironment env) {
     VersioningIterator copy = new VersioningIterator();
@@ -51,7 +51,7 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
     copy.maxVersions = maxVersions;
     return copy;
   }
-  
+
   @Override
   public void next() throws IOException {
     if (numVersions >= maxVersions) {
@@ -59,7 +59,7 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
       resetVersionCount();
       return;
     }
-    
+
     super.next();
     if (getSource().hasTop()) {
       if (getSource().getTopKey().equals(currentKey, PartialKey.ROW_COLFAM_COLQUAL_COLVIS)) {
@@ -69,7 +69,7 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
       }
     }
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     // do not want to seek to the middle of a row
@@ -77,25 +77,25 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
     this.range = seekRange;
     this.columnFamilies = columnFamilies;
     this.inclusive = inclusive;
-    
+
     super.seek(seekRange, columnFamilies, inclusive);
     resetVersionCount();
-    
+
     if (range.getStartKey() != null)
       while (hasTop() && range.beforeStartKey(getTopKey()))
         next();
   }
-  
+
   private void resetVersionCount() {
     if (super.hasTop())
       currentKey.set(getSource().getTopKey());
     numVersions = 1;
   }
-  
+
   private void skipRowColumn() throws IOException {
     Key keyToSkip = currentKey;
     super.next();
-    
+
     int count = 0;
     while (getSource().hasTop() && getSource().getTopKey().equals(keyToSkip, PartialKey.ROW_COLFAM_COLQUAL_COLVIS)) {
       if (count < maxCount) {
@@ -109,7 +109,7 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
       }
     }
   }
-  
+
   protected void reseek(Key key) throws IOException {
     if (key == null)
       return;
@@ -121,30 +121,30 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
       getSource().seek(range, columnFamilies, inclusive);
     }
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     this.numVersions = 0;
-    
+
     String maxVerString = options.get("maxVersions");
     if (maxVerString != null)
       this.maxVersions = Integer.parseInt(maxVerString);
     else
       this.maxVersions = 1;
-    
+
     if (maxVersions < 1)
       throw new IllegalArgumentException("maxVersions for versioning iterator must be >= 1");
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     return new IteratorOptions("vers", "The VersioningIterator keeps a fixed number of versions for each key", Collections.singletonMap("maxVersions",
         "number of versions to keep for a particular key (with differing timestamps)"), null);
   }
-  
+
   private static final String MAXVERSIONS_OPT = "maxVersions";
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     int i;
@@ -157,7 +157,7 @@ public class VersioningIterator extends WrappingIterator implements OptionDescri
       throw new IllegalArgumentException(MAXVERSIONS_OPT + " for versioning iterator must be >= 1");
     return true;
   }
-  
+
   /**
    * Encode the maximum number of versions to return onto the ScanIterator
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/VisibilityFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/VisibilityFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/VisibilityFilter.java
index 878aa8e..6e55aec 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/VisibilityFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/VisibilityFilter.java
@@ -35,28 +35,28 @@ import org.apache.commons.collections.map.LRUMap;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class VisibilityFilter extends org.apache.accumulo.core.iterators.system.VisibilityFilter implements OptionDescriber {
-  
+
   private static final String AUTHS = "auths";
   private static final String FILTER_INVALID_ONLY = "filterInvalid";
-  
+
   private boolean filterInvalid;
-  
+
   /**
-   * 
+   *
    */
   public VisibilityFilter() {
     super();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     validateOptions(options);
     this.filterInvalid = Boolean.parseBoolean(options.get(FILTER_INVALID_ONLY));
-    
+
     if (!filterInvalid) {
       String auths = options.get(AUTHS);
       Authorizations authObj = auths == null || auths.isEmpty() ? new Authorizations() : new Authorizations(auths.getBytes(UTF_8));
@@ -66,7 +66,7 @@ public class VisibilityFilter extends org.apache.accumulo.core.iterators.system.
     this.cache = new LRUMap(1000);
     this.tmpVis = new Text();
   }
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     if (filterInvalid) {
@@ -86,7 +86,7 @@ public class VisibilityFilter extends org.apache.accumulo.core.iterators.system.
       return super.accept(k, v);
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -97,13 +97,13 @@ public class VisibilityFilter extends org.apache.accumulo.core.iterators.system.
     io.addNamedOption(AUTHS, "the serialized set of authorizations to filter against (default: empty string, accepts only entries visible by all)");
     return io;
   }
-  
+
   public static void setAuthorizations(IteratorSetting setting, Authorizations auths) {
     setting.addOption(AUTHS, auths.serialize());
   }
-  
+
   public static void filterInvalidLabelsOnly(IteratorSetting setting, boolean featureEnabled) {
     setting.addOption(FILTER_INVALID_ONLY, Boolean.toString(featureEnabled));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIterator.java
index 037f9a5..25f30a8 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIterator.java
@@ -39,31 +39,31 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  * The WholeColumnFamilyIterator is designed to provide row/cf-isolation so that queries see mutations as atomic. It does so by grouping row/Column family (as
  * key) and rest of data as Value into a single key/value pair, which is returned through the client as an atomic operation.
- * 
+ *
  * To regain the original key/value pairs of the row, call the decodeRow function on the key/value pair that this iterator returned.
- * 
+ *
  * @since 1.6.0
  */
 public class WholeColumnFamilyIterator implements SortedKeyValueIterator<Key,Value>, OptionDescriber {
-  
+
   private SortedKeyValueIterator<Key,Value> sourceIter;
   private Key topKey = null;
   private Value topValue = null;
-  
+
   public WholeColumnFamilyIterator() {
 
   }
-  
+
   WholeColumnFamilyIterator(SortedKeyValueIterator<Key,Value> source) {
     this.sourceIter = source;
   }
-  
+
   /**
    * Decode whole row/cf out of value. decode key value pairs that have been encoded into a single // value
-   * 
+   *
    * @param rowKey
    *          the row key to decode
    * @param rowValue
@@ -105,11 +105,11 @@ public class WholeColumnFamilyIterator implements SortedKeyValueIterator<Key,Val
     }
     return map;
   }
-  
+
   /**
    * Encode row/cf. Take a stream of keys and values and output a value that encodes everything but their row and column families keys and values must be paired
    * one for one
-   * 
+   *
    * @param keys
    *          the row keys to encode into value
    * @param values
@@ -144,25 +144,25 @@ public class WholeColumnFamilyIterator implements SortedKeyValueIterator<Key,Val
       dout.writeInt(valBytes.length);
       dout.write(valBytes);
     }
-    
+
     return new Value(out.toByteArray());
   }
-  
+
   List<Key> keys = new ArrayList<Key>();
   List<Value> values = new ArrayList<Value>();
-  
+
   private void prepKeys() throws IOException {
     if (topKey != null)
       return;
     Text currentRow;
     Text currentCf;
-    
+
     do {
       if (sourceIter.hasTop() == false)
         return;
       currentRow = new Text(sourceIter.getTopKey().getRow());
       currentCf = new Text(sourceIter.getTopKey().getColumnFamily());
-      
+
       keys.clear();
       values.clear();
       while (sourceIter.hasTop() && sourceIter.getTopKey().getRow().equals(currentRow) && sourceIter.getTopKey().getColumnFamily().equals(currentCf)) {
@@ -171,14 +171,14 @@ public class WholeColumnFamilyIterator implements SortedKeyValueIterator<Key,Val
         sourceIter.next();
       }
     } while (!filter(currentRow, keys, values));
-    
+
     topKey = new Key(currentRow, currentCf);
     topValue = encodeColumnFamily(keys, values);
-    
+
   }
-  
+
   /**
-   * 
+   *
    * @param currentRow
    *          All keys & cf have this in their row portion (do not modify!).
    * @param keys
@@ -190,48 +190,48 @@ public class WholeColumnFamilyIterator implements SortedKeyValueIterator<Key,Val
   protected boolean filter(Text currentRow, List<Key> keys, List<Value> values) {
     return true;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     if (sourceIter != null)
       return new WholeColumnFamilyIterator(sourceIter.deepCopy(env));
     return new WholeColumnFamilyIterator();
   }
-  
+
   @Override
   public Key getTopKey() {
     return topKey;
   }
-  
+
   @Override
   public Value getTopValue() {
     return topValue;
   }
-  
+
   @Override
   public boolean hasTop() {
     return topKey != null || sourceIter.hasTop();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     sourceIter = source;
   }
-  
+
   @Override
   public void next() throws IOException {
     topKey = null;
     topValue = null;
     prepKeys();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     topKey = null;
     topValue = null;
-    
+
     Key sk = range.getStartKey();
-    
+
     if (sk != null && sk.getColumnQualifierData().length() == 0 && sk.getColumnVisibilityData().length() == 0 && sk.getTimestamp() == Long.MAX_VALUE
         && !range.isStartKeyInclusive()) {
       // assuming that we are seeking using a key previously returned by
@@ -240,22 +240,22 @@ public class WholeColumnFamilyIterator implements SortedKeyValueIterator<Key,Val
       Key followingRowKey = sk.followingKey(PartialKey.ROW_COLFAM);
       if (range.getEndKey() != null && followingRowKey.compareTo(range.getEndKey()) > 0)
         return;
-      
+
       range = new Range(sk.followingKey(PartialKey.ROW_COLFAM), true, range.getEndKey(), range.isEndKeyInclusive());
     }
-    
+
     sourceIter.seek(range, columnFamilies, inclusive);
     prepKeys();
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     return new IteratorOptions("wholecolumnfamilyiterator", "WholeColumnFamilyIterator. Group equal row & column family into single row entry.", null, null);
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     return true;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeRowIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeRowIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeRowIterator.java
index 4b7802d..7c47ec3 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeRowIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/WholeRowIterator.java
@@ -32,26 +32,26 @@ import org.apache.accumulo.core.iterators.IteratorEnvironment;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 /**
- * 
+ *
  * The WholeRowIterator is designed to provide row-isolation so that queries see mutations as atomic. It does so by encapsulating an entire row of key/value
  * pairs into a single key/value pair, which is returned through the client as an atomic operation.
- * 
+ *
  * <p>
  * This iterator extends the {@link RowEncodingIterator}, providing implementations for rowEncoder and rowDecoder which serializes all column and value
  * information from a given row into a single ByteStream in a value.
- * 
+ *
  * <p>
  * As with the RowEncodingIterator, when seeking in the WholeRowIterator using a range that starts at a non-inclusive first key in a row, this iterator will
  * skip to the next row.
- * 
+ *
  * <p>
  * To regain the original key/value pairs of the row, call the decodeRow function on the key/value pair that this iterator returned.
- * 
+ *
  * @see RowFilter
  */
 public class WholeRowIterator extends RowEncodingIterator {
   public WholeRowIterator() {}
-  
+
   WholeRowIterator(SortedKeyValueIterator<Key,Value> source) {
     this.sourceIter = source;
   }
@@ -74,8 +74,8 @@ public class WholeRowIterator extends RowEncodingIterator {
   }
 
   /**
-   * Returns the byte array containing the field of row key from the given DataInputStream din.
-   * Assumes that din first has the length of the field, followed by the field itself.
+   * Returns the byte array containing the field of row key from the given DataInputStream din. Assumes that din first has the length of the field, followed by
+   * the field itself.
    */
   private static byte[] readField(DataInputStream din) throws IOException {
     int len = din.readInt();
@@ -85,8 +85,7 @@ public class WholeRowIterator extends RowEncodingIterator {
     // We ignore the zero length case because DataInputStream.read can return -1
     // if zero length was expected and end of stream has been reached.
     if (len > 0 && len != readLen) {
-      throw new IOException(String.format("Expected to read %d bytes but read %d",
-          len, readLen));
+      throw new IOException(String.format("Expected to read %d bytes but read %d", len, readLen));
     }
     return b;
   }
@@ -107,7 +106,7 @@ public class WholeRowIterator extends RowEncodingIterator {
     }
     return map;
   }
-  
+
   // take a stream of keys and values and output a value that encodes everything but their row
   // keys and values must be paired one for one
   public static final Value encodeRow(List<Key> keys, List<Value> values) throws IOException {
@@ -142,7 +141,7 @@ public class WholeRowIterator extends RowEncodingIterator {
       dout.writeInt(valBytes.length);
       dout.write(valBytes);
     }
-    
+
     return new Value(out.toByteArray());
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/master/state/tables/TableState.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/master/state/tables/TableState.java b/core/src/main/java/org/apache/accumulo/core/master/state/tables/TableState.java
index 8cac10c..cbf6f4c 100644
--- a/core/src/main/java/org/apache/accumulo/core/master/state/tables/TableState.java
+++ b/core/src/main/java/org/apache/accumulo/core/master/state/tables/TableState.java
@@ -19,16 +19,16 @@ package org.apache.accumulo.core.master.state.tables;
 public enum TableState {
   // NEW while making directories and tablets;
   NEW,
-  
+
   // ONLINE tablets will be assigned
   ONLINE,
-  
+
   // OFFLINE tablets will be taken offline
   OFFLINE,
-  
+
   // DELETING waiting for tablets to go offline and table will be removed
   DELETING,
-  
+
   // UNKNOWN is NOT a valid state; it is reserved for unrecognized serialized
   // representations of table state
   UNKNOWN;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/MetadataLocationObtainer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/MetadataLocationObtainer.java b/core/src/main/java/org/apache/accumulo/core/metadata/MetadataLocationObtainer.java
index b40a6bd..224aafd 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/MetadataLocationObtainer.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/MetadataLocationObtainer.java
@@ -62,30 +62,30 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
   private static final Logger log = Logger.getLogger(MetadataLocationObtainer.class);
   private SortedSet<Column> locCols;
   private ArrayList<Column> columns;
-  
+
   public MetadataLocationObtainer() {
-    
+
     locCols = new TreeSet<Column>();
     locCols.add(new Column(TextUtil.getBytes(TabletsSection.CurrentLocationColumnFamily.NAME), null, null));
     locCols.add(TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.toColumn());
     columns = new ArrayList<Column>(locCols);
   }
-  
+
   @Override
   public TabletLocations lookupTablet(ClientContext context, TabletLocation src, Text row, Text stopRow, TabletLocator parent)
       throws AccumuloSecurityException, AccumuloException {
-    
+
     try {
       OpTimer opTimer = null;
       if (log.isTraceEnabled())
         opTimer = new OpTimer(log, Level.TRACE).start("Looking up in " + src.tablet_extent.getTableId() + " row=" + TextUtil.truncate(row) + "  extent="
             + src.tablet_extent + " tserver=" + src.tablet_location);
-      
+
       Range range = new Range(row, true, stopRow, true);
-      
+
       TreeMap<Key,Value> encodedResults = new TreeMap<Key,Value>();
       TreeMap<Key,Value> results = new TreeMap<Key,Value>();
-      
+
       // Use the whole row iterator so that a partial mutations is not read. The code that extracts locations for tablets does a sanity check to ensure there is
       // only one location. Reading a partial mutation could make it appear there are multiple locations when there are not.
       List<IterInfo> serverSideIteratorList = new ArrayList<IterInfo>();
@@ -93,25 +93,25 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
       Map<String,Map<String,String>> serverSideIteratorOptions = Collections.emptyMap();
       boolean more = ThriftScanner.getBatchFromServer(context, range, src.tablet_extent, src.tablet_location, encodedResults, locCols, serverSideIteratorList,
           serverSideIteratorOptions, Constants.SCAN_BATCH_SIZE, Authorizations.EMPTY, false);
-      
+
       decodeRows(encodedResults, results);
-      
+
       if (more && results.size() == 1) {
         range = new Range(results.lastKey().followingKey(PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME), true, new Key(stopRow).followingKey(PartialKey.ROW), false);
         encodedResults.clear();
         more = ThriftScanner.getBatchFromServer(context, range, src.tablet_extent, src.tablet_location, encodedResults, locCols, serverSideIteratorList,
             serverSideIteratorOptions, Constants.SCAN_BATCH_SIZE, Authorizations.EMPTY, false);
-        
+
         decodeRows(encodedResults, results);
       }
-      
+
       if (opTimer != null)
         opTimer.stop("Got " + results.size() + " results  from " + src.tablet_extent + " in %DURATION%");
-      
-      //if (log.isTraceEnabled()) log.trace("results "+results);
-      
+
+      // if (log.isTraceEnabled()) log.trace("results "+results);
+
       return MetadataLocationObtainer.getMetadataLocationEntries(results);
-      
+
     } catch (AccumuloServerException ase) {
       if (log.isTraceEnabled())
         log.trace(src.tablet_extent.getTableId() + " lookup failed, " + src.tablet_location + " server side exception");
@@ -125,10 +125,10 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
         log.trace(src.tablet_extent.getTableId() + " lookup failed", e);
       parent.invalidateCache(context.getInstance(), src.tablet_location);
     }
-    
+
     return null;
   }
-  
+
   private void decodeRows(TreeMap<Key,Value> encodedResults, TreeMap<Key,Value> results) throws AccumuloException {
     for (Entry<Key,Value> entry : encodedResults.entrySet()) {
       try {
@@ -138,15 +138,15 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
       }
     }
   }
-  
+
   @Override
   public List<TabletLocation> lookupTablets(ClientContext context, String tserver, Map<KeyExtent,List<Range>> tabletsRanges, TabletLocator parent)
       throws AccumuloSecurityException, AccumuloException {
-    
+
     final TreeMap<Key,Value> results = new TreeMap<Key,Value>();
 
     ResultReceiver rr = new ResultReceiver() {
-      
+
       @Override
       public void receive(List<Entry<Key,Value>> entries) {
         for (Entry<Key,Value> entry : entries) {
@@ -158,7 +158,7 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
         }
       }
     };
-    
+
     ScannerOptions opts = new ScannerOptions() {
       ScannerOptions setOpts() {
         this.fetchedColumns = locCols;
@@ -168,7 +168,7 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
         return this;
       }
     }.setOpts();
-    
+
     Map<KeyExtent,List<Range>> unscanned = new HashMap<KeyExtent,List<Range>>();
     Map<KeyExtent,List<Range>> failures = new HashMap<KeyExtent,List<Range>>();
     try {
@@ -186,10 +186,10 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
       log.trace("lookupTablets failed server=" + tserver, e);
       throw e;
     }
-    
+
     return MetadataLocationObtainer.getMetadataLocationEntries(results).getLocations();
   }
-  
+
   public static TabletLocations getMetadataLocationEntries(SortedMap<Key,Value> entries) {
     Key key;
     Value val;
@@ -197,30 +197,30 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
     Text session = null;
     Value prevRow = null;
     KeyExtent ke;
-    
+
     List<TabletLocation> results = new ArrayList<TabletLocation>();
     ArrayList<KeyExtent> locationless = new ArrayList<KeyExtent>();
-    
+
     Text lastRowFromKey = new Text();
-    
+
     // text obj below is meant to be reused in loop for efficiency
     Text colf = new Text();
     Text colq = new Text();
-    
+
     for (Entry<Key,Value> entry : entries.entrySet()) {
       key = entry.getKey();
       val = entry.getValue();
-      
+
       if (key.compareRow(lastRowFromKey) != 0) {
         prevRow = null;
         location = null;
         session = null;
         key.getRow(lastRowFromKey);
       }
-      
+
       colf = key.getColumnFamily(colf);
       colq = key.getColumnQualifier(colq);
-      
+
       // interpret the row id as a key extent
       if (colf.equals(TabletsSection.CurrentLocationColumnFamily.NAME) || colf.equals(TabletsSection.FutureLocationColumnFamily.NAME)) {
         if (location != null) {
@@ -231,19 +231,19 @@ public class MetadataLocationObtainer implements TabletLocationObtainer {
       } else if (TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.equals(colf, colq)) {
         prevRow = new Value(val);
       }
-      
+
       if (prevRow != null) {
         ke = new KeyExtent(key.getRow(), prevRow);
         if (location != null)
           results.add(new TabletLocation(ke, location.toString(), session.toString()));
         else
           locationless.add(ke);
-        
+
         location = null;
         prevRow = null;
       }
     }
-    
+
     return new TabletLocations(results, locationless);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/MetadataServicer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/MetadataServicer.java b/core/src/main/java/org/apache/accumulo/core/metadata/MetadataServicer.java
index 7d9592b..3b443f4 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/MetadataServicer.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/MetadataServicer.java
@@ -30,12 +30,12 @@ import org.apache.accumulo.core.data.KeyExtent;
  * Provides a consolidated API for handling table metadata
  */
 public abstract class MetadataServicer {
-  
+
   public static MetadataServicer forTableName(ClientContext context, String tableName) throws AccumuloException, AccumuloSecurityException {
     checkArgument(tableName != null, "tableName is null");
     return forTableId(context, context.getConnector().tableOperations().tableIdMap().get(tableName));
   }
-  
+
   public static MetadataServicer forTableId(ClientContext context, String tableId) {
     checkArgument(tableId != null, "tableId is null");
     if (RootTable.ID.equals(tableId))
@@ -45,19 +45,19 @@ public abstract class MetadataServicer {
     else
       return new ServicerForUserTables(context, tableId);
   }
-  
+
   /**
-   * 
+   *
    * @return the table id of the table currently being serviced
    */
   public abstract String getServicedTableId();
-  
+
   /**
    * Populate the provided data structure with the known tablets for the table being serviced
-   * 
+   *
    * @param tablets
    *          A mapping of all known tablets to their location (if available, null otherwise)
    */
   public abstract void getTabletLocations(SortedMap<KeyExtent,String> tablets) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/RootTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/RootTable.java b/core/src/main/java/org/apache/accumulo/core/metadata/RootTable.java
index 85219eb..24148b1 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/RootTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/RootTable.java
@@ -21,7 +21,7 @@ import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class RootTable {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForMetadataTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForMetadataTable.java b/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForMetadataTable.java
index 29f7027..525e2a2 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForMetadataTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForMetadataTable.java
@@ -23,9 +23,9 @@ import org.apache.accumulo.core.client.impl.ClientContext;
  * The metadata table's metadata is serviced in the root table.
  */
 class ServicerForMetadataTable extends TableMetadataServicer {
-  
+
   public ServicerForMetadataTable(ClientContext context) {
     super(context, RootTable.NAME, MetadataTable.ID);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForRootTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForRootTable.java b/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForRootTable.java
index b13149e..d02242c 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForRootTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForRootTable.java
@@ -30,18 +30,18 @@ import org.apache.accumulo.core.data.KeyExtent;
  * The root table's metadata is serviced in zookeeper.
  */
 class ServicerForRootTable extends MetadataServicer {
-  
+
   private final Instance instance;
-  
+
   public ServicerForRootTable(ClientContext context) {
     this.instance = context.getInstance();
   }
-  
+
   @Override
   public String getServicedTableId() {
     return RootTable.ID;
   }
-  
+
   @Override
   public void getTabletLocations(SortedMap<KeyExtent,String> tablets) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     tablets.put(RootTable.EXTENT, instance.getRootTabletLocation());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForUserTables.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForUserTables.java b/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForUserTables.java
index c9e2ede..5efa8a6 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForUserTables.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/ServicerForUserTables.java
@@ -23,9 +23,9 @@ import org.apache.accumulo.core.client.impl.ClientContext;
  * Metadata for user tables are serviced in the metadata table.
  */
 class ServicerForUserTables extends TableMetadataServicer {
-  
+
   public ServicerForUserTables(ClientContext context, String tableId) {
     super(context, MetadataTable.NAME, tableId);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/TableMetadataServicer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/TableMetadataServicer.java b/core/src/main/java/org/apache/accumulo/core/metadata/TableMetadataServicer.java
index fbba279..7e2ae0a 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/TableMetadataServicer.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/TableMetadataServicer.java
@@ -37,40 +37,40 @@ import org.apache.hadoop.io.Text;
  * A {@link MetadataServicer} that is backed by a table
  */
 abstract class TableMetadataServicer extends MetadataServicer {
-  
+
   private final ClientContext context;
   private String tableIdBeingServiced;
   private String serviceTableName;
-  
+
   public TableMetadataServicer(ClientContext context, String serviceTableName, String tableIdBeingServiced) {
     this.context = context;
     this.serviceTableName = serviceTableName;
     this.tableIdBeingServiced = tableIdBeingServiced;
   }
-  
+
   @Override
   public String getServicedTableId() {
     return tableIdBeingServiced;
   }
-  
+
   public String getServicingTableName() {
     return serviceTableName;
   }
-  
+
   @Override
   public void getTabletLocations(SortedMap<KeyExtent,String> tablets) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
-    
+
     Scanner scanner = context.getConnector().createScanner(getServicingTableName(), Authorizations.EMPTY);
-    
+
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.fetch(scanner);
     scanner.fetchColumnFamily(TabletsSection.CurrentLocationColumnFamily.NAME);
-    
+
     // position at first entry in metadata table for given table
     scanner.setRange(TabletsSection.getRange(getServicedTableId()));
-    
+
     Text colf = new Text();
     Text colq = new Text();
-    
+
     KeyExtent currentKeyExtent = null;
     String location = null;
     Text row = null;
@@ -85,10 +85,10 @@ abstract class TableMetadataServicer extends MetadataServicer {
       } else {
         row = entry.getKey().getRow();
       }
-      
+
       colf = entry.getKey().getColumnFamily(colf);
       colq = entry.getKey().getColumnQualifier(colq);
-      
+
       if (TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.equals(colf, colq)) {
         currentKeyExtent = new KeyExtent(entry.getKey().getRow(), entry.getValue());
         tablets.put(currentKeyExtent, location);
@@ -97,42 +97,42 @@ abstract class TableMetadataServicer extends MetadataServicer {
       } else if (colf.equals(TabletsSection.CurrentLocationColumnFamily.NAME)) {
         location = entry.getValue().toString();
       }
-      
+
     }
-    
+
     validateEntries(tablets);
   }
-  
+
   private void validateEntries(SortedMap<KeyExtent,String> tablets) throws AccumuloException {
     SortedSet<KeyExtent> tabletsKeys = (SortedSet<KeyExtent>) tablets.keySet();
     // sanity check of metadata table entries
     // make sure tablets has no holes, and that it starts and ends w/ null
     if (tabletsKeys.size() == 0)
       throw new AccumuloException("No entries found in metadata table for table " + getServicedTableId());
-    
+
     if (tabletsKeys.first().getPrevEndRow() != null)
       throw new AccumuloException("Problem with metadata table, first entry for table " + getServicedTableId() + "- " + tabletsKeys.first()
           + " - has non null prev end row");
-    
+
     if (tabletsKeys.last().getEndRow() != null)
       throw new AccumuloException("Problem with metadata table, last entry for table " + getServicedTableId() + "- " + tabletsKeys.first()
           + " - has non null end row");
-    
+
     Iterator<KeyExtent> tabIter = tabletsKeys.iterator();
     Text lastEndRow = tabIter.next().getEndRow();
     while (tabIter.hasNext()) {
       KeyExtent tabke = tabIter.next();
-      
+
       if (tabke.getPrevEndRow() == null)
         throw new AccumuloException("Problem with metadata table, it has null prev end row in middle of table " + tabke);
-      
+
       if (!tabke.getPrevEndRow().equals(lastEndRow))
         throw new AccumuloException("Problem with metadata table, it has a hole " + tabke.getPrevEndRow() + " != " + lastEndRow);
-      
+
       lastEndRow = tabke.getEndRow();
     }
-    
+
     // end METADATA table sanity check
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/schema/DataFileValue.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/schema/DataFileValue.java b/core/src/main/java/org/apache/accumulo/core/metadata/schema/DataFileValue.java
index d3323a4..cebe041 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/schema/DataFileValue.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/schema/DataFileValue.java
@@ -20,74 +20,74 @@ public class DataFileValue {
   private long size;
   private long numEntries;
   private long time = -1;
-  
+
   public DataFileValue(long size, long numEntries, long time) {
     this.size = size;
     this.numEntries = numEntries;
     this.time = time;
   }
-  
+
   public DataFileValue(long size, long numEntries) {
     this.size = size;
     this.numEntries = numEntries;
     this.time = -1;
   }
-  
+
   public DataFileValue(byte[] encodedDFV) {
     String[] ba = new String(encodedDFV).split(",");
-    
+
     size = Long.parseLong(ba[0]);
     numEntries = Long.parseLong(ba[1]);
-    
+
     if (ba.length == 3)
       time = Long.parseLong(ba[2]);
     else
       time = -1;
   }
-  
+
   public long getSize() {
     return size;
   }
-  
+
   public long getNumEntries() {
     return numEntries;
   }
-  
+
   public boolean isTimeSet() {
     return time >= 0;
   }
-  
+
   public long getTime() {
     return time;
   }
-  
+
   public byte[] encode() {
     if (time >= 0)
       return ("" + size + "," + numEntries + "," + time).getBytes();
     return ("" + size + "," + numEntries).getBytes();
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof DataFileValue) {
       DataFileValue odfv = (DataFileValue) o;
-      
+
       return size == odfv.size && numEntries == odfv.numEntries;
     }
-    
+
     return false;
   }
-  
+
   @Override
   public int hashCode() {
     return Long.valueOf(size + numEntries).hashCode();
   }
-  
+
   @Override
   public String toString() {
     return size + " " + numEntries;
   }
-  
+
   public void setTime(long time) {
     if (time < 0)
       throw new IllegalArgumentException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/metadata/schema/MetadataSchema.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/metadata/schema/MetadataSchema.java b/core/src/main/java/org/apache/accumulo/core/metadata/schema/MetadataSchema.java
index 0d8a0dc..534dd7f 100644
--- a/core/src/main/java/org/apache/accumulo/core/metadata/schema/MetadataSchema.java
+++ b/core/src/main/java/org/apache/accumulo/core/metadata/schema/MetadataSchema.java
@@ -31,26 +31,26 @@ import com.google.common.base.Preconditions;
  * Describes the table schema used for metadata tables
  */
 public class MetadataSchema {
-  
+
   public static final String RESERVED_PREFIX = "~";
-  
+
   /**
    * Used for storing information about tablets
    */
   public static class TabletsSection {
     private static final Section section = new Section(null, false, RESERVED_PREFIX, false);
-    
+
     public static Range getRange() {
       return section.getRange();
     }
-    
+
     public static Range getRange(String tableId) {
       return new Range(new Key(tableId + ';'), true, new Key(tableId + '<').followingKey(PartialKey.ROW), false);
     }
-    
+
     public static Text getRow(Text tableId, Text endRow) {
       Text entry = new Text(tableId);
-      
+
       if (endRow == null) {
         // append delimiter for default tablet
         entry.append(new byte[] {'<'}, 0, 1);
@@ -59,10 +59,10 @@ public class MetadataSchema {
         entry.append(new byte[] {';'}, 0, 1);
         entry.append(endRow.getBytes(), 0, endRow.getLength());
       }
-      
+
       return entry;
     }
-    
+
     /**
      * Column family for storing the tablet information needed by clients
      */
@@ -86,7 +86,7 @@ public class MetadataSchema {
        */
       public static final ColumnFQ SPLIT_RATIO_COLUMN = new ColumnFQ(NAME, new Text("splitRatio"));
     }
-    
+
     /**
      * Column family for recording information used by the TServer
      */
@@ -113,63 +113,63 @@ public class MetadataSchema {
        */
       public static final ColumnFQ LOCK_COLUMN = new ColumnFQ(NAME, new Text("lock"));
     }
-    
+
     /**
      * Column family for storing entries created by the TServer to indicate it has loaded a tablet that it was assigned
      */
     public static class CurrentLocationColumnFamily {
       public static final Text NAME = new Text("loc");
     }
-    
+
     /**
      * Column family for storing the assigned location
      */
     public static class FutureLocationColumnFamily {
       public static final Text NAME = new Text("future");
     }
-    
+
     /**
      * Column family for storing last location, as a hint for assignment
      */
     public static class LastLocationColumnFamily {
       public static final Text NAME = new Text("last");
     }
-    
+
     /**
      * Temporary markers that indicate a tablet loaded a bulk file
      */
     public static class BulkFileColumnFamily {
       public static final Text NAME = new Text("loaded");
     }
-    
+
     /**
      * Temporary marker that indicates a tablet was successfully cloned
      */
     public static class ClonedColumnFamily {
       public static final Text NAME = new Text("!cloned");
     }
-    
+
     /**
      * Column family for storing files used by a tablet
      */
     public static class DataFileColumnFamily {
       public static final Text NAME = new Text("file");
     }
-    
+
     /**
      * Column family for storing the set of files scanned with an isolated scanner, to prevent them from being deleted
      */
     public static class ScanFileColumnFamily {
       public static final Text NAME = new Text("scan");
     }
-    
+
     /**
      * Column family for storing write-ahead log entries
      */
     public static class LogColumnFamily {
       public static final Text NAME = new Text("log");
     }
-    
+
     /**
      * Column family for indicating that the files in a tablet have been trimmed to only include data for the current tablet, so that they are safe to merge
      */
@@ -178,53 +178,53 @@ public class MetadataSchema {
       public static final ColumnFQ CHOPPED_COLUMN = new ColumnFQ(NAME, new Text("chopped"));
     }
   }
-  
+
   /**
    * Contains additional metadata in a reserved area not for tablets
    */
   public static class ReservedSection {
     private static final Section section = new Section(RESERVED_PREFIX, true, null, false);
-    
+
     public static Range getRange() {
       return section.getRange();
     }
-    
+
     public static String getRowPrefix() {
       return section.getRowPrefix();
     }
-    
+
   }
-  
+
   /**
    * Holds delete markers for potentially unused files/directories
    */
   public static class DeletesSection {
     private static final Section section = new Section(RESERVED_PREFIX + "del", true, RESERVED_PREFIX + "dem", false);
-    
+
     public static Range getRange() {
       return section.getRange();
     }
-    
+
     public static String getRowPrefix() {
       return section.getRowPrefix();
     }
-    
+
   }
-  
+
   /**
    * Holds bulk-load-in-progress processing flags
    */
   public static class BlipSection {
     private static final Section section = new Section(RESERVED_PREFIX + "blip", true, RESERVED_PREFIX + "bliq", false);
-    
+
     public static Range getRange() {
       return section.getRange();
     }
-    
+
     public static String getRowPrefix() {
       return section.getRowPrefix();
     }
-    
+
   }
 
   /**
@@ -247,7 +247,7 @@ public class MetadataSchema {
 
     /**
      * Extract the table ID from the colfam (inefficiently if called repeatedly)
-     * 
+     *
      * @param k
      *          Key to extract from
      * @return The table ID
@@ -261,7 +261,7 @@ public class MetadataSchema {
 
     /**
      * Extract the table ID from the colfam into the given {@link Text}
-     * 
+     *
      * @param k
      *          Key to extract from
      * @param buff
@@ -276,7 +276,7 @@ public class MetadataSchema {
 
     /**
      * Extract the file name from the row suffix into the given {@link Text}
-     * 
+     *
      * @param k
      *          Key to extract from
      * @param buff

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/protobuf/ProtobufUtil.java b/core/src/main/java/org/apache/accumulo/core/protobuf/ProtobufUtil.java
index 60eb840..e2489f9 100644
--- a/core/src/main/java/org/apache/accumulo/core/protobuf/ProtobufUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/protobuf/ProtobufUtil.java
@@ -25,7 +25,7 @@ import com.google.protobuf.TextFormat;
  * Helper methods for interacting with Protocol Buffers and Accumulo
  */
 public class ProtobufUtil {
-  private static final char LEFT_BRACKET = '[', RIGHT_BRACKET = ']'; 
+  private static final char LEFT_BRACKET = '[', RIGHT_BRACKET = ']';
 
   public static Value toValue(GeneratedMessage msg) {
     return new Value(msg.toByteArray());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/AccumuloReplicationReplayer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/AccumuloReplicationReplayer.java b/core/src/main/java/org/apache/accumulo/core/replication/AccumuloReplicationReplayer.java
index ff89196..fccafc5 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/AccumuloReplicationReplayer.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/AccumuloReplicationReplayer.java
@@ -24,7 +24,7 @@ import org.apache.accumulo.core.replication.thrift.RemoteReplicationException;
 import org.apache.accumulo.core.replication.thrift.WalEdits;
 
 /**
- * 
+ *
  */
 public interface AccumuloReplicationReplayer {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtil.java b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtil.java
index ae94c2a..0817d3b 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtil.java
@@ -27,8 +27,11 @@ public class ReplicationConfigurationUtil {
 
   /**
    * Determines if the replication is enabled for the given {@link KeyExtent}
-   * @param extent The {@link KeyExtent} for the Tablet in question
-   * @param conf The {@link AccumuloConfiguration} for that Tablet (table or namespace)
+   *
+   * @param extent
+   *          The {@link KeyExtent} for the Tablet in question
+   * @param conf
+   *          The {@link AccumuloConfiguration} for that Tablet (table or namespace)
    * @return True if this extent is a candidate for replication at the given point in time.
    */
   public static boolean isEnabled(KeyExtent extent, AccumuloConfiguration conf) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConstants.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConstants.java b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConstants.java
index 3d71681..3f96eca 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConstants.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationConstants.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.core.replication;
 
 /**
- * 
+ *
  */
 public class ReplicationConstants {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/ReplicationSchema.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationSchema.java b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationSchema.java
index 491fda5..ed46130 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationSchema.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationSchema.java
@@ -19,6 +19,7 @@ package org.apache.accumulo.core.replication;
 import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.nio.charset.CharacterCodingException;
+
 import org.apache.accumulo.core.client.ScannerBase;
 import org.apache.accumulo.core.client.lexicoder.ULongLexicoder;
 import org.apache.accumulo.core.data.ArrayByteSequence;
@@ -34,7 +35,7 @@ import org.slf4j.LoggerFactory;
 import com.google.common.base.Preconditions;
 
 /**
- * 
+ *
  */
 public class ReplicationSchema {
   private static final Logger log = LoggerFactory.getLogger(ReplicationSchema.class);
@@ -90,7 +91,7 @@ public class ReplicationSchema {
 
     /**
      * Extract the table ID from the key (inefficiently if called repeatedly)
-     * 
+     *
      * @param k
      *          Key to extract from
      * @return The table ID
@@ -104,7 +105,7 @@ public class ReplicationSchema {
 
     /**
      * Extract the table ID from the key into the given {@link Text}
-     * 
+     *
      * @param k
      *          Key to extract from
      * @param buff
@@ -119,7 +120,7 @@ public class ReplicationSchema {
 
     /**
      * Extract the file name from the row suffix into the given {@link Text}
-     * 
+     *
      * @param k
      *          Key to extract from
      * @param buff
@@ -154,12 +155,12 @@ public class ReplicationSchema {
    */
   public static class OrderSection {
     public static final Text NAME = new Text("order");
-    public static final Text ROW_SEPARATOR = new Text(new byte[]{0});
+    public static final Text ROW_SEPARATOR = new Text(new byte[] {0});
     private static final ULongLexicoder longEncoder = new ULongLexicoder();
 
     /**
      * Extract the table ID from the given key (inefficiently if called repeatedly)
-     * 
+     *
      * @param k
      *          OrderSection Key
      * @return source table id
@@ -172,7 +173,7 @@ public class ReplicationSchema {
 
     /**
      * Extract the table ID from the given key
-     * 
+     *
      * @param k
      *          OrderSection key
      * @param buff
@@ -194,7 +195,7 @@ public class ReplicationSchema {
 
     /**
      * Creates the Mutation for the Order section for the given file and time
-     * 
+     *
      * @param file
      *          Filename
      * @param timeInMillis
@@ -224,7 +225,7 @@ public class ReplicationSchema {
 
     /**
      * Add a column update to the given mutation with the provided tableId and value
-     * 
+     *
      * @param m
      *          Mutation for OrderSection
      * @param tableId

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/replication/ReplicationTarget.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationTarget.java b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationTarget.java
index 9c57ad9..eb7833e 100644
--- a/core/src/main/java/org/apache/accumulo/core/replication/ReplicationTarget.java
+++ b/core/src/main/java/org/apache/accumulo/core/replication/ReplicationTarget.java
@@ -21,6 +21,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.IOException;
+
 import org.apache.hadoop.io.DataInputBuffer;
 import org.apache.hadoop.io.DataOutputBuffer;
 import org.apache.hadoop.io.Text;
@@ -36,7 +37,7 @@ public class ReplicationTarget implements Writable {
   private String remoteIdentifier;
   private String sourceTableId;
 
-  public ReplicationTarget() { }
+  public ReplicationTarget() {}
 
   public ReplicationTarget(String peerName, String remoteIdentifier, String sourceTableId) {
     this.peerName = peerName;
@@ -130,7 +131,9 @@ public class ReplicationTarget implements Writable {
 
   /**
    * Deserialize a ReplicationTarget
-   * @param t Serialized copy
+   *
+   * @param t
+   *          Serialized copy
    * @return the deserialized version
    */
   public static ReplicationTarget from(Text t) {
@@ -149,7 +152,9 @@ public class ReplicationTarget implements Writable {
 
   /**
    * Deserialize a ReplicationTarget
-   * @param s Serialized copy
+   *
+   * @param s
+   *          Serialized copy
    * @return the deserialized version
    */
   public static ReplicationTarget from(String s) {
@@ -167,8 +172,9 @@ public class ReplicationTarget implements Writable {
   }
 
   /**
-   * Convenience method to serialize a ReplicationTarget to {@link Text} using the {@link Writable} methods without caring about
-   * performance penalties due to excessive object creation
+   * Convenience method to serialize a ReplicationTarget to {@link Text} using the {@link Writable} methods without caring about performance penalties due to
+   * excessive object creation
+   *
    * @return The serialized representation of the object
    */
   public Text toText() {


[51/66] [abbrv] accumulo git commit: Merge branch '1.5' into 1.6 with -sours

Posted by ct...@apache.org.
Merge branch '1.5' into 1.6 with -sours


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/dff686b4
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/dff686b4
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/dff686b4

Branch: refs/heads/master
Commit: dff686b454e67e74150d967bdc3eb94dcb5e39a6
Parents: 100ffd1 c2155f4
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:21:16 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:21:16 2015 -0500

----------------------------------------------------------------------

----------------------------------------------------------------------



[30/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/IteratorSettingTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/IteratorSettingTest.java b/core/src/test/java/org/apache/accumulo/core/client/IteratorSettingTest.java
index 72342f9..eed3dea 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/IteratorSettingTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/IteratorSettingTest.java
@@ -31,7 +31,7 @@ import com.google.common.collect.Maps;
  * Test cases for the IteratorSetting class
  */
 public class IteratorSettingTest {
-  
+
   IteratorSetting setting1 = new IteratorSetting(500, "combiner", Combiner.class.getName());
   IteratorSetting setting2 = new IteratorSetting(500, "combiner", Combiner.class.getName());
   IteratorSetting setting3 = new IteratorSetting(500, "combiner", Combiner.class.getName());
@@ -40,45 +40,45 @@ public class IteratorSettingTest {
   IteratorSetting setting4 = new IteratorSetting(300, "combiner", Combiner.class.getName());
   IteratorSetting setting5 = new IteratorSetting(500, "foocombiner", Combiner.class.getName());
   IteratorSetting setting6 = new IteratorSetting(500, "combiner", "MySuperCombiner");
-  
+
   @Test
   public final void testHashCodeSameObject() {
     assertEquals(setting1.hashCode(), setting1.hashCode());
   }
-  
+
   @Test
   public final void testHashCodeEqualObjects() {
     assertEquals(setting1.hashCode(), setting2.hashCode());
   }
-  
+
   @Test
   public final void testEqualsObjectReflexive() {
     assertEquals(setting1, setting1);
   }
-  
+
   @Test
   public final void testEqualsObjectSymmetric() {
     assertEquals(setting1, setting2);
     assertEquals(setting2, setting1);
   }
-  
+
   @Test
   public final void testEqualsObjectTransitive() {
     assertEquals(setting1, setting2);
     assertEquals(setting2, setting3);
     assertEquals(setting1, setting3);
   }
-  
+
   @Test
   public final void testEqualsNullSetting() {
     assertNotEquals(setting1, nullsetting);
   }
-  
+
   @Test
   public final void testEqualsObjectNotEqual() {
     assertNotEquals(setting1, devnull);
   }
-  
+
   @Test
   public final void testEqualsObjectProperties() {
     IteratorSetting mysettings = new IteratorSetting(500, "combiner", Combiner.class.getName());
@@ -86,29 +86,29 @@ public class IteratorSettingTest {
     mysettings.addOption("myoption1", "myvalue1");
     assertNotEquals(setting1, mysettings);
   }
-  
+
   @Test
   public final void testEqualsDifferentMembers() {
     assertNotEquals(setting1, setting4);
     assertNotEquals(setting1, setting5);
     assertNotEquals(setting1, setting6);
   }
-  
+
   @Test
   public void testEquivalentConstructor() {
     IteratorSetting setting1 = new IteratorSetting(100, Combiner.class);
     IteratorSetting setting2 = new IteratorSetting(100, "Combiner", Combiner.class, Maps.<String,String> newHashMap());
-    
+
     assertEquals(setting1, setting2);
-    
+
     IteratorSetting notEqual1 = new IteratorSetting(100, "FooCombiner", Combiner.class, Maps.<String,String> newHashMap());
-    
+
     assertNotEquals(setting1, notEqual1);
-    
+
     Map<String,String> props = Maps.newHashMap();
     props.put("foo", "bar");
     IteratorSetting notEquals2 = new IteratorSetting(100, "Combiner", Combiner.class, props);
-    
+
     assertNotEquals(setting1, notEquals2);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/RowIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/RowIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/client/RowIteratorTest.java
index e12e22f..55e21d1 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/RowIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/RowIteratorTest.java
@@ -34,7 +34,7 @@ import org.apache.accumulo.core.data.Value;
 import org.junit.Test;
 
 public class RowIteratorTest {
-  
+
   Iterator<Entry<Key,Value>> makeIterator(final String... args) {
     final Map<Key,Value> result = new TreeMap<Key,Value>();
     for (String s : args) {
@@ -45,7 +45,7 @@ public class RowIteratorTest {
     }
     return result.entrySet().iterator();
   }
-  
+
   List<List<Entry<Key,Value>>> getRows(final Iterator<Entry<Key,Value>> iter) {
     final List<List<Entry<Key,Value>>> result = new ArrayList<List<Entry<Key,Value>>>();
     final RowIterator riter = new RowIterator(iter);
@@ -58,7 +58,7 @@ public class RowIteratorTest {
     }
     return result;
   }
-  
+
   @Test
   public void testRowIterator() {
     List<List<Entry<Key,Value>>> rows = getRows(makeIterator());
@@ -70,13 +70,13 @@ public class RowIteratorTest {
     assertEquals(2, rows.size());
     assertEquals(3, rows.get(0).size());
     assertEquals(1, rows.get(1).size());
-    
+
     RowIterator i = new RowIterator(makeIterator());
     try {
       i.next();
       fail();
     } catch (NoSuchElementException ex) {}
-    
+
     i = new RowIterator(makeIterator("a b c d", "a 1 2 3"));
     assertTrue(i.hasNext());
     Iterator<Entry<Key,Value>> row = i.next();
@@ -97,7 +97,7 @@ public class RowIteratorTest {
       fail();
     } catch (NoSuchElementException ex) {}
   }
-  
+
   @Test
   public void testUnreadRow() {
     final RowIterator i = new RowIterator(makeIterator("a b c d", "a 1 2 3", "b 1 2 3"));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/TestThrift1474.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/TestThrift1474.java b/core/src/test/java/org/apache/accumulo/core/client/TestThrift1474.java
index 4c732c0..24c2bbc 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/TestThrift1474.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/TestThrift1474.java
@@ -37,7 +37,7 @@ import org.apache.thrift.transport.TTransport;
 import org.junit.Test;
 
 public class TestThrift1474 {
-  
+
   static class TestServer implements ThriftTest.Iface {
 
     @Override
@@ -54,9 +54,9 @@ public class TestThrift1474 {
     public boolean throwsError() throws ThriftSecurityException, TException {
       throw new ThriftSecurityException();
     }
-    
+
   }
-  
+
   @Test
   public void test() throws IOException, TException, InterruptedException {
     TServerSocket serverTransport = new TServerSocket(0);
@@ -64,7 +64,7 @@ public class TestThrift1474 {
     int port = serverTransport.getServerSocket().getLocalPort();
     TestServer handler = new TestServer();
     ThriftTest.Processor<ThriftTest.Iface> processor = new ThriftTest.Processor<ThriftTest.Iface>(handler);
-    
+
     TThreadPoolServer.Args args = new TThreadPoolServer.Args(serverTransport);
     args.stopTimeoutVal = 10;
     args.stopTimeoutUnit = TimeUnit.MILLISECONDS;
@@ -78,7 +78,7 @@ public class TestThrift1474 {
     while (!server.isServing()) {
       UtilWaitThread.sleep(10);
     }
-    
+
     TTransport transport = new TSocket("localhost", port);
     transport.open();
     TProtocol protocol = new TBinaryProtocol(transport);
@@ -94,5 +94,5 @@ public class TestThrift1474 {
     server.stop();
     thread.join();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/admin/FindMaxTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/admin/FindMaxTest.java b/core/src/test/java/org/apache/accumulo/core/client/admin/FindMaxTest.java
index d9fca47..2dc6ba5 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/admin/FindMaxTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/admin/FindMaxTest.java
@@ -34,27 +34,27 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.hadoop.io.Text;
 
 public class FindMaxTest extends TestCase {
-  
+
   private static Mutation nm(byte[] row) {
     Mutation m = new Mutation(new Text(row));
     m.put("cf", "cq", "v");
     return m;
   }
-  
+
   private static Mutation nm(String row) {
     Mutation m = new Mutation(row);
     m.put("cf", "cq", "v");
     return m;
   }
-  
+
   public void test1() throws Exception {
     MockInstance mi = new MockInstance();
-    
+
     Connector conn = mi.getConnector("root", new PasswordToken(""));
     conn.tableOperations().create("foo");
-    
+
     BatchWriter bw = conn.createBatchWriter("foo", new BatchWriterConfig());
-    
+
     bw.addMutation(nm(new byte[] {0}));
     bw.addMutation(nm(new byte[] {0, 0}));
     bw.addMutation(nm(new byte[] {0, 1}));
@@ -63,48 +63,48 @@ public class FindMaxTest extends TestCase {
     bw.addMutation(nm(new byte[] {'a', 'b', 'c'}));
     bw.addMutation(nm(new byte[] {(byte) 0xff}));
     bw.addMutation(nm(new byte[] {(byte) 0xff, (byte) 0xff, (byte) 0xff, (byte) 0xff, (byte) 0xff, (byte) 0xff}));
-    
+
     for (int i = 0; i < 1000; i++) {
       bw.addMutation(nm(String.format("r%05d", i)));
     }
-    
+
     bw.close();
-    
+
     Scanner scanner = conn.createScanner("foo", Authorizations.EMPTY);
-    
+
     ArrayList<Text> rows = new ArrayList<Text>();
-    
+
     for (Entry<Key,Value> entry : scanner) {
       rows.add(entry.getKey().getRow());
     }
-    
+
     for (int i = rows.size() - 1; i > 0; i--) {
       Text max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), null, true, rows.get(i), false);
       assertEquals(rows.get(i - 1), max);
-      
+
       max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), rows.get(i - 1), true, rows.get(i), false);
       assertEquals(rows.get(i - 1), max);
-      
+
       max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), rows.get(i - 1), false, rows.get(i), false);
       assertNull(max);
-      
+
       max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), null, true, rows.get(i), true);
       assertEquals(rows.get(i), max);
-      
+
       max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), rows.get(i), true, rows.get(i), true);
       assertEquals(rows.get(i), max);
-      
+
       max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), rows.get(i - 1), false, rows.get(i), true);
       assertEquals(rows.get(i), max);
-      
+
     }
-    
+
     Text max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), null, true, null, true);
     assertEquals(rows.get(rows.size() - 1), max);
-    
+
     max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), null, true, new Text(new byte[] {0}), false);
     assertNull(max);
-    
+
     max = FindMax.findMax(conn.createScanner("foo", Authorizations.EMPTY), null, true, new Text(new byte[] {0}), true);
     assertEquals(rows.get(0), max);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/impl/ClientContextTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/impl/ClientContextTest.java b/core/src/test/java/org/apache/accumulo/core/client/impl/ClientContextTest.java
index 2e3951d..494eb50 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/impl/ClientContextTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/impl/ClientContextTest.java
@@ -35,7 +35,7 @@ public class ClientContextTest {
   private static boolean isCredentialProviderAvailable = false;
   private static final String keystoreName = "/site-cfg.jceks";
 
-  //site-cfg.jceks={'ignored.property'=>'ignored', 'instance.secret'=>'mysecret', 'general.rpc.timeout'=>'timeout'}
+  // site-cfg.jceks={'ignored.property'=>'ignored', 'instance.secret'=>'mysecret', 'general.rpc.timeout'=>'timeout'}
   private static File keystore;
 
   @BeforeClass

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerImplTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerImplTest.java b/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerImplTest.java
index be4d467..dec6748 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerImplTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerImplTest.java
@@ -26,7 +26,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ScannerImplTest {
 
@@ -38,10 +38,10 @@ public class ScannerImplTest {
     s.setReadaheadThreshold(0);
     s.setReadaheadThreshold(10);
     s.setReadaheadThreshold(Long.MAX_VALUE);
-    
+
     Assert.assertEquals(Long.MAX_VALUE, s.getReadaheadThreshold());
   }
-  
+
   @Test(expected = IllegalArgumentException.class)
   public void testInValidReadaheadValues() {
     MockInstance instance = new MockInstance();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerOptionsTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerOptionsTest.java b/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerOptionsTest.java
index 463822a..123509a 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerOptionsTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/impl/ScannerOptionsTest.java
@@ -28,7 +28,7 @@ import org.junit.Test;
  * Test that scanner options are set/unset correctly
  */
 public class ScannerOptionsTest {
-  
+
   /**
    * Test that you properly add and remove iterators from a scanner
    */
@@ -40,7 +40,7 @@ public class ScannerOptionsTest {
     options.removeScanIterator("NAME");
     assertEquals(0, options.serverSideIteratorList.size());
   }
-  
+
   @Test
   public void testIteratorConflict() {
     ScannerOptions options = new ScannerOptions();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/impl/TableOperationsHelperTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/impl/TableOperationsHelperTest.java b/core/src/test/java/org/apache/accumulo/core/client/impl/TableOperationsHelperTest.java
index 2885002..dab7226 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/impl/TableOperationsHelperTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/impl/TableOperationsHelperTest.java
@@ -28,14 +28,13 @@ import java.util.Set;
 import java.util.SortedSet;
 import java.util.TreeMap;
 
-import org.apache.accumulo.core.client.admin.CompactionConfig;
-
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.NewTableConfiguration;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
+import org.apache.accumulo.core.client.admin.CompactionConfig;
 import org.apache.accumulo.core.client.admin.DiskUsage;
 import org.apache.accumulo.core.client.admin.TimeType;
 import org.apache.accumulo.core.data.Range;
@@ -234,16 +233,16 @@ public class TableOperationsHelperTest {
   }
 
   void check(TableOperationsHelper t, String tablename, String[] values) throws Exception {
-      Map<String,String> expected = new TreeMap<String,String>();
-      for (String value : values) {
-        String parts[] = value.split("=", 2);
-        expected.put(parts[0], parts[1]);
-      }
-      Map<String,String> actual = new TreeMap<String,String>();
-      for (Entry<String,String> entry : t.getProperties(tablename)) {
-        actual.put(entry.getKey(), entry.getValue());
-      }
-      Assert.assertEquals(expected, actual);
+    Map<String,String> expected = new TreeMap<String,String>();
+    for (String value : values) {
+      String parts[] = value.split("=", 2);
+      expected.put(parts[0], parts[1]);
+    }
+    Map<String,String> actual = new TreeMap<String,String>();
+    for (Entry<String,String> entry : t.getProperties(tablename)) {
+      actual.put(entry.getKey(), entry.getValue());
+    }
+    Assert.assertEquals(expected, actual);
   }
 
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/impl/ZookeeperLockCheckerTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/impl/ZookeeperLockCheckerTest.java b/core/src/test/java/org/apache/accumulo/core/client/impl/ZookeeperLockCheckerTest.java
index 1749a4b..8e2e79c 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/impl/ZookeeperLockCheckerTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/impl/ZookeeperLockCheckerTest.java
@@ -16,6 +16,11 @@
  */
 package org.apache.accumulo.core.client.impl;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.zookeeper.ZooUtil;
@@ -23,10 +28,6 @@ import org.apache.accumulo.fate.zookeeper.ZooCache;
 import org.apache.accumulo.fate.zookeeper.ZooCacheFactory;
 import org.junit.Before;
 import org.junit.Test;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
 
 public class ZookeeperLockCheckerTest {
   private Instance instance;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoderTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoderTest.java b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoderTest.java
index 347649c..9d453e3 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoderTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoderTest.java
@@ -20,7 +20,7 @@ import java.math.BigInteger;
 import java.util.Arrays;
 
 /**
- * 
+ *
  */
 public class BigIntegerLexicoderTest extends LexicoderTest {
   public void testSortOrder() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoderTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoderTest.java b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoderTest.java
index b299896..fc37f04 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoderTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoderTest.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.core.client.lexicoder;
 import java.util.Arrays;
 
 /**
- * 
+ *
  */
 public class DoubleLexicoderTest extends LexicoderTest {
   public void testSortOrder() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/lexicoder/ListLexicoderTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/ListLexicoderTest.java b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/ListLexicoderTest.java
index be1001d..60a9f79 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/ListLexicoderTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/ListLexicoderTest.java
@@ -28,47 +28,47 @@ public class ListLexicoderTest extends LexicoderTest {
     List<Long> data1 = new ArrayList<Long>();
     data1.add(1l);
     data1.add(2l);
-    
+
     List<Long> data2 = new ArrayList<Long>();
     data2.add(1l);
-    
+
     List<Long> data3 = new ArrayList<Long>();
     data3.add(1l);
     data3.add(3l);
-    
+
     List<Long> data4 = new ArrayList<Long>();
     data4.add(1l);
     data4.add(2l);
     data4.add(3l);
-    
+
     List<Long> data5 = new ArrayList<Long>();
     data5.add(2l);
     data5.add(1l);
-    
+
     List<List<Long>> data = new ArrayList<List<Long>>();
-    
+
     // add list in expected sort order
     data.add(data2);
     data.add(data1);
     data.add(data4);
     data.add(data3);
     data.add(data5);
-    
+
     TreeSet<Text> sortedEnc = new TreeSet<Text>();
-    
+
     ListLexicoder<Long> listLexicoder = new ListLexicoder<Long>(new LongLexicoder());
-    
+
     for (List<Long> list : data) {
       sortedEnc.add(new Text(listLexicoder.encode(list)));
     }
-    
+
     List<List<Long>> unenc = new ArrayList<List<Long>>();
-    
+
     for (Text enc : sortedEnc) {
       unenc.add(listLexicoder.decode(TextUtil.getBytes(enc)));
     }
-    
+
     assertEquals(data, unenc);
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/lexicoder/PairLexicoderTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/PairLexicoderTest.java b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/PairLexicoderTest.java
index a653259..434a603 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/PairLexicoderTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/PairLexicoderTest.java
@@ -21,7 +21,7 @@ import java.util.Arrays;
 import org.apache.accumulo.core.util.ComparablePair;
 
 /**
- * 
+ *
  */
 public class PairLexicoderTest extends LexicoderTest {
   public void testSortOrder() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoderTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoderTest.java b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoderTest.java
index 58cec6d..bd59237 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoderTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoderTest.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.core.client.lexicoder;
 public class UIntegerLexicoderTest extends LexicoderTest {
   public void testEncoding() {
     UIntegerLexicoder uil = new UIntegerLexicoder();
-    
+
     assertEqualsB(uil.encode(0), new byte[] {0x00});
     assertEqualsB(uil.encode(0x01), new byte[] {0x01, 0x01});
     assertEqualsB(uil.encode(0x0102), new byte[] {0x02, 0x01, 0x02});

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormatTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormatTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormatTest.java
index 9991206..5ed26d4 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormatTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormatTest.java
@@ -130,7 +130,7 @@ public class AccumuloInputFormatTest {
   /**
    * Test adding iterator options where the keys and values contain both the FIELD_SEPARATOR character (':') and ITERATOR_SEPARATOR (',') characters. There
    * should be no exceptions thrown when trying to parse these types of option entries.
-   * 
+   *
    * This test makes sure that the expected raw values, as appears in the Job, are equal to what's expected.
    */
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormatTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormatTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormatTest.java
index 36054c8..be5b452 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormatTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormatTest.java
@@ -51,7 +51,7 @@ import org.apache.hadoop.util.ToolRunner;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class AccumuloOutputFormatTest {
   private static AssertionError e1 = null;
@@ -59,13 +59,13 @@ public class AccumuloOutputFormatTest {
   private static final String INSTANCE_NAME = PREFIX + "_mapred_instance";
   private static final String TEST_TABLE_1 = PREFIX + "_mapred_table_1";
   private static final String TEST_TABLE_2 = PREFIX + "_mapred_table_2";
-  
+
   private static class MRTester extends Configured implements Tool {
     private static class TestMapper implements Mapper<Key,Value,Text,Mutation> {
       Key key = null;
       int count = 0;
       OutputCollector<Text,Mutation> finalOutput;
-      
+
       @Override
       public void map(Key k, Value v, OutputCollector<Text,Mutation> output, Reporter reporter) throws IOException {
         finalOutput = output;
@@ -80,102 +80,102 @@ public class AccumuloOutputFormatTest {
         key = new Key(k);
         count++;
       }
-      
+
       @Override
       public void configure(JobConf job) {}
-      
+
       @Override
       public void close() throws IOException {
         Mutation m = new Mutation("total");
         m.put("", "", Integer.toString(count));
         finalOutput.collect(new Text(), m);
       }
-      
+
     }
-    
+
     @Override
     public int run(String[] args) throws Exception {
-      
+
       if (args.length != 4) {
         throw new IllegalArgumentException("Usage : " + MRTester.class.getName() + " <user> <pass> <inputtable> <outputtable>");
       }
-      
+
       String user = args[0];
       String pass = args[1];
       String table1 = args[2];
       String table2 = args[3];
-      
+
       JobConf job = new JobConf(getConf());
       job.setJarByClass(this.getClass());
-      
+
       job.setInputFormat(AccumuloInputFormat.class);
-      
+
       AccumuloInputFormat.setConnectorInfo(job, user, new PasswordToken(pass));
       AccumuloInputFormat.setInputTableName(job, table1);
       AccumuloInputFormat.setMockInstance(job, INSTANCE_NAME);
-      
+
       job.setMapperClass(TestMapper.class);
       job.setMapOutputKeyClass(Key.class);
       job.setMapOutputValueClass(Value.class);
       job.setOutputFormat(AccumuloOutputFormat.class);
       job.setOutputKeyClass(Text.class);
       job.setOutputValueClass(Mutation.class);
-      
+
       AccumuloOutputFormat.setConnectorInfo(job, user, new PasswordToken(pass));
       AccumuloOutputFormat.setCreateTables(job, false);
       AccumuloOutputFormat.setDefaultTableName(job, table2);
       AccumuloOutputFormat.setMockInstance(job, INSTANCE_NAME);
-      
+
       job.setNumReduceTasks(0);
-      
+
       return JobClient.runJob(job).isSuccessful() ? 0 : 1;
     }
-    
+
     public static void main(String[] args) throws Exception {
       assertEquals(0, ToolRunner.run(CachedConfiguration.getInstance(), new MRTester(), args));
     }
   }
-  
+
   @Test
   public void testBWSettings() throws IOException {
     JobConf job = new JobConf();
-    
+
     // make sure we aren't testing defaults
     final BatchWriterConfig bwDefaults = new BatchWriterConfig();
     assertNotEquals(7654321l, bwDefaults.getMaxLatency(TimeUnit.MILLISECONDS));
     assertNotEquals(9898989l, bwDefaults.getTimeout(TimeUnit.MILLISECONDS));
     assertNotEquals(42, bwDefaults.getMaxWriteThreads());
     assertNotEquals(1123581321l, bwDefaults.getMaxMemory());
-    
+
     final BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxLatency(7654321l, TimeUnit.MILLISECONDS);
     bwConfig.setTimeout(9898989l, TimeUnit.MILLISECONDS);
     bwConfig.setMaxWriteThreads(42);
     bwConfig.setMaxMemory(1123581321l);
     AccumuloOutputFormat.setBatchWriterOptions(job, bwConfig);
-    
+
     AccumuloOutputFormat myAOF = new AccumuloOutputFormat() {
       @Override
       public void checkOutputSpecs(FileSystem ignored, JobConf job) throws IOException {
         BatchWriterConfig bwOpts = getBatchWriterOptions(job);
-        
+
         // passive check
         assertEquals(bwConfig.getMaxLatency(TimeUnit.MILLISECONDS), bwOpts.getMaxLatency(TimeUnit.MILLISECONDS));
         assertEquals(bwConfig.getTimeout(TimeUnit.MILLISECONDS), bwOpts.getTimeout(TimeUnit.MILLISECONDS));
         assertEquals(bwConfig.getMaxWriteThreads(), bwOpts.getMaxWriteThreads());
         assertEquals(bwConfig.getMaxMemory(), bwOpts.getMaxMemory());
-        
+
         // explicit check
         assertEquals(7654321l, bwOpts.getMaxLatency(TimeUnit.MILLISECONDS));
         assertEquals(9898989l, bwOpts.getTimeout(TimeUnit.MILLISECONDS));
         assertEquals(42, bwOpts.getMaxWriteThreads());
         assertEquals(1123581321l, bwOpts.getMaxMemory());
-        
+
       }
     };
     myAOF.checkOutputSpecs(null, job);
   }
-  
+
   @Test
   public void testMR() throws Exception {
     MockInstance mockInstance = new MockInstance(INSTANCE_NAME);
@@ -189,10 +189,10 @@ public class AccumuloOutputFormatTest {
       bw.addMutation(m);
     }
     bw.close();
-    
+
     MRTester.main(new String[] {"root", "", TEST_TABLE_1, TEST_TABLE_2});
     assertNull(e1);
-    
+
     Scanner scanner = c.createScanner(TEST_TABLE_2, new Authorizations());
     Iterator<Entry<Key,Value>> iter = scanner.iterator();
     assertTrue(iter.hasNext());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormatTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormatTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormatTest.java
index a0ae0b3..219e4f0 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormatTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormatTest.java
@@ -55,7 +55,7 @@ public class AccumuloRowInputFormatTest {
   private static final String PREFIX = AccumuloRowInputFormatTest.class.getSimpleName();
   private static final String INSTANCE_NAME = PREFIX + "_mapred_instance";
   private static final String TEST_TABLE_1 = PREFIX + "_mapred_table_1";
-  
+
   private static final String ROW1 = "row1";
   private static final String ROW2 = "row2";
   private static final String ROW3 = "row3";
@@ -65,7 +65,7 @@ public class AccumuloRowInputFormatTest {
   private static List<Entry<Key,Value>> row3;
   private static AssertionError e1 = null;
   private static AssertionError e2 = null;
-  
+
   public AccumuloRowInputFormatTest() {
     row1 = new ArrayList<Entry<Key,Value>>();
     row1.add(new KeyValue(new Key(ROW1, COLF1, "colq1"), "v1".getBytes()));
@@ -76,7 +76,7 @@ public class AccumuloRowInputFormatTest {
     row3 = new ArrayList<Entry<Key,Value>>();
     row3.add(new KeyValue(new Key(ROW3, COLF1, "colq5"), "v5".getBytes()));
   }
-  
+
   public static void checkLists(final List<Entry<Key,Value>> first, final List<Entry<Key,Value>> second) {
     assertEquals("Sizes should be the same.", first.size(), second.size());
     for (int i = 0; i < first.size(); i++) {
@@ -84,7 +84,7 @@ public class AccumuloRowInputFormatTest {
       assertEquals("Values should be equal.", first.get(i).getValue(), second.get(i).getValue());
     }
   }
-  
+
   public static void checkLists(final List<Entry<Key,Value>> first, final Iterator<Entry<Key,Value>> second) {
     int entryIndex = 0;
     while (second.hasNext()) {
@@ -94,7 +94,7 @@ public class AccumuloRowInputFormatTest {
       entryIndex++;
     }
   }
-  
+
   public static void insertList(final BatchWriter writer, final List<Entry<Key,Value>> list) throws MutationsRejectedException {
     for (Entry<Key,Value> e : list) {
       final Key key = e.getKey();
@@ -104,11 +104,11 @@ public class AccumuloRowInputFormatTest {
       writer.addMutation(mutation);
     }
   }
-  
+
   private static class MRTester extends Configured implements Tool {
     public static class TestMapper implements Mapper<Text,PeekingIterator<Entry<Key,Value>>,Key,Value> {
       int count = 0;
-      
+
       @Override
       public void map(Text k, PeekingIterator<Entry<Key,Value>> v, OutputCollector<Key,Value> output, Reporter reporter) throws IOException {
         try {
@@ -133,10 +133,10 @@ public class AccumuloRowInputFormatTest {
         }
         count++;
       }
-      
+
       @Override
       public void configure(JobConf job) {}
-      
+
       @Override
       public void close() throws IOException {
         try {
@@ -145,44 +145,44 @@ public class AccumuloRowInputFormatTest {
           e2 = e;
         }
       }
-      
+
     }
-    
+
     @Override
     public int run(String[] args) throws Exception {
-      
+
       if (args.length != 3) {
         throw new IllegalArgumentException("Usage : " + MRTester.class.getName() + " <user> <pass> <table>");
       }
-      
+
       String user = args[0];
       String pass = args[1];
       String table = args[2];
-      
+
       JobConf job = new JobConf(getConf());
       job.setJarByClass(this.getClass());
-      
+
       job.setInputFormat(AccumuloRowInputFormat.class);
-      
+
       AccumuloInputFormat.setConnectorInfo(job, user, new PasswordToken(pass));
       AccumuloInputFormat.setInputTableName(job, table);
       AccumuloRowInputFormat.setMockInstance(job, INSTANCE_NAME);
-      
+
       job.setMapperClass(TestMapper.class);
       job.setMapOutputKeyClass(Key.class);
       job.setMapOutputValueClass(Value.class);
       job.setOutputFormat(NullOutputFormat.class);
-      
+
       job.setNumReduceTasks(0);
-      
+
       return JobClient.runJob(job).isSuccessful() ? 0 : 1;
     }
-    
+
     public static void main(String[] args) throws Exception {
       assertEquals(0, ToolRunner.run(CachedConfiguration.getInstance(), new MRTester(), args));
     }
   }
-  
+
   @Test
   public void test() throws Exception {
     final MockInstance instance = new MockInstance(INSTANCE_NAME);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapred/RangeInputSplitTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapred/RangeInputSplitTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapred/RangeInputSplitTest.java
index 88f5527..f567454 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapred/RangeInputSplitTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapred/RangeInputSplitTest.java
@@ -43,28 +43,28 @@ public class RangeInputSplitTest {
 
   @Test
   public void testSimpleWritable() throws IOException {
-    RangeInputSplit split = new RangeInputSplit("table", "1", new Range(new Key("a"), new Key("b")), new String[]{"localhost"});
-    
+    RangeInputSplit split = new RangeInputSplit("table", "1", new Range(new Key("a"), new Key("b")), new String[] {"localhost"});
+
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
     split.write(dos);
-    
+
     RangeInputSplit newSplit = new RangeInputSplit();
-    
+
     ByteArrayInputStream bais = new ByteArrayInputStream(baos.toByteArray());
     DataInputStream dis = new DataInputStream(bais);
     newSplit.readFields(dis);
-    
+
     Assert.assertEquals(split.getRange(), newSplit.getRange());
     Assert.assertTrue(Arrays.equals(split.getLocations(), newSplit.getLocations()));
   }
 
   @Test
   public void testAllFieldsWritable() throws IOException {
-    RangeInputSplit split = new RangeInputSplit("table", "1", new Range(new Key("a"), new Key("b")), new String[]{"localhost"});
-    
+    RangeInputSplit split = new RangeInputSplit("table", "1", new Range(new Key("a"), new Key("b")), new String[] {"localhost"});
+
     Set<Pair<Text,Text>> fetchedColumns = new HashSet<Pair<Text,Text>>();
-    
+
     fetchedColumns.add(new Pair<Text,Text>(new Text("colf1"), new Text("colq1")));
     fetchedColumns.add(new Pair<Text,Text>(new Text("colf2"), new Text("colq2")));
 
@@ -77,7 +77,7 @@ public class RangeInputSplitTest {
     setting = new IteratorSetting(100, WholeRowIterator.class);
     setting.addOption("bar", "foo");
     iterators.add(setting);
-    
+
     split.setAuths(new Authorizations("foo"));
     split.setOffline(true);
     split.setIsolatedScan(true);
@@ -90,20 +90,20 @@ public class RangeInputSplitTest {
     split.setZooKeepers("localhost");
     split.setIterators(iterators);
     split.setLogLevel(Level.WARN);
-    
+
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
     split.write(dos);
-    
+
     RangeInputSplit newSplit = new RangeInputSplit();
-    
+
     ByteArrayInputStream bais = new ByteArrayInputStream(baos.toByteArray());
     DataInputStream dis = new DataInputStream(bais);
     newSplit.readFields(dis);
-    
+
     Assert.assertEquals(split.getRange(), newSplit.getRange());
     Assert.assertArrayEquals(split.getLocations(), newSplit.getLocations());
-    
+
     Assert.assertEquals(split.getAuths(), newSplit.getAuths());
     Assert.assertEquals(split.isOffline(), newSplit.isOffline());
     Assert.assertEquals(split.isIsolatedScan(), newSplit.isOffline());
@@ -117,5 +117,5 @@ public class RangeInputSplitTest {
     Assert.assertEquals(split.getIterators(), newSplit.getIterators());
     Assert.assertEquals(split.getLogLevel(), newSplit.getLogLevel());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapred/TokenFileTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapred/TokenFileTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapred/TokenFileTest.java
index 0e1fe39..2fcaf34 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapred/TokenFileTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapred/TokenFileTest.java
@@ -54,7 +54,7 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
 /**
- * 
+ *
  */
 public class TokenFileTest {
   private static AssertionError e1 = null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormatTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormatTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormatTest.java
index 869ae9d..9d74182 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormatTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormatTest.java
@@ -122,7 +122,7 @@ public class AccumuloInputFormatTest {
   /**
    * Test adding iterator options where the keys and values contain both the FIELD_SEPARATOR character (':') and ITERATOR_SEPARATOR (',') characters. There
    * should be no exceptions thrown when trying to parse these types of option entries.
-   * 
+   *
    * This test makes sure that the expected raw values, as appears in the Job, are equal to what's expected.
    */
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormatTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormatTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormatTest.java
index a0cb4e3..0be7b0a 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormatTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormatTest.java
@@ -48,7 +48,7 @@ import org.apache.hadoop.util.ToolRunner;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class AccumuloOutputFormatTest {
   private static AssertionError e1 = null;
@@ -56,12 +56,12 @@ public class AccumuloOutputFormatTest {
   private static final String INSTANCE_NAME = PREFIX + "_mapreduce_instance";
   private static final String TEST_TABLE_1 = PREFIX + "_mapreduce_table_1";
   private static final String TEST_TABLE_2 = PREFIX + "_mapreduce_table_2";
-  
+
   private static class MRTester extends Configured implements Tool {
     private static class TestMapper extends Mapper<Key,Value,Text,Mutation> {
       Key key = null;
       int count = 0;
-      
+
       @Override
       protected void map(Key k, Value v, Context context) throws IOException, InterruptedException {
         try {
@@ -75,7 +75,7 @@ public class AccumuloOutputFormatTest {
         key = new Key(k);
         count++;
       }
-      
+
       @Override
       protected void cleanup(Context context) throws IOException, InterruptedException {
         Mutation m = new Mutation("total");
@@ -83,14 +83,14 @@ public class AccumuloOutputFormatTest {
         context.write(new Text(), m);
       }
     }
-    
+
     @Override
     public int run(String[] args) throws Exception {
-      
+
       if (args.length != 4) {
         throw new IllegalArgumentException("Usage : " + MRTester.class.getName() + " <user> <pass> <inputtable> <outputtable>");
       }
-      
+
       String user = args[0];
       String pass = args[1];
       String table1 = args[2];
@@ -99,78 +99,78 @@ public class AccumuloOutputFormatTest {
       @SuppressWarnings("deprecation")
       Job job = new Job(getConf(), this.getClass().getSimpleName() + "_" + System.currentTimeMillis());
       job.setJarByClass(this.getClass());
-      
+
       job.setInputFormatClass(AccumuloInputFormat.class);
-      
+
       AccumuloInputFormat.setConnectorInfo(job, user, new PasswordToken(pass));
       AccumuloInputFormat.setInputTableName(job, table1);
       AccumuloInputFormat.setMockInstance(job, INSTANCE_NAME);
-      
+
       job.setMapperClass(TestMapper.class);
       job.setMapOutputKeyClass(Key.class);
       job.setMapOutputValueClass(Value.class);
       job.setOutputFormatClass(AccumuloOutputFormat.class);
       job.setOutputKeyClass(Text.class);
       job.setOutputValueClass(Mutation.class);
-      
+
       AccumuloOutputFormat.setConnectorInfo(job, user, new PasswordToken(pass));
       AccumuloOutputFormat.setCreateTables(job, false);
       AccumuloOutputFormat.setDefaultTableName(job, table2);
       AccumuloOutputFormat.setMockInstance(job, INSTANCE_NAME);
-      
+
       job.setNumReduceTasks(0);
-      
+
       job.waitForCompletion(true);
-      
+
       return job.isSuccessful() ? 0 : 1;
     }
-    
+
     public static void main(String[] args) throws Exception {
       assertEquals(0, ToolRunner.run(CachedConfiguration.getInstance(), new MRTester(), args));
     }
   }
-  
+
   @Test
   public void testBWSettings() throws IOException {
     @SuppressWarnings("deprecation")
     Job job = new Job();
-    
+
     // make sure we aren't testing defaults
     final BatchWriterConfig bwDefaults = new BatchWriterConfig();
     assertNotEquals(7654321l, bwDefaults.getMaxLatency(TimeUnit.MILLISECONDS));
     assertNotEquals(9898989l, bwDefaults.getTimeout(TimeUnit.MILLISECONDS));
     assertNotEquals(42, bwDefaults.getMaxWriteThreads());
     assertNotEquals(1123581321l, bwDefaults.getMaxMemory());
-    
+
     final BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxLatency(7654321l, TimeUnit.MILLISECONDS);
     bwConfig.setTimeout(9898989l, TimeUnit.MILLISECONDS);
     bwConfig.setMaxWriteThreads(42);
     bwConfig.setMaxMemory(1123581321l);
     AccumuloOutputFormat.setBatchWriterOptions(job, bwConfig);
-    
+
     AccumuloOutputFormat myAOF = new AccumuloOutputFormat() {
       @Override
       public void checkOutputSpecs(JobContext job) throws IOException {
         BatchWriterConfig bwOpts = getBatchWriterOptions(job);
-        
+
         // passive check
         assertEquals(bwConfig.getMaxLatency(TimeUnit.MILLISECONDS), bwOpts.getMaxLatency(TimeUnit.MILLISECONDS));
         assertEquals(bwConfig.getTimeout(TimeUnit.MILLISECONDS), bwOpts.getTimeout(TimeUnit.MILLISECONDS));
         assertEquals(bwConfig.getMaxWriteThreads(), bwOpts.getMaxWriteThreads());
         assertEquals(bwConfig.getMaxMemory(), bwOpts.getMaxMemory());
-        
+
         // explicit check
         assertEquals(7654321l, bwOpts.getMaxLatency(TimeUnit.MILLISECONDS));
         assertEquals(9898989l, bwOpts.getTimeout(TimeUnit.MILLISECONDS));
         assertEquals(42, bwOpts.getMaxWriteThreads());
         assertEquals(1123581321l, bwOpts.getMaxMemory());
-        
+
       }
     };
     myAOF.checkOutputSpecs(job);
   }
-  
+
   @Test
   public void testMR() throws Exception {
     MockInstance mockInstance = new MockInstance(INSTANCE_NAME);
@@ -184,10 +184,10 @@ public class AccumuloOutputFormatTest {
       bw.addMutation(m);
     }
     bw.close();
-    
+
     MRTester.main(new String[] {"root", "", TEST_TABLE_1, TEST_TABLE_2});
     assertNull(e1);
-    
+
     Scanner scanner = c.createScanner(TEST_TABLE_2, new Authorizations());
     Iterator<Entry<Key,Value>> iter = scanner.iterator();
     assertTrue(iter.hasNext());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormatTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormatTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormatTest.java
index 2207437..e8920eb 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormatTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormatTest.java
@@ -52,7 +52,7 @@ public class AccumuloRowInputFormatTest {
   private static final String PREFIX = AccumuloRowInputFormatTest.class.getSimpleName();
   private static final String INSTANCE_NAME = PREFIX + "_mapreduce_instance";
   private static final String TEST_TABLE_1 = PREFIX + "_mapreduce_table_1";
-  
+
   private static final String ROW1 = "row1";
   private static final String ROW2 = "row2";
   private static final String ROW3 = "row3";
@@ -62,7 +62,7 @@ public class AccumuloRowInputFormatTest {
   private static List<Entry<Key,Value>> row3;
   private static AssertionError e1 = null;
   private static AssertionError e2 = null;
-  
+
   public AccumuloRowInputFormatTest() {
     row1 = new ArrayList<Entry<Key,Value>>();
     row1.add(new KeyValue(new Key(ROW1, COLF1, "colq1"), "v1".getBytes()));
@@ -73,7 +73,7 @@ public class AccumuloRowInputFormatTest {
     row3 = new ArrayList<Entry<Key,Value>>();
     row3.add(new KeyValue(new Key(ROW3, COLF1, "colq5"), "v5".getBytes()));
   }
-  
+
   public static void checkLists(final List<Entry<Key,Value>> first, final List<Entry<Key,Value>> second) {
     assertEquals("Sizes should be the same.", first.size(), second.size());
     for (int i = 0; i < first.size(); i++) {
@@ -81,7 +81,7 @@ public class AccumuloRowInputFormatTest {
       assertEquals("Values should be equal.", first.get(i).getValue(), second.get(i).getValue());
     }
   }
-  
+
   public static void checkLists(final List<Entry<Key,Value>> first, final Iterator<Entry<Key,Value>> second) {
     int entryIndex = 0;
     while (second.hasNext()) {
@@ -91,7 +91,7 @@ public class AccumuloRowInputFormatTest {
       entryIndex++;
     }
   }
-  
+
   public static void insertList(final BatchWriter writer, final List<Entry<Key,Value>> list) throws MutationsRejectedException {
     for (Entry<Key,Value> e : list) {
       final Key key = e.getKey();
@@ -101,11 +101,11 @@ public class AccumuloRowInputFormatTest {
       writer.addMutation(mutation);
     }
   }
-  
+
   private static class MRTester extends Configured implements Tool {
     private static class TestMapper extends Mapper<Text,PeekingIterator<Entry<Key,Value>>,Key,Value> {
       int count = 0;
-      
+
       @Override
       protected void map(Text k, PeekingIterator<Entry<Key,Value>> v, Context context) throws IOException, InterruptedException {
         try {
@@ -130,7 +130,7 @@ public class AccumuloRowInputFormatTest {
         }
         count++;
       }
-      
+
       @Override
       protected void cleanup(Context context) throws IOException, InterruptedException {
         try {
@@ -140,14 +140,14 @@ public class AccumuloRowInputFormatTest {
         }
       }
     }
-    
+
     @Override
     public int run(String[] args) throws Exception {
-      
+
       if (args.length != 3) {
         throw new IllegalArgumentException("Usage : " + MRTester.class.getName() + " <user> <pass> <table>");
       }
-      
+
       String user = args[0];
       String pass = args[1];
       String table = args[2];
@@ -155,30 +155,30 @@ public class AccumuloRowInputFormatTest {
       @SuppressWarnings("deprecation")
       Job job = new Job(getConf(), this.getClass().getSimpleName() + "_" + System.currentTimeMillis());
       job.setJarByClass(this.getClass());
-      
+
       job.setInputFormatClass(AccumuloRowInputFormat.class);
-      
+
       AccumuloInputFormat.setConnectorInfo(job, user, new PasswordToken(pass));
       AccumuloInputFormat.setInputTableName(job, table);
       AccumuloRowInputFormat.setMockInstance(job, INSTANCE_NAME);
-      
+
       job.setMapperClass(TestMapper.class);
       job.setMapOutputKeyClass(Key.class);
       job.setMapOutputValueClass(Value.class);
       job.setOutputFormatClass(NullOutputFormat.class);
-      
+
       job.setNumReduceTasks(0);
-      
+
       job.waitForCompletion(true);
-      
+
       return job.isSuccessful() ? 0 : 1;
     }
-    
+
     public static void main(String[] args) throws Exception {
       assertEquals(0, ToolRunner.run(CachedConfiguration.getInstance(), new MRTester(), args));
     }
   }
-  
+
   @Test
   public void test() throws Exception {
     final MockInstance instance = new MockInstance(INSTANCE_NAME);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/BadPasswordSplitsAccumuloInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/BadPasswordSplitsAccumuloInputFormat.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/BadPasswordSplitsAccumuloInputFormat.java
index fce7781..9028d94 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/BadPasswordSplitsAccumuloInputFormat.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/BadPasswordSplitsAccumuloInputFormat.java
@@ -27,16 +27,16 @@ import org.apache.hadoop.mapreduce.JobContext;
  * AccumuloInputFormat which returns an "empty" RangeInputSplit
  */
 public class BadPasswordSplitsAccumuloInputFormat extends AccumuloInputFormat {
-  
+
   @Override
   public List<InputSplit> getSplits(JobContext context) throws IOException {
     List<InputSplit> splits = super.getSplits(context);
-    
+
     for (InputSplit split : splits) {
       org.apache.accumulo.core.client.mapreduce.RangeInputSplit rangeSplit = (org.apache.accumulo.core.client.mapreduce.RangeInputSplit) split;
       rangeSplit.setToken(new PasswordToken("anythingelse"));
     }
-    
+
     return splits;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/InputTableConfigTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/InputTableConfigTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/InputTableConfigTest.java
index 4855094..4953654 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/InputTableConfigTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/InputTableConfigTest.java
@@ -16,6 +16,8 @@
  */
 package org.apache.accumulo.core.client.mapreduce;
 
+import static org.junit.Assert.assertEquals;
+
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
 import java.io.DataInputStream;
@@ -33,22 +35,20 @@ import org.apache.hadoop.io.Text;
 import org.junit.Before;
 import org.junit.Test;
 
-import static org.junit.Assert.assertEquals;
-
 public class InputTableConfigTest {
-  
+
   private InputTableConfig tableQueryConfig;
-  
+
   @Before
   public void setUp() {
     tableQueryConfig = new InputTableConfig();
   }
-  
+
   @Test
   public void testSerialization_OnlyTable() throws IOException {
     byte[] serialized = serialize(tableQueryConfig);
     InputTableConfig actualConfig = deserialize(serialized);
-    
+
     assertEquals(tableQueryConfig, actualConfig);
   }
 
@@ -64,33 +64,32 @@ public class InputTableConfigTest {
     assertEquals(tableQueryConfig, actualConfig);
   }
 
-
   @Test
   public void testSerialization_ranges() throws IOException {
     List<Range> ranges = new ArrayList<Range>();
     ranges.add(new Range("a", "b"));
     ranges.add(new Range("c", "d"));
     tableQueryConfig.setRanges(ranges);
-    
+
     byte[] serialized = serialize(tableQueryConfig);
     InputTableConfig actualConfig = deserialize(serialized);
-    
+
     assertEquals(ranges, actualConfig.getRanges());
   }
-  
+
   @Test
   public void testSerialization_columns() throws IOException {
     Set<Pair<Text,Text>> columns = new HashSet<Pair<Text,Text>>();
     columns.add(new Pair<Text,Text>(new Text("cf1"), new Text("cq1")));
     columns.add(new Pair<Text,Text>(new Text("cf2"), null));
     tableQueryConfig.fetchColumns(columns);
-    
+
     byte[] serialized = serialize(tableQueryConfig);
     InputTableConfig actualConfig = deserialize(serialized);
-    
+
     assertEquals(actualConfig.getFetchedColumns(), columns);
   }
-  
+
   @Test
   public void testSerialization_iterators() throws IOException {
     List<IteratorSetting> settings = new ArrayList<IteratorSetting>();
@@ -100,16 +99,16 @@ public class InputTableConfigTest {
     byte[] serialized = serialize(tableQueryConfig);
     InputTableConfig actualConfig = deserialize(serialized);
     assertEquals(actualConfig.getIterators(), settings);
-    
+
   }
-  
+
   private byte[] serialize(InputTableConfig tableQueryConfig) throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     tableQueryConfig.write(new DataOutputStream(baos));
     baos.close();
     return baos.toByteArray();
   }
-  
+
   private InputTableConfig deserialize(byte[] bytes) throws IOException {
     ByteArrayInputStream bais = new ByteArrayInputStream(bytes);
     InputTableConfig actualConfig = new InputTableConfig(new DataInputStream(bais));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/TokenFileTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/TokenFileTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/TokenFileTest.java
index fd207a1..8359b89 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/TokenFileTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/TokenFileTest.java
@@ -51,7 +51,7 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
 /**
- * 
+ *
  */
 public class TokenFileTest {
   private static AssertionError e1 = null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBaseTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBaseTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBaseTest.java
index 7c1f98b..bcbc236 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBaseTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBaseTest.java
@@ -37,7 +37,7 @@ import org.apache.log4j.Logger;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ConfiguratorBaseTest {
 
@@ -128,10 +128,10 @@ public class ConfiguratorBaseTest {
   }
 
   @Test
-  public void testSetVisibiltyCacheSize(){
-	 Configuration conf = new Configuration();
-	 assertEquals(Constants.DEFAULT_VISIBILITY_CACHE_SIZE,ConfiguratorBase.getVisibilityCacheSize(conf));
-	 ConfiguratorBase.setVisibilityCacheSize(conf, 2000);
-	 assertEquals(2000,ConfiguratorBase.getVisibilityCacheSize(conf));
+  public void testSetVisibiltyCacheSize() {
+    Configuration conf = new Configuration();
+    assertEquals(Constants.DEFAULT_VISIBILITY_CACHE_SIZE, ConfiguratorBase.getVisibilityCacheSize(conf));
+    ConfiguratorBase.setVisibilityCacheSize(conf, 2000);
+    assertEquals(2000, ConfiguratorBase.getVisibilityCacheSize(conf));
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/partition/RangePartitionerTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/partition/RangePartitionerTest.java b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/partition/RangePartitionerTest.java
index 8fca169..13a09d6 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/partition/RangePartitionerTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mapreduce/lib/partition/RangePartitionerTest.java
@@ -25,9 +25,9 @@ import org.apache.hadoop.mapreduce.Job;
 import org.junit.Test;
 
 public class RangePartitionerTest {
-  
+
   private static Text[] cutArray = new Text[] {new Text("A"), new Text("B"), new Text("C")};
-  
+
   @Test
   public void testNoSubBins() throws IOException {
     for (int i = -2; i < 2; ++i) {
@@ -36,22 +36,22 @@ public class RangePartitionerTest {
       checkExpectedBins(i, new String[] {"", "AA", "BB", "CC"}, new int[] {0, 1, 2, 3});
     }
   }
-  
+
   @Test
   public void testSubBins() throws IOException {
     checkExpectedRangeBins(2, new String[] {"A", "B", "C"}, new int[] {1, 3, 5});
     checkExpectedRangeBins(2, new String[] {"C", "A", "B"}, new int[] {5, 1, 3});
     checkExpectedRangeBins(2, new String[] {"", "AA", "BB", "CC"}, new int[] {1, 3, 5, 7});
-    
+
     checkExpectedRangeBins(3, new String[] {"A", "B", "C"}, new int[] {2, 5, 8});
     checkExpectedRangeBins(3, new String[] {"C", "A", "B"}, new int[] {8, 2, 5});
     checkExpectedRangeBins(3, new String[] {"", "AA", "BB", "CC"}, new int[] {2, 5, 8, 11});
-    
+
     checkExpectedRangeBins(10, new String[] {"A", "B", "C"}, new int[] {9, 19, 29});
     checkExpectedRangeBins(10, new String[] {"C", "A", "B"}, new int[] {29, 9, 19});
     checkExpectedRangeBins(10, new String[] {"", "AA", "BB", "CC"}, new int[] {9, 19, 29, 39});
   }
-  
+
   private RangePartitioner prepPartitioner(int numSubBins) throws IOException {
     @SuppressWarnings("deprecation")
     Job job = new Job();
@@ -60,7 +60,7 @@ public class RangePartitionerTest {
     rp.setConf(job.getConfiguration());
     return rp;
   }
-  
+
   private void checkExpectedRangeBins(int numSubBins, String[] strings, int[] rangeEnds) throws IOException {
     assertTrue(strings.length == rangeEnds.length);
     for (int i = 0; i < strings.length; ++i) {
@@ -71,7 +71,7 @@ public class RangePartitionerTest {
       assertTrue(part <= endRange);
     }
   }
-  
+
   private void checkExpectedBins(int numSubBins, String[] strings, int[] bins) throws IOException {
     assertTrue(strings.length == bins.length);
     for (int i = 0; i < strings.length; ++i) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mock/MockConnectorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mock/MockConnectorTest.java b/core/src/test/java/org/apache/accumulo/core/client/mock/MockConnectorTest.java
index 7ccae0a..4b90d72 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mock/MockConnectorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mock/MockConnectorTest.java
@@ -53,11 +53,11 @@ import org.junit.Test;
 
 public class MockConnectorTest {
   Random random = new Random();
-  
+
   static Text asText(int i) {
     return new Text(Integer.toHexString(i));
   }
-  
+
   @Test
   public void testSunnyDay() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
@@ -83,7 +83,7 @@ public class MockConnectorTest {
     }
     assertEquals(100, count);
   }
-  
+
   @Test
   public void testChangeAuths() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
@@ -95,14 +95,14 @@ public class MockConnectorTest {
     assertTrue(c.securityOperations().getUserAuthorizations("greg").contains("X".getBytes()));
     assertFalse(c.securityOperations().getUserAuthorizations("greg").contains("A".getBytes()));
   }
-  
+
   @Test
   public void testBadMutations() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
     c.tableOperations().create("test");
     BatchWriter bw = c
         .createBatchWriter("test", new BatchWriterConfig().setMaxMemory(10000L).setMaxLatency(1000L, TimeUnit.MILLISECONDS).setMaxWriteThreads(4));
-    
+
     try {
       bw.addMutation(null);
       Assert.fail("addMutation should throw IAE for null mutation");
@@ -111,15 +111,15 @@ public class MockConnectorTest {
       bw.addMutations(null);
       Assert.fail("addMutations should throw IAE for null iterable");
     } catch (IllegalArgumentException iae) {}
-    
+
     bw.addMutations(Collections.<Mutation> emptyList());
-    
+
     Mutation bad = new Mutation("bad");
     try {
       bw.addMutation(bad);
       Assert.fail("addMutation should throw IAE for empty mutation");
     } catch (IllegalArgumentException iae) {}
-    
+
     Mutation good = new Mutation("good");
     good.put(asText(random.nextInt()), asText(random.nextInt()), new Value("good".getBytes()));
     List<Mutation> mutations = new ArrayList<Mutation>();
@@ -129,10 +129,10 @@ public class MockConnectorTest {
       bw.addMutations(mutations);
       Assert.fail("addMutations should throw IAE if it contains empty mutation");
     } catch (IllegalArgumentException iae) {}
-    
+
     bw.close();
   }
-  
+
   @Test
   public void testAggregation() throws Exception {
     MockInstance mockInstance = new MockInstance();
@@ -152,7 +152,7 @@ public class MockConnectorTest {
       bw.addMutation(m);
     }
     bw.close();
-    
+
     Scanner s = c.createScanner("perDayCounts", Authorizations.EMPTY);
     Iterator<Entry<Key,Value>> iterator = s.iterator();
     assertTrue(iterator.hasNext());
@@ -163,75 +163,75 @@ public class MockConnectorTest {
     checkEntry(iterator.next(), "foo", "day", "20080103", "1");
     assertFalse(iterator.hasNext());
   }
-  
+
   @Test
   public void testDelete() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
     c.tableOperations().create("test");
     BatchWriter bw = c.createBatchWriter("test", new BatchWriterConfig());
-    
+
     Mutation m1 = new Mutation("r1");
-    
+
     m1.put("cf1", "cq1", 1, "v1");
-    
+
     bw.addMutation(m1);
     bw.flush();
-    
+
     Mutation m2 = new Mutation("r1");
-    
+
     m2.putDelete("cf1", "cq1", 2);
-    
+
     bw.addMutation(m2);
     bw.flush();
-    
+
     Scanner scanner = c.createScanner("test", Authorizations.EMPTY);
-    
+
     int count = 0;
     for (@SuppressWarnings("unused")
     Entry<Key,Value> entry : scanner) {
       count++;
     }
-    
+
     assertEquals(0, count);
-    
+
     try {
       c.tableOperations().create("test_this_$tableName");
       assertTrue(false);
-      
+
     } catch (IllegalArgumentException iae) {
-      
+
     }
   }
-  
+
   @Test
   public void testDeletewithBatchDeleter() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
-    
+
     // make sure we are using a clean table
     if (c.tableOperations().exists("test"))
       c.tableOperations().delete("test");
     c.tableOperations().create("test");
-    
+
     BatchDeleter deleter = c.createBatchDeleter("test", Authorizations.EMPTY, 2, new BatchWriterConfig());
     // first make sure it deletes fine when its empty
     deleter.setRanges(Collections.singletonList(new Range(("r1"))));
     deleter.delete();
     this.checkRemaining(c, "test", 0);
-    
+
     // test deleting just one row
     BatchWriter writer = c.createBatchWriter("test", new BatchWriterConfig());
     Mutation m = new Mutation("r1");
     m.put("fam", "qual", "value");
     writer.addMutation(m);
-    
+
     // make sure the write goes through
     writer.flush();
     writer.close();
-    
+
     deleter.setRanges(Collections.singletonList(new Range(("r1"))));
     deleter.delete();
     this.checkRemaining(c, "test", 0);
-    
+
     // test multi row deletes
     writer = c.createBatchWriter("test", new BatchWriterConfig());
     m = new Mutation("r1");
@@ -240,19 +240,19 @@ public class MockConnectorTest {
     Mutation m2 = new Mutation("r2");
     m2.put("fam", "qual", "value");
     writer.addMutation(m2);
-    
+
     // make sure the write goes through
     writer.flush();
     writer.close();
-    
+
     deleter.setRanges(Collections.singletonList(new Range(("r1"))));
     deleter.delete();
     checkRemaining(c, "test", 1);
   }
-  
+
   /**
    * Test to make sure that a certain number of rows remain
-   * 
+   *
    * @param c
    *          connector to the {@link MockInstance}
    * @param tableName
@@ -262,7 +262,7 @@ public class MockConnectorTest {
    */
   private void checkRemaining(Connector c, String tableName, int count) throws Exception {
     Scanner scanner = c.createScanner(tableName, Authorizations.EMPTY);
-    
+
     int total = 0;
     for (@SuppressWarnings("unused")
     Entry<Key,Value> entry : scanner) {
@@ -270,24 +270,24 @@ public class MockConnectorTest {
     }
     assertEquals(count, total);
   }
-  
+
   @Test
   public void testCMod() throws Exception {
     // test writing to a table that the is being scanned
     Connector c = new MockConnector("root", new MockInstance());
     c.tableOperations().create("test");
     BatchWriter bw = c.createBatchWriter("test", new BatchWriterConfig());
-    
+
     for (int i = 0; i < 10; i++) {
       Mutation m1 = new Mutation("r" + i);
       m1.put("cf1", "cq1", 1, "v" + i);
       bw.addMutation(m1);
     }
-    
+
     bw.flush();
-    
+
     int count = 10;
-    
+
     Scanner scanner = c.createScanner("test", Authorizations.EMPTY);
     for (Entry<Key,Value> entry : scanner) {
       Key key = entry.getKey();
@@ -296,33 +296,33 @@ public class MockConnectorTest {
       count++;
       bw.addMutation(m);
     }
-    
+
     bw.flush();
-    
+
     count = 10;
-    
+
     for (Entry<Key,Value> entry : scanner) {
       assertEquals(entry.getValue().toString(), "v" + (count++));
     }
-    
+
     assertEquals(count, 20);
-    
+
     try {
       c.tableOperations().create("test_this_$tableName");
       assertTrue(false);
-      
+
     } catch (IllegalArgumentException iae) {
-      
+
     }
   }
-  
+
   private void checkEntry(Entry<Key,Value> next, String row, String cf, String cq, String value) {
     assertEquals(row, next.getKey().getRow().toString());
     assertEquals(cf, next.getKey().getColumnFamily().toString());
     assertEquals(cq, next.getKey().getColumnQualifier().toString());
     assertEquals(value, next.getValue().toString());
   }
-  
+
   @Test
   public void testMockMultiTableBatchWriter() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
@@ -337,7 +337,7 @@ public class MockConnectorTest {
     b = bw.getBatchWriter("b");
     b.addMutation(m1);
     b.flush();
-    
+
     Scanner scanner = c.createScanner("a", Authorizations.EMPTY);
     int count = 0;
     for (@SuppressWarnings("unused")
@@ -352,31 +352,31 @@ public class MockConnectorTest {
       count++;
     }
     assertEquals(1, count);
-    
+
   }
-  
+
   @Test
   public void testUpdate() throws Exception {
     Connector c = new MockConnector("root", new MockInstance());
     c.tableOperations().create("test");
     BatchWriter bw = c.createBatchWriter("test", new BatchWriterConfig());
-    
+
     for (int i = 0; i < 10; i++) {
       Mutation m = new Mutation("r1");
       m.put("cf1", "cq1", "" + i);
       bw.addMutation(m);
     }
-    
+
     bw.close();
-    
+
     Scanner scanner = c.createScanner("test", Authorizations.EMPTY);
-    
+
     Entry<Key,Value> entry = scanner.iterator().next();
-    
+
     assertEquals("9", entry.getValue().toString());
-    
+
   }
-  
+
   @Test
   public void testMockConnectorReturnsCorrectInstance() throws AccumuloException, AccumuloSecurityException {
     String name = "an-interesting-instance-name";
@@ -384,5 +384,5 @@ public class MockConnectorTest {
     assertEquals(mockInstance, mockInstance.getConnector("foo", new PasswordToken("bar")).getInstance());
     assertEquals(name, mockInstance.getConnector("foo", new PasswordToken("bar")).getInstance().getInstanceName());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mock/MockNamespacesTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mock/MockNamespacesTest.java b/core/src/test/java/org/apache/accumulo/core/client/mock/MockNamespacesTest.java
index b45054a..77b989f 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mock/MockNamespacesTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mock/MockNamespacesTest.java
@@ -115,9 +115,9 @@ public class MockNamespacesTest {
   /**
    * This test creates a namespace, modifies it's properties, and checks to make sure that those properties are applied to its tables. To do something on a
    * namespace-wide level, use {@link NamespaceOperations}.
-   * 
+   *
    * Checks to make sure namespace-level properties are overridden by table-level properties.
-   * 
+   *
    * Checks to see if the default namespace's properties work as well.
    */
 
@@ -212,7 +212,7 @@ public class MockNamespacesTest {
     // TODO implement clone in mock
     /*
      * c.tableOperations().clone(tableName1, tableName2, false, null, null);
-     * 
+     *
      * assertTrue(c.tableOperations().exists(tableName1)); assertTrue(c.tableOperations().exists(tableName2));
      */
     return;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mock/MockTableOperationsTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mock/MockTableOperationsTest.java b/core/src/test/java/org/apache/accumulo/core/client/mock/MockTableOperationsTest.java
index c15fded..193973a 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mock/MockTableOperationsTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mock/MockTableOperationsTest.java
@@ -334,7 +334,6 @@ public class MockTableOperationsTest {
 
   }
 
-
   @Test
   public void testTableIdMap() throws Exception {
     Instance inst = new MockInstance("testTableIdMap");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mock/TestBatchScanner821.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mock/TestBatchScanner821.java b/core/src/test/java/org/apache/accumulo/core/client/mock/TestBatchScanner821.java
index 368bc29..b03bda9 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mock/TestBatchScanner821.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mock/TestBatchScanner821.java
@@ -35,7 +35,7 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.junit.Test;
 
 public class TestBatchScanner821 {
-  
+
   @Test
   public void test() throws Exception {
     MockInstance inst = new MockInstance();
@@ -61,5 +61,5 @@ public class TestBatchScanner821 {
     }
     assertEquals("a,b,c,d", sb.toString());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/mock/TransformIterator.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/mock/TransformIterator.java b/core/src/test/java/org/apache/accumulo/core/client/mock/TransformIterator.java
index bac9a17..a7e7eef 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/mock/TransformIterator.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/mock/TransformIterator.java
@@ -28,5 +28,3 @@ public class TransformIterator extends WrappingIterator {
     return new Key(new Text(k.getRow().toString().toLowerCase()), k.getColumnFamily(), k.getColumnQualifier(), k.getColumnVisibility(), k.getTimestamp());
   }
 }
-
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/security/SecurityErrorCodeTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/security/SecurityErrorCodeTest.java b/core/src/test/java/org/apache/accumulo/core/client/security/SecurityErrorCodeTest.java
index 7554a49..75d4c35 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/security/SecurityErrorCodeTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/security/SecurityErrorCodeTest.java
@@ -22,21 +22,21 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class SecurityErrorCodeTest {
-  
+
   @Test
   public void testEnumsSame() {
     HashSet<String> secNames1 = new HashSet<String>();
     HashSet<String> secNames2 = new HashSet<String>();
-    
+
     for (SecurityErrorCode sec : SecurityErrorCode.values())
       secNames1.add(sec.name());
-    
+
     for (org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode sec : org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode.values())
       secNames2.add(sec.name());
-    
+
     Assert.assertEquals(secNames1, secNames2);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/security/tokens/CredentialProviderTokenTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/security/tokens/CredentialProviderTokenTest.java b/core/src/test/java/org/apache/accumulo/core/client/security/tokens/CredentialProviderTokenTest.java
index bbed50c..d6c7025 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/security/tokens/CredentialProviderTokenTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/security/tokens/CredentialProviderTokenTest.java
@@ -21,6 +21,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 import java.io.File;
 import java.io.IOException;
 import java.net.URL;
+
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken.Properties;
 import org.apache.accumulo.core.conf.CredentialProviderFactoryShim;
 import org.junit.Assert;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/security/tokens/PasswordTokenTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/security/tokens/PasswordTokenTest.java b/core/src/test/java/org/apache/accumulo/core/client/security/tokens/PasswordTokenTest.java
index 3751f02..2d8a603 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/security/tokens/PasswordTokenTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/security/tokens/PasswordTokenTest.java
@@ -24,10 +24,10 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class PasswordTokenTest {
-  
+
   @Test
   public void testMultiByte() throws DestroyFailedException {
     PasswordToken pt = new PasswordToken();
@@ -37,7 +37,7 @@ public class PasswordTokenTest {
     props.destroy();
     String s = new String(pt.getPassword(), UTF_8);
     Assert.assertEquals("五六", s);
-    
+
     pt = new PasswordToken("五六");
     s = new String(pt.getPassword(), UTF_8);
     Assert.assertEquals("五六", s);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/conf/AccumuloConfigurationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/conf/AccumuloConfigurationTest.java b/core/src/test/java/org/apache/accumulo/core/conf/AccumuloConfigurationTest.java
index a115215..c8d17f5 100644
--- a/core/src/test/java/org/apache/accumulo/core/conf/AccumuloConfigurationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/conf/AccumuloConfigurationTest.java
@@ -33,16 +33,16 @@ public class AccumuloConfigurationTest {
     assertEquals(42l * 1024l * 1024l, AccumuloConfiguration.getMemoryInBytes("42m"));
     assertEquals(42l * 1024l * 1024l * 1024l, AccumuloConfiguration.getMemoryInBytes("42G"));
     assertEquals(42l * 1024l * 1024l * 1024l, AccumuloConfiguration.getMemoryInBytes("42g"));
-    
+
   }
-  
-  @Test(expected = IllegalArgumentException.class)  
+
+  @Test(expected = IllegalArgumentException.class)
   public void testGetMemoryInBytesFailureCases1() throws Exception {
     AccumuloConfiguration.getMemoryInBytes("42x");
   }
-  
-  @Test(expected = IllegalArgumentException.class)  
+
+  @Test(expected = IllegalArgumentException.class)
   public void testGetMemoryInBytesFailureCases2() throws Exception {
     AccumuloConfiguration.getMemoryInBytes("FooBar");
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/conf/ConfigSanityCheckTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/conf/ConfigSanityCheckTest.java b/core/src/test/java/org/apache/accumulo/core/conf/ConfigSanityCheckTest.java
index 88bbe64..f34bf3b 100644
--- a/core/src/test/java/org/apache/accumulo/core/conf/ConfigSanityCheckTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/conf/ConfigSanityCheckTest.java
@@ -17,6 +17,7 @@
 package org.apache.accumulo.core.conf;
 
 import java.util.Map;
+
 import org.apache.accumulo.core.conf.ConfigSanityCheck.SanityCheckException;
 import org.junit.Before;
 import org.junit.Test;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/conf/DefaultConfigurationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/conf/DefaultConfigurationTest.java b/core/src/test/java/org/apache/accumulo/core/conf/DefaultConfigurationTest.java
index beb63a7..3823dda 100644
--- a/core/src/test/java/org/apache/accumulo/core/conf/DefaultConfigurationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/conf/DefaultConfigurationTest.java
@@ -16,11 +16,13 @@
  */
 package org.apache.accumulo.core.conf;
 
-import org.apache.accumulo.core.conf.AccumuloConfiguration.AllFilter;
+import static org.junit.Assert.assertEquals;
+
 import java.util.Map;
+
+import org.apache.accumulo.core.conf.AccumuloConfiguration.AllFilter;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.assertEquals;
 
 public class DefaultConfigurationTest {
   private DefaultConfiguration c;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/conf/ObservableConfigurationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/conf/ObservableConfigurationTest.java b/core/src/test/java/org/apache/accumulo/core/conf/ObservableConfigurationTest.java
index 3534b6c..b92fac6 100644
--- a/core/src/test/java/org/apache/accumulo/core/conf/ObservableConfigurationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/conf/ObservableConfigurationTest.java
@@ -16,15 +16,17 @@
  */
 package org.apache.accumulo.core.conf;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+
 import java.util.Collection;
 import java.util.Map;
+
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.easymock.EasyMock.createMock;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.verify;
 
 public class ObservableConfigurationTest {
   private static class TestObservableConfig extends ObservableConfiguration {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/conf/PropertyTypeTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/conf/PropertyTypeTest.java b/core/src/test/java/org/apache/accumulo/core/conf/PropertyTypeTest.java
index e1accb5..b70806d 100644
--- a/core/src/test/java/org/apache/accumulo/core/conf/PropertyTypeTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/conf/PropertyTypeTest.java
@@ -16,8 +16,11 @@
  */
 package org.apache.accumulo.core.conf;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
+
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class PropertyTypeTest {
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/conf/SiteConfigurationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/conf/SiteConfigurationTest.java b/core/src/test/java/org/apache/accumulo/core/conf/SiteConfigurationTest.java
index 1783a28..f54adb1 100644
--- a/core/src/test/java/org/apache/accumulo/core/conf/SiteConfigurationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/conf/SiteConfigurationTest.java
@@ -31,7 +31,6 @@ import org.junit.Test;
 public class SiteConfigurationTest {
   private static boolean isCredentialProviderAvailable;
 
-
   @BeforeClass
   public static void checkCredentialProviderAvailable() {
     try {
@@ -52,25 +51,25 @@ public class SiteConfigurationTest {
         .withConstructor(AccumuloConfiguration.class).withArgs(DefaultConfiguration.getInstance()).createMock();
 
     siteCfg.set(Property.INSTANCE_SECRET, "ignored");
-    
+
     // site-cfg.jceks={'ignored.property'=>'ignored', 'instance.secret'=>'mysecret', 'general.rpc.timeout'=>'timeout'}
     URL keystore = SiteConfigurationTest.class.getResource("/site-cfg.jceks");
     Assert.assertNotNull(keystore);
     String keystorePath = new File(keystore.getFile()).getAbsolutePath();
-    
+
     Configuration hadoopConf = new Configuration();
     hadoopConf.set(CredentialProviderFactoryShim.CREDENTIAL_PROVIDER_PATH, "jceks://file" + keystorePath);
-    
+
     EasyMock.expect(siteCfg.getHadoopConfiguration()).andReturn(hadoopConf).once();
 
     EasyMock.replay(siteCfg);
-    
+
     Map<String,String> props = new HashMap<String,String>();
     siteCfg.getProperties(props, new AllFilter());
-    
+
     Assert.assertEquals("mysecret", props.get(Property.INSTANCE_SECRET.getKey()));
     Assert.assertEquals(null, props.get("ignored.property"));
     Assert.assertEquals(Property.GENERAL_RPC_TIMEOUT.getDefaultValue(), props.get(Property.GENERAL_RPC_TIMEOUT.getKey()));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraintTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraintTest.java b/core/src/test/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraintTest.java
index 915188f..dce0ae1 100644
--- a/core/src/test/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraintTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraintTest.java
@@ -26,33 +26,33 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class DefaultKeySizeConstraintTest {
-  
+
   Constraint constraint = new DefaultKeySizeConstraint();
-  
+
   byte[] oversized = new byte[1048577];
   byte[] large = new byte[419430];
-  
+
   @Test
   public void testConstraint() {
 
     // pass constraints
     Mutation m = new Mutation("rowId");
-    m.put("colf", "colq", new Value(new byte[]{}));
+    m.put("colf", "colq", new Value(new byte[] {}));
     assertEquals(Collections.emptyList(), constraint.check(null, m));
 
     // test with row id > 1mb
     m = new Mutation(oversized);
-    m.put("colf", "colq", new Value(new byte[]{}));
+    m.put("colf", "colq", new Value(new byte[] {}));
     assertEquals(Collections.singletonList(DefaultKeySizeConstraint.MAX__KEY_SIZE_EXCEEDED_VIOLATION), constraint.check(null, m));
 
     // test with colf > 1mb
     m = new Mutation("rowid");
-    m.put(new Text(oversized), new Text("colq"), new Value(new byte[]{}));
+    m.put(new Text(oversized), new Text("colq"), new Value(new byte[] {}));
     assertEquals(Collections.singletonList(DefaultKeySizeConstraint.MAX__KEY_SIZE_EXCEEDED_VIOLATION), constraint.check(null, m));
 
     // test with colf > 1mb
     m = new Mutation("rowid");
-    m.put(new Text(oversized), new Text("colq"), new Value(new byte[]{}));
+    m.put(new Text(oversized), new Text("colq"), new Value(new byte[] {}));
     assertEquals(Collections.singletonList(DefaultKeySizeConstraint.MAX__KEY_SIZE_EXCEEDED_VIOLATION), constraint.check(null, m));
 
     // test sum of smaller sizes violates 1mb constraint

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/data/ColumnTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/data/ColumnTest.java b/core/src/test/java/org/apache/accumulo/core/data/ColumnTest.java
index 9071248..aceddb4 100644
--- a/core/src/test/java/org/apache/accumulo/core/data/ColumnTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/data/ColumnTest.java
@@ -44,7 +44,6 @@ public class ColumnTest {
     col[4] = new Column("colfam".getBytes(), "cq".getBytes(), "cv".getBytes());
   }
 
-
   @Test
   public void testEquals() {
     for (int i = 0; i < col.length; i++) {
@@ -56,7 +55,7 @@ public class ColumnTest {
       }
     }
   }
-  
+
   @Test
   public void testCompare() {
     for (int i = 0; i < col.length; i++) {
@@ -75,7 +74,7 @@ public class ColumnTest {
       for (int j = 0; j < col.length; j++)
         assertEquals(col[i].equals(col[j]), col[i].compareTo(col[j]) == 0);
   }
-  
+
   @Test
   public void testWriteReadFields() throws IOException {
     for (Column c : col) {


[32/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleFactory.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleFactory.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleFactory.java
index bca42c9..0a3f557 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleFactory.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleFactory.java
@@ -27,8 +27,8 @@ import org.apache.log4j.Logger;
 
 /**
  * This factory module exists to assist other classes in loading crypto modules.
- * 
- * 
+ *
+ *
  */
 public class CryptoModuleFactory {
 
@@ -38,7 +38,7 @@ public class CryptoModuleFactory {
 
   /**
    * This method returns a crypto module based on settings in the given configuration parameter.
-   * 
+   *
    * @return a class implementing the CryptoModule interface. It will *never* return null; rather, it will return a class which obeys the interface but makes no
    *         changes to the underlying data.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
index 244a877..573d64b 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
@@ -30,22 +30,22 @@ import javax.crypto.CipherOutputStream;
  * This class defines several parameters needed by by a module providing cryptographic stream support in Accumulo. The following Javadoc details which
  * parameters are used for which operations (encryption vs. decryption), which ones return values (i.e. are "out" parameters from the {@link CryptoModule}), and
  * which ones are required versus optional in certain situations.
- * 
+ *
  * Most of the time, these classes can be constructed using
  * {@link CryptoModuleFactory#createParamsObjectFromAccumuloConfiguration(org.apache.accumulo.core.conf.AccumuloConfiguration)}.
  */
 public class CryptoModuleParameters {
-  
+
   /**
    * Gets the name of the symmetric algorithm to use for encryption.
-   * 
+   *
    * @see CryptoModuleParameters#setAlgorithmName(String)
    */
-  
+
   public String getAlgorithmName() {
     return algorithmName;
   }
-  
+
   /**
    * Sets the name of the symmetric algorithm to use for an encryption stream.
    * <p>
@@ -55,27 +55,27 @@ public class CryptoModuleParameters {
    * For <b>encryption</b>, this value is <b>required</b> and is always used. Its value should be prepended or otherwise included with the ciphertext for future
    * decryption. <br>
    * For <b>decryption</b>, this value is often disregarded in favor of the value encoded with the ciphertext.
-   * 
+   *
    * @param algorithmName
    *          the name of the cryptographic algorithm to use.
    * @see <a href="http://docs.oracle.com/javase/1.5.0/docs/guide/security/jce/JCERefGuide.html#AppA">Standard Algorithm Names in JCE</a>
-   * 
+   *
    */
-  
+
   public void setAlgorithmName(String algorithmName) {
     this.algorithmName = algorithmName;
   }
-  
+
   /**
    * Gets the name of the encryption mode to use for encryption.
-   * 
+   *
    * @see CryptoModuleParameters#setEncryptionMode(String)
    */
-  
+
   public String getEncryptionMode() {
     return encryptionMode;
   }
-  
+
   /**
    * Sets the name of the encryption mode to use for an encryption stream.
    * <p>
@@ -85,27 +85,27 @@ public class CryptoModuleParameters {
    * For <b>encryption</b>, this value is <b>required</b> and is always used. Its value should be prepended or otherwise included with the ciphertext for future
    * decryption. <br>
    * For <b>decryption</b>, this value is often disregarded in favor of the value encoded with the ciphertext.
-   * 
+   *
    * @param encryptionMode
    *          the name of the encryption mode to use.
    * @see <a href="http://docs.oracle.com/javase/1.5.0/docs/guide/security/jce/JCERefGuide.html#AppA">Standard Mode Names in JCE</a>
-   * 
+   *
    */
-  
+
   public void setEncryptionMode(String encryptionMode) {
     this.encryptionMode = encryptionMode;
   }
-  
+
   /**
    * Gets the name of the padding type to use for encryption.
-   * 
+   *
    * @see CryptoModuleParameters#setPadding(String)
    */
-  
+
   public String getPadding() {
     return padding;
   }
-  
+
   /**
    * Sets the name of the padding type to use for an encryption stream.
    * <p>
@@ -115,29 +115,29 @@ public class CryptoModuleParameters {
    * For <b>encryption</b>, this value is <b>required</b> and is always used. Its value should be prepended or otherwise included with the ciphertext for future
    * decryption. <br>
    * For <b>decryption</b>, this value is often disregarded in favor of the value encoded with the ciphertext.
-   * 
+   *
    * @param padding
    *          the name of the padding type to use.
    * @see <a href="http://docs.oracle.com/javase/1.5.0/docs/guide/security/jce/JCERefGuide.html#AppA">Standard Padding Names in JCE</a>
-   * 
+   *
    */
   public void setPadding(String padding) {
     this.padding = padding;
   }
-  
+
   /**
    * Gets the plaintext secret key.
    * <p>
    * For <b>decryption</b>, this value is often the out parameter of using a secret key encryption strategy to decrypt an encrypted version of this secret key.
    * (See {@link CryptoModuleParameters#setKeyEncryptionStrategyClass(String)}.)
-   * 
-   * 
+   *
+   *
    * @see CryptoModuleParameters#setPlaintextKey(byte[])
    */
   public byte[] getPlaintextKey() {
     return plaintextKey;
   }
-  
+
   /**
    * Sets the plaintext secret key that will be used to encrypt and decrypt bytes.
    * <p>
@@ -146,24 +146,24 @@ public class CryptoModuleParameters {
    * For <b>encryption</b>, this value is <b>optional</b>. If it is not provided, it will be automatically generated by the underlying cryptographic module. <br>
    * For <b>decryption</b>, this value is often obtained from the underlying cipher stream, or derived from the encrypted version of the key (see
    * {@link CryptoModuleParameters#setEncryptedKey(byte[])}).
-   * 
+   *
    * @param plaintextKey
    *          the value of the plaintext secret key
    */
-  
+
   public void setPlaintextKey(byte[] plaintextKey) {
     this.plaintextKey = plaintextKey;
   }
-  
+
   /**
    * Gets the length of the secret key.
-   * 
+   *
    * @see CryptoModuleParameters#setKeyLength(int)
    */
   public int getKeyLength() {
     return keyLength;
   }
-  
+
   /**
    * Sets the length of the secret key that will be used to encrypt and decrypt bytes.
    * <p>
@@ -173,50 +173,50 @@ public class CryptoModuleParameters {
    * For <b>encryption</b>, this value is <b>required if the secret key is not set</b>. <br>
    * For <b>decryption</b>, this value is often obtained from the underlying cipher stream, or derived from the encrypted version of the key (see
    * {@link CryptoModuleParameters#setEncryptedKey(byte[])}).
-   * 
+   *
    * @param keyLength
    *          the length of the secret key to be generated
    */
-  
+
   public void setKeyLength(int keyLength) {
     this.keyLength = keyLength;
   }
-  
+
   /**
    * Gets the random number generator name.
-   * 
+   *
    * @see CryptoModuleParameters#setRandomNumberGenerator(String)
    */
-  
+
   public String getRandomNumberGenerator() {
     return randomNumberGenerator;
   }
-  
+
   /**
    * Sets the name of the random number generator to use. The default for this for the baseline JCE implementation is "SHA1PRNG".
    * <p>
-   * 
+   *
    * <p>
    * For <b>encryption</b>, this value is <b>required</b>. <br>
    * For <b>decryption</b>, this value is often obtained from the underlying cipher stream.
-   * 
+   *
    * @param randomNumberGenerator
    *          the name of the random number generator to use
    */
-  
+
   public void setRandomNumberGenerator(String randomNumberGenerator) {
     this.randomNumberGenerator = randomNumberGenerator;
   }
-  
+
   /**
    * Gets the random number generator provider name.
-   * 
+   *
    * @see CryptoModuleParameters#setRandomNumberGeneratorProvider(String)
    */
   public String getRandomNumberGeneratorProvider() {
     return randomNumberGeneratorProvider;
   }
-  
+
   /**
    * Sets the name of the random number generator provider to use. The default for this for the baseline JCE implementation is "SUN".
    * <p>
@@ -224,55 +224,55 @@ public class CryptoModuleParameters {
    * <p>
    * For <b>encryption</b>, this value is <b>required</b>. <br>
    * For <b>decryption</b>, this value is often obtained from the underlying cipher stream.
-   * 
+   *
    * @param randomNumberGeneratorProvider
    *          the name of the provider to use
    */
-  
+
   public void setRandomNumberGeneratorProvider(String randomNumberGeneratorProvider) {
     this.randomNumberGeneratorProvider = randomNumberGeneratorProvider;
   }
-  
+
   /**
    * Gets the key encryption strategy class.
-   * 
+   *
    * @see CryptoModuleParameters#setKeyEncryptionStrategyClass(String)
    */
-  
+
   public String getKeyEncryptionStrategyClass() {
     return keyEncryptionStrategyClass;
   }
-  
+
   /**
    * Sets the class name of the key encryption strategy class. The class obeys the {@link SecretKeyEncryptionStrategy} interface. It instructs the
    * {@link DefaultCryptoModule} on how to encrypt the keys it uses to secure the streams.
    * <p>
-   * The default implementation of this interface, {@link CachingHDFSSecretKeyEncryptionStrategy}, creates a random key encryption key (KEK) as another symmetric
-   * key and places the KEK into HDFS. <i>This is not really very secure.</i> Users of the crypto modules are encouraged to either safeguard that KEK carefully
-   * or to obtain and use another {@link SecretKeyEncryptionStrategy} class.
+   * The default implementation of this interface, {@link CachingHDFSSecretKeyEncryptionStrategy}, creates a random key encryption key (KEK) as another
+   * symmetric key and places the KEK into HDFS. <i>This is not really very secure.</i> Users of the crypto modules are encouraged to either safeguard that KEK
+   * carefully or to obtain and use another {@link SecretKeyEncryptionStrategy} class.
    * <p>
    * For <b>encryption</b>, this value is <b>optional</b>. If it is not specified, then it assumed that the secret keys used for encrypting files will not be
    * encrypted. This is not a secure approach, thus setting this is highly recommended.<br>
    * For <b>decryption</b>, this value is often obtained from the underlying cipher stream. However, the underlying stream's value can be overridden (at least
    * when using {@link DefaultCryptoModule}) by setting the {@link CryptoModuleParameters#setOverrideStreamsSecretKeyEncryptionStrategy(boolean)} to true.
-   * 
+   *
    * @param keyEncryptionStrategyClass
    *          the name of the key encryption strategy class to use
    */
   public void setKeyEncryptionStrategyClass(String keyEncryptionStrategyClass) {
     this.keyEncryptionStrategyClass = keyEncryptionStrategyClass;
   }
-  
+
   /**
    * Gets the encrypted version of the plaintext key. This parameter is generally either obtained from an underlying stream or computed in the process of
    * employed the {@link CryptoModuleParameters#getKeyEncryptionStrategyClass()}.
-   * 
+   *
    * @see CryptoModuleParameters#setEncryptedKey(byte[])
    */
   public byte[] getEncryptedKey() {
     return encryptedKey;
   }
-  
+
   /**
    * Sets the encrypted version of the plaintext key ({@link CryptoModuleParameters#getPlaintextKey()}). Generally this operation will be done either by:
    * <p>
@@ -284,25 +284,25 @@ public class CryptoModuleParameters {
    * <p>
    * For <b>encryption</b>, this value is generally not required, but is usually set by the underlying module during encryption. <br>
    * For <b>decryption</b>, this value is <b>usually required</b>.
-   * 
-   * 
+   *
+   *
    * @param encryptedKey
    *          the encrypted value of the plaintext key
    */
-  
+
   public void setEncryptedKey(byte[] encryptedKey) {
     this.encryptedKey = encryptedKey;
   }
-  
+
   /**
    * Gets the opaque ID associated with the encrypted version of the plaintext key.
-   * 
+   *
    * @see CryptoModuleParameters#setOpaqueKeyEncryptionKeyID(String)
    */
   public String getOpaqueKeyEncryptionKeyID() {
     return opaqueKeyEncryptionKeyID;
   }
-  
+
   /**
    * Sets an opaque ID assocaited with the encrypted version of the plaintext key.
    * <p>
@@ -315,259 +315,259 @@ public class CryptoModuleParameters {
    * For <b>encryption</b>, this value is generally not required, but will be typically generated and set by the {@link SecretKeyEncryptionStrategy} class (see
    * {@link CryptoModuleParameters#getKeyEncryptionStrategyClass()}). <br>
    * For <b>decryption</b>, this value is <b>required</b>, though it will typically be read from the underlying stream.
-   * 
+   *
    * @param opaqueKeyEncryptionKeyID
    *          the opaque ID assoicated with the encrypted version of the plaintext key (see {@link CryptoModuleParameters#getEncryptedKey()}).
    */
-  
+
   public void setOpaqueKeyEncryptionKeyID(String opaqueKeyEncryptionKeyID) {
     this.opaqueKeyEncryptionKeyID = opaqueKeyEncryptionKeyID;
   }
-  
+
   /**
    * Gets the flag that indicates whether or not the module should record its cryptographic parameters to the stream automatically, or rely on the calling code
    * to do so.
-   * 
+   *
    * @see CryptoModuleParameters#setRecordParametersToStream(boolean)
    */
   public boolean getRecordParametersToStream() {
     return recordParametersToStream;
   }
-  
+
   /**
    * Gets the flag that indicates whether or not the module should record its cryptographic parameters to the stream automatically, or rely on the calling code
    * to do so.
-   * 
+   *
    * <p>
-   * 
+   *
    * If this is set to <i>true</i>, then the stream passed to {@link CryptoModule#getEncryptingOutputStream(CryptoModuleParameters)} will be <i>written to by
    * the module</i> before it is returned to the caller. There are situations where it is easier to let the crypto module do this writing on behalf of the
    * caller, and other times where it is not appropriate (if the format of the underlying stream must be carefully maintained, for instance).
-   * 
+   *
    * @param recordParametersToStream
    *          whether or not to require the module to record its parameters to the stream by itself
    */
   public void setRecordParametersToStream(boolean recordParametersToStream) {
     this.recordParametersToStream = recordParametersToStream;
   }
-  
+
   /**
    * Gets the flag that indicates whether or not to close the underlying stream when the cipher stream is closed.
-   * 
+   *
    * @see CryptoModuleParameters#setCloseUnderylingStreamAfterCryptoStreamClose(boolean)
    */
   public boolean getCloseUnderylingStreamAfterCryptoStreamClose() {
     return closeUnderylingStreamAfterCryptoStreamClose;
   }
-  
+
   /**
    * Sets the flag that indicates whether or not to close the underlying stream when the cipher stream is closed.
-   * 
+   *
    * <p>
-   * 
+   *
    * {@link CipherOutputStream} will only output its padding bytes when its {@link CipherOutputStream#close()} method is called. However, there are times when a
    * caller doesn't want its underlying stream closed at the time that the {@link CipherOutputStream} is closed. This flag indicates that the
    * {@link CryptoModule} should wrap the underlying stream in a basic {@link FilterOutputStream} which will swallow any close() calls and prevent them from
    * propogating to the underlying stream.
-   * 
+   *
    * @param closeUnderylingStreamAfterCryptoStreamClose
    *          the flag that indicates whether or not to close the underlying stream when the cipher stream is closed
    */
   public void setCloseUnderylingStreamAfterCryptoStreamClose(boolean closeUnderylingStreamAfterCryptoStreamClose) {
     this.closeUnderylingStreamAfterCryptoStreamClose = closeUnderylingStreamAfterCryptoStreamClose;
   }
-  
+
   /**
    * Gets the flag that indicates if the underlying stream's key encryption strategy should be overridden by the currently configured key encryption strategy.
-   * 
+   *
    * @see CryptoModuleParameters#setOverrideStreamsSecretKeyEncryptionStrategy(boolean)
    */
   public boolean getOverrideStreamsSecretKeyEncryptionStrategy() {
     return overrideStreamsSecretKeyEncryptionStrategy;
   }
-  
+
   /**
    * Sets the flag that indicates if the underlying stream's key encryption strategy should be overridden by the currently configured key encryption strategy.
-   * 
+   *
    * <p>
-   * 
+   *
    * So, why is this important? Say you started out with the default secret key encryption strategy. So, now you have a secret key in HDFS that encrypts all the
    * other secret keys. <i>Then</i> you deploy a key management solution. You want to move that secret key up to the key management server. Great! No problem.
    * Except, all your encrypted files now contain a setting that says
    * "hey I was encrypted by the default strategy, so find decrypt my key using that, not the key management server". This setting signals the
    * {@link CryptoModule} that it should ignore the setting in the file and prefer the one from the configuration.
-   * 
+   *
    * @param overrideStreamsSecretKeyEncryptionStrategy
    *          the flag that indicates if the underlying stream's key encryption strategy should be overridden by the currently configured key encryption
    *          strategy
    */
-  
+
   public void setOverrideStreamsSecretKeyEncryptionStrategy(boolean overrideStreamsSecretKeyEncryptionStrategy) {
     this.overrideStreamsSecretKeyEncryptionStrategy = overrideStreamsSecretKeyEncryptionStrategy;
   }
-  
+
   /**
    * Gets the plaintext output stream to wrap for encryption.
-   * 
+   *
    * @see CryptoModuleParameters#setPlaintextOutputStream(OutputStream)
    */
   public OutputStream getPlaintextOutputStream() {
     return plaintextOutputStream;
   }
-  
+
   /**
    * Sets the plaintext output stream to wrap for encryption.
-   * 
+   *
    * <p>
-   * 
+   *
    * For <b>encryption</b>, this parameter is <b>required</b>. <br>
    * For <b>decryption</b>, this parameter is ignored.
    */
   public void setPlaintextOutputStream(OutputStream plaintextOutputStream) {
     this.plaintextOutputStream = plaintextOutputStream;
   }
-  
+
   /**
    * Gets the encrypted output stream, which is nearly always a wrapped version of the output stream from
    * {@link CryptoModuleParameters#getPlaintextOutputStream()}.
-   * 
+   *
    * <p>
-   * 
+   *
    * Generally this method is used by {@link CryptoModule} classes as an <i>out</i> parameter from calling
    * {@link CryptoModule#getEncryptingOutputStream(CryptoModuleParameters)}.
-   * 
+   *
    * @see CryptoModuleParameters#setEncryptedOutputStream(OutputStream)
    */
-  
+
   public OutputStream getEncryptedOutputStream() {
     return encryptedOutputStream;
   }
-  
+
   /**
    * Sets the encrypted output stream. This method should really only be called by {@link CryptoModule} implementations unless something very unusual is going
    * on.
-   * 
+   *
    * @param encryptedOutputStream
    *          the encrypted version of the stream from output stream from {@link CryptoModuleParameters#getPlaintextOutputStream()}.
    */
   public void setEncryptedOutputStream(OutputStream encryptedOutputStream) {
     this.encryptedOutputStream = encryptedOutputStream;
   }
-  
+
   /**
    * Gets the plaintext input stream, which is nearly always a wrapped version of the output from {@link CryptoModuleParameters#getEncryptedInputStream()}.
-   * 
+   *
    * <p>
-   * 
+   *
    * Generally this method is used by {@link CryptoModule} classes as an <i>out</i> parameter from calling
    * {@link CryptoModule#getDecryptingInputStream(CryptoModuleParameters)}.
-   * 
-   * 
+   *
+   *
    * @see CryptoModuleParameters#setPlaintextInputStream(InputStream)
    */
   public InputStream getPlaintextInputStream() {
     return plaintextInputStream;
   }
-  
+
   /**
    * Sets the plaintext input stream, which is nearly always a wrapped version of the output from {@link CryptoModuleParameters#getEncryptedInputStream()}.
-   * 
+   *
    * <p>
-   * 
+   *
    * This method should really only be called by {@link CryptoModule} implementations.
    */
-  
+
   public void setPlaintextInputStream(InputStream plaintextInputStream) {
     this.plaintextInputStream = plaintextInputStream;
   }
-  
+
   /**
    * Gets the encrypted input stream to wrap for decryption.
-   * 
+   *
    * @see CryptoModuleParameters#setEncryptedInputStream(InputStream)
    */
   public InputStream getEncryptedInputStream() {
     return encryptedInputStream;
   }
-  
+
   /**
    * Sets the encrypted input stream to wrap for decryption.
    */
-  
+
   public void setEncryptedInputStream(InputStream encryptedInputStream) {
     this.encryptedInputStream = encryptedInputStream;
   }
-  
+
   /**
    * Gets the initialized cipher object.
-   * 
-   * 
+   *
+   *
    * @see CryptoModuleParameters#setCipher(Cipher)
    */
   public Cipher getCipher() {
     return cipher;
   }
-  
+
   /**
    * Sets the initialized cipher object. Generally speaking, callers do not have to create and set this object. There may be circumstances where the cipher
    * object is created outside of the module (to determine IV lengths, for one). If it is created and you want the module to use the cipher you already
    * initialized, set it here.
-   * 
+   *
    * @param cipher
    *          the cipher object
    */
   public void setCipher(Cipher cipher) {
     this.cipher = cipher;
   }
-  
+
   /**
    * Gets the initialized secure random object.
-   * 
+   *
    * @see CryptoModuleParameters#setSecureRandom(SecureRandom)
    */
   public SecureRandom getSecureRandom() {
     return secureRandom;
   }
-  
+
   /**
    * Sets the initialized secure random object. Generally speaking, callers do not have to create and set this object. There may be circumstances where the
    * random object is created outside of the module (for instance, to create a random secret key). If it is created outside the module and you want the module
    * to use the random object you already created, set it here.
-   * 
+   *
    * @param secureRandom
    *          the {@link SecureRandom} object
    */
-  
+
   public void setSecureRandom(SecureRandom secureRandom) {
     this.secureRandom = secureRandom;
   }
-  
+
   /**
    * Gets the initialization vector to use for this crypto module.
-   * 
+   *
    * @see CryptoModuleParameters#setInitializationVector(byte[])
    */
   public byte[] getInitializationVector() {
     return initializationVector;
   }
-  
+
   /**
    * Sets the initialization vector to use for this crypto module.
-   * 
+   *
    * <p>
-   * 
+   *
    * For <b>encryption</b>, this parameter is <i>optional</i>. If the initialization vector is created by the caller, for whatever reasons, it can be set here
    * and the crypto module will use it. <br>
-   * 
+   *
    * For <b>decryption</b>, this parameter is <b>required</b>. It should be read from the underlying stream that contains the encrypted data.
-   * 
+   *
    * @param initializationVector
    *          the initialization vector to use for this crypto operation.
    */
   public void setInitializationVector(byte[] initializationVector) {
     this.initializationVector = initializationVector;
   }
-  
+
   /**
    * Gets the size of the buffering stream that sits above the cipher stream
    */
@@ -584,29 +584,29 @@ public class CryptoModuleParameters {
 
   /**
    * Gets the overall set of options for the {@link CryptoModule}.
-   * 
+   *
    * @see CryptoModuleParameters#setAllOptions(Map)
    */
   public Map<String,String> getAllOptions() {
     return allOptions;
   }
-  
+
   /**
    * Sets the overall set of options for the {@link CryptoModule}.
-   * 
+   *
    * <p>
-   * 
+   *
    * Often, options for the cryptographic modules will be encoded as key/value pairs in a configuration file. This map represents those values. It may include
    * some of the parameters already called out as members of this class. It may contain any number of additional parameters which may be required by different
    * module or key encryption strategy implementations.
-   * 
+   *
    * @param allOptions
    *          the set of key/value pairs that confiure a module, based on a configuration file
    */
   public void setAllOptions(Map<String,String> allOptions) {
     this.allOptions = allOptions;
   }
-  
+
   private String algorithmName = null;
   private String encryptionMode = null;
   private String padding = null;
@@ -614,24 +614,24 @@ public class CryptoModuleParameters {
   private int keyLength = 0;
   private String randomNumberGenerator = null;
   private String randomNumberGeneratorProvider = null;
-  
+
   private String keyEncryptionStrategyClass;
   private byte[] encryptedKey;
   private String opaqueKeyEncryptionKeyID;
-  
+
   private boolean recordParametersToStream = true;
   private boolean closeUnderylingStreamAfterCryptoStreamClose = true;
   private boolean overrideStreamsSecretKeyEncryptionStrategy = false;
-  
+
   private OutputStream plaintextOutputStream;
   private OutputStream encryptedOutputStream;
   private InputStream plaintextInputStream;
   private InputStream encryptedInputStream;
-  
+
   private Cipher cipher;
   private SecureRandom secureRandom;
   private byte[] initializationVector;
-  
+
   private Map<String,String> allOptions;
   private int blockStreamSize;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModule.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModule.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModule.java
index 1e21535..bde5e64 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModule.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModule.java
@@ -38,39 +38,37 @@ import org.apache.accumulo.core.conf.Property;
 import org.apache.log4j.Logger;
 
 /**
- * This class implements the {@link CryptoModule} interface, defining how calling applications can receive encrypted 
- * input and output streams.  While the default implementation given here allows for a lot of flexibility in terms of 
- * choices of algorithm, key encryption strategies, and so on, some Accumulo users may choose to swap out this implementation
- * for others, and can base their implementation details off of this class's work.
- * 
- * In general, the module is quite straightforward: provide it with crypto-related settings and an input/output stream, and
- * it will hand back those streams wrapped in encrypting (or decrypting) streams.
- * 
+ * This class implements the {@link CryptoModule} interface, defining how calling applications can receive encrypted input and output streams. While the default
+ * implementation given here allows for a lot of flexibility in terms of choices of algorithm, key encryption strategies, and so on, some Accumulo users may
+ * choose to swap out this implementation for others, and can base their implementation details off of this class's work.
+ *
+ * In general, the module is quite straightforward: provide it with crypto-related settings and an input/output stream, and it will hand back those streams
+ * wrapped in encrypting (or decrypting) streams.
+ *
  */
 public class DefaultCryptoModule implements CryptoModule {
-  
+
   private static final String ENCRYPTION_HEADER_MARKER_V1 = "---Log File Encrypted (v1)---";
   private static final String ENCRYPTION_HEADER_MARKER_V2 = "---Log File Encrypted (v2)---";
   private static final Logger log = Logger.getLogger(DefaultCryptoModule.class);
-  
+
   public DefaultCryptoModule() {}
-  
-  
+
   @Override
   public CryptoModuleParameters initializeCipher(CryptoModuleParameters params) {
-    String cipherTransformation = getCipherTransformation(params); 
-    
+    String cipherTransformation = getCipherTransformation(params);
+
     log.trace(String.format("Using cipher suite \"%s\" with key length %d with RNG \"%s\" and RNG provider \"%s\" and key encryption strategy \"%s\"",
         cipherTransformation, params.getKeyLength(), params.getRandomNumberGenerator(), params.getRandomNumberGeneratorProvider(),
         params.getKeyEncryptionStrategyClass()));
-    
+
     if (params.getSecureRandom() == null) {
       SecureRandom secureRandom = DefaultCryptoModuleUtils.getSecureRandom(params.getRandomNumberGenerator(), params.getRandomNumberGeneratorProvider());
       params.setSecureRandom(secureRandom);
     }
-    
+
     Cipher cipher = DefaultCryptoModuleUtils.getCipher(cipherTransformation);
-    
+
     if (params.getInitializationVector() == null) {
       try {
         cipher.init(Cipher.ENCRYPT_MODE, new SecretKeySpec(params.getPlaintextKey(), params.getAlgorithmName()), params.getSecureRandom());
@@ -78,13 +76,13 @@ public class DefaultCryptoModule implements CryptoModule {
         log.error("Accumulo encountered an unknown error in generating the secret key object (SecretKeySpec) for an encrypted stream");
         throw new RuntimeException(e);
       }
-      
+
       params.setInitializationVector(cipher.getIV());
-      
-      
+
     } else {
       try {
-        cipher.init(Cipher.ENCRYPT_MODE, new SecretKeySpec(params.getPlaintextKey(), params.getAlgorithmName()), new IvParameterSpec(params.getInitializationVector()));
+        cipher.init(Cipher.ENCRYPT_MODE, new SecretKeySpec(params.getPlaintextKey(), params.getAlgorithmName()),
+            new IvParameterSpec(params.getInitializationVector()));
       } catch (InvalidKeyException e) {
         log.error("Accumulo encountered an unknown error in generating the secret key object (SecretKeySpec) for an encrypted stream");
         throw new RuntimeException(e);
@@ -95,27 +93,27 @@ public class DefaultCryptoModule implements CryptoModule {
     }
 
     params.setCipher(cipher);
-    
+
     return params;
-    
+
   }
 
   private String getCipherTransformation(CryptoModuleParameters params) {
     String cipherSuite = params.getAlgorithmName() + "/" + params.getEncryptionMode() + "/" + params.getPadding();
     return cipherSuite;
   }
-  
+
   private String[] parseCipherSuite(String cipherSuite) {
     return cipherSuite.split("/");
   }
-  
+
   private boolean validateNotEmpty(String givenValue, boolean allIsWell, StringBuffer buf, String errorMessage) {
     if (givenValue == null || givenValue.equals("")) {
       buf.append(errorMessage);
       buf.append("\n");
       return false;
     }
-    
+
     return true && allIsWell;
   }
 
@@ -125,120 +123,118 @@ public class DefaultCryptoModule implements CryptoModule {
       buf.append("\n");
       return false;
     }
-    
+
     return true && allIsWell;
   }
 
-  
   private boolean validateNotZero(int givenValue, boolean allIsWell, StringBuffer buf, String errorMessage) {
     if (givenValue == 0) {
       buf.append(errorMessage);
       buf.append("\n");
       return false;
     }
-    
+
     return true && allIsWell;
   }
 
   private boolean validateParamsObject(CryptoModuleParameters params, int cipherMode) {
-    
+
     if (cipherMode == Cipher.ENCRYPT_MODE) {
-      
-      StringBuffer errorBuf = new StringBuffer("The following problems were found with the CryptoModuleParameters object you provided for an encrypt operation:\n");
+
+      StringBuffer errorBuf = new StringBuffer(
+          "The following problems were found with the CryptoModuleParameters object you provided for an encrypt operation:\n");
       boolean allIsWell = true;
-      
+
       allIsWell = validateNotEmpty(params.getAlgorithmName(), allIsWell, errorBuf, "No algorithm name was specified.");
-      
+
       if (allIsWell && params.getAlgorithmName().equals("NullCipher")) {
         return true;
       }
-      
-      allIsWell = validateNotEmpty(params.getPadding(),                       allIsWell, errorBuf, "No padding was specified.");
-      allIsWell = validateNotZero (params.getKeyLength(),                     allIsWell, errorBuf, "No key length was specified.");
-      allIsWell = validateNotEmpty(params.getEncryptionMode(),                allIsWell, errorBuf, "No encryption mode was specified.");
-      allIsWell = validateNotEmpty(params.getRandomNumberGenerator(),         allIsWell, errorBuf, "No random number generator was specified.");
+
+      allIsWell = validateNotEmpty(params.getPadding(), allIsWell, errorBuf, "No padding was specified.");
+      allIsWell = validateNotZero(params.getKeyLength(), allIsWell, errorBuf, "No key length was specified.");
+      allIsWell = validateNotEmpty(params.getEncryptionMode(), allIsWell, errorBuf, "No encryption mode was specified.");
+      allIsWell = validateNotEmpty(params.getRandomNumberGenerator(), allIsWell, errorBuf, "No random number generator was specified.");
       allIsWell = validateNotEmpty(params.getRandomNumberGeneratorProvider(), allIsWell, errorBuf, "No random number generate provider was specified.");
-      allIsWell = validateNotNull (params.getPlaintextOutputStream(),         allIsWell, errorBuf, "No plaintext output stream was specified.");
+      allIsWell = validateNotNull(params.getPlaintextOutputStream(), allIsWell, errorBuf, "No plaintext output stream was specified.");
 
       if (!allIsWell) {
         log.error("CryptoModulesParameters object is not valid.");
         log.error(errorBuf.toString());
         throw new RuntimeException("CryptoModulesParameters object is not valid.");
       }
-      
+
       return allIsWell;
-      
+
     } else if (cipherMode == Cipher.DECRYPT_MODE) {
-      StringBuffer errorBuf = new StringBuffer("The following problems were found with the CryptoModuleParameters object you provided for a decrypt operation:\n");
+      StringBuffer errorBuf = new StringBuffer(
+          "The following problems were found with the CryptoModuleParameters object you provided for a decrypt operation:\n");
       boolean allIsWell = true;
 
-      allIsWell = validateNotEmpty(params.getPadding(),                       allIsWell, errorBuf, "No padding was specified.");
-      allIsWell = validateNotZero (params.getKeyLength(),                     allIsWell, errorBuf, "No key length was specified.");
-      allIsWell = validateNotEmpty(params.getEncryptionMode(),                allIsWell, errorBuf, "No encryption mode was specified.");
-      allIsWell = validateNotEmpty(params.getRandomNumberGenerator(),         allIsWell, errorBuf, "No random number generator was specified.");
+      allIsWell = validateNotEmpty(params.getPadding(), allIsWell, errorBuf, "No padding was specified.");
+      allIsWell = validateNotZero(params.getKeyLength(), allIsWell, errorBuf, "No key length was specified.");
+      allIsWell = validateNotEmpty(params.getEncryptionMode(), allIsWell, errorBuf, "No encryption mode was specified.");
+      allIsWell = validateNotEmpty(params.getRandomNumberGenerator(), allIsWell, errorBuf, "No random number generator was specified.");
       allIsWell = validateNotEmpty(params.getRandomNumberGeneratorProvider(), allIsWell, errorBuf, "No random number generate provider was specified.");
-      allIsWell = validateNotNull (params.getEncryptedInputStream(),          allIsWell, errorBuf, "No encrypted input stream was specified.");
-      allIsWell = validateNotNull (params.getInitializationVector(),          allIsWell, errorBuf, "No initialization vector was specified.");
-      allIsWell = validateNotNull (params.getEncryptedKey(),                  allIsWell, errorBuf, "No encrypted key was specified.");
-      
+      allIsWell = validateNotNull(params.getEncryptedInputStream(), allIsWell, errorBuf, "No encrypted input stream was specified.");
+      allIsWell = validateNotNull(params.getInitializationVector(), allIsWell, errorBuf, "No initialization vector was specified.");
+      allIsWell = validateNotNull(params.getEncryptedKey(), allIsWell, errorBuf, "No encrypted key was specified.");
+
       if (params.getKeyEncryptionStrategyClass() != null && !params.getKeyEncryptionStrategyClass().equals("NullSecretKeyEncryptionStrategy")) {
         allIsWell = validateNotEmpty(params.getOpaqueKeyEncryptionKeyID(), allIsWell, errorBuf, "No opqaue key encryption ID was specified.");
       }
-      
-      
+
       if (!allIsWell) {
         log.error("CryptoModulesParameters object is not valid.");
         log.error(errorBuf.toString());
         throw new RuntimeException("CryptoModulesParameters object is not valid.");
       }
-      
+
       return allIsWell;
-      
-    } 
-    
+
+    }
+
     return false;
   }
-  
-  
+
   @Override
   public CryptoModuleParameters getEncryptingOutputStream(CryptoModuleParameters params) throws IOException {
-    
+
     log.trace("Initializing crypto output stream (new style)");
-    
+
     boolean allParamsOK = validateParamsObject(params, Cipher.ENCRYPT_MODE);
     if (!allParamsOK) {
       // This would be weird because the above call should throw an exception, but if they don't we'll check and throw.
-      
+
       log.error("CryptoModuleParameters was not valid.");
       throw new RuntimeException("Invalid CryptoModuleParameters");
     }
-    
-    
+
     // If they want a null output stream, just return their plaintext stream as the encrypted stream
     if (params.getAlgorithmName().equals("NullCipher")) {
       params.setEncryptedOutputStream(params.getPlaintextOutputStream());
       return params;
     }
-    
+
     // Get the secret key
-    
+
     SecureRandom secureRandom = DefaultCryptoModuleUtils.getSecureRandom(params.getRandomNumberGenerator(), params.getRandomNumberGeneratorProvider());
-    
+
     if (params.getPlaintextKey() == null) {
       byte[] randomKey = new byte[params.getKeyLength() / 8];
       secureRandom.nextBytes(randomKey);
       params.setPlaintextKey(randomKey);
     }
-    
+
     // Encrypt the secret key
-    
+
     SecretKeyEncryptionStrategy keyEncryptionStrategy = CryptoModuleFactory.getSecretKeyEncryptionStrategy(params.getKeyEncryptionStrategyClass());
     params = keyEncryptionStrategy.encryptSecretKey(params);
-    
-    // Now the encrypted version of the key and any opaque ID are within the params object.  Initialize the cipher.
-    
+
+    // Now the encrypted version of the key and any opaque ID are within the params object. Initialize the cipher.
+
     // Check if the caller wants us to close the downstream stream when close() is called on the
-    // cipher object.  Calling close() on a CipherOutputStream is necessary for it to write out
+    // cipher object. Calling close() on a CipherOutputStream is necessary for it to write out
     // padding bytes.
     if (!params.getCloseUnderylingStreamAfterCryptoStreamClose()) {
       params.setPlaintextOutputStream(new DiscardCloseOutputStream(params.getPlaintextOutputStream()));
@@ -253,20 +249,19 @@ public class DefaultCryptoModule implements CryptoModule {
     if (0 == cipher.getBlockSize()) {
       throw new RuntimeException("Encryption cipher must be a block cipher");
     }
-    
+
     CipherOutputStream cipherOutputStream = new CipherOutputStream(params.getPlaintextOutputStream(), cipher);
     BlockedOutputStream blockedOutputStream = new BlockedOutputStream(cipherOutputStream, cipher.getBlockSize(), params.getBlockStreamSize());
 
     params.setEncryptedOutputStream(blockedOutputStream);
-    
+
     if (params.getRecordParametersToStream()) {
       DataOutputStream dataOut = new DataOutputStream(params.getPlaintextOutputStream());
-      
+
       // Write a marker to indicate this is an encrypted log file (in case we read it a plain one and need to
-      // not try to decrypt it. Can happen during a failure when the log's encryption settings are changing.      
+      // not try to decrypt it. Can happen during a failure when the log's encryption settings are changing.
       dataOut.writeUTF(ENCRYPTION_HEADER_MARKER_V2);
-      
-      
+
       // Write out all the parameters
       dataOut.writeInt(params.getAllOptions().size());
       for (String key : params.getAllOptions().keySet()) {
@@ -278,43 +273,43 @@ public class DefaultCryptoModule implements CryptoModule {
       // decode the old format.
       dataOut.writeUTF(getCipherTransformation(params));
       dataOut.writeUTF(params.getAlgorithmName());
-      
+
       // Write the init vector to the log file
       dataOut.writeInt(params.getInitializationVector().length);
       dataOut.write(params.getInitializationVector());
-      
+
       // Write out the encrypted session key and the opaque ID
       dataOut.writeUTF(params.getOpaqueKeyEncryptionKeyID());
       dataOut.writeInt(params.getEncryptedKey().length);
       dataOut.write(params.getEncryptedKey());
       dataOut.writeInt(params.getBlockStreamSize());
     }
-    
+
     return params;
   }
-  
+
   @Override
   public CryptoModuleParameters getDecryptingInputStream(CryptoModuleParameters params) throws IOException {
     log.trace("About to initialize decryption stream (new style)");
-        
+
     if (params.getRecordParametersToStream()) {
       DataInputStream dataIn = new DataInputStream(params.getEncryptedInputStream());
       log.trace("About to read encryption parameters from underlying stream");
-      
+
       String marker = dataIn.readUTF();
       if (marker.equals(ENCRYPTION_HEADER_MARKER_V1) || marker.equals(ENCRYPTION_HEADER_MARKER_V2)) {
-        
-        Map<String, String> paramsFromFile = new HashMap<String, String>();
-        
+
+        Map<String,String> paramsFromFile = new HashMap<String,String>();
+
         // Read in the bulk of parameters
         int paramsCount = dataIn.readInt();
         for (int i = 0; i < paramsCount; i++) {
           String key = dataIn.readUTF();
           String value = dataIn.readUTF();
-          
+
           paramsFromFile.put(key, value);
         }
-                
+
         // Set the cipher parameters
         String cipherSuiteFromFile = dataIn.readUTF();
         String algorithmNameFromFile = dataIn.readUTF();
@@ -322,25 +317,23 @@ public class DefaultCryptoModule implements CryptoModule {
         params.setAlgorithmName(algorithmNameFromFile);
         params.setEncryptionMode(cipherSuiteParts[1]);
         params.setPadding(cipherSuiteParts[2]);
-        
-        
+
         // Read the secret key and initialization vector from the file
         int initVectorLength = dataIn.readInt();
         byte[] initVector = new byte[initVectorLength];
         dataIn.readFully(initVector);
-        
+
         params.setInitializationVector(initVector);
-        
+
         // Read the opaque ID and encrypted session key
         String opaqueId = dataIn.readUTF();
         params.setOpaqueKeyEncryptionKeyID(opaqueId);
-        
+
         int encryptedSecretKeyLength = dataIn.readInt();
-        byte[] encryptedSecretKey = new byte[encryptedSecretKeyLength]; 
+        byte[] encryptedSecretKey = new byte[encryptedSecretKeyLength];
         dataIn.readFully(encryptedSecretKey);
         params.setEncryptedKey(encryptedSecretKey);
-        
-        
+
         if (params.getOverrideStreamsSecretKeyEncryptionStrategy()) {
           // Merge in options from file selectively
           for (String name : paramsFromFile.keySet()) {
@@ -352,62 +345,63 @@ public class DefaultCryptoModule implements CryptoModule {
         } else {
           params = CryptoModuleFactory.fillParamsObjectFromStringMap(params, paramsFromFile);
         }
-             
+
         SecretKeyEncryptionStrategy keyEncryptionStrategy = CryptoModuleFactory.getSecretKeyEncryptionStrategy(params.getKeyEncryptionStrategyClass());
-        
+
         params = keyEncryptionStrategy.decryptSecretKey(params);
-        
+
         if (marker.equals(ENCRYPTION_HEADER_MARKER_V2))
           params.setBlockStreamSize(dataIn.readInt());
         else
           params.setBlockStreamSize(0);
       } else {
-        
+
         log.trace("Read something off of the encrypted input stream that was not the encryption header marker, so pushing back bytes and returning the given stream");
         // Push these bytes back on to the stream. This method is a bit roundabout but isolates our code
         // from having to understand the format that DataOuputStream uses for its bytes.
         ByteArrayOutputStream tempByteOut = new ByteArrayOutputStream();
         DataOutputStream tempOut = new DataOutputStream(tempByteOut);
         tempOut.writeUTF(marker);
-        
+
         byte[] bytesToPutBack = tempByteOut.toByteArray();
-        
+
         PushbackInputStream pushbackStream = new PushbackInputStream(params.getEncryptedInputStream(), bytesToPutBack.length);
         pushbackStream.unread(bytesToPutBack);
-        
+
         params.setPlaintextInputStream(pushbackStream);
-        
+
         return params;
-      }      
+      }
     }
-    
+
     // We validate here after reading parameters from the stream, not at the top of the function.
     boolean allParamsOK = validateParamsObject(params, Cipher.DECRYPT_MODE);
-    
+
     if (!allParamsOK) {
       log.error("CryptoModuleParameters object failed validation for decrypt");
       throw new RuntimeException("CryptoModuleParameters object failed validation for decrypt");
     }
-    
+
     Cipher cipher = DefaultCryptoModuleUtils.getCipher(getCipherTransformation(params));
-    
+
     try {
-      cipher.init(Cipher.DECRYPT_MODE, new SecretKeySpec(params.getPlaintextKey(), params.getAlgorithmName()), new IvParameterSpec(params.getInitializationVector()));
+      cipher.init(Cipher.DECRYPT_MODE, new SecretKeySpec(params.getPlaintextKey(), params.getAlgorithmName()),
+          new IvParameterSpec(params.getInitializationVector()));
     } catch (InvalidKeyException e) {
       log.error("Error when trying to initialize cipher with secret key");
       throw new RuntimeException(e);
     } catch (InvalidAlgorithmParameterException e) {
       log.error("Error when trying to initialize cipher with initialization vector");
       throw new RuntimeException(e);
-    }   
-    
+    }
+
     InputStream blockedDecryptingInputStream = new CipherInputStream(params.getEncryptedInputStream(), cipher);
-    
+
     if (params.getBlockStreamSize() > 0)
       blockedDecryptingInputStream = new BlockedInputStream(blockedDecryptingInputStream, cipher.getBlockSize(), params.getBlockStreamSize());
 
-    log.trace("Initialized cipher input stream with transformation ["+getCipherTransformation(params)+"]");
-    
+    log.trace("Initialized cipher input stream with transformation [" + getCipherTransformation(params) + "]");
+
     params.setPlaintextInputStream(blockedDecryptingInputStream);
 
     return params;
@@ -423,8 +417,8 @@ public class DefaultCryptoModule implements CryptoModule {
 
     params.getSecureRandom().nextBytes(newSessionKey);
     params.setPlaintextKey(newSessionKey);
-    
+
     return params;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModuleUtils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModuleUtils.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModuleUtils.java
index 34ec1f3..21c9a09 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModuleUtils.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultCryptoModuleUtils.java
@@ -29,16 +29,16 @@ import org.apache.log4j.Logger;
 public class DefaultCryptoModuleUtils {
 
   private static final Logger log = Logger.getLogger(DefaultCryptoModuleUtils.class);
-  
+
   public static SecureRandom getSecureRandom(String secureRNG, String secureRNGProvider) {
     SecureRandom secureRandom = null;
     try {
       secureRandom = SecureRandom.getInstance(secureRNG, secureRNGProvider);
-      
+
       // Immediately seed the generator
       byte[] throwAway = new byte[16];
       secureRandom.nextBytes(throwAway);
-      
+
     } catch (NoSuchAlgorithmException e) {
       log.error(String.format("Accumulo configuration file specified a secure random generator \"%s\" that was not found by any provider.", secureRNG));
       throw new RuntimeException(e);
@@ -51,7 +51,7 @@ public class DefaultCryptoModuleUtils {
 
   public static Cipher getCipher(String cipherSuite) {
     Cipher cipher = null;
-    
+
     if (cipherSuite.equals("NullCipher")) {
       cipher = new NullCipher();
     } else {
@@ -67,5 +67,5 @@ public class DefaultCryptoModuleUtils {
     }
     return cipher;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/DiscardCloseOutputStream.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/DiscardCloseOutputStream.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/DiscardCloseOutputStream.java
index 846cf35..261e224 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/DiscardCloseOutputStream.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/DiscardCloseOutputStream.java
@@ -26,14 +26,14 @@ import org.apache.log4j.Logger;
 public class DiscardCloseOutputStream extends FilterOutputStream {
 
   private static final Logger log = Logger.getLogger(DiscardCloseOutputStream.class);
-  
+
   public DiscardCloseOutputStream(OutputStream out) {
     super(out);
   }
-  
+
   public void close() throws IOException {
     // Discard
     log.trace("Discarded close");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
index 67278bf..f42f9ff 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
@@ -37,7 +37,7 @@ import org.apache.log4j.Logger;
 
 //TODO ACCUMULO-2530 Update properties to use a URI instead of a relative path to secret key
 public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptionStrategy {
-  
+
   private static final Logger log = Logger.getLogger(NonCachingSecretKeyEncryptionStrategy.class);
 
   private void doKeyEncryptionOperation(int encryptionMode, CryptoModuleParameters params, String pathToKeyName, Path pathToKey, FileSystem fs)
@@ -45,10 +45,10 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
     DataInputStream in = null;
     try {
       if (!fs.exists(pathToKey)) {
-        
+
         if (encryptionMode == Cipher.UNWRAP_MODE) {
           log.error("There was a call to decrypt the session key but no key encryption key exists.  Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
-          throw new RuntimeException("Could not find key encryption key file in configured location in HDFS ("+pathToKeyName+")");
+          throw new RuntimeException("Could not find key encryption key file in configured location in HDFS (" + pathToKeyName + ")");
         } else {
           DataOutputStream out = null;
           try {
@@ -64,18 +64,18 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
             out.flush();
           } finally {
             if (out != null) {
-              out.close();        
+              out.close();
             }
           }
 
         }
       }
       in = fs.open(pathToKey);
-            
+
       int keyEncryptionKeyLength = in.readInt();
       byte[] keyEncryptionKey = new byte[keyEncryptionKeyLength];
       in.read(keyEncryptionKey);
-      
+
       Cipher cipher = DefaultCryptoModuleUtils.getCipher(params.getAllOptions().get(Property.CRYPTO_DEFAULT_KEY_STRATEGY_CIPHER_SUITE.getKey()));
 
       try {
@@ -83,8 +83,8 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
       } catch (InvalidKeyException e) {
         log.error(e);
         throw new RuntimeException(e);
-      }      
-      
+      }
+
       if (Cipher.UNWRAP_MODE == encryptionMode) {
         try {
           Key plaintextKey = cipher.unwrap(params.getEncryptedKey(), params.getAlgorithmName(), Cipher.SECRET_KEY);
@@ -109,9 +109,9 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
           log.error(e);
           throw new RuntimeException(e);
         }
-        
+
       }
-      
+
     } finally {
       if (in != null) {
         in.close();
@@ -123,20 +123,19 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
   private String getFullPathToKey(CryptoModuleParameters params) {
     String pathToKeyName = params.getAllOptions().get(Property.CRYPTO_DEFAULT_KEY_STRATEGY_KEY_LOCATION.getKey());
     String instanceDirectory = params.getAllOptions().get(Property.INSTANCE_DFS_DIR.getKey());
-    
-    
+
     if (pathToKeyName == null) {
       pathToKeyName = Property.CRYPTO_DEFAULT_KEY_STRATEGY_KEY_LOCATION.getDefaultValue();
     }
-    
+
     if (instanceDirectory == null) {
       instanceDirectory = Property.INSTANCE_DFS_DIR.getDefaultValue();
     }
-    
+
     if (!pathToKeyName.startsWith("/")) {
       pathToKeyName = "/" + pathToKeyName;
     }
-    
+
     String fullPath = instanceDirectory + pathToKeyName;
     return fullPath;
   }
@@ -148,20 +147,20 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
     if (hdfsURI == null) {
       hdfsURI = Property.INSTANCE_DFS_URI.getDefaultValue();
     }
-    
+
     String fullPath = getFullPathToKey(params);
     Path pathToKey = new Path(fullPath);
-    
+
     try {
       // TODO ACCUMULO-2530 Ensure volumes a properly supported
-      FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());   
+      FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
       doKeyEncryptionOperation(Cipher.WRAP_MODE, params, fullPath, pathToKey, fs);
-      
+
     } catch (IOException e) {
       log.error(e);
       throw new RuntimeException(e);
     }
-    
+
     return params;
   }
 
@@ -170,24 +169,23 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
   public CryptoModuleParameters decryptSecretKey(CryptoModuleParameters params) {
     String hdfsURI = params.getAllOptions().get(Property.INSTANCE_DFS_URI.getKey());
     if (hdfsURI == null) {
-      hdfsURI = Property.INSTANCE_DFS_URI.getDefaultValue(); 
+      hdfsURI = Property.INSTANCE_DFS_URI.getDefaultValue();
     }
-    
+
     String pathToKeyName = getFullPathToKey(params);
     Path pathToKey = new Path(pathToKeyName);
-    
+
     try {
       // TODO ACCUMULO-2530 Ensure volumes a properly supported
-      FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());   
+      FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
       doKeyEncryptionOperation(Cipher.UNWRAP_MODE, params, pathToKeyName, pathToKey, fs);
-      
-      
+
     } catch (IOException e) {
       log.error(e);
       throw new RuntimeException(e);
     }
-        
+
     return params;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/security/crypto/SecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/SecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/SecretKeyEncryptionStrategy.java
index 0cd9bd0..7d3c333 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/SecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/SecretKeyEncryptionStrategy.java
@@ -17,12 +17,12 @@
 package org.apache.accumulo.core.security.crypto;
 
 /**
- * 
+ *
  */
-public interface SecretKeyEncryptionStrategy {  
-  
+public interface SecretKeyEncryptionStrategy {
+
   CryptoModuleParameters encryptSecretKey(CryptoModuleParameters params);
+
   CryptoModuleParameters decryptSecretKey(CryptoModuleParameters params);
-  
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/tabletserver/log/LogEntry.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/tabletserver/log/LogEntry.java b/core/src/main/java/org/apache/accumulo/core/tabletserver/log/LogEntry.java
index 926e42e..25d0f32 100644
--- a/core/src/main/java/org/apache/accumulo/core/tabletserver/log/LogEntry.java
+++ b/core/src/main/java/org/apache/accumulo/core/tabletserver/log/LogEntry.java
@@ -38,7 +38,7 @@ public class LogEntry {
   public String filename;
   public int tabletId;
   public Collection<String> logSet;
-  
+
   public LogEntry() {}
 
   public LogEntry(LogEntry le) {
@@ -53,11 +53,11 @@ public class LogEntry {
   public String toString() {
     return extent.toString() + " " + filename + " (" + tabletId + ")";
   }
-  
+
   public String getName() {
     return server + "/" + filename;
   }
-  
+
   public byte[] toBytes() throws IOException {
     DataOutputBuffer out = new DataOutputBuffer();
     extent.write(out);
@@ -71,7 +71,7 @@ public class LogEntry {
     }
     return Arrays.copyOf(out.getData(), out.getLength());
   }
-  
+
   public void fromBytes(byte bytes[]) throws IOException {
     DataInputBuffer inp = new DataInputBuffer();
     inp.reset(bytes, bytes.length);
@@ -87,7 +87,7 @@ public class LogEntry {
       logSet.add(inp.readUTF());
     this.logSet = logSet;
   }
-  
+
   static private final Text EMPTY_TEXT = new Text();
 
   public static LogEntry fromKeyValue(Key key, Value value) {
@@ -102,19 +102,19 @@ public class LogEntry {
     result.timestamp = key.getTimestamp();
     return result;
   }
-  
+
   public Text getRow() {
     return extent.getMetadataEntry();
   }
-  
+
   public Text getColumnFamily() {
     return MetadataSchema.TabletsSection.LogColumnFamily.NAME;
   }
-  
+
   public Text getColumnQualifier() {
     return new Text(server + "/" + filename);
   }
-  
+
   public Value getValue() {
     return new Value((Joiner.on(";").join(logSet) + "|" + tabletId).getBytes());
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/DistributedTrace.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/DistributedTrace.java b/core/src/main/java/org/apache/accumulo/core/trace/DistributedTrace.java
index 5aa026b..ab81f30 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/DistributedTrace.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/DistributedTrace.java
@@ -35,7 +35,6 @@ import org.apache.zookeeper.KeeperException;
 import org.htrace.HTraceConfiguration;
 import org.htrace.SpanReceiver;
 
-
 /**
  * Utility class to enable tracing for Accumulo server processes.
  *
@@ -70,27 +69,24 @@ public class DistributedTrace {
   }
 
   /**
-   * Enable tracing by setting up SpanReceivers for the current process.
-   * If service name is null, the simple name of the class will be used.
+   * Enable tracing by setting up SpanReceivers for the current process. If service name is null, the simple name of the class will be used.
    */
   public static void enable(String service) {
     enable(null, service);
   }
 
   /**
-   * Enable tracing by setting up SpanReceivers for the current process.
-   * If host name is null, it will be determined.
-   * If service name is null, the simple name of the class will be used.
+   * Enable tracing by setting up SpanReceivers for the current process. If host name is null, it will be determined. If service name is null, the simple name
+   * of the class will be used.
    */
   public static void enable(String hostname, String service) {
     enable(hostname, service, ClientConfiguration.loadDefault());
   }
 
   /**
-   * Enable tracing by setting up SpanReceivers for the current process.
-   * If host name is null, it will be determined.
-   * If service name is null, the simple name of the class will be used.
-   * Properties required in the client configuration include {@link org.apache.accumulo.core.client.ClientConfiguration.ClientProperty#TRACE_SPAN_RECEIVERS} and any properties specific to the span receiver.
+   * Enable tracing by setting up SpanReceivers for the current process. If host name is null, it will be determined. If service name is null, the simple name
+   * of the class will be used. Properties required in the client configuration include
+   * {@link org.apache.accumulo.core.client.ClientConfiguration.ClientProperty#TRACE_SPAN_RECEIVERS} and any properties specific to the span receiver.
    */
   public static void enable(String hostname, String service, ClientConfiguration conf) {
     String spanReceivers = conf.get(ClientProperty.TRACE_SPAN_RECEIVERS);
@@ -102,9 +98,8 @@ public class DistributedTrace {
   }
 
   /**
-   * Enable tracing by setting up SpanReceivers for the current process.
-   * If host name is null, it will be determined.
-   * If service name is null, the simple name of the class will be used.
+   * Enable tracing by setting up SpanReceivers for the current process. If host name is null, it will be determined. If service name is null, the simple name
+   * of the class will be used.
    */
   public static void enable(String hostname, String service, AccumuloConfiguration conf) {
     String spanReceivers = conf.get(Property.TRACE_SPAN_RECEIVERS);
@@ -180,7 +175,7 @@ public class DistributedTrace {
     SpanReceiver impl;
     try {
       Object o = ReflectionUtils.newInstance(implClass, conf);
-      impl = (SpanReceiver)o;
+      impl = (SpanReceiver) o;
       impl.configure(wrapHadoopConf(conf));
     } catch (SecurityException e) {
       throw new IOException(e);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/Span.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/Span.java b/core/src/main/java/org/apache/accumulo/core/trace/Span.java
index 3427a2e..b72e6ee 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/Span.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/Span.java
@@ -18,16 +18,15 @@ package org.apache.accumulo.core.trace;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
 
+import java.util.List;
+import java.util.Map;
+
 import org.htrace.NullScope;
 import org.htrace.TimelineAnnotation;
 import org.htrace.TraceScope;
 
-import java.util.List;
-import java.util.Map;
-
 /**
- * This is a wrapper for a TraceScope object, which is a wrapper for a Span and its parent.
- * Not recommended for client use.
+ * This is a wrapper for a TraceScope object, which is a wrapper for a Span and its parent. Not recommended for client use.
  */
 public class Span implements org.htrace.Span {
   public static final Span NULL_SPAN = new Span(NullScope.INSTANCE);
@@ -127,7 +126,7 @@ public class Span implements org.htrace.Span {
   }
 
   @Override
-  public Map<byte[], byte[]> getKVAnnotations() {
+  public Map<byte[],byte[]> getKVAnnotations() {
     return span.getKVAnnotations();
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/Trace.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/Trace.java b/core/src/main/java/org/apache/accumulo/core/trace/Trace.java
index fc2df17..eba19ab 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/Trace.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/Trace.java
@@ -16,16 +16,16 @@
  */
 package org.apache.accumulo.core.trace;
 
+import static java.nio.charset.StandardCharsets.UTF_8;
+
 import org.apache.accumulo.core.trace.thrift.TInfo;
 import org.apache.accumulo.core.trace.wrappers.TraceRunnable;
 import org.htrace.Sampler;
 import org.htrace.TraceInfo;
 import org.htrace.wrappers.TraceProxy;
 
-import static java.nio.charset.StandardCharsets.UTF_8;
-
 /**
- * Utility class for tracing within Accumulo.  Not intended for client use!
+ * Utility class for tracing within Accumulo. Not intended for client use!
  *
  */
 public class Trace {
@@ -71,6 +71,7 @@ public class Trace {
 
   /**
    * Return the current span.
+   *
    * @deprecated since 1.7 -- it is better to save the span you create in a local variable and call its methods, rather than retrieving the current span
    */
   @Deprecated
@@ -129,7 +130,7 @@ public class Trace {
   }
 
   // Sample trace all calls to the given object
-  public static <T, V> T wrapAll(T instance, Sampler<V> dist) {
+  public static <T,V> T wrapAll(T instance, Sampler<V> dist) {
     return TraceProxy.trace(instance, dist);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcClientInvocationHandler.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcClientInvocationHandler.java b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcClientInvocationHandler.java
index 8b0156d..9871d90 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcClientInvocationHandler.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcClientInvocationHandler.java
@@ -16,15 +16,15 @@
  */
 package org.apache.accumulo.core.trace.wrappers;
 
+import java.lang.reflect.InvocationHandler;
+import java.lang.reflect.InvocationTargetException;
+import java.lang.reflect.Method;
+
 import org.apache.accumulo.core.trace.Span;
 import org.apache.accumulo.core.trace.Trace;
 import org.apache.accumulo.core.trace.Tracer;
 import org.apache.accumulo.core.trace.thrift.TInfo;
 
-import java.lang.reflect.InvocationHandler;
-import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.Method;
-
 public class RpcClientInvocationHandler<I> implements InvocationHandler {
 
   private final I instance;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcServerInvocationHandler.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcServerInvocationHandler.java b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcServerInvocationHandler.java
index 700e543..b7c578d 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcServerInvocationHandler.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/RpcServerInvocationHandler.java
@@ -16,14 +16,14 @@
  */
 package org.apache.accumulo.core.trace.wrappers;
 
-import org.apache.accumulo.core.trace.Span;
-import org.apache.accumulo.core.trace.Trace;
-import org.apache.accumulo.core.trace.thrift.TInfo;
-
 import java.lang.reflect.InvocationHandler;
 import java.lang.reflect.InvocationTargetException;
 import java.lang.reflect.Method;
 
+import org.apache.accumulo.core.trace.Span;
+import org.apache.accumulo.core.trace.Trace;
+import org.apache.accumulo.core.trace.thrift.TInfo;
+
 public class RpcServerInvocationHandler<I> implements InvocationHandler {
 
   private final I instance;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceCallable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceCallable.java b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceCallable.java
index 0d0dc56..d8b9b78 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceCallable.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceCallable.java
@@ -16,25 +16,25 @@
  */
 package org.apache.accumulo.core.trace.wrappers;
 
+import java.util.concurrent.Callable;
+
 import org.htrace.Span;
 import org.htrace.Trace;
 import org.htrace.TraceScope;
 
-import java.util.concurrent.Callable;
-
 /**
  * Wrap a Callable with a Span that survives a change in threads.
- * 
+ *
  */
 public class TraceCallable<V> implements Callable<V> {
   private final Callable<V> impl;
   private final Span parent;
   private final String description;
-  
+
   TraceCallable(Callable<V> impl) {
     this(Trace.currentSpan(), impl);
   }
-  
+
   TraceCallable(Span parent, Callable<V> impl) {
     this(parent, impl, null);
   }
@@ -44,7 +44,7 @@ public class TraceCallable<V> implements Callable<V> {
     this.parent = parent;
     this.description = description;
   }
-  
+
   @Override
   public V call() throws Exception {
     if (parent != null) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceExecutorService.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceExecutorService.java b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceExecutorService.java
index 5ba61bf..0ff5751 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceExecutorService.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceExecutorService.java
@@ -16,9 +16,6 @@
  */
 package org.apache.accumulo.core.trace.wrappers;
 
-import org.htrace.Span;
-import org.htrace.Tracer;
-
 import java.util.ArrayList;
 import java.util.Collection;
 import java.util.List;
@@ -29,59 +26,62 @@ import java.util.concurrent.Future;
 import java.util.concurrent.TimeUnit;
 import java.util.concurrent.TimeoutException;
 
+import org.htrace.Span;
+import org.htrace.Tracer;
+
 public class TraceExecutorService implements ExecutorService {
-  
+
   private final ExecutorService impl;
-  
+
   public TraceExecutorService(ExecutorService impl) {
     this.impl = impl;
   }
-  
+
   @Override
   public void execute(Runnable command) {
     impl.execute(new TraceRunnable(command));
   }
-  
+
   @Override
   public void shutdown() {
     impl.shutdown();
   }
-  
+
   @Override
   public List<Runnable> shutdownNow() {
     return impl.shutdownNow();
   }
-  
+
   @Override
   public boolean isShutdown() {
     return impl.isShutdown();
   }
-  
+
   @Override
   public boolean isTerminated() {
     return impl.isTerminated();
   }
-  
+
   @Override
   public boolean awaitTermination(long timeout, TimeUnit unit) throws InterruptedException {
     return impl.awaitTermination(timeout, unit);
   }
-  
+
   @Override
   public <T> Future<T> submit(Callable<T> task) {
     return impl.submit(new TraceCallable<T>(task));
   }
-  
+
   @Override
   public <T> Future<T> submit(Runnable task, T result) {
     return impl.submit(new TraceRunnable(task), result);
   }
-  
+
   @Override
   public Future<?> submit(Runnable task) {
     return impl.submit(new TraceRunnable(task));
   }
-  
+
   private <T> Collection<? extends Callable<T>> wrapCollection(Collection<? extends Callable<T>> tasks) {
     List<Callable<T>> result = new ArrayList<Callable<T>>();
     for (Callable<T> task : tasks) {
@@ -89,28 +89,28 @@ public class TraceExecutorService implements ExecutorService {
     }
     return result;
   }
-  
+
   @Override
   public <T> List<Future<T>> invokeAll(Collection<? extends Callable<T>> tasks) throws InterruptedException {
     return impl.invokeAll(wrapCollection(tasks));
   }
-  
+
   @Override
   public <T> List<Future<T>> invokeAll(Collection<? extends Callable<T>> tasks, long timeout, TimeUnit unit) throws InterruptedException {
     return impl.invokeAll(wrapCollection(tasks), timeout, unit);
   }
-  
+
   @Override
   public <T> T invokeAny(Collection<? extends Callable<T>> tasks) throws InterruptedException, ExecutionException {
     return impl.invokeAny(wrapCollection(tasks));
   }
-  
+
   @Override
   public <T> T invokeAny(Collection<? extends Callable<T>> tasks, long timeout, TimeUnit unit) throws InterruptedException, ExecutionException,
       TimeoutException {
     return impl.invokeAny(wrapCollection(tasks), timeout, unit);
   }
-  
+
   /**
    * Finish a given trace and set the span for the current thread to null.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceRunnable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceRunnable.java b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceRunnable.java
index 5723b17..a870cbf 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceRunnable.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceRunnable.java
@@ -22,18 +22,18 @@ import org.htrace.TraceScope;
 
 /**
  * Wrap a Runnable with a Span that survives a change in threads.
- * 
+ *
  */
 public class TraceRunnable implements Runnable, Comparable<TraceRunnable> {
-  
+
   private final Span parent;
   private final Runnable runnable;
   private final String description;
-  
+
   public TraceRunnable(Runnable runnable) {
     this(Trace.currentSpan(), runnable);
   }
-  
+
   public TraceRunnable(Span parent, Runnable runnable) {
     this(parent, runnable, null);
   }
@@ -43,7 +43,7 @@ public class TraceRunnable implements Runnable, Comparable<TraceRunnable> {
     this.runnable = runnable;
     this.description = description;
   }
-  
+
   @Override
   public void run() {
     if (parent != null) {
@@ -61,16 +61,16 @@ public class TraceRunnable implements Runnable, Comparable<TraceRunnable> {
   private String getDescription() {
     return this.description == null ? Thread.currentThread().getName() : description;
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof TraceRunnable) {
       return 0 == this.compareTo((TraceRunnable) o);
     }
-    
+
     return false;
   }
-  
+
   @SuppressWarnings({"rawtypes", "unchecked"})
   @Override
   public int compareTo(TraceRunnable o) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceWrap.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceWrap.java b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceWrap.java
index f675c53..bbfa386 100644
--- a/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceWrap.java
+++ b/core/src/main/java/org/apache/accumulo/core/trace/wrappers/TraceWrap.java
@@ -22,7 +22,7 @@ import java.lang.reflect.Proxy;
 /**
  * To move trace data from client to server, the RPC call must be annotated to take a TInfo object as its first argument. The user can simply pass null, so long
  * as they wrap their Client and Service objects with these functions.
- * 
+ *
  * <pre>
  * Trace.on(&quot;remoteMethod&quot;);
  * Iface c = new Client();
@@ -30,11 +30,11 @@ import java.lang.reflect.Proxy;
  * c.remoteMethod(null, arg2, arg3);
  * Trace.off();
  * </pre>
- * 
+ *
  * The wrapper will see the annotated method and send or re-establish the trace information.
- * 
+ *
  * Note that the result of these calls is a Proxy object that conforms to the basic interfaces, but is not your concrete instance.
- * 
+ *
  */
 public class TraceWrap {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/AddressUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/AddressUtil.java b/core/src/main/java/org/apache/accumulo/core/util/AddressUtil.java
index bd9a5ca..1426239 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/AddressUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/AddressUtil.java
@@ -20,7 +20,6 @@ import com.google.common.net.HostAndPort;
 
 public class AddressUtil extends org.apache.accumulo.fate.util.AddressUtil {
 
-
   static public HostAndPort parseAddress(String address) throws NumberFormatException {
     return parseAddress(address, false);
   }
@@ -30,7 +29,7 @@ public class AddressUtil extends org.apache.accumulo.fate.util.AddressUtil {
     HostAndPort hap = HostAndPort.fromString(address);
     if (!ignoreMissingPort && !hap.hasPort())
       throw new IllegalArgumentException("Address was expected to contain port. address=" + address);
-    
+
     return hap;
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/BadArgumentException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/BadArgumentException.java b/core/src/main/java/org/apache/accumulo/core/util/BadArgumentException.java
index c2f79ba..80c9beb 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/BadArgumentException.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/BadArgumentException.java
@@ -20,7 +20,7 @@ import java.util.regex.PatternSyntaxException;
 
 public final class BadArgumentException extends PatternSyntaxException {
   private static final long serialVersionUID = 1L;
-  
+
   public BadArgumentException(String desc, String badarg, int index) {
     super(desc, badarg, index);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/Base64.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/Base64.java b/core/src/main/java/org/apache/accumulo/core/util/Base64.java
index 76de4ed..ce54aae 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/Base64.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/Base64.java
@@ -19,17 +19,15 @@ package org.apache.accumulo.core.util;
 import org.apache.commons.codec.binary.StringUtils;
 
 /**
- * A wrapper around commons-codec's Base64 to make sure we get the non-chunked behavior that
- * became the default in commons-codec version 1.5+ while relying on the commons-codec version 1.4
- * that Hadoop Client provides.
+ * A wrapper around commons-codec's Base64 to make sure we get the non-chunked behavior that became the default in commons-codec version 1.5+ while relying on
+ * the commons-codec version 1.4 that Hadoop Client provides.
  */
 public final class Base64 {
 
   /**
    * Private to prevent instantiation.
    */
-  private Base64() {
-  }
+  private Base64() {}
 
   /**
    * Serialize to Base64 byte array, non-chunked.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/ByteArrayComparator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/ByteArrayComparator.java b/core/src/main/java/org/apache/accumulo/core/util/ByteArrayComparator.java
index 461ef38..1d7372c 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/ByteArrayComparator.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/ByteArrayComparator.java
@@ -21,21 +21,21 @@ import java.util.Comparator;
 
 public class ByteArrayComparator implements Comparator<byte[]>, Serializable {
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   public int compare(byte[] o1, byte[] o2) {
-    
+
     int minLen = Math.min(o1.length, o2.length);
-    
+
     for (int i = 0; i < minLen; i++) {
       int a = (o1[i] & 0xff);
       int b = (o2[i] & 0xff);
-      
+
       if (a != b) {
         return a - b;
       }
     }
-    
+
     return o1.length - o2.length;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/ByteArraySet.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/ByteArraySet.java b/core/src/main/java/org/apache/accumulo/core/util/ByteArraySet.java
index 4b7db82..ca43469 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/ByteArraySet.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/ByteArraySet.java
@@ -25,31 +25,31 @@ import java.util.List;
 import java.util.TreeSet;
 
 public class ByteArraySet extends TreeSet<byte[]> {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   public ByteArraySet() {
     super(new ByteArrayComparator());
   }
-  
+
   public ByteArraySet(Collection<? extends byte[]> c) {
     this();
     addAll(c);
   }
-  
+
   public static ByteArraySet fromStrings(Collection<String> c) {
     List<byte[]> lst = new ArrayList<byte[]>();
     for (String s : c)
       lst.add(s.getBytes(UTF_8));
     return new ByteArraySet(lst);
   }
-  
+
   public static ByteArraySet fromStrings(String... c) {
     return ByteArraySet.fromStrings(Arrays.asList(c));
   }
-  
+
   public List<byte[]> toList() {
     return new ArrayList<byte[]>(this);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/ByteBufferUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/ByteBufferUtil.java b/core/src/main/java/org/apache/accumulo/core/util/ByteBufferUtil.java
index dbf04dc..f849d5b 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/ByteBufferUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/ByteBufferUtil.java
@@ -33,7 +33,7 @@ public class ByteBufferUtil {
       return null;
     return Arrays.copyOfRange(buffer.array(), buffer.position(), buffer.limit());
   }
-  
+
   public static List<ByteBuffer> toByteBuffers(Collection<byte[]> bytesList) {
     if (bytesList == null)
       return null;
@@ -43,7 +43,7 @@ public class ByteBufferUtil {
     }
     return result;
   }
-  
+
   public static List<byte[]> toBytesList(Collection<ByteBuffer> bytesList) {
     if (bytesList == null)
       return null;
@@ -53,7 +53,7 @@ public class ByteBufferUtil {
     }
     return result;
   }
-  
+
   public static Text toText(ByteBuffer bytes) {
     if (bytes == null)
       return null;
@@ -61,11 +61,11 @@ public class ByteBufferUtil {
     result.set(bytes.array(), bytes.position(), bytes.remaining());
     return result;
   }
-  
+
   public static String toString(ByteBuffer bytes) {
     return new String(bytes.array(), bytes.position(), bytes.remaining(), UTF_8);
   }
-  
+
   public static ByteBuffer toByteBuffers(ByteSequence bs) {
     if (bs == null)
       return null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/CachedConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/CachedConfiguration.java b/core/src/main/java/org/apache/accumulo/core/util/CachedConfiguration.java
index daf165c..00cc063 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/CachedConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/CachedConfiguration.java
@@ -20,13 +20,13 @@ import org.apache.hadoop.conf.Configuration;
 
 public class CachedConfiguration {
   private static Configuration configuration = null;
-  
+
   public synchronized static Configuration getInstance() {
     if (configuration == null)
       setInstance(new Configuration());
     return configuration;
   }
-  
+
   public synchronized static Configuration setInstance(Configuration update) {
     Configuration result = configuration;
     configuration = update;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/CleanUp.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/CleanUp.java b/core/src/main/java/org/apache/accumulo/core/util/CleanUp.java
index be5a41a..a632fe3 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/CleanUp.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/CleanUp.java
@@ -23,12 +23,12 @@ import org.apache.accumulo.fate.zookeeper.ZooSession;
 import org.apache.log4j.Logger;
 
 /**
- * 
+ *
  */
 public class CleanUp {
-  
+
   private static final Logger log = Logger.getLogger(CleanUp.class);
-  
+
   /**
    * kills all threads created by internal Accumulo singleton resources. After this method is called, no accumulo client will work in the current classloader.
    */
@@ -37,17 +37,17 @@ public class CleanUp {
     ZooSession.shutdown();
     waitForZooKeeperClientThreads();
   }
-  
+
   /**
-   * As documented in https://issues.apache.org/jira/browse/ZOOKEEPER-1816, ZooKeeper.close()
-   * is a non-blocking call. This method will wait on the ZooKeeper internal threads to exit.
+   * As documented in https://issues.apache.org/jira/browse/ZOOKEEPER-1816, ZooKeeper.close() is a non-blocking call. This method will wait on the ZooKeeper
+   * internal threads to exit.
    */
   private static void waitForZooKeeperClientThreads() {
     Set<Thread> threadSet = Thread.getAllStackTraces().keySet();
-    for (Thread thread : threadSet) {    
+    for (Thread thread : threadSet) {
       // find ZooKeeper threads that were created in the same ClassLoader as the current thread.
-      if (thread.getClass().getName().startsWith("org.apache.zookeeper.ClientCnxn") &&
-          thread.getContextClassLoader().equals(Thread.currentThread().getContextClassLoader())) {
+      if (thread.getClass().getName().startsWith("org.apache.zookeeper.ClientCnxn")
+          && thread.getContextClassLoader().equals(Thread.currentThread().getContextClassLoader())) {
 
         // wait for the thread the die
         while (thread.isAlive()) {


[41/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/ComparableBytes.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/ComparableBytes.java b/core/src/main/java/org/apache/accumulo/core/data/ComparableBytes.java
index 6bdc7e6..67e4bb1 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/ComparableBytes.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/ComparableBytes.java
@@ -19,24 +19,22 @@ package org.apache.accumulo.core.data;
 import org.apache.hadoop.io.BinaryComparable;
 
 /**
- * An array of bytes wrapped so as to extend Hadoop's
- * <code>BinaryComparable</code> class.
+ * An array of bytes wrapped so as to extend Hadoop's <code>BinaryComparable</code> class.
  */
 public class ComparableBytes extends BinaryComparable {
-  
+
   public byte[] data;
-  
+
   /**
-   * Creates a new byte wrapper. The given byte array is used directly as a
-   * backing array, so later changes made to the array reflect into the new
-   * object.
+   * Creates a new byte wrapper. The given byte array is used directly as a backing array, so later changes made to the array reflect into the new object.
    *
-   * @param b bytes to wrap
+   * @param b
+   *          bytes to wrap
    */
   public ComparableBytes(byte[] b) {
     this.data = b;
   }
-  
+
   /**
    * Gets the wrapped bytes in this object.
    *
@@ -45,10 +43,10 @@ public class ComparableBytes extends BinaryComparable {
   public byte[] getBytes() {
     return data;
   }
-  
+
   @Override
   public int getLength() {
     return data.length;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/Condition.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Condition.java b/core/src/main/java/org/apache/accumulo/core/data/Condition.java
index e4b089d..2dc2a0f 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Condition.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Condition.java
@@ -28,11 +28,11 @@ import org.apache.hadoop.io.Text;
 
 /**
  * Conditions that must be met on a particular column in a row.
- * 
+ *
  * @since 1.6.0
  */
 public class Condition {
-  
+
   private ByteSequence cf;
   private ByteSequence cq;
   private ByteSequence cv;
@@ -40,16 +40,17 @@ public class Condition {
   private Long ts;
   private IteratorSetting iterators[] = new IteratorSetting[0];
   private static final ByteSequence EMPTY = new ArrayByteSequence(new byte[0]);
-  
 
   /**
-   * Creates a new condition. The initial column value and timestamp are null,
-   * and the initial column visibility is empty. Characters in the column family
-   * and column qualifier are encoded as bytes in the condition using UTF-8.
+   * Creates a new condition. The initial column value and timestamp are null, and the initial column visibility is empty. Characters in the column family and
+   * column qualifier are encoded as bytes in the condition using UTF-8.
    *
-   * @param cf column family
-   * @param cq column qualifier
-   * @throws IllegalArgumentException if any argument is null
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @throws IllegalArgumentException
+   *           if any argument is null
    */
   public Condition(CharSequence cf, CharSequence cq) {
     checkArgument(cf != null, "cf is null");
@@ -58,14 +59,16 @@ public class Condition {
     this.cq = new ArrayByteSequence(cq.toString().getBytes(UTF_8));
     this.cv = EMPTY;
   }
-  
+
   /**
-   * Creates a new condition. The initial column value and timestamp are null,
-   * and the initial column visibility is empty.
+   * Creates a new condition. The initial column value and timestamp are null, and the initial column visibility is empty.
    *
-   * @param cf column family
-   * @param cq column qualifier
-   * @throws IllegalArgumentException if any argument is null
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @throws IllegalArgumentException
+   *           if any argument is null
    */
   public Condition(byte[] cf, byte[] cq) {
     checkArgument(cf != null, "cf is null");
@@ -76,12 +79,14 @@ public class Condition {
   }
 
   /**
-   * Creates a new condition. The initial column value and timestamp are null,
-   * and the initial column visibility is empty.
+   * Creates a new condition. The initial column value and timestamp are null, and the initial column visibility is empty.
    *
-   * @param cf column family
-   * @param cq column qualifier
-   * @throws IllegalArgumentException if any argument is null
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @throws IllegalArgumentException
+   *           if any argument is null
    */
   public Condition(Text cf, Text cq) {
     checkArgument(cf != null, "cf is null");
@@ -92,12 +97,14 @@ public class Condition {
   }
 
   /**
-   * Creates a new condition. The initial column value and timestamp are null,
-   * and the initial column visibility is empty.
+   * Creates a new condition. The initial column value and timestamp are null, and the initial column visibility is empty.
    *
-   * @param cf column family
-   * @param cq column qualifier
-   * @throws IllegalArgumentException if any argument is null
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @throws IllegalArgumentException
+   *           if any argument is null
    */
   public Condition(ByteSequence cf, ByteSequence cq) {
     checkArgument(cf != null, "cf is null");
@@ -115,7 +122,7 @@ public class Condition {
   public ByteSequence getFamily() {
     return cf;
   }
-  
+
   /**
    * Gets the column qualifier of this condition.
    *
@@ -128,14 +135,15 @@ public class Condition {
   /**
    * Sets the version for the column to check. If this is not set then the latest column will be checked, unless iterators do something different.
    *
-   * @param ts timestamp
+   * @param ts
+   *          timestamp
    * @return this condition
    */
   public Condition setTimestamp(long ts) {
     this.ts = ts;
     return this;
   }
-  
+
   /**
    * Gets the timestamp of this condition.
    *
@@ -147,12 +155,13 @@ public class Condition {
 
   /**
    * This method sets the expected value of a column. In order for the condition to pass the column must exist and have this value. If a value is not set, then
-   * the column must be absent for the condition to pass. The passed-in character sequence is encoded as UTF-8.
-   * See {@link #setValue(byte[])}.
-   * 
-   * @param value value
+   * the column must be absent for the condition to pass. The passed-in character sequence is encoded as UTF-8. See {@link #setValue(byte[])}.
+   *
+   * @param value
+   *          value
    * @return this condition
-   * @throws IllegalArgumentException if value is null
+   * @throws IllegalArgumentException
+   *           if value is null
    */
   public Condition setValue(CharSequence value) {
     checkArgument(value != null, "value is null");
@@ -163,40 +172,44 @@ public class Condition {
   /**
    * This method sets the expected value of a column. In order for the condition to pass the column must exist and have this value. If a value is not set, then
    * the column must be absent for the condition to pass.
-   * 
-   * @param value value
+   *
+   * @param value
+   *          value
    * @return this condition
-   * @throws IllegalArgumentException if value is null
+   * @throws IllegalArgumentException
+   *           if value is null
    */
   public Condition setValue(byte[] value) {
     checkArgument(value != null, "value is null");
     this.val = new ArrayByteSequence(value);
     return this;
   }
-  
+
   /**
    * This method sets the expected value of a column. In order for the condition to pass the column must exist and have this value. If a value is not set, then
-   * the column must be absent for the condition to pass.
-   * See {@link #setValue(byte[])}.
+   * the column must be absent for the condition to pass. See {@link #setValue(byte[])}.
    *
-   * @param value value
+   * @param value
+   *          value
    * @return this condition
-   * @throws IllegalArgumentException if value is null
+   * @throws IllegalArgumentException
+   *           if value is null
    */
   public Condition setValue(Text value) {
     checkArgument(value != null, "value is null");
     this.val = new ArrayByteSequence(value.getBytes(), 0, value.getLength());
     return this;
   }
-  
+
   /**
    * This method sets the expected value of a column. In order for the condition to pass the column must exist and have this value. If a value is not set, then
-   * the column must be absent for the condition to pass.
-   * See {@link #setValue(byte[])}.
+   * the column must be absent for the condition to pass. See {@link #setValue(byte[])}.
    *
-   * @param value value
+   * @param value
+   *          value
    * @return this condition
-   * @throws IllegalArgumentException if value is null
+   * @throws IllegalArgumentException
+   *           if value is null
    */
   public Condition setValue(ByteSequence value) {
     checkArgument(value != null, "value is null");
@@ -216,8 +229,10 @@ public class Condition {
   /**
    * Sets the visibility for the column to check. If not set it defaults to empty visibility.
    *
-   * @param cv column visibility
-   * @throws IllegalArgumentException if cv is null
+   * @param cv
+   *          column visibility
+   * @throws IllegalArgumentException
+   *           if cv is null
    */
   public Condition setVisibility(ColumnVisibility cv) {
     checkArgument(cv != null, "cv is null");
@@ -240,18 +255,19 @@ public class Condition {
    * covers only the family, qualifier, and visibility (if the timestamp is set then it will be used to narrow the range). Value equality will be tested using
    * the first entry returned by the iterator stack.
    *
-   * @param iterators iterators
+   * @param iterators
+   *          iterators
    * @return this condition
-   * @throws IllegalArgumentException if iterators or any of its elements are null,
-   * or if any two iterators share the same name or priority
+   * @throws IllegalArgumentException
+   *           if iterators or any of its elements are null, or if any two iterators share the same name or priority
    */
   public Condition setIterators(IteratorSetting... iterators) {
     checkArgument(iterators != null, "iterators is null");
-    
+
     if (iterators.length > 1) {
       HashSet<String> names = new HashSet<String>();
       HashSet<Integer> prios = new HashSet<Integer>();
-      
+
       for (IteratorSetting iteratorSetting : iterators) {
         if (!names.add(iteratorSetting.getName()))
           throw new IllegalArgumentException("iterator name used more than once " + iteratorSetting.getName());
@@ -259,7 +275,7 @@ public class Condition {
           throw new IllegalArgumentException("iterator priority used more than once " + iteratorSetting.getPriority());
       }
     }
-    
+
     this.iterators = iterators;
     return this;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/ConditionalMutation.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/ConditionalMutation.java b/core/src/main/java/org/apache/accumulo/core/data/ConditionalMutation.java
index 4772c9c..ccec325 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/ConditionalMutation.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/ConditionalMutation.java
@@ -29,39 +29,39 @@ import org.apache.hadoop.io.Text;
 
 /**
  * A Mutation that contains a list of conditions that must all be met before the mutation is applied.
- * 
+ *
  * @since 1.6.0
  */
 public class ConditionalMutation extends Mutation {
-  
+
   private List<Condition> conditions = new ArrayList<Condition>();
 
   public ConditionalMutation(byte[] row, Condition... conditions) {
     super(row);
     init(conditions);
   }
-  
+
   public ConditionalMutation(byte[] row, int start, int length, Condition... conditions) {
     super(row, start, length);
     init(conditions);
   }
-  
+
   public ConditionalMutation(Text row, Condition... conditions) {
     super(row);
     init(conditions);
   }
-  
+
   public ConditionalMutation(CharSequence row, Condition... conditions) {
     super(row);
     init(conditions);
   }
-  
+
   public ConditionalMutation(ByteSequence row, Condition... conditions) {
     // TODO add ByteSequence methods to mutations
     super(row.toArray());
     init(conditions);
   }
-  
+
   public ConditionalMutation(ConditionalMutation cm) {
     super(cm);
     this.conditions = new ArrayList<Condition>(cm.conditions);
@@ -71,12 +71,12 @@ public class ConditionalMutation extends Mutation {
     checkArgument(conditions != null, "conditions is null");
     this.conditions.addAll(Arrays.asList(conditions));
   }
-  
+
   public void addCondition(Condition condition) {
     checkArgument(condition != null, "condition is null");
     this.conditions.add(condition);
   }
-  
+
   public List<Condition> getConditions() {
     return Collections.unmodifiableList(conditions);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/ConstraintViolationSummary.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/ConstraintViolationSummary.java b/core/src/main/java/org/apache/accumulo/core/data/ConstraintViolationSummary.java
index f84d1f6..6109ac8 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/ConstraintViolationSummary.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/ConstraintViolationSummary.java
@@ -22,20 +22,23 @@ import org.apache.accumulo.core.data.thrift.TConstraintViolationSummary;
  * A summary of constraint violations across some number of mutations.
  */
 public class ConstraintViolationSummary {
-  
+
   public String constrainClass;
   public short violationCode;
   public String violationDescription;
   public long numberOfViolatingMutations;
-  
+
   /**
    * Creates a new summary.
    *
-   * @param constrainClass class of constraint that was violated
-   * @param violationCode violation code
-   * @param violationDescription description of violation
-   * @param numberOfViolatingMutations number of mutations that produced this
-   * particular violation
+   * @param constrainClass
+   *          class of constraint that was violated
+   * @param violationCode
+   *          violation code
+   * @param violationDescription
+   *          description of violation
+   * @param numberOfViolatingMutations
+   *          number of mutations that produced this particular violation
    */
   public ConstraintViolationSummary(String constrainClass, short violationCode, String violationDescription, long numberOfViolatingMutations) {
     this.constrainClass = constrainClass;
@@ -43,39 +46,39 @@ public class ConstraintViolationSummary {
     this.violationDescription = violationDescription;
     this.numberOfViolatingMutations = numberOfViolatingMutations;
   }
-  
+
   /**
    * Creates a new summary from Thrift.
    *
-   * @param tcvs Thrift summary
+   * @param tcvs
+   *          Thrift summary
    */
   public ConstraintViolationSummary(TConstraintViolationSummary tcvs) {
     this(tcvs.constrainClass, tcvs.violationCode, tcvs.violationDescription, tcvs.numberOfViolatingMutations);
   }
-  
+
   public String getConstrainClass() {
     return this.constrainClass;
   }
-  
+
   public short getViolationCode() {
     return this.violationCode;
   }
-  
+
   public String getViolationDescription() {
     return this.violationDescription;
   }
-  
+
   public long getNumberOfViolatingMutations() {
     return this.numberOfViolatingMutations;
   }
-  
+
   @Override
   public String toString() {
-    return String.format(
-        "ConstraintViolationSummary(constrainClass:%s, violationCode:%d, violationDescription:%s, numberOfViolatingMutations:%d)",
+    return String.format("ConstraintViolationSummary(constrainClass:%s, violationCode:%d, violationDescription:%s, numberOfViolatingMutations:%d)",
         constrainClass, violationCode, violationDescription, numberOfViolatingMutations);
   }
-  
+
   /**
    * Converts this summary to Thrift.
    *
@@ -84,5 +87,5 @@ public class ConstraintViolationSummary {
   public TConstraintViolationSummary toThrift() {
     return new TConstraintViolationSummary(this.constrainClass, violationCode, violationDescription, numberOfViolatingMutations);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/Key.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Key.java b/core/src/main/java/org/apache/accumulo/core/data/Key.java
index 23b48ff..8ebba43 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Key.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Key.java
@@ -17,11 +17,11 @@
 package org.apache.accumulo.core.data;
 
 /**
- * This is the Key used to store and access individual values in Accumulo.  A Key is a tuple composed of a row, column family, column qualifier, 
+ * This is the Key used to store and access individual values in Accumulo.  A Key is a tuple composed of a row, column family, column qualifier,
  * column visibility, timestamp, and delete marker.
- * 
+ *
  * Keys are comparable and therefore have a sorted order defined by {@link #compareTo(Key)}.
- * 
+ *
  */
 
 import static org.apache.accumulo.core.util.ByteBufferUtil.toBytes;
@@ -43,35 +43,35 @@ import org.apache.hadoop.io.WritableComparator;
 import org.apache.hadoop.io.WritableUtils;
 
 public class Key implements WritableComparable<Key>, Cloneable {
-  
+
   protected byte[] row;
   protected byte[] colFamily;
   protected byte[] colQualifier;
   protected byte[] colVisibility;
   protected long timestamp;
   protected boolean deleted;
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof Key)
       return this.equals((Key) o, PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME_DEL);
     return false;
   }
-  
+
   private static final byte EMPTY_BYTES[] = new byte[0];
-  
+
   private byte[] copyIfNeeded(byte ba[], int off, int len, boolean copyData) {
     if (len == 0)
       return EMPTY_BYTES;
-    
+
     if (!copyData && ba.length == len && off == 0)
       return ba;
-    
+
     byte[] copy = new byte[len];
     System.arraycopy(ba, off, copy, 0, len);
     return copy;
   }
-  
+
   private final void init(byte r[], int rOff, int rLen, byte cf[], int cfOff, int cfLen, byte cq[], int cqOff, int cqLen, byte cv[], int cvOff, int cvLen,
       long ts, boolean del, boolean copy) {
     row = copyIfNeeded(r, rOff, rLen, copy);
@@ -81,7 +81,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
     timestamp = ts;
     deleted = del;
   }
-  
+
   /**
    * Creates a key with empty row, empty column family, empty column qualifier, empty column visibility, timestamp {@link Long#MAX_VALUE}, and delete marker
    * false.
@@ -94,92 +94,126 @@ public class Key implements WritableComparable<Key>, Cloneable {
     timestamp = Long.MAX_VALUE;
     deleted = false;
   }
-  
+
   /**
    * Creates a key with the specified row, empty column family, empty column qualifier, empty column visibility, timestamp {@link Long#MAX_VALUE}, and delete
    * marker false.
    *
-   * @param row row ID
+   * @param row
+   *          row ID
    */
   public Key(Text row) {
     init(row.getBytes(), 0, row.getLength(), EMPTY_BYTES, 0, 0, EMPTY_BYTES, 0, 0, EMPTY_BYTES, 0, 0, Long.MAX_VALUE, false, true);
   }
-  
+
   /**
    * Creates a key with the specified row, empty column family, empty column qualifier, empty column visibility, the specified timestamp, and delete marker
    * false.
    *
-   * @param row row ID
-   * @param ts timestamp
+   * @param row
+   *          row ID
+   * @param ts
+   *          timestamp
    */
   public Key(Text row, long ts) {
     this(row);
     timestamp = ts;
   }
-  
+
   /**
    * Creates a key. The delete marker defaults to false.
    *
-   * @param row bytes containing row ID
-   * @param rOff offset into row where key's row ID begins (inclusive)
-   * @param rLen length of row ID in row
-   * @param cf bytes containing column family
-   * @param cfOff offset into cf where key's column family begins (inclusive)
-   * @param cfLen length of column family in cf
-   * @param cq bytes containing column qualifier
-   * @param cqOff offset into cq where key's column qualifier begins (inclusive)
-   * @param cqLen length of column qualifier in cq
-   * @param cv bytes containing column visibility
-   * @param cvOff offset into cv where key's column visibility begins (inclusive)
-   * @param cvLen length of column visibility in cv
-   * @param ts timestamp
+   * @param row
+   *          bytes containing row ID
+   * @param rOff
+   *          offset into row where key's row ID begins (inclusive)
+   * @param rLen
+   *          length of row ID in row
+   * @param cf
+   *          bytes containing column family
+   * @param cfOff
+   *          offset into cf where key's column family begins (inclusive)
+   * @param cfLen
+   *          length of column family in cf
+   * @param cq
+   *          bytes containing column qualifier
+   * @param cqOff
+   *          offset into cq where key's column qualifier begins (inclusive)
+   * @param cqLen
+   *          length of column qualifier in cq
+   * @param cv
+   *          bytes containing column visibility
+   * @param cvOff
+   *          offset into cv where key's column visibility begins (inclusive)
+   * @param cvLen
+   *          length of column visibility in cv
+   * @param ts
+   *          timestamp
    */
   public Key(byte row[], int rOff, int rLen, byte cf[], int cfOff, int cfLen, byte cq[], int cqOff, int cqLen, byte cv[], int cvOff, int cvLen, long ts) {
     init(row, rOff, rLen, cf, cfOff, cfLen, cq, cqOff, cqLen, cv, cvOff, cvLen, ts, false, true);
   }
-  
+
   /**
    * Creates a key. The delete marker defaults to false.
    *
-   * @param row row ID
-   * @param colFamily column family
-   * @param colQualifier column qualifier
-   * @param colVisibility column visibility
-   * @param timestamp timestamp
+   * @param row
+   *          row ID
+   * @param colFamily
+   *          column family
+   * @param colQualifier
+   *          column qualifier
+   * @param colVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
    */
   public Key(byte[] row, byte[] colFamily, byte[] colQualifier, byte[] colVisibility, long timestamp) {
     this(row, colFamily, colQualifier, colVisibility, timestamp, false, true);
   }
-  
+
   /**
    * Creates a key.
    *
-   * @param row row ID
-   * @param cf column family
-   * @param cq column qualifier
-   * @param cv column visibility
-   * @param ts timestamp
-   * @param deleted delete marker
+   * @param row
+   *          row ID
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @param cv
+   *          column visibility
+   * @param ts
+   *          timestamp
+   * @param deleted
+   *          delete marker
    */
   public Key(byte[] row, byte[] cf, byte[] cq, byte[] cv, long ts, boolean deleted) {
     this(row, cf, cq, cv, ts, deleted, true);
   }
-  
+
   /**
    * Creates a key.
    *
-   * @param row row ID
-   * @param cf column family
-   * @param cq column qualifier
-   * @param cv column visibility
-   * @param ts timestamp
-   * @param deleted delete marker
-   * @param copy if true, forces copy of byte array values into key
+   * @param row
+   *          row ID
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @param cv
+   *          column visibility
+   * @param ts
+   *          timestamp
+   * @param deleted
+   *          delete marker
+   * @param copy
+   *          if true, forces copy of byte array values into key
    */
   public Key(byte[] row, byte[] cf, byte[] cq, byte[] cv, long ts, boolean deleted, boolean copy) {
     init(row, 0, row.length, cf, 0, cf.length, cq, 0, cq.length, cv, 0, cv.length, ts, deleted, copy);
   }
-  
+
   /**
    * Creates a key with the specified row, the specified column family, empty column qualifier, empty column visibility, timestamp {@link Long#MAX_VALUE}, and
    * delete marker false.
@@ -187,7 +221,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public Key(Text row, Text cf) {
     init(row.getBytes(), 0, row.getLength(), cf.getBytes(), 0, cf.getLength(), EMPTY_BYTES, 0, 0, EMPTY_BYTES, 0, 0, Long.MAX_VALUE, false, true);
   }
-  
+
   /**
    * Creates a key with the specified row, the specified column family, the specified column qualifier, empty column visibility, timestamp
    * {@link Long#MAX_VALUE}, and delete marker false.
@@ -195,7 +229,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public Key(Text row, Text cf, Text cq) {
     init(row.getBytes(), 0, row.getLength(), cf.getBytes(), 0, cf.getLength(), cq.getBytes(), 0, cq.getLength(), EMPTY_BYTES, 0, 0, Long.MAX_VALUE, false, true);
   }
-  
+
   /**
    * Creates a key with the specified row, the specified column family, the specified column qualifier, the specified column visibility, timestamp
    * {@link Long#MAX_VALUE}, and delete marker false.
@@ -204,7 +238,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
     init(row.getBytes(), 0, row.getLength(), cf.getBytes(), 0, cf.getLength(), cq.getBytes(), 0, cq.getLength(), cv.getBytes(), 0, cv.getLength(),
         Long.MAX_VALUE, false, true);
   }
-  
+
   /**
    * Creates a key with the specified row, the specified column family, the specified column qualifier, empty column visibility, the specified timestamp, and
    * delete marker false.
@@ -212,7 +246,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public Key(Text row, Text cf, Text cq, long ts) {
     init(row.getBytes(), 0, row.getLength(), cf.getBytes(), 0, cf.getLength(), cq.getBytes(), 0, cq.getLength(), EMPTY_BYTES, 0, 0, ts, false, true);
   }
-  
+
   /**
    * Creates a key with the specified row, the specified column family, the specified column qualifier, the specified column visibility, the specified
    * timestamp, and delete marker false.
@@ -221,7 +255,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
     init(row.getBytes(), 0, row.getLength(), cf.getBytes(), 0, cf.getLength(), cq.getBytes(), 0, cq.getLength(), cv.getBytes(), 0, cv.getLength(), ts, false,
         true);
   }
-  
+
   /**
    * Creates a key with the specified row, the specified column family, the specified column qualifier, the specified column visibility, the specified
    * timestamp, and delete marker false.
@@ -230,66 +264,66 @@ public class Key implements WritableComparable<Key>, Cloneable {
     byte[] expr = cv.getExpression();
     init(row.getBytes(), 0, row.getLength(), cf.getBytes(), 0, cf.getLength(), cq.getBytes(), 0, cq.getLength(), expr, 0, expr.length, ts, false, true);
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text)}.
    */
   public Key(CharSequence row) {
     this(new Text(row.toString()));
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text,Text)}.
    */
   public Key(CharSequence row, CharSequence cf) {
     this(new Text(row.toString()), new Text(cf.toString()));
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text,Text,Text)}.
    */
   public Key(CharSequence row, CharSequence cf, CharSequence cq) {
     this(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()));
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text,Text,Text,Text)}.
    */
   public Key(CharSequence row, CharSequence cf, CharSequence cq, CharSequence cv) {
     this(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()), new Text(cv.toString()));
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text,Text,Text,long)}.
    */
   public Key(CharSequence row, CharSequence cf, CharSequence cq, long ts) {
     this(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()), ts);
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text,Text,Text,Text,long)}.
    */
   public Key(CharSequence row, CharSequence cf, CharSequence cq, CharSequence cv, long ts) {
     this(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()), new Text(cv.toString()), ts);
   }
-  
+
   /**
    * Converts CharSequence to Text and creates a Key using {@link #Key(Text,Text,Text,ColumnVisibility,long)}.
    */
   public Key(CharSequence row, CharSequence cf, CharSequence cq, ColumnVisibility cv, long ts) {
     this(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()), new Text(cv.getExpression()), ts);
   }
-  
+
   private byte[] followingArray(byte ba[]) {
     byte[] fba = new byte[ba.length + 1];
     System.arraycopy(ba, 0, fba, 0, ba.length);
     fba[ba.length] = (byte) 0x00;
     return fba;
   }
-  
+
   /**
    * Returns a key that will sort immediately after this key.
-   * 
+   *
    * @param part
    *          PartialKey except {@link PartialKey#ROW_COLFAM_COLQUAL_COLVIS_TIME_DEL}
    */
@@ -328,18 +362,19 @@ public class Key implements WritableComparable<Key>, Cloneable {
     }
     return returnKey;
   }
-  
+
   /**
    * Creates a key with the same row, column family, column qualifier, column visibility, timestamp, and delete marker as the given key.
    */
   public Key(Key other) {
     set(other);
   }
-  
+
   /**
    * Creates a key from Thrift.
    *
-   * @param tkey Thrift key
+   * @param tkey
+   *          Thrift key
    */
   public Key(TKey tkey) {
     this.row = toBytes(tkey.row);
@@ -362,143 +397,143 @@ public class Key implements WritableComparable<Key>, Cloneable {
       throw new IllegalArgumentException("null column visibility");
     }
   }
-  
+
   /**
-   * Writes the row ID into the given <code>Text</code>. This method gives
-   * users control over allocation of Text objects by copying into the passed in
-   * text.
-   * 
-   * @param r <code>Text</code> object to copy into
+   * Writes the row ID into the given <code>Text</code>. This method gives users control over allocation of Text objects by copying into the passed in text.
+   *
+   * @param r
+   *          <code>Text</code> object to copy into
    * @return the <code>Text</code> that was passed in
    */
   public Text getRow(Text r) {
     r.set(row, 0, row.length);
     return r;
   }
-  
+
   /**
-   * Returns the row ID as a byte sequence. This method returns a pointer to the
-   * key's internal data and does not copy it.
-   * 
+   * Returns the row ID as a byte sequence. This method returns a pointer to the key's internal data and does not copy it.
+   *
    * @return ByteSequence that points to the internal key row ID data
    */
   public ByteSequence getRowData() {
     return new ArrayByteSequence(row);
   }
-  
+
   /**
    * Gets the row ID as a <code>Text</code> object.
-   * 
+   *
    * @return Text containing the row ID
    */
   public Text getRow() {
     return getRow(new Text());
   }
-  
+
   /**
    * Compares this key's row ID with another.
-   * 
-   * @param r row ID to compare
+   *
+   * @param r
+   *          row ID to compare
    * @return same as {@link #getRow()}.compareTo(r)
    */
   public int compareRow(Text r) {
     return WritableComparator.compareBytes(row, 0, row.length, r.getBytes(), 0, r.getLength());
   }
-  
+
   /**
-   * Returns the column family as a byte sequence. This method returns a pointer to the
-   * key's internal data and does not copy it.
-   * 
+   * Returns the column family as a byte sequence. This method returns a pointer to the key's internal data and does not copy it.
+   *
    * @return ByteSequence that points to the internal key column family data
    */
   public ByteSequence getColumnFamilyData() {
     return new ArrayByteSequence(colFamily);
   }
-  
+
   /**
-   * Writes the column family into the given <code>Text</code>. This method gives
-   * users control over allocation of Text objects by copying into the passed in
+   * Writes the column family into the given <code>Text</code>. This method gives users control over allocation of Text objects by copying into the passed in
    * text.
-   * 
-   * @param cf <code>Text</code> object to copy into
+   *
+   * @param cf
+   *          <code>Text</code> object to copy into
    * @return the <code>Text</code> that was passed in
    */
   public Text getColumnFamily(Text cf) {
     cf.set(colFamily, 0, colFamily.length);
     return cf;
   }
-  
+
   /**
    * Gets the column family as a <code>Text</code> object.
-   * 
+   *
    * @return Text containing the column family
    */
   public Text getColumnFamily() {
     return getColumnFamily(new Text());
   }
-  
+
   /**
    * Compares this key's column family with another.
-   * 
-   * @param cf column family to compare
+   *
+   * @param cf
+   *          column family to compare
    * @return same as {@link #getColumnFamily()}.compareTo(cf)
    */
-  
+
   public int compareColumnFamily(Text cf) {
     return WritableComparator.compareBytes(colFamily, 0, colFamily.length, cf.getBytes(), 0, cf.getLength());
   }
-  
+
   /**
-   * Returns the column qualifier as a byte sequence. This method returns a pointer to the
-   * key's internal data and does not copy it.
-   * 
+   * Returns the column qualifier as a byte sequence. This method returns a pointer to the key's internal data and does not copy it.
+   *
    * @return ByteSequence that points to the internal key column qualifier data
    */
   public ByteSequence getColumnQualifierData() {
     return new ArrayByteSequence(colQualifier);
   }
-  
+
   /**
-   * Writes the column qualifier into the given <code>Text</code>. This method gives
-   * users control over allocation of Text objects by copying into the passed in
+   * Writes the column qualifier into the given <code>Text</code>. This method gives users control over allocation of Text objects by copying into the passed in
    * text.
-   * 
-   * @param cq <code>Text</code> object to copy into
+   *
+   * @param cq
+   *          <code>Text</code> object to copy into
    * @return the <code>Text</code> that was passed in
    */
   public Text getColumnQualifier(Text cq) {
     cq.set(colQualifier, 0, colQualifier.length);
     return cq;
   }
-  
+
   /**
    * Gets the column qualifier as a <code>Text</code> object.
-   * 
+   *
    * @return Text containing the column qualifier
    */
   public Text getColumnQualifier() {
     return getColumnQualifier(new Text());
   }
-  
+
   /**
    * Compares this key's column qualifier with another.
-   * 
-   * @param cq column qualifier to compare
+   *
+   * @param cq
+   *          column qualifier to compare
    * @return same as {@link #getColumnQualifier()}.compareTo(cq)
    */
   public int compareColumnQualifier(Text cq) {
     return WritableComparator.compareBytes(colQualifier, 0, colQualifier.length, cq.getBytes(), 0, cq.getLength());
   }
-  
+
   /**
    * Sets the timestamp.
    *
-   * @param ts timestamp
+   * @param ts
+   *          timestamp
    */
   public void setTimestamp(long ts) {
     this.timestamp = ts;
   }
-  
+
   /**
    * Gets the timestamp.
    *
@@ -507,7 +542,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public long getTimestamp() {
     return timestamp;
   }
-  
+
   /**
    * Determines if this key is deleted (i.e., has a delete marker = true).
    *
@@ -516,64 +551,64 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public boolean isDeleted() {
     return deleted;
   }
-  
+
   /**
    * Sets the delete marker on this key.
    *
-   * @param deleted delete marker (true to delete)
+   * @param deleted
+   *          delete marker (true to delete)
    */
   public void setDeleted(boolean deleted) {
     this.deleted = deleted;
   }
-  
+
   /**
-   * Returns the column visibility as a byte sequence. This method returns a pointer to the
-   * key's internal data and does not copy it.
-   * 
+   * Returns the column visibility as a byte sequence. This method returns a pointer to the key's internal data and does not copy it.
+   *
    * @return ByteSequence that points to the internal key column visibility data
    */
   public ByteSequence getColumnVisibilityData() {
     return new ArrayByteSequence(colVisibility);
   }
-  
+
   /**
    * Gets the column visibility as a <code>Text</code> object.
-   * 
+   *
    * @return Text containing the column visibility
    */
   public final Text getColumnVisibility() {
     return getColumnVisibility(new Text());
   }
-  
+
   /**
-   * Writes the column visibvility into the given <code>Text</code>. This method gives
-   * users control over allocation of Text objects by copying into the passed in
-   * text.
-   * 
-   * @param cv <code>Text</code> object to copy into
+   * Writes the column visibvility into the given <code>Text</code>. This method gives users control over allocation of Text objects by copying into the passed
+   * in text.
+   *
+   * @param cv
+   *          <code>Text</code> object to copy into
    * @return the <code>Text</code> that was passed in
    */
   public final Text getColumnVisibility(Text cv) {
     cv.set(colVisibility, 0, colVisibility.length);
     return cv;
   }
-  
+
   /**
-   * Gets the column visibility. <b>WARNING:</b> using this method may inhibit
-   * performance since a new ColumnVisibility object is created on every call.
-   * 
+   * Gets the column visibility. <b>WARNING:</b> using this method may inhibit performance since a new ColumnVisibility object is created on every call.
+   *
    * @return ColumnVisibility representing the column visibility
    * @since 1.5.0
    */
   public final ColumnVisibility getColumnVisibilityParsed() {
     return new ColumnVisibility(colVisibility);
   }
-  
+
   /**
-   * Sets this key's row, column family, column qualifier, column visibility, timestamp, and delete marker to be the same as another key's.
-   * This method does not copy data from the other key, but only references to it.
+   * Sets this key's row, column family, column qualifier, column visibility, timestamp, and delete marker to be the same as another key's. This method does not
+   * copy data from the other key, but only references to it.
    *
-   * @param k key to set from
+   * @param k
+   *          key to set from
    */
   public void set(Key k) {
     row = k.row;
@@ -582,61 +617,63 @@ public class Key implements WritableComparable<Key>, Cloneable {
     colVisibility = k.colVisibility;
     timestamp = k.timestamp;
     deleted = k.deleted;
-    
+
   }
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     // this method is a little screwy so it will be compatible with older
     // code that serialized data
-    
+
     int colFamilyOffset = WritableUtils.readVInt(in);
     int colQualifierOffset = WritableUtils.readVInt(in);
     int colVisibilityOffset = WritableUtils.readVInt(in);
     int totalLen = WritableUtils.readVInt(in);
-    
+
     row = new byte[colFamilyOffset];
     colFamily = new byte[colQualifierOffset - colFamilyOffset];
     colQualifier = new byte[colVisibilityOffset - colQualifierOffset];
     colVisibility = new byte[totalLen - colVisibilityOffset];
-    
+
     in.readFully(row);
     in.readFully(colFamily);
     in.readFully(colQualifier);
     in.readFully(colVisibility);
-    
+
     timestamp = WritableUtils.readVLong(in);
     deleted = in.readBoolean();
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
-    
+
     int colFamilyOffset = row.length;
     int colQualifierOffset = colFamilyOffset + colFamily.length;
     int colVisibilityOffset = colQualifierOffset + colQualifier.length;
     int totalLen = colVisibilityOffset + colVisibility.length;
-    
+
     WritableUtils.writeVInt(out, colFamilyOffset);
     WritableUtils.writeVInt(out, colQualifierOffset);
     WritableUtils.writeVInt(out, colVisibilityOffset);
-    
+
     WritableUtils.writeVInt(out, totalLen);
-    
+
     out.write(row);
     out.write(colFamily);
     out.write(colQualifier);
     out.write(colVisibility);
-    
+
     WritableUtils.writeVLong(out, timestamp);
     out.writeBoolean(deleted);
   }
-  
+
   /**
    * Compares part of a key. For example, compares just the row and column family, and if those are equal then return true.
-   * 
-   * @param other key to compare to
-   * @param part part of key to compare
+   *
+   * @param other
+   *          key to compare to
+   * @param part
+   *          part of key to compare
    * @return true if specified parts of keys match, false otherwise
    */
   public boolean equals(Key other, PartialKey part) {
@@ -660,18 +697,20 @@ public class Key implements WritableComparable<Key>, Cloneable {
         throw new IllegalArgumentException("Unrecognized partial key specification " + part);
     }
   }
-  
+
   /**
    * Compares elements of a key given by a {@link PartialKey}. The corresponding elements (row, column family, column qualifier, column visibility, timestamp,
-   * and delete marker) are compared in order until unequal elements are found. The row, column family, column qualifier, and column
-   * visibility are compared lexographically and sorted ascending. The timestamps are compared numerically and sorted descending so that the most recent data
-   * comes first. Lastly, a delete marker of true sorts before a delete marker of false. The result of the first unequal comparison is returned.
+   * and delete marker) are compared in order until unequal elements are found. The row, column family, column qualifier, and column visibility are compared
+   * lexographically and sorted ascending. The timestamps are compared numerically and sorted descending so that the most recent data comes first. Lastly, a
+   * delete marker of true sorts before a delete marker of false. The result of the first unequal comparison is returned.
+   *
+   * For example, for {@link PartialKey#ROW_COLFAM}, this method compares just the row and column family. If the row IDs are not equal, return the result of the
+   * row comparison; otherwise, returns the result of the column family comparison.
    *
-   * For example, for {@link PartialKey#ROW_COLFAM}, this method compares just the row and column family. If the
-   * row IDs are not equal, return the result of the row comparison; otherwise, returns the result of the column family comparison.
-   * 
-   * @param other key to compare to
-   * @param part part of key to compare
+   * @param other
+   *          key to compare to
+   * @param part
+   *          part of key to compare
    * @return comparison result
    * @see #compareTo(Key)
    */
@@ -680,22 +719,22 @@ public class Key implements WritableComparable<Key>, Cloneable {
     int result = WritableComparator.compareBytes(row, 0, row.length, other.row, 0, other.row.length);
     if (result != 0 || part.equals(PartialKey.ROW))
       return result;
-    
+
     // check for matching column family
     result = WritableComparator.compareBytes(colFamily, 0, colFamily.length, other.colFamily, 0, other.colFamily.length);
     if (result != 0 || part.equals(PartialKey.ROW_COLFAM))
       return result;
-    
+
     // check for matching column qualifier
     result = WritableComparator.compareBytes(colQualifier, 0, colQualifier.length, other.colQualifier, 0, other.colQualifier.length);
     if (result != 0 || part.equals(PartialKey.ROW_COLFAM_COLQUAL))
       return result;
-    
+
     // check for matching column visibility
     result = WritableComparator.compareBytes(colVisibility, 0, colVisibility.length, other.colVisibility, 0, other.colVisibility.length);
     if (result != 0 || part.equals(PartialKey.ROW_COLFAM_COLQUAL_COLVIS))
       return result;
-    
+
     // check for matching timestamp
     if (timestamp < other.timestamp)
       result = 1;
@@ -703,23 +742,24 @@ public class Key implements WritableComparable<Key>, Cloneable {
       result = -1;
     else
       result = 0;
-    
+
     if (result != 0 || part.equals(PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME))
       return result;
-    
+
     // check for matching deleted flag
     if (deleted)
       result = other.deleted ? 0 : -1;
     else
       result = other.deleted ? 1 : 0;
-    
+
     return result;
   }
-  
+
   /**
    * Compares this key with another.
    *
-   * @param other key to compare to
+   * @param other
+   *          key to compare to
    * @return comparison result
    * @see #compareTo(Key, PartialKey)
    */
@@ -727,48 +767,53 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public int compareTo(Key other) {
     return compareTo(other, PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME_DEL);
   }
-  
+
   @Override
   public int hashCode() {
     return WritableComparator.hashBytes(row, row.length) + WritableComparator.hashBytes(colFamily, colFamily.length)
         + WritableComparator.hashBytes(colQualifier, colQualifier.length) + WritableComparator.hashBytes(colVisibility, colVisibility.length)
         + (int) (timestamp ^ (timestamp >>> 32));
   }
-  
+
   /**
-   * Returns an ASCII printable string form of the given byte array, treating
-   * the bytes as ASCII characters. See
-   * {@link #appendPrintableString(byte[], int, int, int, StringBuilder)}
-   * for caveats.
+   * Returns an ASCII printable string form of the given byte array, treating the bytes as ASCII characters. See
+   * {@link #appendPrintableString(byte[], int, int, int, StringBuilder)} for caveats.
    *
-   * @param ba byte array
-   * @param offset offset to start with in byte array (inclusive)
-   * @param len number of bytes to print
-   * @param maxLen maximum number of bytes to convert to printable form
+   * @param ba
+   *          byte array
+   * @param offset
+   *          offset to start with in byte array (inclusive)
+   * @param len
+   *          number of bytes to print
+   * @param maxLen
+   *          maximum number of bytes to convert to printable form
    * @return printable string
    * @see #appendPrintableString(byte[], int, int, int, StringBuilder)
    */
   public static String toPrintableString(byte ba[], int offset, int len, int maxLen) {
     return appendPrintableString(ba, offset, len, maxLen, new StringBuilder()).toString();
   }
-  
+
   /**
-   * Appends ASCII printable characters to a string, based on the given byte
-   * array, treating the bytes as ASCII characters. If a byte can be converted
-   * to a ASCII printable character it is appended as is; otherwise, it is
-   * appended as a character code, e.g., %05; for byte value 5. If len > maxlen,
-   * the string includes a "TRUNCATED" note at the end.
+   * Appends ASCII printable characters to a string, based on the given byte array, treating the bytes as ASCII characters. If a byte can be converted to a
+   * ASCII printable character it is appended as is; otherwise, it is appended as a character code, e.g., %05; for byte value 5. If len > maxlen, the string
+   * includes a "TRUNCATED" note at the end.
    *
-   * @param ba byte array
-   * @param offset offset to start with in byte array (inclusive)
-   * @param len number of bytes to print
-   * @param maxLen maximum number of bytes to convert to printable form
-   * @param sb <code>StringBuilder</code> to append to
+   * @param ba
+   *          byte array
+   * @param offset
+   *          offset to start with in byte array (inclusive)
+   * @param len
+   *          number of bytes to print
+   * @param maxLen
+   *          maximum number of bytes to convert to printable form
+   * @param sb
+   *          <code>StringBuilder</code> to append to
    * @return given <code>StringBuilder</code>
    */
   public static StringBuilder appendPrintableString(byte ba[], int offset, int len, int maxLen, StringBuilder sb) {
     int plen = Math.min(len, maxLen);
-    
+
     for (int i = 0; i < plen; i++) {
       int c = 0xff & ba[offset + i];
       if (c >= 32 && c <= 126)
@@ -776,18 +821,18 @@ public class Key implements WritableComparable<Key>, Cloneable {
       else
         sb.append("%" + String.format("%02x;", c));
     }
-    
+
     if (len > maxLen) {
       sb.append("... TRUNCATED");
     }
-    
+
     return sb;
   }
 
   private StringBuilder rowColumnStringBuilder() {
     return rowColumnStringBuilder(Constants.MAX_DATA_TO_PRINT);
   }
-  
+
   private StringBuilder rowColumnStringBuilder(int maxComponentLength) {
     StringBuilder sb = new StringBuilder();
     appendPrintableString(row, 0, row.length, maxComponentLength, sb);
@@ -800,7 +845,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
     sb.append("]");
     return sb;
   }
-  
+
   @Override
   public String toString() {
     StringBuilder sb = rowColumnStringBuilder();
@@ -812,8 +857,8 @@ public class Key implements WritableComparable<Key>, Cloneable {
   }
 
   /**
-   * Stringify this {@link Key}, avoiding truncation of each component, only limiting
-   * each component to a length of {@link Integer#MAX_VALUE}
+   * Stringify this {@link Key}, avoiding truncation of each component, only limiting each component to a length of {@link Integer#MAX_VALUE}
+   *
    * @since 1.7.0
    */
   public String toStringNoTruncate() {
@@ -833,16 +878,16 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public String toStringNoTime() {
     return rowColumnStringBuilder().toString();
   }
-  
+
   /**
    * Returns the sums of the lengths of the row, column family, column qualifier, and column visibility.
-   * 
+   *
    * @return sum of key field lengths
    */
   public int getLength() {
     return row.length + colFamily.length + colQualifier.length + colVisibility.length;
   }
-  
+
   /**
    * Same as {@link #getLength()}.
    *
@@ -851,19 +896,19 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public int getSize() {
     return getLength();
   }
-  
+
   private static boolean isEqual(byte a1[], byte a2[]) {
     if (a1 == a2)
       return true;
-    
+
     int last = a1.length;
-    
+
     if (last != a2.length)
       return false;
-    
+
     if (last == 0)
       return true;
-    
+
     // since sorted data is usually compared in accumulo,
     // the prefixes will normally be the same... so compare
     // the last two charachters first.. the most likely place
@@ -873,9 +918,9 @@ public class Key implements WritableComparable<Key>, Cloneable {
     // (compiler and cpu optimized for reading data forward)..
     // do not want slower comparisons when data is equal...
     // sorting brings equals data together
-    
+
     last--;
-    
+
     if (a1[last] == a2[last]) {
       for (int i = 0; i < last; i++)
         if (a1[i] != a2[i])
@@ -883,21 +928,22 @@ public class Key implements WritableComparable<Key>, Cloneable {
     } else {
       return false;
     }
-    
+
     return true;
-    
+
   }
-  
+
   /**
    * Compresses a list of key/value pairs before sending them via thrift.
-   * 
-   * @param param list of key/value pairs
+   *
+   * @param param
+   *          list of key/value pairs
    * @return list of Thrift key/value pairs
    */
   public static List<TKeyValue> compress(List<? extends KeyValue> param) {
-    
+
     List<TKeyValue> tkvl = Arrays.asList(new TKeyValue[param.size()]);
-    
+
     if (param.size() > 0)
       tkvl.set(0, new TKeyValue(param.get(0).getKey().toThrift(), ByteBuffer.wrap(param.get(0).getValue().get())));
 
@@ -907,50 +953,50 @@ public class Key implements WritableComparable<Key>, Cloneable {
       Key key = kv.getKey();
 
       TKey newKey = null;
-      
+
       if (isEqual(prevKey.row, key.row)) {
         newKey = key.toThrift();
         newKey.row = null;
       }
-      
+
       if (isEqual(prevKey.colFamily, key.colFamily)) {
         if (newKey == null)
           newKey = key.toThrift();
         newKey.colFamily = null;
       }
-      
+
       if (isEqual(prevKey.colQualifier, key.colQualifier)) {
         if (newKey == null)
           newKey = key.toThrift();
         newKey.colQualifier = null;
       }
-      
+
       if (isEqual(prevKey.colVisibility, key.colVisibility)) {
         if (newKey == null)
           newKey = key.toThrift();
         newKey.colVisibility = null;
       }
-      
+
       if (newKey == null)
         newKey = key.toThrift();
 
       tkvl.set(i, new TKeyValue(newKey, ByteBuffer.wrap(kv.getValue().get())));
     }
-    
+
     return tkvl;
   }
-  
+
   /**
-   * Decompresses a list of key/value pairs received from thrift. Decompression
-   * occurs in place, in the list.
-   * 
-   * @param param list of Thrift key/value pairs
+   * Decompresses a list of key/value pairs received from thrift. Decompression occurs in place, in the list.
+   *
+   * @param param
+   *          list of Thrift key/value pairs
    */
   public static void decompress(List<TKeyValue> param) {
     for (int i = 1; i < param.size(); i++) {
       TKey prevKey = param.get(i - 1).key;
       TKey key = param.get(i).key;
-      
+
       if (key.row == null) {
         key.row = prevKey.row;
       }
@@ -965,7 +1011,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
       }
     }
   }
-  
+
   /**
    * Gets the row ID as a byte array.
    *
@@ -974,7 +1020,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   byte[] getRowBytes() {
     return row;
   }
-  
+
   /**
    * Gets the column family as a byte array.
    *
@@ -983,7 +1029,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   byte[] getColFamily() {
     return colFamily;
   }
-  
+
   /**
    * Gets the column qualifier as a byte array.
    *
@@ -992,7 +1038,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   byte[] getColQualifier() {
     return colQualifier;
   }
-  
+
   /**
    * Gets the column visibility as a byte array.
    *
@@ -1001,7 +1047,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   byte[] getColVisibility() {
     return colVisibility;
   }
-  
+
   /**
    * Converts this key to Thrift.
    *
@@ -1010,7 +1056,7 @@ public class Key implements WritableComparable<Key>, Cloneable {
   public TKey toThrift() {
     return new TKey(ByteBuffer.wrap(row), ByteBuffer.wrap(colFamily), ByteBuffer.wrap(colQualifier), ByteBuffer.wrap(colVisibility), timestamp);
   }
-  
+
   /**
    * Performs a deep copy of this key.
    *

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/KeyExtent.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/KeyExtent.java b/core/src/main/java/org/apache/accumulo/core/data/KeyExtent.java
index 4c9978f..bc1513e 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/KeyExtent.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/KeyExtent.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.core.data;
 /**
  * keeps track of information needed to identify a tablet
  * apparently, we only need the endKey and not the start as well
- * 
+ *
  */
 
 import java.io.ByteArrayOutputStream;
@@ -89,7 +89,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Default constructor
-   * 
+   *
    */
   public KeyExtent() {
     this.setTableId(new Text());
@@ -124,7 +124,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Returns a String representing this extent's entry in the Metadata table
-   * 
+   *
    */
   public Text getMetadataEntry() {
     return getMetadataEntry(getTableId(), getEndRow());
@@ -159,7 +159,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Sets the extents table id
-   * 
+   *
    */
   public void setTableId(Text tId) {
 
@@ -173,7 +173,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Returns the extent's table id
-   * 
+   *
    */
   public Text getTableId() {
     return textTableId;
@@ -195,7 +195,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Sets this extent's end row
-   * 
+   *
    */
   public void setEndRow(Text endRow) {
     setEndRow(endRow, true, true);
@@ -203,7 +203,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Returns this extent's end row
-   * 
+   *
    */
   public Text getEndRow() {
     return textEndRow;
@@ -211,7 +211,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Return the previous extent's end row
-   * 
+   *
    */
   public Text getPrevEndRow() {
     return textPrevEndRow;
@@ -233,7 +233,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Sets the previous extent's end row
-   * 
+   *
    */
   public void setPrevEndRow(Text prevEndRow) {
     setPrevEndRow(prevEndRow, true, true);
@@ -241,7 +241,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Populates the extents data fields from a DataInput object
-   * 
+   *
    */
   @Override
   public void readFields(DataInput in) throws IOException {
@@ -271,7 +271,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Writes this extent's data fields to a DataOutput object
-   * 
+   *
    */
   @Override
   public void write(DataOutput out) throws IOException {
@@ -292,7 +292,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Returns a String representing the previous extent's entry in the Metadata table
-   * 
+   *
    */
   public Mutation getPrevRowUpdateMutation() {
     return getPrevRowUpdateMutation(this);
@@ -300,7 +300,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Empty start or end rows tell the method there are no start or end rows, and to use all the keyextents that are before the end row if no start row etc.
-   * 
+   *
    * @deprecated this method not intended for public use and is likely to be removed in a future version.
    * @return all the key extents that the rows cover
    */
@@ -397,7 +397,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
 
   /**
    * Compares extents based on rows
-   * 
+   *
    */
   @Override
   public int compareTo(KeyExtent other) {
@@ -508,7 +508,7 @@ public class KeyExtent implements WritableComparable<KeyExtent> {
   // note: this is only the encoding of the table id and the last row, not the prev row
   /**
    * Populates the extent's fields based on a flatted extent
-   * 
+   *
    */
   private void decodeMetadataRow(Text flattenedExtent) {
     int semiPos = -1;


[10/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ImportTableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ImportTableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ImportTableCommand.java
index 46c941f..893c913 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ImportTableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ImportTableCommand.java
@@ -25,25 +25,25 @@ import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 
 public class ImportTableCommand extends Command {
-  
+
   @Override
-  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, TableNotFoundException,
-      TableExistsException {
-    
+  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException, TableExistsException {
+
     shellState.getConnector().tableOperations().importTable(cl.getArgs()[0], cl.getArgs()[1]);
     return 0;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <table name> <import dir>";
   }
-  
+
   @Override
   public String description() {
     return "imports a table";
   }
-  
+
   @Override
   public int numArgs() {
     return 2;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/InsertCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/InsertCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/InsertCommand.java
index 16afc9e..05037d5 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/InsertCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/InsertCommand.java
@@ -48,29 +48,29 @@ public class InsertCommand extends Command {
   private Option insertOptAuths, timestampOpt;
   private Option timeoutOption;
   private Option durabilityOption;
-  
+
   protected long getTimeout(final CommandLine cl) {
     if (cl.hasOption(timeoutOption.getLongOpt())) {
       return AccumuloConfiguration.getTimeInMillis(cl.getOptionValue(timeoutOption.getLongOpt()));
     }
-    
+
     return Long.MAX_VALUE;
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
       TableNotFoundException, IOException, ConstraintViolationException {
     shellState.checkTableState();
-    
+
     final Mutation m = new Mutation(new Text(cl.getArgs()[0].getBytes(Shell.CHARSET)));
     final Text colf = new Text(cl.getArgs()[1].getBytes(Shell.CHARSET));
     final Text colq = new Text(cl.getArgs()[2].getBytes(Shell.CHARSET));
     final Value val = new Value(cl.getArgs()[3].getBytes(Shell.CHARSET));
-    
+
     if (cl.hasOption(insertOptAuths.getOpt())) {
       final ColumnVisibility le = new ColumnVisibility(cl.getOptionValue(insertOptAuths.getOpt()));
       Shell.log.debug("Authorization label will be set to: " + le.toString());
-      
+
       if (cl.hasOption(timestampOpt.getOpt()))
         m.put(colf, colq, le, Long.parseLong(cl.getOptionValue(timestampOpt.getOpt())), val);
       else
@@ -79,8 +79,9 @@ public class InsertCommand extends Command {
       m.put(colf, colq, Long.parseLong(cl.getOptionValue(timestampOpt.getOpt())), val);
     else
       m.put(colf, colq, val);
-    
-    final BatchWriterConfig cfg = new BatchWriterConfig().setMaxMemory(Math.max(m.estimatedMemoryUsed(), 1024)).setMaxWriteThreads(1).setTimeout(getTimeout(cl), TimeUnit.MILLISECONDS);
+
+    final BatchWriterConfig cfg = new BatchWriterConfig().setMaxMemory(Math.max(m.estimatedMemoryUsed(), 1024)).setMaxWriteThreads(1)
+        .setTimeout(getTimeout(cl), TimeUnit.MILLISECONDS);
     if (cl.hasOption(durabilityOption.getOpt())) {
       String userDurability = cl.getOptionValue(durabilityOption.getOpt());
       switch (userDurability) {
@@ -107,64 +108,64 @@ public class InsertCommand extends Command {
     } catch (MutationsRejectedException e) {
       final ArrayList<String> lines = new ArrayList<String>();
       if (e.getAuthorizationFailuresMap().isEmpty() == false) {
-        lines.add("	Authorization Failures:");
+        lines.add("\tAuthorization Failures:");
       }
       for (Entry<KeyExtent,Set<SecurityErrorCode>> entry : e.getAuthorizationFailuresMap().entrySet()) {
-        lines.add("		" + entry);
+        lines.add("\t\t" + entry);
       }
       if (e.getConstraintViolationSummaries().isEmpty() == false) {
-        lines.add("	Constraint Failures:");
+        lines.add("\tConstraint Failures:");
       }
       for (ConstraintViolationSummary cvs : e.getConstraintViolationSummaries()) {
-        lines.add("		" + cvs.toString());
+        lines.add("\t\t" + cvs.toString());
       }
-      
+
       if (lines.size() == 0 || e.getUnknownExceptions() > 0) {
         // must always print something
         lines.add(" " + e.getClass().getName() + " : " + e.getMessage());
         if (e.getCause() != null)
           lines.add("   Caused by : " + e.getCause().getClass().getName() + " : " + e.getCause().getMessage());
       }
-      
+
       shellState.printLines(lines.iterator(), false);
-      
+
       return 1;
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "inserts a record";
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <row> <colfamily> <colqualifier> <value>";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
     insertOptAuths = new Option("l", "visibility-label", true, "formatted visibility");
     insertOptAuths.setArgName("expression");
     o.addOption(insertOptAuths);
-    
+
     timestampOpt = new Option("ts", "timestamp", true, "timestamp to use for insert");
     timestampOpt.setArgName("timestamp");
     o.addOption(timestampOpt);
-    
+
     timeoutOption = new Option(null, "timeout", true,
         "time before insert should fail if no data is written. If no unit is given assumes seconds.  Units d,h,m,s,and ms are supported.  e.g. 30s or 100ms");
     timeoutOption.setArgName("timeout");
     o.addOption(timeoutOption);
-    
+
     durabilityOption = new Option("d", "durability", true, "durability to use for insert, should be one of \"none\" \"log\" \"flush\" or \"sync\"");
     o.addOption(durabilityOption);
-    
+
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 4;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/InterpreterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/InterpreterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/InterpreterCommand.java
index 9d79601..5478c75 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/InterpreterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/InterpreterCommand.java
@@ -21,19 +21,19 @@ import org.apache.accumulo.core.util.interpret.ScanInterpreter;
 import org.apache.accumulo.shell.Shell;
 
 /**
- * 
+ *
  */
 public class InterpreterCommand extends ShellPluginConfigurationCommand {
-  
+
   public InterpreterCommand() {
     super("interpreter", Property.TABLE_INTERPRETER_CLASS, "i");
   }
-  
+
   @Override
   public String description() {
     return "specifies a scan interpreter to interpret scan range and column arguments";
   }
-  
+
   public static Class<? extends ScanInterpreter> getCurrentInterpreter(final String tableName, final Shell shellState) {
     return ShellPluginConfigurationCommand.getPluginClass(tableName, shellState, ScanInterpreter.class, Property.TABLE_INTERPRETER_CLASS);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ListCompactionsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ListCompactionsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ListCompactionsCommand.java
index 809ef8c..e4b3410 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ListCompactionsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ListCompactionsCommand.java
@@ -27,52 +27,52 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class ListCompactionsCommand extends Command {
-  
+
   private Option tserverOption, disablePaginationOpt;
-  
+
   @Override
   public String description() {
     return "lists what compactions are currently running in accumulo. See the accumulo.core.client.admin.ActiveCompaciton javadoc for more information about columns.";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
-    
+
     List<String> tservers;
-    
+
     final InstanceOperations instanceOps = shellState.getConnector().instanceOperations();
-    
+
     final boolean paginate = !cl.hasOption(disablePaginationOpt.getOpt());
-    
+
     if (cl.hasOption(tserverOption.getOpt())) {
       tservers = new ArrayList<String>();
       tservers.add(cl.getOptionValue(tserverOption.getOpt()));
     } else {
       tservers = instanceOps.getTabletServers();
     }
-    
+
     shellState.printLines(new ActiveCompactionIterator(tservers, instanceOps), paginate);
-    
+
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();
-    
+
     tserverOption = new Option("ts", "tabletServer", true, "tablet server to list compactions for");
     tserverOption.setArgName("tablet server");
     opts.addOption(tserverOption);
-    
+
     disablePaginationOpt = new Option("np", "no-pagination", false, "disable pagination of output");
     opts.addOption(disablePaginationOpt);
-    
+
     return opts;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ListIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ListIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ListIterCommand.java
index fcebd1f..6187300 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ListIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ListIterCommand.java
@@ -119,7 +119,7 @@ public class ListIterCommand extends Command {
 
     allScopesOpt = new Option("all", "all-scopes", false, "list from all scopes");
     o.addOption(allScopesOpt);
-    
+
     scopeOpts = new EnumMap<IteratorScope,Option>(IteratorScope.class);
     scopeOpts.put(IteratorScope.minc, new Option(IteratorScope.minc.name(), "minor-compaction", false, "list iterator for minor compaction scope"));
     scopeOpts.put(IteratorScope.majc, new Option(IteratorScope.majc.name(), "major-compaction", false, "list iterator for major compaction scope"));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ListScansCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ListScansCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ListScansCommand.java
index 598503e..f89b6fc 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ListScansCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ListScansCommand.java
@@ -27,52 +27,52 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class ListScansCommand extends Command {
-  
+
   private Option tserverOption, disablePaginationOpt;
-  
+
   @Override
   public String description() {
     return "lists what scans are currently running in accumulo. See the accumulo.core.client.admin.ActiveScan javadoc for more information about columns.";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
-    
+
     List<String> tservers;
-    
+
     final InstanceOperations instanceOps = shellState.getConnector().instanceOperations();
-    
+
     final boolean paginate = !cl.hasOption(disablePaginationOpt.getOpt());
-    
+
     if (cl.hasOption(tserverOption.getOpt())) {
       tservers = new ArrayList<String>();
       tservers.add(cl.getOptionValue(tserverOption.getOpt()));
     } else {
       tservers = instanceOps.getTabletServers();
     }
-    
+
     shellState.printLines(new ActiveScanIterator(tservers, instanceOps), paginate);
-    
+
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();
-    
+
     tserverOption = new Option("ts", "tabletServer", true, "tablet server to list scans for");
     tserverOption.setArgName("tablet server");
     opts.addOption(tserverOption);
-    
+
     disablePaginationOpt = new Option("np", "no-pagination", false, "disable pagination of output");
     opts.addOption(disablePaginationOpt);
-    
+
     return opts;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ListShellIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ListShellIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ListShellIterCommand.java
index 59f8f46..d899d1d 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ListShellIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ListShellIterCommand.java
@@ -28,17 +28,17 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 /**
- * 
+ *
  */
 public class ListShellIterCommand extends Command {
-  
+
   private Option nameOpt, profileOpt;
 
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     if (shellState.iteratorProfiles.size() == 0)
       return 0;
-    
+
     final StringBuilder sb = new StringBuilder();
 
     String profile = null;
@@ -68,7 +68,7 @@ public class ListShellIterCommand extends Command {
         }
       }
     }
-    
+
     if (sb.length() > 0) {
       sb.append("-\n");
     }
@@ -77,26 +77,26 @@ public class ListShellIterCommand extends Command {
 
     return 0;
   }
-  
+
   public String description() {
     return "lists iterators profiles configured in shell";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     profileOpt = new Option("pn", "profile", true, "iterator profile name");
     profileOpt.setArgName("profile");
 
     nameOpt = new Option("n", "name", true, "iterator to list");
     nameOpt.setArgName("itername");
-    
+
     o.addOption(profileOpt);
     o.addOption(nameOpt);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/MaxRowCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/MaxRowCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/MaxRowCommand.java
index 7118b7e..de62716 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/MaxRowCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/MaxRowCommand.java
@@ -25,19 +25,19 @@ import org.apache.hadoop.io.Text;
 import org.apache.log4j.Logger;
 
 public class MaxRowCommand extends ScanCommand {
-  
+
   private static final Logger log = Logger.getLogger(MaxRowCommand.class);
-  
+
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     final ScanInterpreter interpeter = getInterpreter(cl, tableName, shellState);
-    
+
     final Range range = getRange(cl, interpeter);
     final Authorizations auths = getAuths(cl, shellState);
     final Text startRow = range.getStartKey() == null ? null : range.getStartKey().getRow();
     final Text endRow = range.getEndKey() == null ? null : range.getEndKey().getRow();
-    
+
     try {
       final Text max = shellState.getConnector().tableOperations()
           .getMaxRow(tableName, auths, startRow, range.isStartKeyInclusive(), endRow, range.isEndKeyInclusive());
@@ -47,10 +47,10 @@ public class MaxRowCommand extends ScanCommand {
     } catch (Exception e) {
       log.debug("Could not get shell state.", e);
     }
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "finds the max row in a table within a given range";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/MergeCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/MergeCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/MergeCommand.java
index 33d63fa..bd4843d 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/MergeCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/MergeCommand.java
@@ -29,7 +29,7 @@ import org.apache.hadoop.io.Text;
 
 public class MergeCommand extends Command {
   private Option verboseOpt, forceOpt, sizeOpt, allOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     boolean verbose = shellState.isVerbose();
@@ -78,17 +78,17 @@ public class MergeCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "merges tablets in a table";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -107,5 +107,5 @@ public class MergeCommand extends Command {
     o.addOption(allOpt);
     return o;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/NoTableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/NoTableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/NoTableCommand.java
index 7ff6358..9cd284f 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/NoTableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/NoTableCommand.java
@@ -24,15 +24,15 @@ public class NoTableCommand extends Command {
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     shellState.setTableName("");
-    
+
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "returns to a tableless shell state";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/OfflineCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/OfflineCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/OfflineCommand.java
index 6ac397c..a3b0105 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/OfflineCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/OfflineCommand.java
@@ -26,15 +26,15 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class OfflineCommand extends TableOperation {
-  
+
   private boolean wait;
   private Option waitOpt;
-  
+
   @Override
   public String description() {
     return "starts the process of taking table offline";
   }
-  
+
   protected void doTableOp(final Shell shellState, final String tableName) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     if (tableName.equals(MetadataTable.NAME)) {
       Shell.log.info("  You cannot take the " + MetadataTable.NAME + " offline.");
@@ -43,19 +43,18 @@ public class OfflineCommand extends TableOperation {
       Shell.log.info("Offline of table " + tableName + (wait ? " completed." : " initiated..."));
     }
   }
-  
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     wait = cl.hasOption(waitOpt.getLongOpt());
     return super.execute(fullCommand, cl, shellState);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = super.getOptions();
     waitOpt = new Option("w", "wait", false, "wait for offline to finish");
-    opts.addOption(waitOpt); 
+    opts.addOption(waitOpt);
     return opts;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/OnlineCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/OnlineCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/OnlineCommand.java
index ace069f..ecafb74 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/OnlineCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/OnlineCommand.java
@@ -26,15 +26,15 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 public class OnlineCommand extends TableOperation {
-  
+
   private boolean wait;
   private Option waitOpt;
-  
+
   @Override
   public String description() {
     return "starts the process of putting a table online";
   }
-  
+
   @Override
   protected void doTableOp(final Shell shellState, final String tableName) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     if (tableName.equals(RootTable.NAME)) {
@@ -44,18 +44,18 @@ public class OnlineCommand extends TableOperation {
       Shell.log.info("Online of table " + tableName + (wait ? " completed." : " initiated..."));
     }
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     wait = cl.hasOption(waitOpt.getLongOpt());
     return super.execute(fullCommand, cl, shellState);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = super.getOptions();
     waitOpt = new Option("w", "wait", false, "wait for online to finish");
-    opts.addOption(waitOpt); 
+    opts.addOption(waitOpt);
     return opts;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/PasswdCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/PasswdCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/PasswdCommand.java
index dd38889..6b8eff8 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/PasswdCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/PasswdCommand.java
@@ -19,6 +19,7 @@ package org.apache.accumulo.shell.commands;
 import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.io.IOException;
+
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode;
@@ -32,25 +33,25 @@ import org.apache.commons.cli.Options;
 
 public class PasswdCommand extends Command {
   private Option userOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, IOException {
     final String currentUser = shellState.getConnector().whoami();
     final String user = cl.getOptionValue(userOpt.getOpt(), currentUser);
-    
+
     String password = null;
     String passwordConfirm = null;
     String oldPassword = null;
-    
+
     oldPassword = shellState.readMaskedLine("Enter current password for '" + currentUser + "': ", '*');
     if (oldPassword == null) {
       shellState.getReader().println();
       return 0;
     } // user canceled
-    
+
     if (!shellState.getConnector().securityOperations().authenticateUser(currentUser, new PasswordToken(oldPassword)))
       throw new AccumuloSecurityException(user, SecurityErrorCode.BAD_CREDENTIALS);
-    
+
     password = shellState.readMaskedLine("Enter new password for '" + user + "': ", '*');
     if (password == null) {
       shellState.getReader().println();
@@ -61,7 +62,7 @@ public class PasswdCommand extends Command {
       shellState.getReader().println();
       return 0;
     } // user canceled
-    
+
     if (!password.equals(passwordConfirm)) {
       throw new IllegalArgumentException("Passwords do not match");
     }
@@ -75,12 +76,12 @@ public class PasswdCommand extends Command {
     Shell.log.debug("Changed password for user " + user);
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "changes a user's password";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -89,7 +90,7 @@ public class PasswdCommand extends Command {
     o.addOption(userOpt);
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/PingCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/PingCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/PingCommand.java
index ef7a9e4..ba6de68 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/PingCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/PingCommand.java
@@ -27,56 +27,55 @@ import org.apache.commons.cli.Option;
 import org.apache.commons.cli.Options;
 
 /**
- * 
+ *
  */
 public class PingCommand extends Command {
-  
+
   private Option tserverOption, disablePaginationOpt;
-  
+
   @Override
   public String description() {
     return "ping tablet servers";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
-    
+
     List<String> tservers;
-    
+
     final InstanceOperations instanceOps = shellState.getConnector().instanceOperations();
-    
+
     final boolean paginate = !cl.hasOption(disablePaginationOpt.getOpt());
-    
+
     if (cl.hasOption(tserverOption.getOpt())) {
       tservers = new ArrayList<String>();
       tservers.add(cl.getOptionValue(tserverOption.getOpt()));
     } else {
       tservers = instanceOps.getTabletServers();
     }
-    
+
     shellState.printLines(new PingIterator(tservers, instanceOps), paginate);
-    
+
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();
-    
+
     tserverOption = new Option("ts", "tabletServer", true, "tablet server to ping");
     tserverOption.setArgName("tablet server");
     opts.addOption(tserverOption);
-    
+
     disablePaginationOpt = new Option("np", "no-pagination", false, "disable pagination of output");
     opts.addOption(disablePaginationOpt);
-    
+
     return opts;
   }
-  
-}
 
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/PingIterator.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/PingIterator.java b/shell/src/main/java/org/apache/accumulo/shell/commands/PingIterator.java
index e414ed4..4072ec0 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/PingIterator.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/PingIterator.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.admin.InstanceOperations;
 
 class PingIterator implements Iterator<String> {
-  
+
   private Iterator<String> iter;
   private InstanceOperations instanceOps;
 
@@ -31,28 +31,28 @@ class PingIterator implements Iterator<String> {
     iter = tservers.iterator();
     this.instanceOps = instanceOps;
   }
-  
+
   @Override
   public boolean hasNext() {
     return iter.hasNext();
   }
-  
+
   @Override
   public String next() {
     String tserver = iter.next();
-    
+
     try {
       instanceOps.ping(tserver);
     } catch (AccumuloException e) {
       return tserver + " ERROR " + e.getMessage();
     }
-    
+
     return tserver + " OK";
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java b/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
index 2fdb7fe..1f3a1ae 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
@@ -27,16 +27,16 @@ import org.apache.accumulo.shell.Shell;
 
 /**
  * A basic tokenizer for generating tokens from a string. It understands quoted strings and escaped quote characters.
- * 
+ *
  * You can use the escape sequence '\' to escape single quotes, double quotes, and spaces only, in addition to the escape character itself.
- * 
+ *
  * The behavior is the same for single and double quoted strings. (i.e. '\'' is the same as "\'")
  */
 
 public class QuotedStringTokenizer implements Iterable<String> {
   private ArrayList<String> tokens;
   private String input;
-  
+
   public QuotedStringTokenizer(final String t) throws BadArgumentException {
     tokens = new ArrayList<String>();
     this.input = t;
@@ -46,23 +46,23 @@ public class QuotedStringTokenizer implements Iterable<String> {
       throw new IllegalArgumentException(e.getMessage());
     }
   }
-  
+
   public String[] getTokens() {
     return tokens.toArray(new String[tokens.size()]);
   }
-  
+
   private void createTokens() throws BadArgumentException, UnsupportedEncodingException {
     boolean inQuote = false;
     boolean inEscapeSequence = false;
     String hexChars = null;
     char inQuoteChar = '"';
-    
+
     final byte[] token = new byte[input.length()];
     int tokenLength = 0;
     final byte[] inputBytes = input.getBytes(UTF_8);
     for (int i = 0; i < input.length(); ++i) {
       final char ch = input.charAt(i);
-      
+
       // if I ended up in an escape sequence, check for valid escapable character, and add it as a literal
       if (inEscapeSequence) {
         inEscapeSequence = false;
@@ -134,7 +134,7 @@ public class QuotedStringTokenizer implements Iterable<String> {
       tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
     }
   }
-  
+
   @Override
   public Iterator<String> iterator() {
     return tokens.iterator();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/RenameNamespaceCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/RenameNamespaceCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/RenameNamespaceCommand.java
index f456a30..6d0cdd3 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/RenameNamespaceCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/RenameNamespaceCommand.java
@@ -29,8 +29,8 @@ import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.impl.Namespaces;
 import org.apache.accumulo.core.client.impl.Tables;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 
 public class RenameNamespaceCommand extends Command {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/RenameTableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/RenameTableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/RenameTableCommand.java
index a810320..8630f81 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/RenameTableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/RenameTableCommand.java
@@ -25,8 +25,8 @@ import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.impl.Tables;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 
 public class RenameTableCommand extends Command {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/RevokeCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/RevokeCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/RevokeCommand.java
index 6737c45..935b6cd 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/RevokeCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/RevokeCommand.java
@@ -24,9 +24,9 @@ import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
 import org.apache.accumulo.core.util.BadArgumentException;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ScanCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ScanCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ScanCommand.java
index 46909e4..5917b1e 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ScanCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ScanCommand.java
@@ -47,41 +47,41 @@ import org.apache.commons.cli.Options;
 import org.apache.hadoop.io.Text;
 
 public class ScanCommand extends Command {
-  
+
   private Option scanOptAuths, scanOptRow, scanOptColumns, disablePaginationOpt, showFewOpt, formatterOpt, interpreterOpt, formatterInterpeterOpt,
       outputFileOpt;
-  
+
   protected Option timestampOpt;
   private Option optStartRowExclusive;
   private Option optEndRowExclusive;
   private Option timeoutOption;
   private Option profileOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final PrintFile printFile = getOutputFile(cl);
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     final Class<? extends Formatter> formatter = getFormatter(cl, tableName, shellState);
     final ScanInterpreter interpeter = getInterpreter(cl, tableName, shellState);
-    
+
     // handle first argument, if present, the authorizations list to
     // scan with
     final Authorizations auths = getAuths(cl, shellState);
     final Scanner scanner = shellState.getConnector().createScanner(tableName, auths);
-    
+
     // handle session-specific scan iterators
     addScanIterators(shellState, cl, scanner, tableName);
-    
+
     // handle remaining optional arguments
     scanner.setRange(getRange(cl, interpeter));
-    
+
     // handle columns
     fetchColumns(cl, scanner, interpeter);
-    
+
     // set timeout
     scanner.setTimeout(getTimeout(cl), TimeUnit.MILLISECONDS);
-    
+
     // output the records
     if (cl.hasOption(showFewOpt.getOpt())) {
       final String showLength = cl.getOptionValue(showFewOpt.getOpt());
@@ -97,32 +97,32 @@ public class ScanCommand extends Command {
       } catch (IllegalArgumentException iae) {
         shellState.getReader().println("Arg must be greater than one.");
       }
-      
+
     } else {
       printRecords(cl, shellState, scanner, formatter, printFile);
     }
     if (printFile != null) {
       printFile.close();
     }
-    
+
     return 0;
   }
-  
+
   protected long getTimeout(final CommandLine cl) {
     if (cl.hasOption(timeoutOption.getLongOpt())) {
       return AccumuloConfiguration.getTimeInMillis(cl.getOptionValue(timeoutOption.getLongOpt()));
     }
-    
+
     return Long.MAX_VALUE;
   }
-  
+
   protected void addScanIterators(final Shell shellState, CommandLine cl, final Scanner scanner, final String tableName) {
-    
+
     List<IteratorSetting> tableScanIterators;
     if (cl.hasOption(profileOpt.getOpt())) {
       String profile = cl.getOptionValue(profileOpt.getOpt());
       tableScanIterators = shellState.iteratorProfiles.get(profile);
-      
+
       if (tableScanIterators == null) {
         throw new IllegalArgumentException("Profile " + profile + " does not exist");
       }
@@ -133,9 +133,9 @@ public class ScanCommand extends Command {
         return;
       }
     }
-    
+
     Shell.log.debug("Found " + tableScanIterators.size() + " scan iterators to set");
-    
+
     for (IteratorSetting setting : tableScanIterators) {
       Shell.log.debug("Setting scan iterator " + setting.getName() + " at priority " + setting.getPriority() + " using class name "
           + setting.getIteratorClass());
@@ -145,12 +145,12 @@ public class ScanCommand extends Command {
       scanner.addScanIterator(setting);
     }
   }
-  
+
   protected void printRecords(final CommandLine cl, final Shell shellState, final Iterable<Entry<Key,Value>> scanner, final Class<? extends Formatter> formatter)
       throws IOException {
     printRecords(cl, shellState, scanner, formatter, null);
   }
-  
+
   protected void printRecords(final CommandLine cl, final Shell shellState, final Iterable<Entry<Key,Value>> scanner,
       final Class<? extends Formatter> formatter, PrintFile outFile) throws IOException {
     if (outFile == null) {
@@ -159,11 +159,11 @@ public class ScanCommand extends Command {
       shellState.printRecords(scanner, cl.hasOption(timestampOpt.getOpt()), !cl.hasOption(disablePaginationOpt.getOpt()), formatter, outFile);
     }
   }
-  
+
   protected void printBinaryRecords(final CommandLine cl, final Shell shellState, final Iterable<Entry<Key,Value>> scanner) throws IOException {
     printBinaryRecords(cl, shellState, scanner, null);
   }
-  
+
   protected void printBinaryRecords(final CommandLine cl, final Shell shellState, final Iterable<Entry<Key,Value>> scanner, PrintFile outFile)
       throws IOException {
     if (outFile == null) {
@@ -172,9 +172,9 @@ public class ScanCommand extends Command {
       shellState.printBinaryRecords(scanner, cl.hasOption(timestampOpt.getOpt()), !cl.hasOption(disablePaginationOpt.getOpt()), outFile);
     }
   }
-  
+
   protected ScanInterpreter getInterpreter(final CommandLine cl, final String tableName, final Shell shellState) throws Exception {
-    
+
     Class<? extends ScanInterpreter> clazz = null;
     try {
       if (cl.hasOption(interpreterOpt.getOpt())) {
@@ -185,18 +185,18 @@ public class ScanCommand extends Command {
     } catch (ClassNotFoundException e) {
       shellState.getReader().println("Interpreter class could not be loaded.\n" + e.getMessage());
     }
-    
+
     if (clazz == null)
       clazz = InterpreterCommand.getCurrentInterpreter(tableName, shellState);
-    
+
     if (clazz == null)
       clazz = DefaultScanInterpreter.class;
-    
+
     return clazz.newInstance();
   }
-  
+
   protected Class<? extends Formatter> getFormatter(final CommandLine cl, final String tableName, final Shell shellState) throws IOException {
-    
+
     try {
       if (cl.hasOption(formatterOpt.getOpt())) {
         return shellState.getClassLoader(cl, shellState).loadClass(cl.getOptionValue(formatterOpt.getOpt())).asSubclass(Formatter.class);
@@ -206,10 +206,10 @@ public class ScanCommand extends Command {
     } catch (Exception e) {
       shellState.getReader().println("Formatter class could not be loaded.\n" + e.getMessage());
     }
-    
+
     return shellState.getFormatter(tableName);
   }
-  
+
   protected void fetchColumns(final CommandLine cl, final ScannerBase scanner, final ScanInterpreter formatter) throws UnsupportedEncodingException {
     if (cl.hasOption(scanOptColumns.getOpt())) {
       for (String a : cl.getOptionValue(scanOptColumns.getOpt()).split(",")) {
@@ -223,14 +223,14 @@ public class ScanCommand extends Command {
       }
     }
   }
-  
+
   protected Range getRange(final CommandLine cl, final ScanInterpreter formatter) throws UnsupportedEncodingException {
     if ((cl.hasOption(OptUtil.START_ROW_OPT) || cl.hasOption(OptUtil.END_ROW_OPT)) && cl.hasOption(scanOptRow.getOpt())) {
       // did not see a way to make commons cli do this check... it has mutually exclusive options but does not support the or
       throw new IllegalArgumentException("Options -" + scanOptRow.getOpt() + " AND (-" + OptUtil.START_ROW_OPT + " OR -" + OptUtil.END_ROW_OPT
           + ") are mutally exclusive ");
     }
-    
+
     if (cl.hasOption(scanOptRow.getOpt())) {
       return new Range(formatter.interpretRow(new Text(cl.getOptionValue(scanOptRow.getOpt()).getBytes(Shell.CHARSET))));
     } else {
@@ -245,7 +245,7 @@ public class ScanCommand extends Command {
       return new Range(startRow, startInclusive, endRow, endInclusive);
     }
   }
-  
+
   protected Authorizations getAuths(final CommandLine cl, final Shell shellState) throws AccumuloSecurityException, AccumuloException {
     final String user = shellState.getConnector().whoami();
     Authorizations auths = shellState.getConnector().securityOperations().getUserAuthorizations(user);
@@ -254,23 +254,23 @@ public class ScanCommand extends Command {
     }
     return auths;
   }
-  
+
   static Authorizations parseAuthorizations(final String field) {
     if (field == null || field.isEmpty()) {
       return Authorizations.EMPTY;
     }
     return new Authorizations(field.split(","));
   }
-  
+
   @Override
   public String description() {
     return "scans the table, and displays the resulting records";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     scanOptAuths = new Option("s", "scan-authorizations", true, "scan authorizations (all user auths are used if this argument is not specified)");
     optStartRowExclusive = new Option("be", "begin-exclusive", false, "make start row exclusive (by default it's inclusive)");
     optStartRowExclusive.setArgName("begin-exclusive");
@@ -287,7 +287,7 @@ public class ScanCommand extends Command {
     timeoutOption = new Option(null, "timeout", true,
         "time before scan should fail if no data is returned. If no unit is given assumes seconds.  Units d,h,m,s,and ms are supported.  e.g. 30s or 100ms");
     outputFileOpt = new Option("o", "output", true, "local file to write the scan output to");
-    
+
     scanOptAuths.setArgName("comma-separated-authorizations");
     scanOptRow.setArgName("row");
     scanOptColumns.setArgName("<columnfamily>[:<columnqualifier>]{,<columnfamily>[:<columnqualifier>]}");
@@ -296,10 +296,10 @@ public class ScanCommand extends Command {
     formatterOpt.setArgName("className");
     timeoutOption.setArgName("timeout");
     outputFileOpt.setArgName("file");
-    
+
     profileOpt = new Option("pn", "profile", true, "iterator profile name");
     profileOpt.setArgName("profile");
-    
+
     o.addOption(scanOptAuths);
     o.addOption(scanOptRow);
     o.addOption(OptUtil.startRowOpt());
@@ -317,15 +317,15 @@ public class ScanCommand extends Command {
     o.addOption(timeoutOption);
     o.addOption(outputFileOpt);
     o.addOption(profileOpt);
-    
+
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
   protected PrintFile getOutputFile(final CommandLine cl) throws FileNotFoundException {
     final String outputFile = cl.getOptionValue(outputFileOpt.getOpt());
     return (outputFile == null ? null : new PrintFile(outputFile));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ScriptCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ScriptCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ScriptCommand.java
index 9bd91cc..7709798 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ScriptCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ScriptCommand.java
@@ -46,18 +46,18 @@ import org.apache.commons.cli.OptionGroup;
 import org.apache.commons.cli.Options;
 
 public class ScriptCommand extends Command {
-  
+
   // Command to allow user to run scripts, see JSR-223
   // http://www.oracle.com/technetwork/articles/javase/scripting-140262.html
-  
+
   protected Option list, engine, script, file, args, out, function, object;
   private static final String DEFAULT_ENGINE = "rhino";
-  
+
   public int execute(String fullCommand, CommandLine cl, Shell shellState) throws Exception {
 
     boolean invoke = false;
     ScriptEngineManager mgr = new ScriptEngineManager();
-    
+
     if (cl.hasOption(list.getOpt())) {
       listJSREngineInfo(mgr, shellState);
     } else if (cl.hasOption(file.getOpt()) || cl.hasOption(script.getOpt())) {
@@ -70,7 +70,7 @@ public class ScriptCommand extends Command {
         shellState.printException(new Exception(engineName + " not found"));
         return 1;
       }
-      
+
       if (cl.hasOption(object.getOpt()) || cl.hasOption(function.getOpt())) {
         if (!(engine instanceof Invocable)) {
           shellState.printException(new Exception(engineName + " does not support invoking functions or methods"));
@@ -78,15 +78,15 @@ public class ScriptCommand extends Command {
         }
         invoke = true;
       }
-      
+
       ScriptContext ctx = new SimpleScriptContext();
-      
+
       // Put the following objects into the context so that they
       // are available to the scripts
       // TODO: What else should go in here?
       Bindings b = engine.getBindings(ScriptContext.ENGINE_SCOPE);
       b.put("connection", shellState.getConnector());
-      
+
       List<Object> argValues = new ArrayList<Object>();
       if (cl.hasOption(args.getOpt())) {
         String[] argList = cl.getOptionValue(args.getOpt()).split(",");
@@ -105,14 +105,14 @@ public class ScriptCommand extends Command {
       }
       ctx.setBindings(b, ScriptContext.ENGINE_SCOPE);
       Object[] argArray = argValues.toArray(new Object[argValues.size()]);
-      
+
       Writer writer = null;
       if (cl.hasOption(out.getOpt())) {
         File f = new File(cl.getOptionValue(out.getOpt()));
         writer = new FileWriter(f);
         ctx.setWriter(writer);
       }
-      
+
       if (cl.hasOption(file.getOpt())) {
         File f = new File(cl.getOptionValue(file.getOpt()));
         if (!f.exists()) {
@@ -165,36 +165,36 @@ public class ScriptCommand extends Command {
       if (null != writer) {
         writer.close();
       }
-      
+
     } else {
       printHelp(shellState);
     }
     return 0;
   }
-  
+
   public String description() {
     return "execute JSR-223 scripts";
   }
-  
+
   public int numArgs() {
     return 0;
   }
-  
+
   @Override
   public String getName() {
     return "script";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
-    
+
     engine = new Option("e", "engine", false, "engine name, defaults to JDK default (Rhino)");
     engine.setArgName("engineName");
     engine.setArgs(1);
     engine.setRequired(false);
     o.addOption(engine);
-    
+
     OptionGroup inputGroup = new OptionGroup();
     list = new Option("l", "list", false, "list available script engines");
     inputGroup.addOption(list);
@@ -204,23 +204,23 @@ public class ScriptCommand extends Command {
     script.setArgs(1);
     script.setRequired(false);
     inputGroup.addOption(script);
-    
+
     file = new Option("f", "file", true, "use script file");
     file.setArgName("fileName");
     file.setArgs(1);
     file.setRequired(false);
-    
+
     inputGroup.addOption(file);
     inputGroup.setRequired(true);
     o.addOptionGroup(inputGroup);
-    
+
     OptionGroup invokeGroup = new OptionGroup();
     object = new Option("obj", "object", true, "name of object");
     object.setArgs(1);
     object.setArgName("objectName:methodName");
     object.setRequired(false);
     invokeGroup.addOption(object);
-    
+
     function = new Option("fx", "function", true, "invoke a script function");
     function.setArgName("functionName");
     function.setArgs(1);
@@ -228,22 +228,22 @@ public class ScriptCommand extends Command {
     invokeGroup.addOption(function);
     invokeGroup.setRequired(false);
     o.addOptionGroup(invokeGroup);
-    
+
     args = new Option("a", "args", true, "comma separated list of key=value arguments");
     args.setArgName("property1=value1,propert2=value2,...");
     args.setArgs(Option.UNLIMITED_VALUES);
     args.setRequired(false);
     o.addOption(args);
-    
+
     out = new Option("o", "output", true, "output file");
     out.setArgName("fileName");
     out.setArgs(1);
     out.setRequired(false);
     o.addOption(out);
-    
+
     return o;
   }
-  
+
   private void listJSREngineInfo(ScriptEngineManager mgr, Shell shellState) throws IOException {
     List<ScriptEngineFactory> factories = mgr.getEngineFactories();
     Set<String> lines = new TreeSet<String>();
@@ -261,9 +261,9 @@ public class ScriptCommand extends Command {
       lines.add("\tLanguage: " + langName + " (" + langVersion + ")");
     }
     shellState.printLines(lines.iterator(), true);
-    
+
   }
-  
+
   private void invokeFunctionOrMethod(Shell shellState, ScriptEngine engine, CommandLine cl, Object[] args) {
     try {
       Invocable inv = (Invocable) engine;
@@ -280,11 +280,11 @@ public class ScriptCommand extends Command {
         String methodName = parts[1];
         Object obj = engine.get(objectName);
         inv.invokeMethod(obj, methodName, args);
-        
+
       }
     } catch (Exception e) {
       shellState.printException(e);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SetAuthsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SetAuthsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SetAuthsCommand.java
index 5c1d73f..3c22c1b 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SetAuthsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SetAuthsCommand.java
@@ -22,9 +22,9 @@ import java.util.Set;
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;
@@ -34,7 +34,7 @@ public class SetAuthsCommand extends Command {
   private Option userOpt;
   private Option scanOptAuths;
   private Option clearOptAuths;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException {
     final String user = cl.getOptionValue(userOpt.getOpt(), shellState.getConnector().whoami());
@@ -43,17 +43,17 @@ public class SetAuthsCommand extends Command {
     Shell.log.debug("Changed record-level authorizations for user " + user);
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "sets the maximum scan authorizations for a user";
   }
-  
+
   @Override
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> completionSet) {
     registerCompletionForUsers(root, completionSet);
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
@@ -70,7 +70,7 @@ public class SetAuthsCommand extends Command {
     o.addOption(userOpt);
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SetGroupsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SetGroupsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SetGroupsCommand.java
index fc8b06e..62ed1a2 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SetGroupsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SetGroupsCommand.java
@@ -31,13 +31,13 @@ public class SetGroupsCommand extends Command {
   public String description() {
     return "sets the locality groups for a given table (for binary or commas, use Java API)";
   }
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     final HashMap<String,Set<Text>> groups = new HashMap<String,Set<Text>>();
-    
+
     for (String arg : cl.getArgs()) {
       final String sa[] = arg.split("=", 2);
       if (sa.length < 2) {
@@ -45,34 +45,34 @@ public class SetGroupsCommand extends Command {
       }
       final String group = sa[0];
       final HashSet<Text> colFams = new HashSet<Text>();
-      
+
       for (String family : sa[1].split(",")) {
         colFams.add(new Text(family.getBytes(Shell.CHARSET)));
       }
-      
+
       groups.put(group, colFams);
     }
-    
+
     shellState.getConnector().tableOperations().setLocalityGroups(tableName, groups);
-    
+
     return 0;
   }
-  
+
   @Override
   public int numArgs() {
     return Shell.NO_FIXED_ARG_LENGTH_CHECK;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <group>=<col fam>{,<col fam>}{ <group>=<col fam>{,<col fam>}}";
   }
-  
+
   @Override
   public Options getOptions() {
     final Options opts = new Options();
     opts.addOption(OptUtil.tableOpt("table to fetch locality groups for"));
     return opts;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SetIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SetIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SetIterCommand.java
index 2bb389b..c7c41b2 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SetIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SetIterCommand.java
@@ -88,7 +88,7 @@ public class SetIterCommand extends Command {
 
     // Try to get the name provided by the setiter command
     String name = cl.getOptionValue(nameOpt.getOpt(), null);
-    
+
     // Cannot continue if no name is provided
     if (null == name && null == configuredName) {
       throw new IllegalArgumentException("No provided or default name for iterator");
@@ -235,7 +235,7 @@ public class SetIterCommand extends Command {
     if (null != iterOptions) {
       final IteratorOptions itopts = iterOptions.describeOptions();
       iteratorName = itopts.getName();
-      
+
       if (iteratorName == null) {
         throw new IllegalArgumentException(className + " described its default distinguishing name as null");
       }
@@ -250,9 +250,9 @@ public class SetIterCommand extends Command {
           options.remove(key);
         }
         localOptions.clear();
-  
+
         reader.println(itopts.getDescription());
-  
+
         String prompt;
         if (itopts.getNamedOptions() != null) {
           for (Entry<String,String> e : itopts.getNamedOptions().entrySet()) {
@@ -269,7 +269,7 @@ public class SetIterCommand extends Command {
             localOptions.put(e.getKey(), input);
           }
         }
-  
+
         if (itopts.getUnnamedOptionDescriptions() != null) {
           for (String desc : itopts.getUnnamedOptionDescriptions()) {
             reader.println(Shell.repeat("-", 10) + "> entering options: " + desc);
@@ -284,20 +284,20 @@ public class SetIterCommand extends Command {
               } else {
                 input = new String(input);
               }
-  
+
               if (input.length() == 0)
                 break;
-  
+
               String[] sa = input.split(" ", 2);
               localOptions.put(sa[0], sa[1]);
             }
           }
         }
-  
+
         options.putAll(localOptions);
         if (!iterOptions.validateOptions(options))
           reader.println("invalid options for " + clazz.getName());
-  
+
       } while (!iterOptions.validateOptions(options));
     } else {
       reader.flush();
@@ -310,12 +310,12 @@ public class SetIterCommand extends Command {
         // Treat whitespace or empty string as no name provided
         iteratorName = null;
       }
-      
+
       reader.flush();
       reader.println("Optional, configure name-value options for iterator:");
       String prompt = Shell.repeat("-", 10) + "> set option (<name> <value>, hit enter to skip): ";
       final HashMap<String,String> localOptions = new HashMap<String,String>();
-      
+
       while (true) {
         reader.flush();
         input = reader.readLine(prompt);
@@ -324,15 +324,15 @@ public class SetIterCommand extends Command {
           throw new IOException("Input stream closed");
         } else if (StringUtils.isWhitespace(input)) {
           break;
-        } 
+        }
 
         String[] sa = input.split(" ", 2);
         localOptions.put(sa[0], sa[1]);
       }
-      
+
       options.putAll(localOptions);
     }
-    
+
     return iteratorName;
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SetScanIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SetScanIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SetScanIterCommand.java
index 3ae42dd..40cdd75 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SetScanIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SetScanIterCommand.java
@@ -42,16 +42,16 @@ import org.apache.commons.cli.Options;
 
 public class SetScanIterCommand extends SetIterCommand {
   @Override
-  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, TableNotFoundException,
-      IOException, ShellCommandException {
+  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException, IOException, ShellCommandException {
     Shell.log.warn("Deprecated, use " + new SetShellIterCommand().getName());
     return super.execute(fullCommand, cl, shellState);
   }
-  
+
   @Override
   protected void setTableProperties(final CommandLine cl, final Shell shellState, final int priority, final Map<String,String> options, final String classname,
       final String name) throws AccumuloException, AccumuloSecurityException, ShellCommandException, TableNotFoundException {
-    
+
     final String tableName = OptUtil.getTableOpt(cl, shellState);
 
     // instead of setting table properties, just put the options in a list to use at scan time
@@ -64,10 +64,9 @@ public class SetScanIterCommand extends SetIterCommand {
     try {
       loadClass.asSubclass(SortedKeyValueIterator.class);
     } catch (ClassCastException ex) {
-      throw new ShellCommandException(ErrorCode.INITIALIZATION_FAILURE, "Unable to load " + classname  + " as type "
-          + SortedKeyValueIterator.class.getName());
+      throw new ShellCommandException(ErrorCode.INITIALIZATION_FAILURE, "Unable to load " + classname + " as type " + SortedKeyValueIterator.class.getName());
     }
-    
+
     for (Iterator<Entry<String,String>> i = options.entrySet().iterator(); i.hasNext();) {
       final Entry<String,String> entry = i.next();
       if (entry.getValue() == null || entry.getValue().isEmpty()) {
@@ -82,7 +81,7 @@ public class SetScanIterCommand extends SetIterCommand {
     }
     final IteratorSetting setting = new IteratorSetting(priority, name, classname);
     setting.addOptions(options);
-    
+
     // initialize a scanner to ensure the new setting does not conflict with existing settings
     final String user = shellState.getConnector().whoami();
     final Authorizations auths = shellState.getConnector().securityOperations().getUserAuthorizations(user);
@@ -91,17 +90,17 @@ public class SetScanIterCommand extends SetIterCommand {
       scanner.addScanIterator(s);
     }
     scanner.addScanIterator(setting);
-    
+
     // if no exception has been thrown, it's safe to add it to the list
     tableScanIterators.add(setting);
     Shell.log.debug("Scan iterators :" + shellState.scanIteratorOptions.get(tableName));
   }
-  
+
   @Override
   public String description() {
     return "sets a table-specific scan iterator for this shell session";
   }
-  
+
   @Override
   public Options getOptions() {
     // Remove the options that specify which type of iterator this is, since
@@ -123,5 +122,5 @@ public class SetScanIterCommand extends SetIterCommand {
     }
     return modifiedOptions;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SetShellIterCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SetShellIterCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SetShellIterCommand.java
index c7fea56..10c67be 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SetShellIterCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SetShellIterCommand.java
@@ -46,12 +46,12 @@ public class SetShellIterCommand extends SetIterCommand {
       TableNotFoundException, IOException, ShellCommandException {
     return super.execute(fullCommand, cl, shellState);
   }
-  
+
   @Override
   protected void setTableProperties(final CommandLine cl, final Shell shellState, final int priority, final Map<String,String> options, final String classname,
       final String name) throws AccumuloException, AccumuloSecurityException, ShellCommandException, TableNotFoundException {
     // instead of setting table properties, just put the options in a list to use at scan time
-    
+
     String profile = cl.getOptionValue(profileOpt.getOpt());
 
     // instead of setting table properties, just put the options in a list to use at scan time
@@ -64,10 +64,9 @@ public class SetShellIterCommand extends SetIterCommand {
     try {
       loadClass.asSubclass(SortedKeyValueIterator.class);
     } catch (ClassCastException ex) {
-      throw new ShellCommandException(ErrorCode.INITIALIZATION_FAILURE, "xUnable to load " + classname  + " as type "
-          + SortedKeyValueIterator.class.getName());
+      throw new ShellCommandException(ErrorCode.INITIALIZATION_FAILURE, "xUnable to load " + classname + " as type " + SortedKeyValueIterator.class.getName());
     }
-    
+
     for (Iterator<Entry<String,String>> i = options.entrySet().iterator(); i.hasNext();) {
       final Entry<String,String> entry = i.next();
       if (entry.getValue() == null || entry.getValue().isEmpty()) {
@@ -92,12 +91,12 @@ public class SetShellIterCommand extends SetIterCommand {
 
     tableScanIterators.add(setting);
   }
-  
+
   @Override
   public String description() {
     return "adds an iterator to a profile for this shell session";
   }
-  
+
   @Override
   public Options getOptions() {
     // Remove the options that specify which type of iterator this is, since
@@ -118,14 +117,14 @@ public class SetShellIterCommand extends SetIterCommand {
     for (OptionGroup group : groups) {
       modifiedOptions.addOptionGroup(group);
     }
-    
+
     profileOpt = new Option("pn", "profile", true, "iterator profile name");
     profileOpt.setRequired(true);
     profileOpt.setArgName("profile");
-    
+
     modifiedOptions.addOption(profileOpt);
 
     return modifiedOptions;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/ShellPluginConfigurationCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/ShellPluginConfigurationCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/ShellPluginConfigurationCommand.java
index 8f99eec..441de2a 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/ShellPluginConfigurationCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/ShellPluginConfigurationCommand.java
@@ -36,13 +36,13 @@ import org.apache.log4j.Logger;
 
 public abstract class ShellPluginConfigurationCommand extends Command {
   private Option removePluginOption, pluginClassOption, listPluginOption;
-  
+
   private String pluginType;
-  
+
   private Property tableProp;
-  
+
   private String classOpt;
-  
+
   ShellPluginConfigurationCommand(final String typeName, final Property tableProp, final String classOpt) {
     this.pluginType = typeName;
     this.tableProp = tableProp;
@@ -52,19 +52,19 @@ public abstract class ShellPluginConfigurationCommand extends Command {
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final String tableName = OptUtil.getTableOpt(cl, shellState);
-    
+
     if (cl.hasOption(removePluginOption.getOpt())) {
       // Remove the property
       removePlugin(cl, shellState, tableName);
-      
-      shellState.getReader().println("Removed "+pluginType+" on " + tableName);
+
+      shellState.getReader().println("Removed " + pluginType + " on " + tableName);
     } else if (cl.hasOption(listPluginOption.getOpt())) {
       // Get the options for this table
       final Iterator<Entry<String,String>> iter = shellState.getConnector().tableOperations().getProperties(tableName).iterator();
-      
+
       while (iter.hasNext()) {
         Entry<String,String> ent = iter.next();
-        
+
         // List all parameters with the property name
         if (ent.getKey().startsWith(tableProp.toString())) {
           shellState.getReader().println(ent.getKey() + ": " + ent.getValue());
@@ -73,22 +73,23 @@ public abstract class ShellPluginConfigurationCommand extends Command {
     } else {
       // Set the plugin with the provided options
       String className = cl.getOptionValue(pluginClassOption.getOpt());
-      
+
       // Set the plugin property on the table
       setPlugin(cl, shellState, tableName, className);
     }
-    
+
     return 0;
   }
 
-  protected void setPlugin(final CommandLine cl, final Shell shellState, final String tableName, final String className) throws AccumuloException, AccumuloSecurityException {
+  protected void setPlugin(final CommandLine cl, final Shell shellState, final String tableName, final String className) throws AccumuloException,
+      AccumuloSecurityException {
     shellState.getConnector().tableOperations().setProperty(tableName, tableProp.toString(), className);
   }
-  
+
   protected void removePlugin(final CommandLine cl, final Shell shellState, final String tableName) throws AccumuloException, AccumuloSecurityException {
     shellState.getConnector().tableOperations().removeProperty(tableName, tableProp.toString());
   }
-  
+
   public static <T> Class<? extends T> getPluginClass(final String tableName, final Shell shellState, final Class<T> clazz, final Property pluginProp) {
     Iterator<Entry<String,String>> props;
     try {
@@ -98,7 +99,7 @@ public abstract class ShellPluginConfigurationCommand extends Command {
     } catch (TableNotFoundException e) {
       return null;
     }
-    
+
     while (props.hasNext()) {
       final Entry<String,String> ent = props.next();
       if (ent.getKey().equals(pluginProp.toString())) {
@@ -124,40 +125,40 @@ public abstract class ShellPluginConfigurationCommand extends Command {
           Logger.getLogger(ShellPluginConfigurationCommand.class).error("Error: " + e.getMessage());
           return null;
         }
-        
+
         return pluginClazz;
       }
     }
-    
+
     return null;
   }
-  
+
   @Override
   public Options getOptions() {
     final Options o = new Options();
     final OptionGroup actionGroup = new OptionGroup();
-    
+
     pluginClassOption = new Option(classOpt, pluginType, true, "fully qualified name of the " + pluginType + " class to use");
     pluginClassOption.setArgName("className");
-    
+
     // Action to take: apply (default), remove, list
-    removePluginOption = new Option("r", "remove", false, "remove the current "+pluginType+"");
-    listPluginOption = new Option("l", "list", false, "display the current "+pluginType+"");
-    
+    removePluginOption = new Option("r", "remove", false, "remove the current " + pluginType + "");
+    listPluginOption = new Option("l", "list", false, "display the current " + pluginType + "");
+
     actionGroup.addOption(pluginClassOption);
     actionGroup.addOption(removePluginOption);
     actionGroup.addOption(listPluginOption);
     actionGroup.setRequired(true);
-    
+
     o.addOptionGroup(actionGroup);
-    o.addOption(OptUtil.tableOpt("table to set the "+pluginType+" on"));
-    
+    o.addOption(OptUtil.tableOpt("table to set the " + pluginType + " on"));
+
     return o;
   }
-  
+
   @Override
   public int numArgs() {
     return 0;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SleepCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SleepCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SleepCommand.java
index 1f7ed5d..db99ccd 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SleepCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SleepCommand.java
@@ -21,24 +21,24 @@ import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 
 public class SleepCommand extends Command {
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws Exception {
     final double secs = Double.parseDouble(cl.getArgs()[0]);
     Thread.sleep((long) (secs * 1000));
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "sleeps for the given number of seconds";
   }
-  
+
   @Override
   public int numArgs() {
     return 1;
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <seconds>";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/SystemPermissionsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/SystemPermissionsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/SystemPermissionsCommand.java
index 7cf4584..6dc9eee 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/SystemPermissionsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/SystemPermissionsCommand.java
@@ -31,12 +31,12 @@ public class SystemPermissionsCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "displays a list of valid system permissions";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/TableCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/TableCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/TableCommand.java
index 6a563ab..dbdf6bb 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/TableCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/TableCommand.java
@@ -23,13 +23,14 @@ import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 
 public class TableCommand extends Command {
   @Override
-  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
+  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException,
+      TableNotFoundException {
     final String tableName = cl.getArgs()[0];
     if (!shellState.getConnector().tableOperations().exists(tableName)) {
       throw new TableNotFoundException(null, tableName, null);
@@ -37,22 +38,22 @@ public class TableCommand extends Command {
     shellState.setTableName(tableName);
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "switches to the specified table";
   }
-  
+
   @Override
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> special) {
     registerCompletionForTables(root, special);
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <tableName>";
   }
-  
+
   @Override
   public int numArgs() {
     return 1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/TableOperation.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/TableOperation.java b/shell/src/main/java/org/apache/accumulo/shell/commands/TableOperation.java
index 565a5e4..b7d0f44 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/TableOperation.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/TableOperation.java
@@ -26,9 +26,9 @@ import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.impl.Namespaces;
 import org.apache.accumulo.core.client.impl.Tables;
 import org.apache.accumulo.shell.Shell;
+import org.apache.accumulo.shell.Shell.Command;
 import org.apache.accumulo.shell.ShellOptions;
 import org.apache.accumulo.shell.Token;
-import org.apache.accumulo.shell.Shell.Command;
 import org.apache.commons.cli.CommandLine;
 import org.apache.commons.cli.Option;
 import org.apache.commons.cli.OptionGroup;
@@ -97,7 +97,7 @@ public abstract class TableOperation extends Command {
 
   /**
    * Allows implementation to remove certain tables from the set of tables to be operated on.
-   * 
+   *
    * @param pattern
    *          The pattern which tables were selected using
    * @param tables

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/TablePermissionsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/TablePermissionsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/TablePermissionsCommand.java
index 75e3fbd..c213d04 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/TablePermissionsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/TablePermissionsCommand.java
@@ -31,12 +31,12 @@ public class TablePermissionsCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "displays a list of valid table permissions";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/TraceCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/TraceCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/TraceCommand.java
index d6fe075..ec1bd42 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/TraceCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/TraceCommand.java
@@ -19,21 +19,21 @@ package org.apache.accumulo.shell.commands;
 import java.io.IOException;
 import java.util.Map;
 
-import org.apache.accumulo.core.trace.Trace;
-import org.apache.accumulo.shell.Shell;
 import org.apache.accumulo.core.client.Scanner;
 import org.apache.accumulo.core.conf.Property;
 import org.apache.accumulo.core.data.Range;
 import org.apache.accumulo.core.security.Authorizations;
+import org.apache.accumulo.core.trace.Trace;
 import org.apache.accumulo.core.util.BadArgumentException;
 import org.apache.accumulo.core.util.UtilWaitThread;
+import org.apache.accumulo.shell.Shell;
 import org.apache.accumulo.tracer.TraceDump;
 import org.apache.accumulo.tracer.TraceDump.Printer;
 import org.apache.commons.cli.CommandLine;
 import org.apache.hadoop.io.Text;
 
 public class TraceCommand extends DebugCommand {
-  
+
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException {
     if (cl.getArgs().length == 1) {
       if (cl.getArgs()[0].equalsIgnoreCase("on")) {
@@ -93,7 +93,7 @@ public class TraceCommand extends DebugCommand {
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "turns trace logging on or off";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/UserCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/UserCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/UserCommand.java
index 6e7247e..d621a4f 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/UserCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/UserCommand.java
@@ -26,8 +26,8 @@ import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.Token;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.Token;
 import org.apache.commons.cli.CommandLine;
 
 public class UserCommand extends Command {
@@ -36,7 +36,7 @@ public class UserCommand extends Command {
     // save old credentials and connection in case of failure
     String user = cl.getArgs()[0];
     byte[] pass;
-    
+
     // We can't let the wrapping try around the execute method deal
     // with the exceptions because we have to do something if one
     // of these methods fails
@@ -49,22 +49,22 @@ public class UserCommand extends Command {
     shellState.updateUser(user, new PasswordToken(pass));
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "switches to the specified user";
   }
-  
+
   @Override
   public void registerCompletion(final Token root, final Map<Command.CompletionSet,Set<String>> special) {
     registerCompletionForUsers(root, special);
   }
-  
+
   @Override
   public String usage() {
     return getName() + " <username>";
   }
-  
+
   @Override
   public int numArgs() {
     return 1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/UserPermissionsCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/UserPermissionsCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/UserPermissionsCommand.java
index 34d7fe7..624a156 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/UserPermissionsCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/UserPermissionsCommand.java
@@ -32,7 +32,7 @@ import org.apache.commons.cli.Options;
 
 public class UserPermissionsCommand extends Command {
   private Option userOpt;
-  
+
   @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws AccumuloException, AccumuloSecurityException, IOException {
     final String user = cl.getOptionValue(userOpt.getOpt(), shellState.getConnector().whoami());
@@ -64,7 +64,6 @@ public class UserPermissionsCommand extends Command {
     }
     shellState.getReader().println();
 
-    
     runOnce = true;
     for (String t : shellState.getConnector().tableOperations().list()) {
       delim = "";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/UsersCommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/UsersCommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/UsersCommand.java
index 5ea61bf..82c05c8 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/UsersCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/UsersCommand.java
@@ -32,12 +32,12 @@ public class UsersCommand extends Command {
     }
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "displays a list of existing users";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/commands/WhoAmICommand.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/commands/WhoAmICommand.java b/shell/src/main/java/org/apache/accumulo/shell/commands/WhoAmICommand.java
index f80a621..45a14e2 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/WhoAmICommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/WhoAmICommand.java
@@ -28,12 +28,12 @@ public class WhoAmICommand extends Command {
     shellState.getReader().println(shellState.getConnector().whoami());
     return 0;
   }
-  
+
   @Override
   public String description() {
     return "reports the current user name";
   }
-  
+
   @Override
   public int numArgs() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/format/DeleterFormatter.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/format/DeleterFormatter.java b/shell/src/main/java/org/apache/accumulo/shell/format/DeleterFormatter.java
index b90b55e..a2a29d0 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/format/DeleterFormatter.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/format/DeleterFormatter.java
@@ -31,14 +31,14 @@ import org.apache.accumulo.shell.Shell;
 import org.apache.log4j.Logger;
 
 public class DeleterFormatter extends DefaultFormatter {
-  
+
   private static final Logger log = Logger.getLogger(DeleterFormatter.class);
   private BatchWriter writer;
   private Shell shellState;
   private boolean printTimestamps;
   private boolean force;
   private boolean more;
-  
+
   public DeleterFormatter(BatchWriter writer, Iterable<Entry<Key,Value>> scanner, boolean printTimestamps, Shell shellState, boolean force) {
     super.initialize(scanner, printTimestamps);
     this.writer = writer;
@@ -47,7 +47,7 @@ public class DeleterFormatter extends DefaultFormatter {
     this.force = force;
     this.more = true;
   }
-  
+
   @Override
   public boolean hasNext() {
     if (!getScannerIterator().hasNext() || !more) {
@@ -63,7 +63,7 @@ public class DeleterFormatter extends DefaultFormatter {
     }
     return true;
   }
-  
+
   /**
    * @return null, because the iteration will provide prompts and handle deletes internally.
    */


[49/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/6bc67602
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/6bc67602
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/6bc67602

Branch: refs/heads/master
Commit: 6bc67602c66f2cb88de5f7d8863701ac76309ab0
Parents: 44e2b2c
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:20:23 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:20:23 2015 -0500

----------------------------------------------------------------------
 .../accumulo/core/bloomfilter/BloomFilter.java  | 106 +--
 .../core/bloomfilter/DynamicBloomFilter.java    | 124 ++--
 .../accumulo/core/bloomfilter/Filter.java       | 102 +--
 .../accumulo/core/cli/BatchScannerOpts.java     |   8 +-
 .../accumulo/core/cli/BatchWriterOpts.java      |  18 +-
 .../accumulo/core/cli/ClientOnDefaultTable.java |   4 +-
 .../core/cli/ClientOnRequiredTable.java         |   1 -
 .../apache/accumulo/core/cli/ClientOpts.java    |  74 ++-
 .../java/org/apache/accumulo/core/cli/Help.java |  10 +-
 .../apache/accumulo/core/cli/ScannerOpts.java   |   4 +-
 .../accumulo/core/client/AccumuloException.java |  12 +-
 .../core/client/AccumuloSecurityException.java  |  34 +-
 .../accumulo/core/client/BatchDeleter.java      |  10 +-
 .../accumulo/core/client/BatchScanner.java      |  16 +-
 .../accumulo/core/client/BatchWriter.java       |  33 +-
 .../accumulo/core/client/BatchWriterConfig.java |  11 +-
 .../core/client/ClientConfiguration.java        |   9 +-
 .../core/client/ClientSideIteratorScanner.java  |  68 +-
 .../accumulo/core/client/ConditionalWriter.java |  29 +-
 .../core/client/ConditionalWriterConfig.java    |   1 +
 .../apache/accumulo/core/client/Connector.java  |  46 +-
 .../apache/accumulo/core/client/Durability.java |   3 +-
 .../accumulo/core/client/IsolatedScanner.java   | 107 +--
 .../accumulo/core/client/IteratorSetting.java   | 117 ++--
 .../core/client/MultiTableBatchWriter.java      |  22 +-
 .../core/client/MutationsRejectedException.java |  36 +-
 .../core/client/NamespaceExistsException.java   |   2 +-
 .../core/client/NamespaceNotFoundException.java |   2 +-
 .../accumulo/core/client/RowIterator.java       |  34 +-
 .../apache/accumulo/core/client/Scanner.java    |  41 +-
 .../accumulo/core/client/ScannerBase.java       |  43 +-
 .../core/client/TableDeletedException.java      |  10 +-
 .../core/client/TableExistsException.java       |   8 +-
 .../core/client/TableNotFoundException.java     |   2 +-
 .../core/client/TableOfflineException.java      |   6 +-
 .../accumulo/core/client/TimedOutException.java |  10 +-
 .../accumulo/core/client/ZooKeeperInstance.java |   6 +-
 .../core/client/admin/ActiveCompaction.java     |  27 +-
 .../accumulo/core/client/admin/ActiveScan.java  |  28 +-
 .../core/client/admin/CompactionConfig.java     |   9 +-
 .../client/admin/CompactionStrategyConfig.java  |   6 +-
 .../accumulo/core/client/admin/DiskUsage.java   |  20 +-
 .../accumulo/core/client/admin/FindMax.java     |  68 +-
 .../core/client/admin/InstanceOperations.java   |  51 +-
 .../core/client/admin/NamespaceOperations.java  |  46 +-
 .../client/admin/ReplicationOperations.java     |  45 +-
 .../core/client/admin/SecurityOperations.java   |  42 +-
 .../core/client/admin/TableOperations.java      | 115 ++--
 .../accumulo/core/client/admin/TimeType.java    |   2 +-
 .../client/impl/AccumuloServerException.java    |  10 +-
 .../core/client/impl/ActiveCompactionImpl.java  |   1 -
 .../core/client/impl/ActiveScanImpl.java        |   3 +-
 .../core/client/impl/BatchWriterImpl.java       |  14 +-
 .../core/client/impl/CompressedIterators.java   |  40 +-
 .../core/client/impl/ConditionalWriterImpl.java | 276 ++++----
 .../core/client/impl/DurabilityImpl.java        |   2 +-
 .../client/impl/InstanceOperationsImpl.java     |   2 +-
 .../core/client/impl/IsolationException.java    |   4 +-
 .../accumulo/core/client/impl/MasterClient.java |  16 +-
 .../client/impl/MultiTableBatchWriterImpl.java  |   4 +-
 .../core/client/impl/OfflineScanner.java        | 157 ++---
 .../core/client/impl/ReplicationClient.java     |   6 +-
 .../core/client/impl/RootTabletLocator.java     |  32 +-
 .../accumulo/core/client/impl/ScannerImpl.java  |  40 +-
 .../core/client/impl/ScannerIterator.java       |  79 ++-
 .../core/client/impl/ScannerOptions.java        |  69 +-
 .../accumulo/core/client/impl/ServerClient.java |  20 +-
 .../core/client/impl/TableOperationsHelper.java |   1 +
 .../core/client/impl/TableOperationsImpl.java   |  17 +-
 .../accumulo/core/client/impl/Tables.java       |   4 +-
 .../core/client/impl/TabletLocator.java         |  70 +-
 .../core/client/impl/TabletLocatorImpl.java     | 246 +++----
 .../client/impl/TabletServerBatchDeleter.java   |   8 +-
 .../client/impl/TabletServerBatchReader.java    |  36 +-
 .../impl/TabletServerBatchReaderIterator.java   | 226 +++----
 .../client/impl/TabletServerBatchWriter.java    | 370 +++++------
 .../accumulo/core/client/impl/TabletType.java   |  10 +-
 .../core/client/impl/ThriftScanner.java         | 140 ++--
 .../core/client/impl/TimeoutTabletLocator.java  |  46 +-
 .../accumulo/core/client/impl/Translator.java   |  50 +-
 .../accumulo/core/client/impl/Writer.java       |  22 +-
 .../core/client/impl/ZookeeperLockChecker.java  |  11 +-
 .../client/lexicoder/BigIntegerLexicoder.java   |  33 +-
 .../core/client/lexicoder/BytesLexicoder.java   |   8 +-
 .../core/client/lexicoder/DoubleLexicoder.java  |  10 +-
 .../accumulo/core/client/lexicoder/Encoder.java |   3 +-
 .../core/client/lexicoder/IntegerLexicoder.java |  13 +-
 .../core/client/lexicoder/Lexicoder.java        |   4 +-
 .../core/client/lexicoder/ListLexicoder.java    |  20 +-
 .../core/client/lexicoder/LongLexicoder.java    |   4 +-
 .../core/client/lexicoder/PairLexicoder.java    |  37 +-
 .../core/client/lexicoder/ReverseLexicoder.java |  20 +-
 .../core/client/lexicoder/StringLexicoder.java  |   8 +-
 .../core/client/lexicoder/TextLexicoder.java    |   8 +-
 .../client/lexicoder/UIntegerLexicoder.java     |  30 +-
 .../core/client/lexicoder/ULongLexicoder.java   |  28 +-
 .../core/client/lexicoder/UUIDLexicoder.java    |  14 +-
 .../core/client/lexicoder/impl/ByteUtils.java   |  38 +-
 .../impl/FixedByteArrayOutputStream.java        |  10 +-
 .../client/mapred/AccumuloFileOutputFormat.java |  48 +-
 .../core/client/mapred/AccumuloInputFormat.java |  13 +-
 .../mapred/AccumuloMultiTableInputFormat.java   |   8 +-
 .../client/mapred/AccumuloOutputFormat.java     |   2 +-
 .../client/mapred/AccumuloRowInputFormat.java   |   6 +-
 .../core/client/mapred/InputFormatBase.java     |  54 +-
 .../core/client/mapred/RangeInputSplit.java     |   2 +-
 .../mapreduce/AccumuloFileOutputFormat.java     |  48 +-
 .../client/mapreduce/AccumuloInputFormat.java   |   6 +-
 .../AccumuloMultiTableInputFormat.java          |  12 +-
 .../client/mapreduce/AccumuloOutputFormat.java  |   2 +-
 .../mapreduce/AccumuloRowInputFormat.java       |   6 +-
 .../core/client/mapreduce/InputFormatBase.java  |  54 +-
 .../core/client/mapreduce/InputTableConfig.java |  44 +-
 .../core/client/mapreduce/RangeInputSplit.java  |   2 +-
 .../mapreduce/lib/impl/ConfiguratorBase.java    |  58 +-
 .../lib/impl/FileOutputConfigurator.java        |  20 +-
 .../mapreduce/lib/impl/InputConfigurator.java   |  68 +-
 .../mapreduce/lib/impl/OutputConfigurator.java  |  25 +-
 .../client/mapreduce/lib/impl/package-info.java |   4 +-
 .../lib/partition/KeyRangePartitioner.java      |  10 +-
 .../core/client/mock/IteratorAdapter.java       |  10 +-
 .../accumulo/core/client/mock/MockAccumulo.java |   2 +-
 .../core/client/mock/MockBatchDeleter.java      |  12 +-
 .../core/client/mock/MockBatchScanner.java      |  12 +-
 .../core/client/mock/MockBatchWriter.java       |  14 +-
 .../core/client/mock/MockConfiguration.java     |   6 +-
 .../core/client/mock/MockConnector.java         |  38 +-
 .../accumulo/core/client/mock/MockInstance.java |   4 +-
 .../client/mock/MockInstanceOperations.java     |   5 +-
 .../client/mock/MockMultiTableBatchWriter.java  |  10 +-
 .../client/mock/MockNamespaceOperations.java    |   4 +-
 .../accumulo/core/client/mock/MockScanner.java  |  34 +-
 .../core/client/mock/MockScannerBase.java       |  24 +-
 .../accumulo/core/client/mock/MockTable.java    |  36 +-
 .../accumulo/core/client/mock/MockUser.java     |   2 +-
 .../core/client/replication/ReplicaSystem.java  |  16 +-
 .../replication/ReplicaSystemFactory.java       |   4 +-
 .../core/client/security/SecurityErrorCode.java |   2 +-
 .../security/tokens/AuthenticationToken.java    |  78 +--
 .../core/client/security/tokens/NullToken.java  |  27 +-
 .../client/security/tokens/PasswordToken.java   |  30 +-
 .../core/conf/AccumuloConfiguration.java        | 203 +++---
 .../accumulo/core/conf/ConfigSanityCheck.java   |  39 +-
 .../accumulo/core/conf/ConfigurationCopy.java   |  37 +-
 .../accumulo/core/conf/ConfigurationDocGen.java |   4 +-
 .../core/conf/ConfigurationObserver.java        |   4 +-
 .../conf/CredentialProviderFactoryShim.java     |   4 +-
 .../core/conf/DefaultConfiguration.java         |   3 +-
 .../apache/accumulo/core/conf/Experimental.java |   2 +-
 .../apache/accumulo/core/conf/Interpolated.java |   2 +-
 .../org/apache/accumulo/core/conf/Property.java |  87 ++-
 .../apache/accumulo/core/conf/PropertyType.java |   5 +-
 .../apache/accumulo/core/conf/Sensitive.java    |   2 +-
 .../accumulo/core/conf/SiteConfiguration.java   |  42 +-
 .../accumulo/core/constraints/Constraint.java   |  22 +-
 .../constraints/DefaultKeySizeConstraint.java   |  14 +-
 .../accumulo/core/constraints/Violations.java   |  48 +-
 .../accumulo/core/data/ArrayByteSequence.java   |  84 +--
 .../apache/accumulo/core/data/ByteSequence.java |  80 +--
 .../org/apache/accumulo/core/data/Column.java   |  82 +--
 .../apache/accumulo/core/data/ColumnUpdate.java |  51 +-
 .../accumulo/core/data/ComparableBytes.java     |  20 +-
 .../apache/accumulo/core/data/Condition.java    | 124 ++--
 .../accumulo/core/data/ConditionalMutation.java |  18 +-
 .../core/data/ConstraintViolationSummary.java   |  39 +-
 .../java/org/apache/accumulo/core/data/Key.java | 520 ++++++++-------
 .../apache/accumulo/core/data/KeyExtent.java    |  30 +-
 .../org/apache/accumulo/core/data/Mutation.java | 663 +++++++++++--------
 .../apache/accumulo/core/data/PartialKey.java   |  19 +-
 .../org/apache/accumulo/core/data/Range.java    | 580 +++++++++-------
 .../org/apache/accumulo/core/data/Value.java    |  26 +-
 .../accumulo/core/file/BloomFilterLayer.java    | 190 +++---
 .../accumulo/core/file/FileOperations.java      |  76 +--
 .../accumulo/core/file/FileSKVIterator.java     |   8 +-
 .../accumulo/core/file/FileSKVWriter.java       |  10 +-
 .../core/file/NoSuchMetaStoreException.java     |  11 +-
 .../core/file/blockfile/ABlockReader.java       |  19 +-
 .../core/file/blockfile/ABlockWriter.java       |  16 +-
 .../core/file/blockfile/BlockFileReader.java    |  18 +-
 .../core/file/blockfile/BlockFileWriter.java    |  14 +-
 .../core/file/blockfile/cache/BlockCache.java   |  14 +-
 .../core/file/blockfile/cache/CacheEntry.java   |   6 +-
 .../core/file/blockfile/cache/CachedBlock.java  |  29 +-
 .../file/blockfile/cache/CachedBlockQueue.java  |  28 +-
 .../core/file/blockfile/cache/ClassSize.java    | 120 ++--
 .../core/file/blockfile/cache/HeapSize.java     |   6 +-
 .../file/blockfile/cache/LruBlockCache.java     | 242 +++----
 .../file/blockfile/cache/SizeConstants.java     |  18 +-
 .../file/blockfile/impl/CachableBlockFile.java  | 250 +++----
 .../file/keyfunctor/ColumnFamilyFunctor.java    |  14 +-
 .../file/keyfunctor/ColumnQualifierFunctor.java |  10 +-
 .../core/file/keyfunctor/KeyFunctor.java        |   4 +-
 .../core/file/keyfunctor/RowFunctor.java        |  16 +-
 .../core/file/map/MapFileOperations.java        |  70 +-
 .../accumulo/core/file/map/MapFileUtil.java     |   2 +-
 .../accumulo/core/file/rfile/BlockIndex.java    |  56 +-
 .../accumulo/core/file/rfile/CreateEmpty.java   |   7 +-
 .../accumulo/core/file/rfile/IndexIterator.java |  20 +-
 .../core/file/rfile/MultiIndexIterator.java     |  28 +-
 .../core/file/rfile/MultiLevelIndex.java        | 336 +++++-----
 .../accumulo/core/file/rfile/PrintInfo.java     |  20 +-
 .../apache/accumulo/core/file/rfile/RFile.java  | 398 +++++------
 .../core/file/rfile/RFileOperations.java        |  32 +-
 .../accumulo/core/file/rfile/RelativeKey.java   | 216 +++---
 .../accumulo/core/file/rfile/SplitLarge.java    |  21 +-
 .../accumulo/core/file/rfile/bcfile/BCFile.java |  60 +-
 .../bcfile/BoundedRangeFileInputStream.java     |  38 +-
 .../core/file/rfile/bcfile/CompareUtils.java    |  26 +-
 .../core/file/rfile/bcfile/Compression.java     |  92 +--
 .../rfile/bcfile/MetaBlockAlreadyExists.java    |   6 +-
 .../rfile/bcfile/MetaBlockDoesNotExist.java     |   6 +-
 .../core/file/rfile/bcfile/PrintInfo.java       |   6 +-
 .../core/file/rfile/bcfile/RawComparable.java   |  16 +-
 .../bcfile/SimpleBufferedOutputStream.java      |  16 +-
 .../accumulo/core/file/rfile/bcfile/Utils.java  | 100 +--
 .../core/iterators/AggregatingIterator.java     |  68 +-
 .../core/iterators/ColumnFamilyCounter.java     |  28 +-
 .../accumulo/core/iterators/Combiner.java       |  33 +-
 .../accumulo/core/iterators/DebugIterator.java  |  25 +-
 .../apache/accumulo/core/iterators/DevNull.java |  32 +-
 .../iterators/FamilyIntersectingIterator.java   |   4 +-
 .../apache/accumulo/core/iterators/Filter.java  |  24 +-
 .../core/iterators/FirstEntryInRowIterator.java |  38 +-
 .../accumulo/core/iterators/GrepIterator.java   |   4 +-
 .../core/iterators/IntersectingIterator.java    |   4 +-
 .../IterationInterruptedException.java          |   6 +-
 .../core/iterators/IteratorEnvironment.java     |  12 +-
 .../accumulo/core/iterators/IteratorUtil.java   | 118 ++--
 .../accumulo/core/iterators/LargeRowFilter.java |   4 +-
 .../accumulo/core/iterators/LongCombiner.java   |  50 +-
 .../core/iterators/OptionDescriber.java         |  36 +-
 .../accumulo/core/iterators/OrIterator.java     |  72 +-
 .../core/iterators/RowDeletingIterator.java     |   4 +-
 .../core/iterators/SortedKeyIterator.java       |  12 +-
 .../core/iterators/SortedKeyValueIterator.java  |  36 +-
 .../core/iterators/SortedMapIterator.java       |  55 +-
 .../core/iterators/TypedValueCombiner.java      |  70 +-
 .../core/iterators/ValueFormatException.java    |   4 +-
 .../core/iterators/VersioningIterator.java      |   4 +-
 .../core/iterators/WholeRowIterator.java        |   4 +-
 .../core/iterators/WrappingIterator.java        |  26 +-
 .../core/iterators/aggregation/Aggregator.java  |   4 +-
 .../iterators/aggregation/LongSummation.java    |  12 +-
 .../aggregation/NumArraySummation.java          |  24 +-
 .../iterators/aggregation/NumSummation.java     |  20 +-
 .../core/iterators/aggregation/StringMax.java   |  10 +-
 .../core/iterators/aggregation/StringMin.java   |  10 +-
 .../iterators/aggregation/StringSummation.java  |  10 +-
 .../conf/AggregatorConfiguration.java           |   4 +-
 .../aggregation/conf/AggregatorSet.java         |   4 +-
 .../accumulo/core/iterators/conf/ColumnSet.java |  42 +-
 .../iterators/conf/ColumnToClassMapping.java    |  36 +-
 .../core/iterators/conf/ColumnUtil.java         |  42 +-
 .../iterators/conf/PerColumnIteratorConfig.java |  22 +-
 .../system/ColumnFamilySkippingIterator.java    |  30 +-
 .../iterators/system/ColumnQualifierFilter.java |  22 +-
 .../core/iterators/system/CountingIterator.java |  14 +-
 .../core/iterators/system/DeletingIterator.java |  28 +-
 .../iterators/system/LocalityGroupIterator.java |  30 +-
 .../core/iterators/system/MapFileIterator.java  |  38 +-
 .../core/iterators/system/MultiIterator.java    |  36 +-
 .../iterators/system/SequenceFileIterator.java  |  42 +-
 .../core/iterators/system/StatsIterator.java    |  16 +-
 .../iterators/system/TimeSettingIterator.java   |  26 +-
 .../core/iterators/system/VisibilityFilter.java |  16 +-
 .../core/iterators/user/AgeOffFilter.java       |  30 +-
 .../core/iterators/user/BigDecimalCombiner.java |  21 +-
 .../core/iterators/user/ColumnAgeOffFilter.java |  28 +-
 .../core/iterators/user/ColumnSliceFilter.java  | 148 ++---
 .../core/iterators/user/GrepIterator.java       |  24 +-
 .../core/iterators/user/IndexedDocIterator.java |  56 +-
 .../iterators/user/IntersectingIterator.java    | 108 +--
 .../core/iterators/user/LargeRowFilter.java     | 114 ++--
 .../core/iterators/user/MaxCombiner.java        |   2 +-
 .../core/iterators/user/MinCombiner.java        |   2 +-
 .../core/iterators/user/RegExFilter.java        |  68 +-
 .../core/iterators/user/ReqVisFilter.java       |   4 +-
 .../iterators/user/RowDeletingIterator.java     |  58 +-
 .../iterators/user/RowEncodingIterator.java     |  12 +-
 .../accumulo/core/iterators/user/RowFilter.java |  40 +-
 .../iterators/user/SummingArrayCombiner.java    |  48 +-
 .../core/iterators/user/SummingCombiner.java    |   2 +-
 .../core/iterators/user/TimestampFilter.java    |  56 +-
 .../iterators/user/TransformingIterator.java    |   2 +-
 .../core/iterators/user/VersioningIterator.java |  38 +-
 .../core/iterators/user/VisibilityFilter.java   |  24 +-
 .../user/WholeColumnFamilyIterator.java         |  66 +-
 .../core/iterators/user/WholeRowIterator.java   |  23 +-
 .../core/master/state/tables/TableState.java    |   8 +-
 .../core/metadata/MetadataLocationObtainer.java |  68 +-
 .../core/metadata/MetadataServicer.java         |  14 +-
 .../accumulo/core/metadata/RootTable.java       |   2 +-
 .../core/metadata/ServicerForMetadataTable.java |   4 +-
 .../core/metadata/ServicerForRootTable.java     |   8 +-
 .../core/metadata/ServicerForUserTables.java    |   4 +-
 .../core/metadata/TableMetadataServicer.java    |  46 +-
 .../core/metadata/schema/DataFileValue.java     |  32 +-
 .../core/metadata/schema/MetadataSchema.java    |  66 +-
 .../accumulo/core/protobuf/ProtobufUtil.java    |   2 +-
 .../AccumuloReplicationReplayer.java            |   2 +-
 .../ReplicationConfigurationUtil.java           |   7 +-
 .../core/replication/ReplicationConstants.java  |   2 +-
 .../core/replication/ReplicationSchema.java     |  19 +-
 .../core/replication/ReplicationTarget.java     |  16 +-
 .../accumulo/core/replication/StatusUtil.java   |  21 +-
 .../core/replication/proto/Replication.java     | 296 +++++----
 .../accumulo/core/rpc/TBufferedSocket.java      |   8 +-
 .../accumulo/core/rpc/TTimeoutTransport.java    |   6 +-
 .../apache/accumulo/core/rpc/TraceProtocol.java |   2 +-
 .../core/security/AuthorizationContainer.java   |   3 +-
 .../accumulo/core/security/Authorizations.java  |  30 +-
 .../core/security/ColumnVisibility.java         | 150 +++--
 .../accumulo/core/security/Credentials.java     |  19 +-
 .../core/security/NamespacePermission.java      |  10 +-
 .../core/security/SystemPermission.java         |  27 +-
 .../accumulo/core/security/TablePermission.java |  29 +-
 .../core/security/VisibilityConstraint.java     |  29 +-
 .../core/security/VisibilityEvaluator.java      |  66 +-
 .../core/security/VisibilityParseException.java |  11 +-
 .../security/crypto/BlockedOutputStream.java    |   2 +-
 .../CachingHDFSSecretKeyEncryptionStrategy.java |  57 +-
 .../core/security/crypto/CryptoModule.java      |  57 +-
 .../security/crypto/CryptoModuleFactory.java    |   6 +-
 .../security/crypto/CryptoModuleParameters.java | 272 ++++----
 .../security/crypto/DefaultCryptoModule.java    | 232 ++++---
 .../crypto/DefaultCryptoModuleUtils.java        |  10 +-
 .../crypto/DiscardCloseOutputStream.java        |   6 +-
 .../NonCachingSecretKeyEncryptionStrategy.java  |  54 +-
 .../crypto/SecretKeyEncryptionStrategy.java     |  10 +-
 .../core/tabletserver/log/LogEntry.java         |  18 +-
 .../accumulo/core/trace/DistributedTrace.java   |  23 +-
 .../org/apache/accumulo/core/trace/Span.java    |  11 +-
 .../org/apache/accumulo/core/trace/Trace.java   |   9 +-
 .../wrappers/RpcClientInvocationHandler.java    |   8 +-
 .../wrappers/RpcServerInvocationHandler.java    |   8 +-
 .../core/trace/wrappers/TraceCallable.java      |  12 +-
 .../trace/wrappers/TraceExecutorService.java    |  40 +-
 .../core/trace/wrappers/TraceRunnable.java      |  16 +-
 .../accumulo/core/trace/wrappers/TraceWrap.java |   8 +-
 .../apache/accumulo/core/util/AddressUtil.java  |   3 +-
 .../core/util/BadArgumentException.java         |   2 +-
 .../org/apache/accumulo/core/util/Base64.java   |   8 +-
 .../accumulo/core/util/ByteArrayComparator.java |  10 +-
 .../apache/accumulo/core/util/ByteArraySet.java |  14 +-
 .../accumulo/core/util/ByteBufferUtil.java      |  10 +-
 .../accumulo/core/util/CachedConfiguration.java |   4 +-
 .../org/apache/accumulo/core/util/CleanUp.java  |  18 +-
 .../org/apache/accumulo/core/util/ColumnFQ.java |  38 +-
 .../accumulo/core/util/ComparablePair.java      |  10 +-
 .../apache/accumulo/core/util/CreateToken.java  |  26 +-
 .../org/apache/accumulo/core/util/Daemon.java   |  18 +-
 .../org/apache/accumulo/core/util/Duration.java |  10 +-
 .../org/apache/accumulo/core/util/Encoding.java |  11 +-
 .../apache/accumulo/core/util/FastFormat.java   |  16 +-
 .../accumulo/core/util/LocalityGroupUtil.java   |  88 +--
 .../accumulo/core/util/LoggingRunnable.java     |   4 +-
 .../apache/accumulo/core/util/MapCounter.java   |  30 +-
 .../apache/accumulo/core/util/MonitorUtil.java  |   2 +-
 .../accumulo/core/util/MutableByteSequence.java |   9 +-
 .../accumulo/core/util/NamingThreadFactory.java |   6 +-
 .../org/apache/accumulo/core/util/NumUtil.java  |  10 +-
 .../org/apache/accumulo/core/util/OpTimer.java  |   6 +-
 .../accumulo/core/util/PeekingIterator.java     |  16 +-
 .../accumulo/core/util/ServerServices.java      |  22 +-
 .../accumulo/core/util/SimpleThreadPool.java    |   5 +-
 .../apache/accumulo/core/util/StopWatch.java    |  26 +-
 .../org/apache/accumulo/core/util/TextUtil.java |   9 +-
 .../core/util/UnsynchronizedBuffer.java         | 111 ++--
 .../accumulo/core/util/UtilWaitThread.java      |   2 +-
 .../apache/accumulo/core/util/Validator.java    |  16 +-
 .../org/apache/accumulo/core/util/Version.java  |  22 +-
 .../core/util/format/BinaryFormatter.java       |  18 +-
 .../core/util/format/DateStringFormatter.java   |  11 +-
 .../core/util/format/DefaultFormatter.java      |  50 +-
 .../core/util/format/FormatterFactory.java      |   4 +-
 .../accumulo/core/util/format/HexFormatter.java |  31 +-
 .../ShardedTableDistributionFormatter.java      |  12 +-
 .../util/format/StatisticsDisplayFormatter.java |  16 +-
 .../util/interpret/DefaultScanInterpreter.java  |  14 +-
 .../core/util/interpret/HexScanInterpreter.java |   2 +-
 .../core/util/interpret/ScanInterpreter.java    |   8 +-
 .../org/apache/accumulo/core/volume/Volume.java |  14 +-
 .../core/volume/VolumeConfiguration.java        |   4 +-
 .../apache/accumulo/core/volume/VolumeImpl.java |   3 +-
 .../org/apache/accumulo/core/cli/TestHelp.java  |   6 +-
 .../core/client/BatchWriterConfigTest.java      |  54 +-
 .../core/client/ClientSideIteratorTest.java     |  18 +-
 .../core/client/IteratorSettingTest.java        |  32 +-
 .../accumulo/core/client/RowIteratorTest.java   |  12 +-
 .../accumulo/core/client/TestThrift1474.java    |  12 +-
 .../accumulo/core/client/admin/FindMaxTest.java |  42 +-
 .../core/client/impl/ClientContextTest.java     |   2 +-
 .../core/client/impl/ScannerImplTest.java       |   6 +-
 .../core/client/impl/ScannerOptionsTest.java    |   4 +-
 .../client/impl/TableOperationsHelperTest.java  |  23 +-
 .../client/impl/ZookeeperLockCheckerTest.java   |   9 +-
 .../lexicoder/BigIntegerLexicoderTest.java      |   2 +-
 .../client/lexicoder/DoubleLexicoderTest.java   |   2 +-
 .../client/lexicoder/ListLexicoderTest.java     |  26 +-
 .../client/lexicoder/PairLexicoderTest.java     |   2 +-
 .../client/lexicoder/UIntegerLexicoderTest.java |   2 +-
 .../client/mapred/AccumuloInputFormatTest.java  |   2 +-
 .../client/mapred/AccumuloOutputFormatTest.java |  54 +-
 .../mapred/AccumuloRowInputFormatTest.java      |  42 +-
 .../core/client/mapred/RangeInputSplitTest.java |  30 +-
 .../core/client/mapred/TokenFileTest.java       |   2 +-
 .../mapreduce/AccumuloInputFormatTest.java      |   2 +-
 .../mapreduce/AccumuloOutputFormatTest.java     |  50 +-
 .../mapreduce/AccumuloRowInputFormatTest.java   |  38 +-
 .../BadPasswordSplitsAccumuloInputFormat.java   |   6 +-
 .../client/mapreduce/InputTableConfigTest.java  |  31 +-
 .../core/client/mapreduce/TokenFileTest.java    |   2 +-
 .../lib/impl/ConfiguratorBaseTest.java          |  12 +-
 .../lib/partition/RangePartitionerTest.java     |  16 +-
 .../core/client/mock/MockConnectorTest.java     | 122 ++--
 .../core/client/mock/MockNamespacesTest.java    |   6 +-
 .../client/mock/MockTableOperationsTest.java    |   1 -
 .../core/client/mock/TestBatchScanner821.java   |   4 +-
 .../core/client/mock/TransformIterator.java     |   2 -
 .../client/security/SecurityErrorCodeTest.java  |  10 +-
 .../tokens/CredentialProviderTokenTest.java     |   1 +
 .../security/tokens/PasswordTokenTest.java      |   6 +-
 .../core/conf/AccumuloConfigurationTest.java    |  12 +-
 .../core/conf/ConfigSanityCheckTest.java        |   1 +
 .../core/conf/DefaultConfigurationTest.java     |   6 +-
 .../core/conf/ObservableConfigurationTest.java  |  12 +-
 .../accumulo/core/conf/PropertyTypeTest.java    |   5 +-
 .../core/conf/SiteConfigurationTest.java        |  13 +-
 .../DefaultKeySizeConstraintTest.java           |  14 +-
 .../apache/accumulo/core/data/ColumnTest.java   |   5 +-
 .../accumulo/core/data/ConditionTest.java       |  14 +-
 .../data/ConstraintViolationSummaryTest.java    |  15 +-
 .../accumulo/core/data/KeyExtentTest.java       | 100 +--
 .../org/apache/accumulo/core/data/KeyTest.java  |  42 +-
 .../apache/accumulo/core/data/MutationTest.java | 233 ++++---
 .../apache/accumulo/core/data/OldMutation.java  | 190 +++---
 .../apache/accumulo/core/data/ValueTest.java    |  22 +-
 .../accumulo/core/file/FileOperationsTest.java  |   8 +-
 .../blockfile/cache/TestCachedBlockQueue.java   |  38 +-
 .../file/blockfile/cache/TestLruBlockCache.java | 228 +++----
 .../core/file/rfile/BlockIndexTest.java         |  59 +-
 .../core/file/rfile/CreateCompatTestFile.java   |  28 +-
 .../core/file/rfile/MultiLevelIndexTest.java    |  36 +-
 .../accumulo/core/file/rfile/RFileTest.java     |  13 +-
 .../core/file/rfile/RelativeKeyTest.java        |  74 +--
 .../core/iterators/AggregatingIteratorTest.java | 280 ++++----
 .../iterators/DefaultIteratorEnvironment.java   |  18 +-
 .../iterators/FirstEntryInRowIteratorTest.java  | 136 ++--
 .../core/iterators/FirstEntryInRowTest.java     |  33 +-
 .../core/iterators/IteratorUtilTest.java        | 158 ++---
 .../iterators/aggregation/NumSummationTest.java |  28 +-
 .../conf/AggregatorConfigurationTest.java       |  16 +-
 .../ColumnFamilySkippingIteratorTest.java       |  94 +--
 .../core/iterators/system/ColumnFilterTest.java |  32 +-
 .../iterators/system/DeletingIteratorTest.java  |  90 +--
 .../iterators/system/MultiIteratorTest.java     | 150 ++---
 .../system/SourceSwitchingIteratorTest.java     |  92 +--
 .../system/TimeSettingIteratorTest.java         |  32 +-
 .../iterators/system/VisibilityFilterTest.java  |  10 +-
 .../iterators/user/BigDecimalCombinerTest.java  |   1 -
 .../iterators/user/ColumnSliceFilterTest.java   | 455 +++++++------
 .../core/iterators/user/CombinerTest.java       | 422 ++++++------
 .../core/iterators/user/FilterTest.java         | 142 ++--
 .../core/iterators/user/GrepIteratorTest.java   |  14 +-
 .../iterators/user/IndexedDocIteratorTest.java  |  56 +-
 .../user/IntersectingIteratorTest.java          |  48 +-
 .../core/iterators/user/LargeRowFilterTest.java |  88 +--
 .../iterators/user/RowDeletingIteratorTest.java |  92 +--
 .../core/iterators/user/RowFilterTest.java      |   8 +-
 .../iterators/user/VersioningIteratorTest.java  | 100 +--
 .../iterators/user/VisibilityFilterTest.java    |  82 +--
 .../user/WholeColumnFamilyIteratorTest.java     |  54 +-
 .../iterators/user/WholeRowIteratorTest.java    |  54 +-
 .../core/metadata/MetadataServicerTest.java     |  16 +-
 .../ReplicationConfigurationUtilTest.java       |   8 +-
 .../core/replication/ReplicationTargetTest.java |   2 +-
 .../core/replication/StatusUtilTest.java        |   2 +-
 .../core/replication/proto/StatusTest.java      |   2 +-
 .../core/security/AuthenticationTokenTest.java  |   8 +-
 .../core/security/AuthorizationsTest.java       |  26 +-
 .../accumulo/core/security/CredentialsTest.java |  22 +-
 .../core/security/VisibilityEvaluatorTest.java  |  34 +-
 .../core/security/crypto/CryptoTest.java        | 168 ++---
 .../accumulo/core/trace/PerformanceTest.java    |   8 +-
 .../core/util/LocalityGroupUtilTest.java        |  29 +-
 .../apache/accumulo/core/util/MergeTest.java    |  36 +-
 .../org/apache/accumulo/core/util/PairTest.java |   6 +-
 .../accumulo/core/util/PartitionerTest.java     |  66 +-
 .../apache/accumulo/core/util/TestVersion.java  |  14 +-
 .../apache/accumulo/core/util/TextUtilTest.java |   4 +-
 .../core/util/format/DefaultFormatterTest.java  |   2 +-
 .../core/util/format/HexFormatterTest.java      |   1 -
 .../accumulo/examples/simple/client/Flush.java  |   2 +-
 .../simple/client/RandomBatchScanner.java       |  84 +--
 .../simple/client/RandomBatchWriter.java        |   8 +-
 .../simple/client/ReadWriteExample.java         |   2 +-
 .../examples/simple/client/RowOperations.java   |  70 +-
 .../simple/client/SequentialBatchWriter.java    |  18 +-
 .../simple/client/TraceDumpExample.java         |  62 +-
 .../examples/simple/client/TracingExample.java  |   6 +-
 .../examples/simple/combiner/StatsCombiner.java |  30 +-
 .../constraints/AlphaNumKeyConstraint.java      |  30 +-
 .../simple/constraints/MaxMutationSize.java     |   6 +-
 .../constraints/NumericValueConstraint.java     |  24 +-
 .../examples/simple/dirlist/FileCount.java      | 113 ++--
 .../examples/simple/dirlist/Ingest.java         |  37 +-
 .../examples/simple/dirlist/QueryUtil.java      |  61 +-
 .../examples/simple/dirlist/Viewer.java         |  47 +-
 .../examples/simple/filedata/ChunkCombiner.java |  56 +-
 .../simple/filedata/ChunkInputFormat.java       |   4 +-
 .../simple/filedata/ChunkInputStream.java       |  52 +-
 .../simple/filedata/FileDataIngest.java         | 105 ++-
 .../examples/simple/filedata/FileDataQuery.java |  12 +-
 .../examples/simple/filedata/KeyUtil.java       |   8 +-
 .../simple/filedata/VisibilityCombiner.java     |  32 +-
 .../helloworld/InsertWithBatchWriter.java       |  10 +-
 .../examples/simple/helloworld/ReadData.java    |  14 +-
 .../simple/isolation/InterferenceTest.java      |  69 +-
 .../examples/simple/mapreduce/TableToFile.java  |   2 +-
 .../simple/mapreduce/TeraSortIngest.java        |  14 +-
 .../simple/mapreduce/TokenFileWordCount.java    |   4 +-
 .../examples/simple/mapreduce/WordCount.java    |   6 +-
 .../simple/mapreduce/bulk/SetupTable.java       |   8 +-
 .../simple/mapreduce/bulk/VerifyIngest.java     |  32 +-
 .../examples/simple/reservations/ARS.java       | 104 +--
 .../examples/simple/shard/ContinuousQuery.java  |  64 +-
 .../accumulo/examples/simple/shard/Index.java   |  48 +-
 .../accumulo/examples/simple/shard/Query.java   |  14 +-
 .../accumulo/examples/simple/shard/Reverse.java |  20 +-
 .../examples/simple/shell/DebugCommand.java     |   2 +-
 .../simple/shell/MyAppShellExtension.java       |  10 +-
 .../examples/simple/dirlist/CountTest.java      |  12 +-
 .../simple/filedata/ChunkCombinerTest.java      |  90 ++-
 .../simple/filedata/ChunkInputStreamTest.java   | 122 ++--
 .../examples/simple/filedata/KeyUtilTest.java   |   5 +-
 .../org/apache/accumulo/fate/AdminUtil.java     |  88 +--
 .../org/apache/accumulo/fate/AgeOffStore.java   |  66 +-
 .../org/apache/accumulo/fate/ReadOnlyRepo.java  |   4 +-
 .../org/apache/accumulo/fate/ReadOnlyStore.java |   6 +-
 .../apache/accumulo/fate/ReadOnlyTStore.java    |  35 +-
 .../java/org/apache/accumulo/fate/Repo.java     |   8 +-
 .../accumulo/fate/StackOverflowException.java   |   6 +-
 .../java/org/apache/accumulo/fate/TStore.java   |  12 +-
 .../java/org/apache/accumulo/fate/ZooStore.java | 127 ++--
 .../apache/accumulo/fate/util/AddressUtil.java  |   2 +-
 .../org/apache/accumulo/fate/util/Daemon.java   |  18 +-
 .../accumulo/fate/util/LoggingRunnable.java     |   6 +-
 .../accumulo/fate/util/UtilWaitThread.java      |   2 +-
 .../zookeeper/DistributedReadWriteLock.java     |  74 +--
 .../fate/zookeeper/IZooReaderWriter.java        |   2 +-
 .../fate/zookeeper/TransactionWatcher.java      |  13 +-
 .../accumulo/fate/zookeeper/ZooCache.java       |  58 +-
 .../fate/zookeeper/ZooCacheFactory.java         |   9 +-
 .../apache/accumulo/fate/zookeeper/ZooLock.java | 240 +++----
 .../fate/zookeeper/ZooReaderWriter.java         |   1 -
 .../accumulo/fate/zookeeper/ZooReservation.java |  22 +-
 .../accumulo/fate/zookeeper/ZooSession.java     |  22 +-
 .../apache/accumulo/fate/AgeOffStoreTest.java   |  66 +-
 .../apache/accumulo/fate/ReadOnlyStoreTest.java |   5 +-
 .../org/apache/accumulo/fate/SimpleStore.java   |  34 +-
 .../accumulo/fate/util/AddressUtilTest.java     |   4 +-
 .../zookeeper/DistributedReadWriteLockTest.java |  28 +-
 .../fate/zookeeper/TransactionWatcherTest.java  |  27 +-
 .../fate/zookeeper/ZooCacheFactoryTest.java     |  10 +-
 .../accumulo/fate/zookeeper/ZooCacheTest.java   |  33 +-
 .../fate/zookeeper/ZooReaderWriterTest.java     |   2 +-
 .../accumulo/fate/zookeeper/ZooSessionTest.java |   4 +-
 .../apache/accumulo/plugin/CustomFilter.java    |   6 +-
 .../org/apache/accumulo/plugin/PluginIT.java    |  18 +-
 .../maven/plugin/AbstractAccumuloMojo.java      |   6 +-
 .../apache/accumulo/maven/plugin/StopMojo.java  |   2 +-
 .../apache/accumulo/cluster/RemoteShell.java    |   3 +-
 .../standalone/StandaloneClusterControl.java    |   2 +-
 .../minicluster/MiniAccumuloConfig.java         |  54 +-
 .../minicluster/MiniAccumuloInstance.java       |   2 +-
 .../minicluster/MiniAccumuloRunner.java         |  16 +-
 .../impl/ZooKeeperBindException.java            |   3 +-
 .../MiniAccumuloClusterStartStopTest.java       |   4 +-
 .../minicluster/impl/CleanShutdownMacTest.java  |   2 +-
 .../impl/MiniAccumuloClusterImplTest.java       |   3 +-
 .../java/org/apache/accumulo/proxy/Proxy.java   |  32 +-
 .../org/apache/accumulo/proxy/ProxyServer.java  | 352 +++++-----
 .../apache/accumulo/proxy/TestProxyClient.java  |  56 +-
 .../java/org/apache/accumulo/proxy/Util.java    |  18 +-
 .../org/apache/accumulo/server/Accumulo.java    |  44 +-
 .../org/apache/accumulo/server/ServerOpts.java  |   6 +-
 .../server/cli/ClientOnDefaultTable.java        |   3 +-
 .../server/cli/ClientOnRequiredTable.java       |   4 +-
 .../apache/accumulo/server/cli/ClientOpts.java  |   2 +-
 .../accumulo/server/client/BulkImporter.java    | 264 ++++----
 .../accumulo/server/client/HdfsZooInstance.java |   2 +-
 .../accumulo/server/conf/ConfigSanityCheck.java |   4 +-
 .../server/conf/NamespaceConfWatcher.java       |   5 +-
 .../server/conf/ServerConfiguration.java        |   4 +-
 .../server/conf/ServerConfigurationFactory.java |  16 +-
 .../accumulo/server/conf/TableConfWatcher.java  |  15 +-
 .../server/conf/TableParentConfiguration.java   |   1 +
 .../server/data/ServerColumnUpdate.java         |   2 +-
 .../accumulo/server/data/ServerMutation.java    |   9 +-
 .../org/apache/accumulo/server/fs/FileRef.java  |  24 +-
 .../server/fs/PerTableVolumeChooser.java        |   8 +-
 .../server/fs/PreferredVolumeChooser.java       |  11 +-
 .../apache/accumulo/server/fs/ViewFSUtils.java  |   2 +-
 .../accumulo/server/fs/VolumeChooser.java       |   4 +-
 .../accumulo/server/fs/VolumeManager.java       |   3 +-
 .../accumulo/server/fs/VolumeManagerImpl.java   |   4 +-
 .../apache/accumulo/server/fs/VolumeUtil.java   |   4 +-
 .../apache/accumulo/server/init/Initialize.java |   2 +-
 .../iterators/MetadataBulkLoadFilter.java       |  24 +-
 .../accumulo/server/log/SortedLogState.java     |   6 +-
 .../master/balancer/ChaoticLoadBalancer.java    |  21 +-
 .../server/master/balancer/TabletBalancer.java  |  51 +-
 .../server/master/recovery/RecoveryPath.java    |  12 +-
 .../server/master/state/Assignment.java         |   2 +-
 .../server/master/state/ClosableIterator.java   |   3 +-
 .../server/master/state/CurrentState.java       |   8 +-
 .../server/master/state/DeadServerList.java     |  10 +-
 .../server/master/state/DistributedStore.java   |  10 +-
 .../master/state/DistributedStoreException.java |   8 +-
 .../accumulo/server/master/state/MergeInfo.java |  30 +-
 .../server/master/state/MergeState.java         |   2 +-
 .../server/master/state/MetaDataStateStore.java |  26 +-
 .../master/state/MetaDataTableScanner.java      |  32 +-
 .../master/state/RootTabletStateStore.java      |   8 +-
 .../server/master/state/TServerInstance.java    |  51 +-
 .../master/state/TabletLocationState.java       |  25 +-
 .../server/master/state/TabletMigration.java    |   4 +-
 .../server/master/state/TabletServerState.java  |  18 +-
 .../master/state/TabletStateChangeIterator.java |  30 +-
 .../server/master/state/TabletStateStore.java   |  22 +-
 .../accumulo/server/master/state/ZooStore.java  |  20 +-
 .../master/state/ZooTabletStateStore.java       |  37 +-
 .../master/tableOps/UserCompactionConfig.java   |   5 +-
 .../server/metrics/AbstractMetricsImpl.java     |   3 +-
 .../server/metrics/MetricsConfiguration.java    |  66 +-
 .../accumulo/server/metrics/ThriftMetrics.java  |   1 -
 .../server/monitor/DedupedLogEvent.java         |  18 +-
 .../accumulo/server/monitor/LogService.java     |   6 +-
 .../accumulo/server/problems/ProblemReport.java |  82 +--
 .../problems/ProblemReportingIterator.java      |  24 +-
 .../server/problems/ProblemReports.java         | 118 ++--
 .../DistributedWorkQueueWorkAssignerHelper.java |   5 +-
 .../server/replication/ReplicationUtil.java     |  11 +-
 .../server/replication/StatusCombiner.java      |   4 +-
 .../server/replication/WorkAssigner.java        |   1 -
 .../replication/ZooKeeperInitialization.java    |   3 +-
 .../server/rpc/ClientInfoProcessorFactory.java  |   2 +-
 .../apache/accumulo/server/rpc/RpcWrapper.java  |   1 +
 .../server/rpc/TNonblockingServerSocket.java    |  14 +-
 .../security/AuditedSecurityOperation.java      |  23 +-
 .../server/security/SecurityOperation.java      |  56 +-
 .../accumulo/server/security/SecurityUtil.java  |  12 +-
 .../server/security/SystemCredentials.java      |   4 +-
 .../server/security/handler/Authenticator.java  |  22 +-
 .../server/security/handler/Authorizor.java     |  14 +-
 .../security/handler/InsecureAuthenticator.java |  37 +-
 .../security/handler/InsecurePermHandler.java   |  70 +-
 .../security/handler/PermissionHandler.java     |   9 +-
 .../server/security/handler/ZKPermHandler.java  |  10 +-
 .../server/security/handler/ZKSecurityTool.java |  32 +-
 .../accumulo/server/tables/TableManager.java    |   2 +-
 .../accumulo/server/tables/TableObserver.java   |   4 +-
 .../accumulo/server/tablets/TabletTime.java     | 104 +--
 .../server/tablets/UniqueNameAllocator.java     |  26 +-
 .../tabletserver/LargestFirstMemoryManager.java |  88 ++-
 .../server/tabletserver/MemoryManager.java      |  16 +-
 .../server/tabletserver/TabletState.java        |   6 +-
 .../server/util/ActionStatsUpdator.java         |   6 +-
 .../org/apache/accumulo/server/util/Admin.java  | 105 +--
 .../accumulo/server/util/ChangeSecret.java      |  22 +-
 .../accumulo/server/util/CleanZookeeper.java    |  19 +-
 .../apache/accumulo/server/util/DefaultMap.java |   8 +-
 .../accumulo/server/util/DeleteZooInstance.java |  14 +-
 .../accumulo/server/util/DumpZookeeper.java     |  23 +-
 .../accumulo/server/util/FileSystemMonitor.java |  44 +-
 .../apache/accumulo/server/util/FileUtil.java   | 236 +++----
 .../org/apache/accumulo/server/util/Halt.java   |  10 +-
 .../accumulo/server/util/ListInstances.java     |  89 +--
 .../accumulo/server/util/ListVolumesUsed.java   |   2 +-
 .../accumulo/server/util/LocalityCheck.java     |  14 +-
 .../accumulo/server/util/LoginProperties.java   |  14 +-
 .../server/util/MasterMetadataUtil.java         | 101 +--
 .../accumulo/server/util/MetadataTableUtil.java |  15 +-
 .../accumulo/server/util/RandomWriter.java      |  30 +-
 .../accumulo/server/util/RandomizeVolumes.java  |   3 +-
 .../util/RemoveEntriesForMissingFiles.java      |   2 +-
 .../accumulo/server/util/RestoreZookeeper.java  |  18 +-
 .../accumulo/server/util/SendLogToChainsaw.java |   2 +-
 .../accumulo/server/util/SystemPropUtil.java    |   6 +-
 .../accumulo/server/util/TableInfoUtil.java     |  10 +-
 .../accumulo/server/util/TablePropUtil.java     |  14 +-
 .../accumulo/server/util/TabletIterator.java    | 136 ++--
 .../accumulo/server/util/TabletOperations.java  |  18 +-
 .../accumulo/server/util/TabletServerLocks.java |  23 +-
 .../server/util/VerifyTabletAssignments.java    |  58 +-
 .../accumulo/server/util/ZooKeeperMain.java     |   8 +-
 .../org/apache/accumulo/server/util/ZooZap.java |  45 +-
 .../server/util/time/BaseRelativeTime.java      |  16 +-
 .../accumulo/server/util/time/ProvidesTime.java |   2 +-
 .../accumulo/server/util/time/RelativeTime.java |  14 +-
 .../accumulo/server/util/time/SimpleTimer.java  |  63 +-
 .../accumulo/server/util/time/SystemTime.java   |   6 +-
 .../server/zookeeper/DistributedWorkQueue.java  |  56 +-
 .../server/zookeeper/TransactionWatcher.java    |  12 +-
 .../accumulo/server/zookeeper/ZooCache.java     |   4 +-
 .../accumulo/server/zookeeper/ZooLock.java      |   6 +-
 .../server/zookeeper/ZooReaderWriter.java       |   6 +-
 .../apache/accumulo/server/AccumuloTest.java    |   2 +
 .../accumulo/server/ServerConstantsTest.java    |   2 +-
 .../apache/accumulo/server/ServerOptsTest.java  |   3 +-
 .../server/client/BulkImporterTest.java         |  24 +-
 .../server/conf/NamespaceConfigurationTest.java |  21 +-
 .../server/conf/TableConfigurationTest.java     |  12 +-
 .../conf/ZooConfigurationFactoryTest.java       |  15 +-
 .../constraints/MetadataConstraintsTest.java    |  89 ++-
 .../server/data/ServerMutationTest.java         |  27 +-
 .../apache/accumulo/server/fs/FileRefTest.java  |   4 +-
 .../apache/accumulo/server/fs/FileTypeTest.java |   2 +-
 .../accumulo/server/fs/ViewFSUtilsTest.java     |  11 +-
 .../server/fs/VolumeManagerImplTest.java        |   3 +-
 .../accumulo/server/fs/VolumeUtilTest.java      |  14 +-
 .../iterators/MetadataBulkLoadFilterTest.java   |  39 +-
 .../balancer/ChaoticLoadBalancerTest.java       |  33 +-
 .../balancer/DefaultLoadBalancerTest.java       |  51 +-
 .../master/balancer/TableLoadBalancerTest.java  |  32 +-
 .../server/master/state/MergeInfoTest.java      |  16 +-
 .../master/state/TabletLocationStateTest.java   |  14 +-
 .../server/problems/ProblemReportTest.java      |  20 +-
 .../server/replication/StatusCombinerTest.java  |   4 +-
 .../ZooKeeperInitializationTest.java            |   2 +-
 .../server/security/SystemCredentialsTest.java  |   8 +-
 .../security/handler/ZKAuthenticatorTest.java   |  22 +-
 .../accumulo/server/tablets/MillisTimeTest.java |  16 +-
 .../accumulo/server/tablets/TabletTimeTest.java |  11 +-
 .../accumulo/server/util/AdminCommandsTest.java |   5 +-
 .../apache/accumulo/server/util/CloneTest.java  | 230 +++----
 .../accumulo/server/util/DefaultMapTest.java    |  12 +-
 .../accumulo/server/util/FileInfoTest.java      |   3 +-
 .../accumulo/server/util/FileUtilTest.java      | 168 ++---
 .../server/util/ReplicationTableUtilTest.java   |   2 +-
 .../accumulo/server/util/TServerUtilsTest.java  |  10 +-
 .../server/util/TabletIteratorTest.java         |  34 +-
 .../server/util/time/BaseRelativeTimeTest.java  |  25 +-
 .../server/util/time/SimpleTimerTest.java       |   9 +-
 .../accumulo/gc/GarbageCollectionAlgorithm.java |   2 +-
 .../gc/GarbageCollectionEnvironment.java        |  19 +-
 .../accumulo/gc/SimpleGarbageCollector.java     |   4 +-
 .../accumulo/gc/GarbageCollectionTest.java      |   3 +-
 .../gc/SimpleGarbageCollectorOptsTest.java      |   3 +-
 .../accumulo/master/EventCoordinator.java       |  16 +-
 .../accumulo/master/FateServiceHandler.java     |   4 +-
 .../accumulo/master/TabletGroupWatcher.java     | 116 ++--
 .../master/metrics/ReplicationMetricsMBean.java |  16 +-
 .../master/recovery/RecoveryManager.java        |   4 +-
 .../MasterReplicationCoordinator.java           |   3 +-
 .../replication/SequentialWorkAssigner.java     |   8 +-
 .../accumulo/master/replication/WorkDriver.java |   2 +-
 .../accumulo/master/state/MergeStats.java       |  28 +-
 .../accumulo/master/state/SetGoalState.java     |   4 +-
 .../accumulo/master/state/TableCounts.java      |   8 +-
 .../accumulo/master/state/TableStats.java       |  14 +-
 .../accumulo/master/tableOps/BulkImport.java    |  13 +-
 .../accumulo/master/tableOps/CloneTable.java    |   4 +-
 .../accumulo/master/tableOps/CompactRange.java  |   4 +-
 .../accumulo/master/tableOps/CreateTable.java   |   4 +-
 .../accumulo/master/tableOps/ExportTable.java   | 112 ++--
 .../accumulo/master/tableOps/MasterRepo.java    |  14 +-
 .../accumulo/master/tableOps/TableRangeOp.java  |   9 +-
 .../accumulo/master/tableOps/TraceRepo.java     |  20 +-
 .../apache/accumulo/master/tableOps/Utils.java  |   6 +-
 .../master/tserverOps/ShutdownTServer.java      |  16 +-
 .../apache/accumulo/master/DefaultMapTest.java  |   9 +-
 .../apache/accumulo/master/TestMergeState.java  |   2 +-
 ...tributedWorkQueueWorkAssignerHelperTest.java |   2 +-
 .../MasterReplicationCoordinatorTest.java       |   2 +-
 .../master/replication/StatusMakerTest.java     |   2 +-
 .../accumulo/master/state/MergeInfoTest.java    |  10 +-
 .../master/state/RootTabletStateStoreTest.java  |  40 +-
 .../org/apache/accumulo/monitor/Monitor.java    |   6 +-
 .../accumulo/monitor/ZooKeeperStatus.java       |  30 +-
 .../monitor/servlets/GcStatusServlet.java       |  12 +-
 .../accumulo/monitor/servlets/JSONServlet.java  |  22 +-
 .../accumulo/monitor/servlets/LogServlet.java   |  12 +-
 .../monitor/servlets/MasterServlet.java         |  27 +-
 .../monitor/servlets/ProblemServlet.java        |  46 +-
 .../accumulo/monitor/servlets/ShellServlet.java |  38 +-
 .../monitor/servlets/TServersServlet.java       |  70 +-
 .../monitor/servlets/TablesServlet.java         |  34 +-
 .../accumulo/monitor/servlets/VisServlet.java   |  52 +-
 .../accumulo/monitor/servlets/XMLServlet.java   |  38 +-
 .../monitor/servlets/trace/ListType.java        |  10 +-
 .../servlets/trace/NullKeyValueIterator.java    |   6 +-
 .../monitor/servlets/trace/NullScanner.java     |  46 +-
 .../monitor/servlets/trace/ShowTrace.java       |  18 +-
 .../servlets/trace/ShowTraceLinkType.java       |   4 +-
 .../monitor/servlets/trace/Summary.java         |  30 +-
 .../accumulo/monitor/util/TableColumn.java      |  10 +-
 .../apache/accumulo/monitor/util/TableRow.java  |  18 +-
 .../monitor/util/celltypes/CellType.java        |   8 +-
 .../monitor/util/celltypes/CompactionsType.java |  12 +-
 .../monitor/util/celltypes/DateTimeType.java    |  12 +-
 .../monitor/util/celltypes/DurationType.java    |  10 +-
 .../monitor/util/celltypes/NumberType.java      |  26 +-
 .../monitor/util/celltypes/PercentageType.java  |  12 +-
 .../util/celltypes/PreciseNumberType.java       |   3 +-
 .../util/celltypes/ProgressChartType.java       |  14 +-
 .../monitor/util/celltypes/StringType.java      |   4 +-
 .../monitor/util/celltypes/TServerLinkType.java |  12 +-
 .../monitor/util/celltypes/TableLinkType.java   |  14 +-
 .../monitor/util/celltypes/TableStateType.java  |   8 +-
 .../accumulo/monitor/ShowTraceLinkTypeTest.java |   8 +-
 .../accumulo/monitor/ZooKeeperStatusTest.java   |  14 +-
 .../monitor/servlets/trace/SummaryTest.java     |   2 +-
 .../util/celltypes/PreciseNumberTypeTest.java   |   3 +-
 .../accumulo/tracer/AsyncSpanReceiver.java      |  46 +-
 .../accumulo/tracer/SendSpansViaThrift.java     |  22 +-
 .../org/apache/accumulo/tracer/SpanTree.java    |  10 +-
 .../org/apache/accumulo/tracer/TraceDump.java   |  18 +-
 .../apache/accumulo/tracer/TraceFormatter.java  |  27 +-
 .../org/apache/accumulo/tracer/TraceServer.java |   6 +-
 .../org/apache/accumulo/tracer/TracerTest.java  |  49 +-
 .../server/GarbageCollectionLogger.java         |  10 +-
 .../tserver/BulkFailedCopyProcessor.java        |   3 +-
 .../accumulo/tserver/CompactionQueue.java       |  54 +-
 .../tserver/ConditionalMutationSet.java         |  10 +-
 .../apache/accumulo/tserver/FileManager.java    | 214 +++---
 .../accumulo/tserver/HoldTimeoutException.java  |   2 +-
 .../apache/accumulo/tserver/InMemoryMap.java    | 280 ++++----
 .../org/apache/accumulo/tserver/MemKey.java     |  26 +-
 .../org/apache/accumulo/tserver/MemValue.java   |  18 +-
 .../accumulo/tserver/MinorCompactionReason.java |   2 +-
 .../org/apache/accumulo/tserver/Mutations.java  |   6 +-
 .../org/apache/accumulo/tserver/NativeMap.java  |  12 +-
 .../org/apache/accumulo/tserver/RowLocks.java   |  68 +-
 .../tserver/TConstraintViolationException.java  |   2 +-
 .../org/apache/accumulo/tserver/TLevel.java     |   8 +-
 .../tserver/TabletIteratorEnvironment.java      |  26 +-
 .../accumulo/tserver/TabletMutations.java       |   2 +-
 .../apache/accumulo/tserver/TabletServer.java   |  16 +-
 .../tserver/TabletServerResourceManager.java    |   4 +-
 .../accumulo/tserver/TabletStatsKeeper.java     |  36 +-
 .../accumulo/tserver/TooManyFilesException.java |   4 +-
 .../accumulo/tserver/TservConstraintEnv.java    |   2 +-
 .../apache/accumulo/tserver/WriteTracker.java   |   8 +-
 .../tserver/compaction/CompactionPlan.java      |   5 +-
 .../tserver/compaction/CompactionStrategy.java  |  10 +-
 .../compaction/SizeLimitCompactionStrategy.java |   2 +-
 .../tserver/constraints/ConstraintChecker.java  |  28 +-
 .../constraints/UnsatisfiableConstraint.java    |  10 +-
 .../tserver/data/ServerConditionalMutation.java |  11 +-
 .../apache/accumulo/tserver/log/DfsLogger.java  |  12 +-
 .../accumulo/tserver/log/LocalWALRecovery.java  |  12 +-
 .../accumulo/tserver/log/SortedLogRecovery.java |  51 +-
 .../tserver/log/TabletServerLogger.java         |   2 +-
 .../accumulo/tserver/logger/LogFileKey.java     |  22 +-
 .../accumulo/tserver/logger/LogFileValue.java   |  25 +-
 .../tserver/mastermessage/MasterMessage.java    |   4 +-
 .../mastermessage/SplitReportMessage.java       |  10 +-
 .../mastermessage/TabletStatusMessage.java      |   8 +-
 .../tserver/metrics/TabletServerMBean.java      |  30 +-
 .../replication/AccumuloReplicaSystem.java      |  12 +-
 .../replication/ReplicationServicerHandler.java |   3 +-
 .../tserver/replication/ReplicationWorker.java  |   2 +-
 .../accumulo/tserver/scan/LookupTask.java       |   8 +-
 .../accumulo/tserver/scan/NextBatchTask.java    |   4 +-
 .../accumulo/tserver/scan/ScanRunState.java     |   2 +-
 .../apache/accumulo/tserver/scan/ScanTask.java  |   2 +-
 .../tserver/session/ConditionalSession.java     |   6 +-
 .../tserver/session/MultiScanSession.java       |   5 +-
 .../accumulo/tserver/session/ScanSession.java   |   7 +-
 .../accumulo/tserver/session/Session.java       |   8 +-
 .../tserver/session/SessionManager.java         |   5 +-
 .../accumulo/tserver/session/UpdateSession.java |   8 +-
 .../apache/accumulo/tserver/tablet/Batch.java   |   4 +-
 .../accumulo/tserver/tablet/CommitSession.java  |   4 +-
 .../accumulo/tserver/tablet/CompactionInfo.java |   6 +-
 .../tserver/tablet/CompactionRunner.java        |   2 +-
 .../tserver/tablet/CompactionStats.java         |  18 +-
 .../tserver/tablet/CompactionWatcher.java       |  18 +-
 .../accumulo/tserver/tablet/Compactor.java      |   6 +-
 .../tserver/tablet/CountingIterator.java        |   2 +-
 .../apache/accumulo/tserver/tablet/KVEntry.java |   2 +-
 .../tserver/tablet/MinorCompactionTask.java     |   6 +-
 .../accumulo/tserver/tablet/MinorCompactor.java |  42 +-
 .../apache/accumulo/tserver/tablet/Rate.java    |  10 +-
 .../accumulo/tserver/tablet/RootFiles.java      |   2 +-
 .../accumulo/tserver/tablet/ScanBatch.java      |   2 +-
 .../accumulo/tserver/tablet/ScanDataSource.java |  10 +-
 .../accumulo/tserver/tablet/ScanOptions.java    |   6 +-
 .../apache/accumulo/tserver/tablet/Scanner.java |   6 +-
 .../accumulo/tserver/tablet/SplitInfo.java      |   8 +-
 .../accumulo/tserver/tablet/SplitRowSpec.java   |   2 +-
 .../apache/accumulo/tserver/tablet/Tablet.java  |   4 +-
 .../tserver/tablet/TabletClosedException.java   |   2 +-
 .../tserver/tablet/TabletCommitter.java         |   2 +-
 .../accumulo/tserver/tablet/TabletMemory.java   |   6 +-
 .../LargestFirstMemoryManagerTest.java          | 142 ++--
 .../tserver/CheckTabletMetadataTest.java        |  53 +-
 .../accumulo/tserver/CountingIteratorTest.java  |   2 +-
 .../accumulo/tserver/InMemoryMapTest.java       |  12 +-
 .../tserver/TabletServerSyncCheckTest.java      |   1 -
 .../tserver/TservConstraintEnvTest.java         |  12 +-
 .../tserver/compaction/CompactionPlanTest.java  |   3 +-
 .../SizeLimitCompactionStrategyTest.java        |   2 +-
 .../constraints/ConstraintCheckerTest.java      |  13 +-
 .../accumulo/tserver/log/DfsLoggerTest.java     |  20 +-
 .../tserver/log/SortedLogRecoveryTest.java      |   4 +-
 .../tserver/log/TestUpgradePathForWALogs.java   |   2 +-
 .../accumulo/tserver/logger/LogFileTest.java    |  13 +-
 .../replication/AccumuloReplicaSystemTest.java  |  14 +-
 .../BatchWriterReplicationReplayerTest.java     |   5 +-
 .../replication/ReplicationProcessorTest.java   |   5 +-
 .../accumulo/tserver/tablet/RootFilesTest.java  |   3 +-
 .../accumulo/tserver/tablet/TabletTest.java     |   4 +-
 .../java/org/apache/accumulo/shell/Shell.java   |   5 +-
 .../accumulo/shell/ShellCommandException.java   |  18 +-
 .../apache/accumulo/shell/ShellExtension.java   |   8 +-
 .../apache/accumulo/shell/ShellOptionsJC.java   |  11 +-
 .../org/apache/accumulo/shell/ShellUtil.java    |   2 +-
 .../java/org/apache/accumulo/shell/Token.java   |  32 +-
 .../accumulo/shell/commands/AboutCommand.java   |   8 +-
 .../commands/ActiveCompactionIterator.java      |  38 +-
 .../shell/commands/ActiveScanIterator.java      |  40 +-
 .../shell/commands/AddAuthsCommand.java         |  12 +-
 .../shell/commands/AddSplitsCommand.java        |  26 +-
 .../shell/commands/AuthenticateCommand.java     |  10 +-
 .../shell/commands/ClasspathCommand.java        |   4 +-
 .../accumulo/shell/commands/ClearCommand.java   |   8 +-
 .../shell/commands/CloneTableCommand.java       |  30 +-
 .../accumulo/shell/commands/CompactCommand.java |  26 +-
 .../accumulo/shell/commands/ConfigCommand.java  |   4 +-
 .../shell/commands/ConstraintCommand.java       |   2 +-
 .../shell/commands/CreateNamespaceCommand.java  |   2 +-
 .../shell/commands/CreateTableCommand.java      |   4 +-
 .../shell/commands/CreateUserCommand.java       |  12 +-
 .../accumulo/shell/commands/DebugCommand.java   |  10 +-
 .../accumulo/shell/commands/DeleteCommand.java  |  26 +-
 .../shell/commands/DeleteManyCommand.java       |  22 +-
 .../shell/commands/DeleteNamespaceCommand.java  |   2 +-
 .../shell/commands/DeleteRowsCommand.java       |   8 +-
 .../shell/commands/DeleteScanIterCommand.java   |  22 +-
 .../shell/commands/DeleteShellIterCommand.java  |  22 +-
 .../accumulo/shell/commands/EGrepCommand.java   |  10 +-
 .../shell/commands/ExecfileCommand.java         |  10 +-
 .../accumulo/shell/commands/ExitCommand.java    |   4 +-
 .../shell/commands/ExportTableCommand.java      |  26 +-
 .../shell/commands/ExtensionCommand.java        |  28 +-
 .../accumulo/shell/commands/FateCommand.java    |  38 +-
 .../accumulo/shell/commands/FlushCommand.java   |  12 +-
 .../shell/commands/FormatterCommand.java        |  17 +-
 .../shell/commands/GetAuthsCommand.java         |   8 +-
 .../shell/commands/GetGroupsCommand.java        |  12 +-
 .../shell/commands/GetSplitsCommand.java        |  36 +-
 .../accumulo/shell/commands/GrantCommand.java   |   2 +-
 .../accumulo/shell/commands/GrepCommand.java    |  28 +-
 .../accumulo/shell/commands/HelpCommand.java    |  16 +-
 .../accumulo/shell/commands/HiddenCommand.java  |  10 +-
 .../accumulo/shell/commands/HistoryCommand.java |  10 +-
 .../shell/commands/ImportDirectoryCommand.java  |  12 +-
 .../shell/commands/ImportTableCommand.java      |  14 +-
 .../accumulo/shell/commands/InsertCommand.java  |  47 +-
 .../shell/commands/InterpreterCommand.java      |   8 +-
 .../shell/commands/ListCompactionsCommand.java  |  30 +-
 .../shell/commands/ListIterCommand.java         |   2 +-
 .../shell/commands/ListScansCommand.java        |  30 +-
 .../shell/commands/ListShellIterCommand.java    |  18 +-
 .../accumulo/shell/commands/MaxRowCommand.java  |  14 +-
 .../accumulo/shell/commands/MergeCommand.java   |  10 +-
 .../accumulo/shell/commands/NoTableCommand.java |   6 +-
 .../accumulo/shell/commands/OfflineCommand.java |  13 +-
 .../accumulo/shell/commands/OnlineCommand.java  |  12 +-
 .../accumulo/shell/commands/PasswdCommand.java  |  19 +-
 .../accumulo/shell/commands/PingCommand.java    |  33 +-
 .../accumulo/shell/commands/PingIterator.java   |  14 +-
 .../shell/commands/QuotedStringTokenizer.java   |  16 +-
 .../shell/commands/RenameNamespaceCommand.java  |   2 +-
 .../shell/commands/RenameTableCommand.java      |   2 +-
 .../accumulo/shell/commands/RevokeCommand.java  |   2 +-
 .../accumulo/shell/commands/ScanCommand.java    |  90 +--
 .../accumulo/shell/commands/ScriptCommand.java  |  58 +-
 .../shell/commands/SetAuthsCommand.java         |  12 +-
 .../shell/commands/SetGroupsCommand.java        |  22 +-
 .../accumulo/shell/commands/SetIterCommand.java |  28 +-
 .../shell/commands/SetScanIterCommand.java      |  23 +-
 .../shell/commands/SetShellIterCommand.java     |  19 +-
 .../ShellPluginConfigurationCommand.java        |  57 +-
 .../accumulo/shell/commands/SleepCommand.java   |   8 +-
 .../commands/SystemPermissionsCommand.java      |   4 +-
 .../accumulo/shell/commands/TableCommand.java   |  13 +-
 .../accumulo/shell/commands/TableOperation.java |   4 +-
 .../shell/commands/TablePermissionsCommand.java |   4 +-
 .../accumulo/shell/commands/TraceCommand.java   |   8 +-
 .../accumulo/shell/commands/UserCommand.java    |  12 +-
 .../shell/commands/UserPermissionsCommand.java  |   3 +-
 .../accumulo/shell/commands/UsersCommand.java   |   4 +-
 .../accumulo/shell/commands/WhoAmICommand.java  |   4 +-
 .../accumulo/shell/format/DeleterFormatter.java |   8 +-
 .../apache/accumulo/shell/mock/MockShell.java   |  31 +-
 .../accumulo/shell/PasswordConverterTest.java   |  24 +-
 .../apache/accumulo/shell/ShellConfigTest.java  |  17 +-
 .../org/apache/accumulo/shell/ShellTest.java    |   9 +-
 .../apache/accumulo/shell/ShellUtilTest.java    |   7 +-
 .../shell/commands/DeleteTableCommandTest.java  |   2 +-
 .../shell/commands/DropUserCommandTest.java     |   9 +-
 .../shell/commands/FormatterCommandTest.java    |  72 +-
 .../shell/commands/HistoryCommandTest.java      |   1 -
 .../shell/format/DeleterFormatterTest.java      |   1 -
 .../java/org/apache/accumulo/start/Main.java    |   6 +-
 .../org/apache/accumulo/start/TestMain.java     |   8 +-
 .../start/classloader/AccumuloClassLoader.java  |  68 +-
 .../vfs/AccumuloReloadingVFSClassLoader.java    |  36 +-
 .../classloader/vfs/AccumuloVFSClassLoader.java |  22 +-
 .../start/classloader/vfs/ContextManager.java   |  82 +--
 .../start/classloader/vfs/MiniDFSUtil.java      |   8 +-
 .../vfs/PostDelegatingVFSClassLoader.java       |   8 +-
 .../classloader/vfs/ReloadingClassLoader.java   |   2 +-
 .../classloader/vfs/UniqueFileReplicator.java   |   2 +-
 .../vfs/providers/HdfsFileAttributes.java       |  61 +-
 .../providers/HdfsFileContentInfoFactory.java   |  36 +-
 .../vfs/providers/HdfsFileObject.java           | 515 +++++++-------
 .../vfs/providers/HdfsFileProvider.java         |  81 +--
 .../vfs/providers/HdfsFileSystem.java           |  87 +--
 .../providers/HdfsFileSystemConfigBuilder.java  |  35 +-
 .../vfs/providers/HdfsRandomAccessContent.java  |  76 +--
 .../AccumuloReloadingVFSClassLoaderTest.java    |   7 +-
 .../classloader/vfs/ContextManagerTest.java     |   2 +-
 .../providers/ReadOnlyHdfsFileProviderTest.java |  76 +--
 .../apache/accumulo/test/AccumuloDFSBase.java   |   6 +-
 start/src/test/java/test/Test.java              |   6 +-
 .../accumulo/test/BulkImportDirectory.java      |   9 +-
 .../org/apache/accumulo/test/CreateRFiles.java  |  57 +-
 .../apache/accumulo/test/CreateRandomRFile.java |  28 +-
 .../accumulo/test/EstimateInMemMapOverhead.java | 152 ++---
 .../accumulo/test/FaultyConditionalWriter.java  |  19 +-
 .../apache/accumulo/test/GetMasterStats.java    |   6 +-
 .../apache/accumulo/test/IMMLGBenchmark.java    |  63 +-
 .../org/apache/accumulo/test/ListTables.java    |   2 +-
 .../accumulo/test/NativeMapConcurrencyTest.java |  94 +--
 .../accumulo/test/NativeMapPerformanceTest.java |  72 +-
 .../accumulo/test/NativeMapStressTest.java      | 124 ++--
 .../apache/accumulo/test/NullBatchWriter.java   |  12 +-
 .../accumulo/test/QueryMetadataTable.java       |  54 +-
 .../apache/accumulo/test/TestBinaryRows.java    |  88 +--
 .../org/apache/accumulo/test/TestIngest.java    | 156 ++---
 .../apache/accumulo/test/TestRandomDeletes.java |  41 +-
 .../org/apache/accumulo/test/VerifyIngest.java  |   1 -
 .../apache/accumulo/test/WrongTabletTest.java   |   4 +-
 .../test/continuous/ContinuousBatchWalker.java  |  82 +--
 .../test/continuous/ContinuousIngest.java       | 121 ++--
 .../test/continuous/ContinuousMoru.java         |  60 +-
 .../test/continuous/ContinuousQuery.java        |  20 +-
 .../test/continuous/ContinuousScanner.java      |  36 +-
 .../continuous/ContinuousStatsCollector.java    |  64 +-
 .../test/continuous/ContinuousVerify.java       |   2 +-
 .../test/continuous/ContinuousWalk.java         |  88 +--
 .../accumulo/test/continuous/GenSplits.java     |  24 +-
 .../accumulo/test/continuous/Histogram.java     |  64 +-
 .../test/continuous/PrintScanTimeHistogram.java |  27 +-
 .../accumulo/test/continuous/TimeBinner.java    |  62 +-
 .../test/continuous/UndefinedAnalyzer.java      |   2 +-
 .../accumulo/test/functional/BadCombiner.java   |   4 +-
 .../accumulo/test/functional/BadIterator.java   |   6 +-
 .../test/functional/CacheTestClean.java         |   8 +-
 .../test/functional/CacheTestReader.java        |  26 +-
 .../test/functional/CacheTestWriter.java        |  68 +-
 .../accumulo/test/functional/DropModIter.java   |  10 +-
 .../test/functional/SlowConstraint.java         |   8 +-
 .../accumulo/test/functional/SlowIterator.java  |   4 +-
 .../accumulo/test/functional/ZombieTServer.java |   4 +-
 .../metadata/MetadataBatchScanTest.java         | 116 ++--
 .../performance/scan/CollectTabletStats.java    | 264 ++++----
 .../test/performance/thrift/NullTserver.java    |   3 +-
 .../accumulo/test/randomwalk/Environment.java   |  20 +-
 .../accumulo/test/randomwalk/Fixture.java       |   6 +-
 .../accumulo/test/randomwalk/Framework.java     |  40 +-
 .../apache/accumulo/test/randomwalk/Module.java |   3 +-
 .../apache/accumulo/test/randomwalk/Node.java   |  16 +-
 .../apache/accumulo/test/randomwalk/State.java  |  45 +-
 .../apache/accumulo/test/randomwalk/Test.java   |   2 +-
 .../accumulo/test/randomwalk/bulk/Setup.java    |  16 +-
 .../accumulo/test/randomwalk/bulk/Verify.java   |  24 +-
 .../test/randomwalk/concurrent/AddSplits.java   |  12 +-
 .../test/randomwalk/concurrent/Apocalypse.java  |   4 +-
 .../test/randomwalk/concurrent/BatchScan.java   |  16 +-
 .../test/randomwalk/concurrent/BatchWrite.java  |  14 +-
 .../test/randomwalk/concurrent/BulkImport.java  |  44 +-
 .../concurrent/ChangeAuthorizations.java        |  12 +-
 .../concurrent/ChangePermissions.java           |  38 +-
 .../randomwalk/concurrent/CheckBalance.java     |   8 +-
 .../test/randomwalk/concurrent/Compact.java     |  14 +-
 .../concurrent/ConcurrentFixture.java           |  28 +-
 .../test/randomwalk/concurrent/CreateTable.java |  10 +-
 .../test/randomwalk/concurrent/CreateUser.java  |   8 +-
 .../test/randomwalk/concurrent/DeleteRange.java |  12 +-
 .../test/randomwalk/concurrent/DeleteTable.java |  10 +-
 .../test/randomwalk/concurrent/DropUser.java    |   8 +-
 .../randomwalk/concurrent/IsolatedScan.java     |  12 +-
 .../test/randomwalk/concurrent/ListSplits.java  |  10 +-
 .../test/randomwalk/concurrent/Merge.java       |  12 +-
 .../randomwalk/concurrent/OfflineTable.java     |  12 +-
 .../test/randomwalk/concurrent/RenameTable.java |  20 +-
 .../test/randomwalk/concurrent/ScanTable.java   |  10 +-
 .../test/randomwalk/concurrent/Setup.java       |  18 +-
 .../test/randomwalk/concurrent/Shutdown.java    |  10 +-
 .../test/randomwalk/concurrent/StartAll.java    |   4 +-
 .../randomwalk/concurrent/StopTabletServer.java |  10 +-
 .../test/randomwalk/conditional/Compact.java    |   2 +-
 .../test/randomwalk/conditional/Flush.java      |   2 +-
 .../test/randomwalk/conditional/Init.java       |  14 +-
 .../test/randomwalk/conditional/Merge.java      |   2 +-
 .../test/randomwalk/conditional/Setup.java      |   6 +-
 .../test/randomwalk/conditional/Split.java      |   2 +-
 .../test/randomwalk/conditional/TearDown.java   |   2 +-
 .../test/randomwalk/conditional/Transfer.java   |   8 +-
 .../test/randomwalk/conditional/Utils.java      |   3 +-
 .../test/randomwalk/conditional/Verify.java     |   3 +-
 .../accumulo/test/randomwalk/image/Commit.java  |   6 +-
 .../test/randomwalk/image/ImageFixture.java     |  44 +-
 .../test/randomwalk/image/ScanMeta.java         |  54 +-
 .../accumulo/test/randomwalk/image/TableOp.java |  14 +-
 .../accumulo/test/randomwalk/image/Verify.java  |  50 +-
 .../accumulo/test/randomwalk/image/Write.java   |  30 +-
 .../test/randomwalk/multitable/Commit.java      |  10 +-
 .../test/randomwalk/multitable/CopyTool.java    |  16 +-
 .../test/randomwalk/multitable/DropTable.java   |  10 +-
 .../multitable/MultiTableFixture.java           |  18 +-
 .../randomwalk/multitable/OfflineTable.java     |  10 +-
 .../test/randomwalk/multitable/Write.java       |  24 +-
 .../randomwalk/security/AlterSystemPerm.java    |  20 +-
 .../test/randomwalk/security/AlterTable.java    |  24 +-
 .../randomwalk/security/AlterTablePerm.java     |  56 +-
 .../test/randomwalk/security/Authenticate.java  |  28 +-
 .../test/randomwalk/security/ChangePass.java    |  36 +-
 .../test/randomwalk/security/CreateTable.java   |  22 +-
 .../test/randomwalk/security/CreateUser.java    |  18 +-
 .../test/randomwalk/security/DropTable.java     |  33 +-
 .../test/randomwalk/security/DropUser.java      |  22 +-
 .../randomwalk/security/SecurityFixture.java    |  74 +--
 .../randomwalk/security/SecurityHelper.java     |  80 +--
 .../test/randomwalk/security/SetAuths.java      |  31 +-
 .../test/randomwalk/security/TableOp.java       |  74 +--
 .../test/randomwalk/security/Validate.java      |  42 +-
 .../randomwalk/security/WalkingSecurity.java    |   2 +-
 .../test/randomwalk/sequential/BatchVerify.java |  40 +-
 .../test/randomwalk/sequential/Commit.java      |   8 +-
 .../randomwalk/sequential/MapRedVerify.java     |  16 +-
 .../randomwalk/sequential/MapRedVerifyTool.java |  24 +-
 .../sequential/SequentialFixture.java           |  26 +-
 .../test/randomwalk/sequential/Write.java       |  10 +-
 .../test/randomwalk/shard/BulkInsert.java       |  73 +-
 .../test/randomwalk/shard/CloneIndex.java       |  12 +-
 .../accumulo/test/randomwalk/shard/Commit.java  |   4 +-
 .../test/randomwalk/shard/CompactFilter.java    |  24 +-
 .../accumulo/test/randomwalk/shard/Delete.java  |  18 +-
 .../test/randomwalk/shard/DeleteSomeDocs.java   |  30 +-
 .../test/randomwalk/shard/DeleteWord.java       |  38 +-
 .../test/randomwalk/shard/ExportIndex.java      |  40 +-
 .../accumulo/test/randomwalk/shard/Flush.java   |  10 +-
 .../accumulo/test/randomwalk/shard/Grep.java    |  36 +-
 .../accumulo/test/randomwalk/shard/Insert.java  |  64 +-
 .../accumulo/test/randomwalk/shard/Merge.java   |   6 +-
 .../accumulo/test/randomwalk/shard/Reindex.java |  24 +-
 .../accumulo/test/randomwalk/shard/Search.java  |  36 +-
 .../test/randomwalk/shard/ShardFixture.java     |   4 +-
 .../test/randomwalk/shard/SortTool.java         |  18 +-
 .../accumulo/test/randomwalk/shard/Split.java   |   6 +-
 .../test/randomwalk/shard/VerifyIndex.java      |  24 +-
 .../test/randomwalk/unit/CreateTable.java       |   2 +-
 .../test/randomwalk/unit/DeleteTable.java       |   2 +-
 .../accumulo/test/randomwalk/unit/Ingest.java   |   2 +-
 .../accumulo/test/randomwalk/unit/Scan.java     |   2 +-
 .../accumulo/test/randomwalk/unit/Verify.java   |   2 +-
 .../ReplicationTablesPrinterThread.java         |   2 +-
 .../test/replication/merkle/MerkleTreeNode.java |   2 +-
 .../replication/merkle/RangeSerialization.java  |   4 +-
 .../replication/merkle/cli/CompareTables.java   |   6 +-
 .../replication/merkle/cli/ComputeRootHash.java |   2 +-
 .../replication/merkle/cli/GenerateHashes.java  |   3 +-
 .../accumulo/test/scalability/Ingest.java       |  47 +-
 .../apache/accumulo/test/scalability/Run.java   |  34 +-
 .../accumulo/test/scalability/ScaleTest.java    |  26 +-
 .../accumulo/test/stress/random/DataWriter.java |   6 +-
 .../test/stress/random/IntArgValidator.java     |   6 +-
 .../test/stress/random/RandomByteArrays.java    |   4 +-
 .../test/stress/random/RandomMutations.java     |   9 +-
 .../test/stress/random/RandomWithinRange.java   |  11 +-
 .../accumulo/test/stress/random/Scan.java       |  48 +-
 .../accumulo/test/stress/random/ScanOpts.java   |  13 +-
 .../accumulo/test/stress/random/Stream.java     |  10 +-
 .../accumulo/test/stress/random/Write.java      |  69 +-
 .../test/stress/random/WriteOptions.java        |  80 ++-
 .../test/stress/random/package-info.java        |   9 +-
 .../accumulo/fate/zookeeper/ZooLockTest.java    |  16 +-
 .../accumulo/proxy/ProxyDurabilityIT.java       |  22 +-
 .../apache/accumulo/proxy/SimpleProxyIT.java    |   9 +-
 .../proxy/TestProxyInstanceOperations.java      |  22 +-
 .../accumulo/proxy/TestProxyReadWrite.java      | 172 ++---
 .../proxy/TestProxySecurityOperations.java      |  42 +-
 .../proxy/TestProxyTableOperations.java         |  56 +-
 .../server/security/SystemCredentialsIT.java    |   7 +-
 .../accumulo/test/AssignmentThreadsIT.java      |  12 +-
 .../apache/accumulo/test/BalanceFasterIT.java   |  12 +-
 .../accumulo/test/BulkImportVolumeIT.java       |   1 -
 .../accumulo/test/ConditionalWriterIT.java      |   8 +-
 .../accumulo/test/InterruptibleScannersIT.java  |   7 +-
 .../accumulo/test/KeyValueEqualityIT.java       |   3 +-
 .../apache/accumulo/test/MetaGetsReadersIT.java |  14 +-
 .../org/apache/accumulo/test/NamespacesIT.java  |   2 +-
 .../accumulo/test/NoMutationRecoveryIT.java     |   8 +-
 .../test/RecoveryCompactionsAreFlushesIT.java   |   9 +-
 .../test/RewriteTabletDirectoriesIT.java        |   2 +-
 .../org/apache/accumulo/test/ShellServerIT.java |   2 +-
 .../test/TableConfigurationUpdateIT.java        |   7 +-
 .../accumulo/test/TabletServerGivesUpIT.java    |  10 +-
 .../org/apache/accumulo/test/TotalQueuedIT.java |  14 +-
 .../test/TraceRepoDeserializationTest.java      |  32 +-
 .../accumulo/test/VerifySerialRecoveryIT.java   |  21 +-
 .../test/functional/BackupMasterIT.java         |   2 +-
 .../BalanceInPresenceOfOfflineTableIT.java      |  12 +-
 .../test/functional/BinaryStressIT.java         |   4 +-
 .../accumulo/test/functional/BloomFilterIT.java |   6 +-
 .../apache/accumulo/test/functional/BulkIT.java |   8 +-
 .../test/functional/ChaoticBalancerIT.java      |   4 +-
 .../test/functional/DeleteEverythingIT.java     |   2 +-
 .../functional/DeleteTableDuringSplitIT.java    |   5 +-
 .../functional/DeletedTablesDontFlushIT.java    |   2 +-
 .../accumulo/test/functional/DurabilityIT.java  |  26 +-
 .../accumulo/test/functional/ExamplesIT.java    |  10 +-
 .../test/functional/FunctionalTestUtils.java    |   7 +-
 .../test/functional/HalfDeadTServerIT.java      |   4 +-
 .../test/functional/LateLastContactIT.java      |   7 +-
 .../accumulo/test/functional/LogicalTimeIT.java |   3 +-
 .../accumulo/test/functional/MaxOpenIT.java     |   2 +-
 .../test/functional/MetadataMaxFilesIT.java     |   8 +-
 .../test/functional/MetadataSplitIT.java        |   6 +-
 .../test/functional/MonitorLoggingIT.java       |   4 +-
 .../accumulo/test/functional/NativeMapIT.java   |   2 +-
 .../functional/RecoveryWithEmptyRFileIT.java    |  16 +-
 .../test/functional/RestartStressIT.java        |   4 +-
 .../accumulo/test/functional/RowDeleteIT.java   |   4 +-
 .../accumulo/test/functional/ScanIdIT.java      |  52 +-
 .../test/functional/SessionDurabilityIT.java    |   2 +-
 .../accumulo/test/functional/ShutdownIT.java    |  18 +-
 .../test/functional/SplitRecoveryIT.java        |  90 +--
 .../test/functional/WatchTheWatchCountIT.java   |   5 +-
 .../test/functional/WriteAheadLogIT.java        |   2 +-
 .../test/functional/ZookeeperRestartIT.java     |  14 +-
 .../accumulo/test/iterator/RegExTest.java       |  52 +-
 .../test/replication/CyclicReplicationIT.java   |   8 +-
 ...bageCollectorCommunicatesWithTServersIT.java |   2 +-
 .../replication/MultiTserverReplicationIT.java  |   2 +-
 ...UnusedWalDoesntCloseReplicationStatusIT.java |   3 +-
 .../apache/accumulo/test/util/CertUtils.java    |   5 +-
 .../trace/instrument/CloudtraceSpan.java        |  30 +-
 .../accumulo/trace/instrument/CountSampler.java |   2 +-
 .../accumulo/trace/instrument/Sampler.java      |   4 +-
 .../apache/accumulo/trace/instrument/Span.java  |   2 +-
 1257 files changed, 19525 insertions(+), 19345 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/bloomfilter/BloomFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/bloomfilter/BloomFilter.java b/core/src/main/java/org/apache/accumulo/core/bloomfilter/BloomFilter.java
index 0cef8d3..4726e83 100644
--- a/core/src/main/java/org/apache/accumulo/core/bloomfilter/BloomFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/bloomfilter/BloomFilter.java
@@ -2,30 +2,30 @@
  *
  * Copyright (c) 2005, European Commission project OneLab under contract 034819 (http://www.one-lab.org)
  * All rights reserved.
- * Redistribution and use in source and binary forms, with or 
- * without modification, are permitted provided that the following 
+ * Redistribution and use in source and binary forms, with or
+ * without modification, are permitted provided that the following
  * conditions are met:
- *  - Redistributions of source code must retain the above copyright 
+ *  - Redistributions of source code must retain the above copyright
  *    notice, this list of conditions and the following disclaimer.
- *  - Redistributions in binary form must reproduce the above copyright 
- *    notice, this list of conditions and the following disclaimer in 
+ *  - Redistributions in binary form must reproduce the above copyright
+ *    notice, this list of conditions and the following disclaimer in
  *    the documentation and/or other materials provided with the distribution.
  *  - Neither the name of the University Catholique de Louvain - UCL
- *    nor the names of its contributors may be used to endorse or 
- *    promote products derived from this software without specific prior 
+ *    nor the names of its contributors may be used to endorse or
+ *    promote products derived from this software without specific prior
  *    written permission.
- *    
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 
- * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE 
- * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 
- * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, 
- * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER 
- * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN 
- * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
+ * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
+ * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
+ * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
+ * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
+ * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
  * POSSIBILITY OF SUCH DAMAGE.
  */
 
@@ -70,29 +70,29 @@ import org.apache.log4j.Logger;
  * of elements. The receiver uses the filter to test whether various elements are members of the set. Though the filter will occasionally return a false
  * positive, it will never return a false negative. When creating the filter, the sender can choose its desired point in a trade-off between the false positive
  * rate and the size.
- * 
+ *
  * <p>
  * Originally created by <a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.
- * 
+ *
  * @see Filter The general behavior of a filter
- * 
+ *
  * @see <a href="http://portal.acm.org/citation.cfm?id=362692&dl=ACM&coll=portal">Space/Time Trade-Offs in Hash Coding with Allowable Errors</a>
  */
 public class BloomFilter extends Filter {
   private static final Logger log = Logger.getLogger(BloomFilter.class);
   private static final byte[] bitvalues = new byte[] {(byte) 0x01, (byte) 0x02, (byte) 0x04, (byte) 0x08, (byte) 0x10, (byte) 0x20, (byte) 0x40, (byte) 0x80};
-  
+
   /** The bit vector. */
   BitSet bits;
-  
+
   /** Default constructor - use with readFields */
   public BloomFilter() {
     super();
   }
-  
+
   /**
    * Constructor
-   * 
+   *
    * @param vectorSize
    *          The vector size of <i>this</i> filter.
    * @param nbHash
@@ -102,44 +102,44 @@ public class BloomFilter extends Filter {
    */
   public BloomFilter(final int vectorSize, final int nbHash, final int hashType) {
     super(vectorSize, nbHash, hashType);
-    
+
     bits = new BitSet(this.vectorSize);
   }
-  
+
   @Override
   public boolean add(final Key key) {
     if (key == null) {
       throw new NullPointerException("key cannot be null");
     }
-    
+
     int[] h = hash.hash(key);
     hash.clear();
-    
+
     boolean bitsSet = false;
 
     for (int i = 0; i < nbHash; i++) {
       bitsSet |= !bits.get(h[i]);
       bits.set(h[i]);
     }
-    
+
     return bitsSet;
   }
-  
+
   @Override
   public void and(final Filter filter) {
     if (filter == null || !(filter instanceof BloomFilter) || filter.vectorSize != this.vectorSize || filter.nbHash != this.nbHash) {
       throw new IllegalArgumentException("filters cannot be and-ed");
     }
-    
+
     this.bits.and(((BloomFilter) filter).bits);
   }
-  
+
   @Override
   public boolean membershipTest(final Key key) {
     if (key == null) {
       throw new NullPointerException("key cannot be null");
     }
-    
+
     int[] h = hash.hash(key);
     hash.clear();
     for (int i = 0; i < nbHash; i++) {
@@ -149,12 +149,12 @@ public class BloomFilter extends Filter {
     }
     return true;
   }
-  
+
   @Override
   public void not() {
     bits.flip(0, vectorSize - 1);
   }
-  
+
   @Override
   public void or(final Filter filter) {
     if (filter == null || !(filter instanceof BloomFilter) || filter.vectorSize != this.vectorSize || filter.nbHash != this.nbHash) {
@@ -162,7 +162,7 @@ public class BloomFilter extends Filter {
     }
     bits.or(((BloomFilter) filter).bits);
   }
-  
+
   @Override
   public void xor(final Filter filter) {
     if (filter == null || !(filter instanceof BloomFilter) || filter.vectorSize != this.vectorSize || filter.nbHash != this.nbHash) {
@@ -170,47 +170,47 @@ public class BloomFilter extends Filter {
     }
     bits.xor(((BloomFilter) filter).bits);
   }
-  
+
   @Override
   public String toString() {
     return bits.toString();
   }
-  
+
   /**
    * @return size of the the bloomfilter
    */
   public int getVectorSize() {
     return this.vectorSize;
   }
-  
+
   // Writable
-  
+
   @Override
   public void write(final DataOutput out) throws IOException {
     super.write(out);
-    
+
     ByteArrayOutputStream boas = new ByteArrayOutputStream();
     ObjectOutputStream oos = new ObjectOutputStream(boas);
-    
+
     oos.writeObject(bits);
     oos.flush();
     oos.close();
     out.write(boas.toByteArray());
   }
-  
+
   @Override
   public void readFields(final DataInput in) throws IOException {
-    
+
     super.readFields(in);
-    
+
     bits = new BitSet(this.vectorSize);
     byte[] bytes = null;
-    
+
     if (super.getSerialVersion() != super.getVersion()) {
       bytes = new byte[getNBytes()];
       in.readFully(bytes);
     }
-    
+
     if (super.getSerialVersion() == super.getVersion()) {
       // ByteArrayInputStream bais = new ByteArrayInputStream(bytes);
       // ObjectInputStream ois = new ObjectInputStream(bais);
@@ -222,7 +222,7 @@ public class BloomFilter extends Filter {
         throw new IOException("BloomFilter tried to deserialize as bitset: " + e);
       }
       // can not close ois, it would close in
-      
+
     } else {
       for (int i = 0, byteIndex = 0, bitIndex = 0; i < vectorSize; i++, bitIndex++) {
         if (bitIndex == 8) {
@@ -234,12 +234,12 @@ public class BloomFilter extends Filter {
         }
       }
     }
-    
+
   }
-  
+
   /* @return number of bytes needed to hold bit vector */
   private int getNBytes() {
     return (vectorSize + 7) / 8;
   }
-  
+
 }// end class


[28/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/rfile/MultiLevelIndexTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/rfile/MultiLevelIndexTest.java b/core/src/test/java/org/apache/accumulo/core/file/rfile/MultiLevelIndexTest.java
index 0487495..6f89454 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/rfile/MultiLevelIndexTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/rfile/MultiLevelIndexTest.java
@@ -39,44 +39,44 @@ import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
 
 public class MultiLevelIndexTest extends TestCase {
-  
+
   public void test1() throws Exception {
-    
+
     runTest(500, 1);
     runTest(500, 10);
     runTest(500, 100);
     runTest(500, 1000);
     runTest(500, 10000);
-    
+
     runTest(1, 100);
   }
-  
+
   private void runTest(int maxBlockSize, int num) throws IOException {
     AccumuloConfiguration aconf = AccumuloConfiguration.getDefaultConfiguration();
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     FSDataOutputStream dos = new FSDataOutputStream(baos, new FileSystem.Statistics("a"));
     CachableBlockFile.Writer _cbw = new CachableBlockFile.Writer(dos, "gz", CachedConfiguration.getInstance(), aconf);
-    
+
     BufferedWriter mliw = new BufferedWriter(new Writer(_cbw, maxBlockSize));
-    
+
     for (int i = 0; i < num; i++)
       mliw.add(new Key(String.format("%05d000", i)), i, 0, 0, 0);
-    
+
     mliw.addLast(new Key(String.format("%05d000", num)), num, 0, 0, 0);
-    
+
     ABlockWriter root = _cbw.prepareMetaBlock("root");
     mliw.close(root);
     root.close();
-    
+
     _cbw.close();
     dos.close();
     baos.close();
-    
+
     byte[] data = baos.toByteArray();
     SeekableByteArrayInputStream bais = new SeekableByteArrayInputStream(data);
     FSDataInputStream in = new FSDataInputStream(bais);
     CachableBlockFile.Reader _cbr = new CachableBlockFile.Reader(in, data.length, CachedConfiguration.getInstance(), aconf);
-    
+
     Reader reader = new Reader(_cbr, RFile.RINDEX_VER_7);
     BlockRead rootIn = _cbr.getMetaBlock("root");
     reader.readFields(rootIn);
@@ -89,22 +89,22 @@ public class MultiLevelIndexTest extends TestCase {
       assertEquals(count, liter.next().getNumEntries());
       count++;
     }
-    
+
     assertEquals(num + 1, count);
-    
+
     while (liter.hasPrevious()) {
       count--;
       assertEquals(count, liter.previousIndex());
       assertEquals(count, liter.peekPrevious().getNumEntries());
       assertEquals(count, liter.previous().getNumEntries());
     }
-    
+
     assertEquals(0, count);
-    
+
     // go past the end
     liter = reader.lookup(new Key(String.format("%05d000", num + 1)));
     assertFalse(liter.hasNext());
-    
+
     Random rand = new Random();
     for (int i = 0; i < 100; i++) {
       int k = rand.nextInt(num * 1000);
@@ -117,7 +117,7 @@ public class MultiLevelIndexTest extends TestCase {
       IndexEntry ie = liter.next();
       assertEquals(expected, ie.getNumEntries());
     }
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/rfile/RFileTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/rfile/RFileTest.java b/core/src/test/java/org/apache/accumulo/core/file/rfile/RFileTest.java
index 969b179..1a83f33 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/rfile/RFileTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/rfile/RFileTest.java
@@ -178,7 +178,7 @@ public class RFileTest {
     private AccumuloConfiguration accumuloConfiguration;
     public Reader reader;
     public SortedKeyValueIterator<Key,Value> iter;
-    
+
     public TestRFile(AccumuloConfiguration accumuloConfiguration) {
       this.accumuloConfiguration = accumuloConfiguration;
       if (this.accumuloConfiguration == null)
@@ -250,7 +250,7 @@ public class RFileTest {
   static String nf(String prefix, int i) {
     return String.format(prefix + "%06d", i);
   }
-  
+
   public AccumuloConfiguration conf = null;
 
   @Test
@@ -1761,15 +1761,16 @@ public class RFileTest {
     trf.closeWriter();
 
     byte[] rfBytes = trf.baos.toByteArray();
-    
+
     // If we get here, we have encrypted bytes
     for (Property prop : Property.values()) {
       if (prop.isSensitive()) {
         byte[] toCheck = prop.getKey().getBytes();
-        assertEquals(-1, Bytes.indexOf(rfBytes, toCheck));  }
-    }    
+        assertEquals(-1, Bytes.indexOf(rfBytes, toCheck));
+      }
+    }
   }
-  
+
   @Test
   public void testRootTabletEncryption() throws Exception {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/file/rfile/RelativeKeyTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/file/rfile/RelativeKeyTest.java b/core/src/test/java/org/apache/accumulo/core/file/rfile/RelativeKeyTest.java
index 8c0e691..e413448 100644
--- a/core/src/test/java/org/apache/accumulo/core/file/rfile/RelativeKeyTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/file/rfile/RelativeKeyTest.java
@@ -37,7 +37,7 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 
 public class RelativeKeyTest {
-  
+
   @Test
   public void testBasicRelativeKey() {
     assertEquals(1, UnsynchronizedBuffer.nextArraySize(0));
@@ -48,11 +48,11 @@ public class RelativeKeyTest {
     assertEquals(8, UnsynchronizedBuffer.nextArraySize(5));
     assertEquals(8, UnsynchronizedBuffer.nextArraySize(8));
     assertEquals(16, UnsynchronizedBuffer.nextArraySize(9));
-    
+
     assertEquals(1 << 16, UnsynchronizedBuffer.nextArraySize((1 << 16) - 1));
     assertEquals(1 << 16, UnsynchronizedBuffer.nextArraySize(1 << 16));
     assertEquals(1 << 17, UnsynchronizedBuffer.nextArraySize((1 << 16) + 1));
-    
+
     assertEquals(1 << 30, UnsynchronizedBuffer.nextArraySize((1 << 30) - 1));
 
     assertEquals(1 << 30, UnsynchronizedBuffer.nextArraySize(1 << 30));
@@ -60,7 +60,7 @@ public class RelativeKeyTest {
     assertEquals(Integer.MAX_VALUE, UnsynchronizedBuffer.nextArraySize(Integer.MAX_VALUE - 1));
     assertEquals(Integer.MAX_VALUE, UnsynchronizedBuffer.nextArraySize(Integer.MAX_VALUE));
   }
-  
+
   @Test
   public void testCommonPrefix() {
     // exact matches
@@ -73,13 +73,13 @@ public class RelativeKeyTest {
     assertEquals(-1, commonPrefixHelper("abab", "abab"));
     assertEquals(-1, commonPrefixHelper(new String("aaa"), new ArrayByteSequence("aaa").toString()));
     assertEquals(-1, commonPrefixHelper("abababababab".substring(3, 6), "ccababababcc".substring(3, 6)));
-    
+
     // no common prefix
     assertEquals(0, commonPrefixHelper("", "a"));
     assertEquals(0, commonPrefixHelper("a", ""));
     assertEquals(0, commonPrefixHelper("a", "b"));
     assertEquals(0, commonPrefixHelper("aaaa", "bbbb"));
-    
+
     // some common prefix
     assertEquals(1, commonPrefixHelper("a", "ab"));
     assertEquals(1, commonPrefixHelper("ab", "ac"));
@@ -87,44 +87,44 @@ public class RelativeKeyTest {
     assertEquals(2, commonPrefixHelper("aa", "aaaa"));
     assertEquals(4, commonPrefixHelper("aaaaa", "aaaab"));
   }
-  
+
   private int commonPrefixHelper(String a, String b) {
     return RelativeKey.getCommonPrefix(new ArrayByteSequence(a), new ArrayByteSequence(b));
   }
-  
+
   @Test
   public void testReadWritePrefix() throws IOException {
     Key prevKey = new Key("row1", "columnfamily1", "columnqualifier1", "columnvisibility1", 1000);
     Key newKey = new Key("row2", "columnfamily2", "columnqualifier2", "columnvisibility2", 3000);
     RelativeKey expected = new RelativeKey(prevKey, newKey);
-    
+
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream out = new DataOutputStream(baos);
     expected.write(out);
-    
+
     RelativeKey actual = new RelativeKey();
     actual.setPrevKey(prevKey);
     actual.readFields(new DataInputStream(new ByteArrayInputStream(baos.toByteArray())));
-    
+
     assertEquals(expected.getKey(), actual.getKey());
   }
-  
+
   private static ArrayList<Key> expectedKeys;
   private static ArrayList<Value> expectedValues;
   private static ArrayList<Integer> expectedPositions;
   private static ByteArrayOutputStream baos;
-  
+
   @BeforeClass
   public static void initSource() throws IOException {
     int initialListSize = 10000;
-    
+
     baos = new ByteArrayOutputStream();
     DataOutputStream out = new DataOutputStream(baos);
-    
+
     expectedKeys = new ArrayList<Key>(initialListSize);
     expectedValues = new ArrayList<Value>(initialListSize);
     expectedPositions = new ArrayList<Integer>(initialListSize);
-    
+
     Key prev = null;
     int val = 0;
     for (int row = 0; row < 4; row++) {
@@ -145,7 +145,7 @@ public class RelativeKeyTest {
               v.write(out);
               expectedKeys.add(k);
               expectedValues.add(v);
-              
+
               k = RFileTest.nk(rowS, cfS, cqS, cvS, ts);
               v = RFileTest.nv("" + val);
               expectedPositions.add(out.size());
@@ -154,7 +154,7 @@ public class RelativeKeyTest {
               v.write(out);
               expectedKeys.add(k);
               expectedValues.add(v);
-              
+
               val++;
             }
           }
@@ -162,34 +162,34 @@ public class RelativeKeyTest {
       }
     }
   }
-  
+
   private DataInputStream in;
-  
+
   @Before
   public void setupDataInputStream() {
     in = new DataInputStream(new ByteArrayInputStream(baos.toByteArray()));
     in.mark(0);
   }
-  
+
   @Test
   public void testSeekBeforeEverything() throws IOException {
     Key seekKey = new Key();
     Key prevKey = new Key();
     Key currKey = null;
     MutableByteSequence value = new MutableByteSequence(new byte[64], 0, 0);
-    
+
     RelativeKey.SkippR skippr = RelativeKey.fastSkip(in, seekKey, value, prevKey, currKey);
     assertEquals(1, skippr.skipped);
     assertEquals(new Key(), skippr.prevKey);
     assertEquals(expectedKeys.get(0), skippr.rk.getKey());
     assertEquals(expectedValues.get(0).toString(), value.toString());
-    
+
     // ensure we can advance after fastskip
     skippr.rk.readFields(in);
     assertEquals(expectedKeys.get(1), skippr.rk.getKey());
-    
+
     in.reset();
-    
+
     seekKey = new Key("a", "b", "c", "d", 1);
     seekKey.setDeleted(true);
     skippr = RelativeKey.fastSkip(in, seekKey, value, prevKey, currKey);
@@ -197,21 +197,21 @@ public class RelativeKeyTest {
     assertEquals(new Key(), skippr.prevKey);
     assertEquals(expectedKeys.get(0), skippr.rk.getKey());
     assertEquals(expectedValues.get(0).toString(), value.toString());
-    
+
     skippr.rk.readFields(in);
     assertEquals(expectedKeys.get(1), skippr.rk.getKey());
   }
-  
+
   @Test(expected = EOFException.class)
   public void testSeekAfterEverything() throws IOException {
     Key seekKey = new Key("s", "t", "u", "v", 1);
     Key prevKey = new Key();
     Key currKey = null;
     MutableByteSequence value = new MutableByteSequence(new byte[64], 0, 0);
-    
+
     RelativeKey.fastSkip(in, seekKey, value, prevKey, currKey);
   }
-  
+
   @Test
   public void testSeekMiddle() throws IOException {
     int seekIndex = expectedKeys.size() / 2;
@@ -219,36 +219,36 @@ public class RelativeKeyTest {
     Key prevKey = new Key();
     Key currKey = null;
     MutableByteSequence value = new MutableByteSequence(new byte[64], 0, 0);
-    
+
     RelativeKey.SkippR skippr = RelativeKey.fastSkip(in, seekKey, value, prevKey, currKey);
-    
+
     assertEquals(seekIndex + 1, skippr.skipped);
     assertEquals(expectedKeys.get(seekIndex - 1), skippr.prevKey);
     assertEquals(expectedKeys.get(seekIndex), skippr.rk.getKey());
     assertEquals(expectedValues.get(seekIndex).toString(), value.toString());
-    
+
     skippr.rk.readFields(in);
     assertEquals(expectedValues.get(seekIndex + 1).toString(), value.toString());
-    
+
     // try fast skipping to a key that does not exist
     in.reset();
     Key fKey = expectedKeys.get(seekIndex).followingKey(PartialKey.ROW_COLFAM_COLQUAL);
     int i;
     for (i = seekIndex; expectedKeys.get(i).compareTo(fKey) < 0; i++) {}
-    
+
     skippr = RelativeKey.fastSkip(in, expectedKeys.get(i), value, prevKey, currKey);
     assertEquals(i + 1, skippr.skipped);
     assertEquals(expectedKeys.get(i - 1), skippr.prevKey);
     assertEquals(expectedKeys.get(i), skippr.rk.getKey());
     assertEquals(expectedValues.get(i).toString(), value.toString());
-    
+
     // try fast skipping to our current location
     skippr = RelativeKey.fastSkip(in, expectedKeys.get(i), value, expectedKeys.get(i - 1), expectedKeys.get(i));
     assertEquals(0, skippr.skipped);
     assertEquals(expectedKeys.get(i - 1), skippr.prevKey);
     assertEquals(expectedKeys.get(i), skippr.rk.getKey());
     assertEquals(expectedValues.get(i).toString(), value.toString());
-    
+
     // try fast skipping 1 column family ahead from our current location, testing fastskip from middle of block as opposed to stating at beginning of block
     fKey = expectedKeys.get(i).followingKey(PartialKey.ROW_COLFAM);
     int j;
@@ -258,6 +258,6 @@ public class RelativeKeyTest {
     assertEquals(expectedKeys.get(j - 1), skippr.prevKey);
     assertEquals(expectedKeys.get(j), skippr.rk.getKey());
     assertEquals(expectedValues.get(j).toString(), value.toString());
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/AggregatingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/AggregatingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/AggregatingIteratorTest.java
index 137a462..788366a 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/AggregatingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/AggregatingIteratorTest.java
@@ -40,413 +40,413 @@ import org.apache.hadoop.io.Text;
  */
 @Deprecated
 public class AggregatingIteratorTest extends TestCase {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-  
+
   public static class SummationAggregator implements Aggregator {
-    
+
     int sum;
-    
+
     public Value aggregate() {
       return new Value((sum + "").getBytes());
     }
-    
+
     public void collect(Value value) {
       int val = Integer.parseInt(value.toString());
-      
+
       sum += val;
     }
-    
+
     public void reset() {
       sum = 0;
-      
+
     }
-    
+
   }
-  
+
   static Key nk(int row, int colf, int colq, long ts, boolean deleted) {
     Key k = nk(row, colf, colq, ts);
     k.setDeleted(true);
     return k;
   }
-  
+
   static Key nk(int row, int colf, int colq, long ts) {
     return new Key(nr(row), new Text(String.format("cf%03d", colf)), new Text(String.format("cq%03d", colq)), ts);
   }
-  
+
   static Range nr(int row, int colf, int colq, long ts, boolean inclusive) {
     return new Range(nk(row, colf, colq, ts), inclusive, null, true);
   }
-  
+
   static Range nr(int row, int colf, int colq, long ts) {
     return nr(row, colf, colq, ts, true);
   }
-  
+
   static void nkv(TreeMap<Key,Value> tm, int row, int colf, int colq, long ts, boolean deleted, String val) {
     Key k = nk(row, colf, colq, ts);
     k.setDeleted(deleted);
     tm.put(k, new Value(val.getBytes()));
   }
-  
+
   static Text nr(int row) {
     return new Text(String.format("r%03d", row));
   }
-  
+
   public void test1() throws IOException {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that do not aggregate
     nkv(tm1, 1, 1, 1, 1, false, "2");
     nkv(tm1, 1, 1, 1, 2, false, "3");
     nkv(tm1, 1, 1, 1, 3, false, "4");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
-    
+
     Map<String,String> emptyMap = Collections.emptyMap();
     ai.init(new SortedMapIterator(tm1), emptyMap, null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("4", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 1), ai.getTopKey());
     assertEquals("2", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking
-    
+
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 1), ai.getTopKey());
     assertEquals("2", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek after everything
     ai.seek(nr(1, 1, 1, 0), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
   }
-  
+
   public void test2() throws IOException {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, "2");
     nkv(tm1, 1, 1, 1, 2, false, "3");
     nkv(tm1, 1, 1, 1, 3, false, "4");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
-    
+
     Map<String,String> opts = new HashMap<String,String>();
-    
+
     opts.put("cf001", SummationAggregator.class.getName());
-    
+
     ai.init(new SortedMapIterator(tm1), opts, null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking to the beginning of a key that aggregates
-    
+
     ai.seek(nr(1, 1, 1, 3), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking the middle of a key the aggregates
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking to the end of a key the aggregates
     ai.seek(nr(1, 1, 1, 1), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking before a key the aggregates
     ai.seek(nr(1, 1, 1, 4), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
   }
-  
+
   public void test3() throws IOException {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, "2");
     nkv(tm1, 1, 1, 1, 2, false, "3");
     nkv(tm1, 1, 1, 1, 3, false, "4");
-    
+
     // keys that do not aggregate
     nkv(tm1, 2, 2, 1, 1, false, "2");
     nkv(tm1, 2, 2, 1, 2, false, "3");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
-    
+
     Map<String,String> opts = new HashMap<String,String>();
-    
+
     opts.put("cf001", SummationAggregator.class.getName());
-    
+
     ai.init(new SortedMapIterator(tm1), opts, null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 1), ai.getTopKey());
     assertEquals("2", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek after key that aggregates
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
     // seek before key that aggregates
     ai.seek(nr(1, 1, 1, 4), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
   }
-  
+
   public void test4() throws IOException {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that do not aggregate
     nkv(tm1, 0, 0, 1, 1, false, "7");
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, "2");
     nkv(tm1, 1, 1, 1, 2, false, "3");
     nkv(tm1, 1, 1, 1, 3, false, "4");
-    
+
     // keys that do not aggregate
     nkv(tm1, 2, 2, 1, 1, false, "2");
     nkv(tm1, 2, 2, 1, 2, false, "3");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
-    
+
     Map<String,String> opts = new HashMap<String,String>();
-    
+
     opts.put("cf001", SummationAggregator.class.getName());
-    
+
     ai.init(new SortedMapIterator(tm1), opts, null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(0, 0, 1, 1), ai.getTopKey());
     assertEquals("7", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 1), ai.getTopKey());
     assertEquals("2", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek test
     ai.seek(nr(0, 0, 1, 0), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
     // seek after key that aggregates
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", ai.getTopValue().toString());
-    
+
   }
-  
+
   public void test5() throws IOException {
     // try aggregating across multiple data sets that contain
     // the exact same keys w/ different values
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     nkv(tm1, 1, 1, 1, 1, false, "2");
-    
+
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     nkv(tm2, 1, 1, 1, 1, false, "3");
-    
+
     TreeMap<Key,Value> tm3 = new TreeMap<Key,Value>();
     nkv(tm3, 1, 1, 1, 1, false, "4");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
     Map<String,String> opts = new HashMap<String,String>();
     opts.put("cf001", SummationAggregator.class.getName());
-    
+
     List<SortedKeyValueIterator<Key,Value>> sources = new ArrayList<SortedKeyValueIterator<Key,Value>>(3);
     sources.add(new SortedMapIterator(tm1));
     sources.add(new SortedMapIterator(tm2));
     sources.add(new SortedMapIterator(tm3));
-    
+
     MultiIterator mi = new MultiIterator(sources, true);
     ai.init(mi, opts, null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 1), ai.getTopKey());
     assertEquals("9", ai.getTopValue().toString());
   }
-  
+
   public void test6() throws IOException {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, "2");
     nkv(tm1, 1, 1, 1, 2, false, "3");
     nkv(tm1, 1, 1, 1, 3, false, "4");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
-    
+
     Map<String,String> opts = new HashMap<String,String>();
-    
+
     opts.put("cf001", SummationAggregator.class.getName());
-    
+
     ai.init(new SortedMapIterator(tm1), opts, new DefaultIteratorEnvironment());
-    
+
     // try seeking to the beginning of a key that aggregates
-    
+
     ai.seek(nr(1, 1, 1, 3, false), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
   }
-  
+
   public void test7() throws IOException {
     // test that delete is not aggregated
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     nkv(tm1, 1, 1, 1, 2, true, "");
     nkv(tm1, 1, 1, 1, 3, false, "4");
     nkv(tm1, 1, 1, 1, 4, false, "3");
-    
+
     AggregatingIterator ai = new AggregatingIterator();
-    
+
     Map<String,String> opts = new HashMap<String,String>();
-    
+
     opts.put("cf001", SummationAggregator.class.getName());
-    
+
     ai.init(new SortedMapIterator(tm1), opts, new DefaultIteratorEnvironment());
-    
+
     ai.seek(nr(1, 1, 1, 4, true), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 4), ai.getTopKey());
     assertEquals("7", ai.getTopValue().toString());
-    
+
     ai.next();
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2, true), ai.getTopKey());
     assertEquals("", ai.getTopValue().toString());
-    
+
     ai.next();
     assertFalse(ai.hasTop());
-    
+
     tm1 = new TreeMap<Key,Value>();
     nkv(tm1, 1, 1, 1, 2, true, "");
     ai = new AggregatingIterator();
     ai.init(new SortedMapIterator(tm1), opts, new DefaultIteratorEnvironment());
-    
+
     ai.seek(nr(1, 1, 1, 4, true), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2, true), ai.getTopKey());
     assertEquals("", ai.getTopValue().toString());
-    
+
     ai.next();
     assertFalse(ai.hasTop());
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/DefaultIteratorEnvironment.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/DefaultIteratorEnvironment.java b/core/src/test/java/org/apache/accumulo/core/iterators/DefaultIteratorEnvironment.java
index 94da7b5..c864fcc 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/DefaultIteratorEnvironment.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/DefaultIteratorEnvironment.java
@@ -28,42 +28,42 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 
 public class DefaultIteratorEnvironment implements IteratorEnvironment {
-  
+
   AccumuloConfiguration conf;
-  
+
   public DefaultIteratorEnvironment(AccumuloConfiguration conf) {
     this.conf = conf;
   }
-  
+
   public DefaultIteratorEnvironment() {
     this.conf = AccumuloConfiguration.getDefaultConfiguration();
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
     Configuration conf = CachedConfiguration.getInstance();
     FileSystem fs = FileSystem.get(conf);
     return new MapFileIterator(this.conf, fs, mapFileName, conf);
   }
-  
+
   @Override
   public AccumuloConfiguration getConfig() {
     return conf;
   }
-  
+
   @Override
   public IteratorScope getIteratorScope() {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public boolean isFullMajorCompaction() {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowIteratorTest.java
index fa46360..21c31e2 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowIteratorTest.java
@@ -16,6 +16,8 @@
  */
 package org.apache.accumulo.core.iterators;
 
+import static org.junit.Assert.assertEquals;
+
 import java.io.IOException;
 import java.util.Collections;
 import java.util.TreeMap;
@@ -29,82 +31,76 @@ import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 import org.apache.accumulo.core.iterators.system.CountingIterator;
 import org.junit.Test;
 
-import static org.junit.Assert.*;
-
 public class FirstEntryInRowIteratorTest {
 
   @SuppressWarnings("unchecked")
-    private static long process(TreeMap<Key,Value> sourceMap, TreeMap<Key,Value> resultMap, Range range, int numScans) throws IOException
-    {
-      org.apache.accumulo.core.iterators.SortedMapIterator source = new SortedMapIterator(sourceMap);
-      CountingIterator counter = new CountingIterator(source);
-      FirstEntryInRowIterator feiri = new FirstEntryInRowIterator();
-      IteratorEnvironment env = new IteratorEnvironment(){
-
-        public AccumuloConfiguration getConfig() {
-          return null;
-        }
-
-        public IteratorScope getIteratorScope() {
-          return null;
-        }
-
-        public boolean isFullMajorCompaction() {
-          return false;
-        }
-
-        public void registerSideChannel(SortedKeyValueIterator<Key, Value> arg0) {
-
-        }
-
-        public SortedKeyValueIterator<Key, Value> reserveMapFileReader(
-            String arg0) throws IOException {
-          return null;
-        }};
-
-        feiri.init(counter, Collections.singletonMap(FirstEntryInRowIterator.NUM_SCANS_STRING_NAME, Integer.toString(numScans)), env);
-
-        feiri.seek(range, Collections.EMPTY_SET, false);
-        while(feiri.hasTop())
-        {
-          resultMap.put(feiri.getTopKey(), feiri.getTopValue());
-          feiri.next();
-        }
-        return counter.getCount();
+  private static long process(TreeMap<Key,Value> sourceMap, TreeMap<Key,Value> resultMap, Range range, int numScans) throws IOException {
+    org.apache.accumulo.core.iterators.SortedMapIterator source = new SortedMapIterator(sourceMap);
+    CountingIterator counter = new CountingIterator(source);
+    FirstEntryInRowIterator feiri = new FirstEntryInRowIterator();
+    IteratorEnvironment env = new IteratorEnvironment() {
+
+      public AccumuloConfiguration getConfig() {
+        return null;
+      }
+
+      public IteratorScope getIteratorScope() {
+        return null;
+      }
+
+      public boolean isFullMajorCompaction() {
+        return false;
+      }
+
+      public void registerSideChannel(SortedKeyValueIterator<Key,Value> arg0) {
+
+      }
+
+      public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String arg0) throws IOException {
+        return null;
+      }
+    };
+
+    feiri.init(counter, Collections.singletonMap(FirstEntryInRowIterator.NUM_SCANS_STRING_NAME, Integer.toString(numScans)), env);
+
+    feiri.seek(range, Collections.EMPTY_SET, false);
+    while (feiri.hasTop()) {
+      resultMap.put(feiri.getTopKey(), feiri.getTopValue());
+      feiri.next();
     }
+    return counter.getCount();
+  }
 
   @Test
-    public void test() throws IOException {
-      TreeMap<Key,Value> sourceMap = new TreeMap<Key, Value>();
-      Value emptyValue = new Value("".getBytes());
-      sourceMap.put(new Key("r1","cf","cq"), emptyValue);
-      sourceMap.put(new Key("r2","cf","cq"), emptyValue);
-      sourceMap.put(new Key("r3","cf","cq"), emptyValue);
-      TreeMap<Key,Value> resultMap = new TreeMap<Key,Value>();
-      long numSourceEntries = sourceMap.size();
-      long numNexts = process(sourceMap,resultMap,new Range(),10);
-      assertEquals(numNexts, numSourceEntries);
-      assertEquals(sourceMap.size(),resultMap.size());
-
-      for(int i = 0; i < 20; i++)
-      {
-        sourceMap.put(new Key("r2","cf","cq"+i),emptyValue);
-      }
-      resultMap.clear();
-      numNexts = process(sourceMap,resultMap,new Range(new Key("r1"), (new Key("r2")).followingKey(PartialKey.ROW)),10);
-      assertEquals(numNexts,resultMap.size()+10);
-      assertEquals(resultMap.size(),2);
-
-      resultMap.clear();
-      numNexts = process(sourceMap,resultMap,new Range(new Key("r1"), new Key("r2","cf2")),10);
-      assertEquals(numNexts,resultMap.size()+10);
-      assertEquals(resultMap.size(),2);
-
-      resultMap.clear();
-      numNexts = process(sourceMap,resultMap,new Range(new Key("r1"), new Key("r4")),10);
-      assertEquals(numNexts,resultMap.size()+10);
-      assertEquals(resultMap.size(),3);
+  public void test() throws IOException {
+    TreeMap<Key,Value> sourceMap = new TreeMap<Key,Value>();
+    Value emptyValue = new Value("".getBytes());
+    sourceMap.put(new Key("r1", "cf", "cq"), emptyValue);
+    sourceMap.put(new Key("r2", "cf", "cq"), emptyValue);
+    sourceMap.put(new Key("r3", "cf", "cq"), emptyValue);
+    TreeMap<Key,Value> resultMap = new TreeMap<Key,Value>();
+    long numSourceEntries = sourceMap.size();
+    long numNexts = process(sourceMap, resultMap, new Range(), 10);
+    assertEquals(numNexts, numSourceEntries);
+    assertEquals(sourceMap.size(), resultMap.size());
+
+    for (int i = 0; i < 20; i++) {
+      sourceMap.put(new Key("r2", "cf", "cq" + i), emptyValue);
     }
+    resultMap.clear();
+    numNexts = process(sourceMap, resultMap, new Range(new Key("r1"), (new Key("r2")).followingKey(PartialKey.ROW)), 10);
+    assertEquals(numNexts, resultMap.size() + 10);
+    assertEquals(resultMap.size(), 2);
+
+    resultMap.clear();
+    numNexts = process(sourceMap, resultMap, new Range(new Key("r1"), new Key("r2", "cf2")), 10);
+    assertEquals(numNexts, resultMap.size() + 10);
+    assertEquals(resultMap.size(), 2);
+
+    resultMap.clear();
+    numNexts = process(sourceMap, resultMap, new Range(new Key("r1"), new Key("r4")), 10);
+    assertEquals(numNexts, resultMap.size() + 10);
+    assertEquals(resultMap.size(), 3);
+  }
 
 }
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowTest.java
index 3afd629..8214c2c 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/FirstEntryInRowTest.java
@@ -36,41 +36,41 @@ import org.junit.Test;
 public class FirstEntryInRowTest {
   private static final Map<String,String> EMPTY_MAP = new HashMap<String,String>();
   private static final Collection<ByteSequence> EMPTY_SET = new HashSet<ByteSequence>();
-  
+
   private Key nk(String row, String cf, String cq, long time) {
     return new Key(new Text(row), new Text(cf), new Text(cq), time);
   }
-  
+
   private Key nk(int row, int cf, int cq, long time) {
     return nk(String.format("%06d", row), String.format("%06d", cf), String.format("%06d", cq), time);
   }
-  
+
   private void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, Value val) {
     tm.put(nk(row, cf, cq, time), val);
   }
-  
+
   private void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, String val) {
     put(tm, row, cf, cq, time, new Value(val.getBytes()));
   }
-  
+
   private void put(TreeMap<Key,Value> tm, int row, int cf, int cq, long time, int val) {
     tm.put(nk(row, cf, cq, time), new Value((val + "").getBytes()));
   }
-  
+
   private void aten(FirstEntryInRowIterator rdi, String row, String cf, String cq, long time, String val) throws Exception {
     assertTrue(rdi.hasTop());
     assertEquals(nk(row, cf, cq, time), rdi.getTopKey());
     assertEquals(val, rdi.getTopValue().toString());
     rdi.next();
   }
-  
+
   private void aten(FirstEntryInRowIterator rdi, int row, int cf, int cq, long time, int val) throws Exception {
     assertTrue(rdi.hasTop());
     assertEquals(nk(row, cf, cq, time), rdi.getTopKey());
     assertEquals(val, Integer.parseInt(rdi.getTopValue().toString()));
     rdi.next();
   }
-  
+
   @Test
   public void test1() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
@@ -80,22 +80,22 @@ public class FirstEntryInRowTest {
     put(tm1, "r2", "cf2", "cq4", 5, "v4");
     put(tm1, "r2", "cf2", "cq5", 5, "v5");
     put(tm1, "r3", "cf3", "cq6", 5, "v6");
-    
+
     FirstEntryInRowIterator fei = new FirstEntryInRowIterator();
     fei.init(new SortedMapIterator(tm1), EMPTY_MAP, null);
-    
+
     fei.seek(new Range(), EMPTY_SET, false);
     aten(fei, "r1", "cf1", "cq1", 5, "v1");
     aten(fei, "r2", "cf1", "cq1", 5, "v3");
     aten(fei, "r3", "cf3", "cq6", 5, "v6");
     assertFalse(fei.hasTop());
-    
+
   }
-  
+
   @Test
   public void test2() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     for (int r = 0; r < 5; r++) {
       for (int cf = r; cf < 100; cf++) {
         for (int cq = 3; cq < 6; cq++) {
@@ -103,7 +103,7 @@ public class FirstEntryInRowTest {
         }
       }
     }
-    
+
     FirstEntryInRowIterator fei = new FirstEntryInRowIterator();
     fei.init(new SortedMapIterator(tm1), EMPTY_MAP, null);
     fei.seek(new Range(nk(0, 10, 0, 0), null), EMPTY_SET, false);
@@ -112,16 +112,15 @@ public class FirstEntryInRowTest {
     aten(fei, 3, 3, 3, 6, 3 * 3 * 3);
     aten(fei, 4, 4, 3, 6, 4 * 4 * 3);
     assertFalse(fei.hasTop());
-    
+
     fei.seek(new Range(nk(1, 1, 3, 6), nk(3, 3, 3, 6)), EMPTY_SET, false);
     aten(fei, 1, 1, 3, 6, 1 * 1 * 3);
     aten(fei, 2, 2, 3, 6, 2 * 2 * 3);
     aten(fei, 3, 3, 3, 6, 3 * 3 * 3);
     assertFalse(fei.hasTop());
-    
+
     fei.seek(new Range(nk(1, 1, 3, 6), false, nk(3, 3, 3, 6), false), EMPTY_SET, false);
     aten(fei, 2, 2, 3, 6, 2 * 2 * 3);
     assertFalse(fei.hasTop());
   }
 }
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/IteratorUtilTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/IteratorUtilTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/IteratorUtilTest.java
index 4d843d3..125268a 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/IteratorUtilTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/IteratorUtilTest.java
@@ -45,248 +45,248 @@ import org.junit.Assert;
 import org.junit.Test;
 
 public class IteratorUtilTest {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-  
+
   static class WrappedIter implements SortedKeyValueIterator<Key,Value> {
-    
+
     protected SortedKeyValueIterator<Key,Value> source;
-    
+
     public WrappedIter deepCopy(IteratorEnvironment env) {
       throw new UnsupportedOperationException();
     }
-    
+
     public Key getTopKey() {
       return source.getTopKey();
     }
-    
+
     public Value getTopValue() {
       return source.getTopValue();
     }
-    
+
     public boolean hasTop() {
       return source.hasTop();
     }
-    
+
     public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
       this.source = source;
     }
-    
+
     public void next() throws IOException {
       source.next();
     }
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
       source.seek(range, columnFamilies, inclusive);
     }
   }
-  
+
   static class AddingIter extends WrappedIter {
-    
+
     int amount = 1;
-    
+
     public Value getTopValue() {
       Value val = super.getTopValue();
-      
+
       int orig = Integer.parseInt(val.toString());
-      
+
       return new Value(((orig + amount) + "").getBytes());
     }
-    
+
     public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
       super.init(source, options, env);
-      
+
       String amount = options.get("amount");
-      
+
       if (amount != null) {
         this.amount = Integer.parseInt(amount);
       }
     }
   }
-  
+
   static class SquaringIter extends WrappedIter {
     public Value getTopValue() {
       Value val = super.getTopValue();
-      
+
       int orig = Integer.parseInt(val.toString());
-      
+
       return new Value(((orig * orig) + "").getBytes());
     }
   }
-  
+
   @Test
   public void test1() throws IOException {
     ConfigurationCopy conf = new ConfigurationCopy();
-    
+
     // create an iterator that adds 1 and then squares
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".addIter", "1," + AddingIter.class.getName());
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".sqIter", "2," + SquaringIter.class.getName());
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     MultiIteratorTest.nkv(tm, 1, 0, false, "1");
     MultiIteratorTest.nkv(tm, 2, 0, false, "2");
-    
+
     SortedMapIterator source = new SortedMapIterator(tm);
-    
+
     SortedKeyValueIterator<Key,Value> iter = IteratorUtil.loadIterators(IteratorScope.minc, source, new KeyExtent(new Text("tab"), null, null), conf,
         new DefaultIteratorEnvironment(conf));
     iter.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(1, 0)));
     assertTrue(iter.getTopValue().toString().equals("4"));
-    
+
     iter.next();
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(2, 0)));
     assertTrue(iter.getTopValue().toString().equals("9"));
-    
+
     iter.next();
-    
+
     assertFalse(iter.hasTop());
   }
-  
+
   @Test
   public void test4() throws IOException {
-    
+
     // try loading for a different scope
     AccumuloConfiguration conf = new ConfigurationCopy();
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     MultiIteratorTest.nkv(tm, 1, 0, false, "1");
     MultiIteratorTest.nkv(tm, 2, 0, false, "2");
-    
+
     SortedMapIterator source = new SortedMapIterator(tm);
-    
+
     SortedKeyValueIterator<Key,Value> iter = IteratorUtil.loadIterators(IteratorScope.majc, source, new KeyExtent(new Text("tab"), null, null), conf,
         new DefaultIteratorEnvironment(conf));
     iter.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(1, 0)));
     assertTrue(iter.getTopValue().toString().equals("1"));
-    
+
     iter.next();
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(2, 0)));
     assertTrue(iter.getTopValue().toString().equals("2"));
-    
+
     iter.next();
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
-  
+
   @Test
   public void test3() throws IOException {
     // change the load order, so it squares and then adds
-    
+
     ConfigurationCopy conf = new ConfigurationCopy();
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     MultiIteratorTest.nkv(tm, 1, 0, false, "1");
     MultiIteratorTest.nkv(tm, 2, 0, false, "2");
-    
+
     SortedMapIterator source = new SortedMapIterator(tm);
-    
+
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".addIter", "2," + AddingIter.class.getName());
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".sqIter", "1," + SquaringIter.class.getName());
-    
+
     SortedKeyValueIterator<Key,Value> iter = IteratorUtil.loadIterators(IteratorScope.minc, source, new KeyExtent(new Text("tab"), null, null), conf,
         new DefaultIteratorEnvironment(conf));
     iter.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(1, 0)));
     assertTrue(iter.getTopValue().toString().equals("2"));
-    
+
     iter.next();
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(2, 0)));
     assertTrue(iter.getTopValue().toString().equals("5"));
-    
+
     iter.next();
-    
+
     assertFalse(iter.hasTop());
   }
-  
+
   @Test
   public void test2() throws IOException {
-    
+
     ConfigurationCopy conf = new ConfigurationCopy();
-    
+
     // create an iterator that adds 1 and then squares
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".addIter", "1," + AddingIter.class.getName());
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".addIter.opt.amount", "7");
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".sqIter", "2," + SquaringIter.class.getName());
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     MultiIteratorTest.nkv(tm, 1, 0, false, "1");
     MultiIteratorTest.nkv(tm, 2, 0, false, "2");
-    
+
     SortedMapIterator source = new SortedMapIterator(tm);
-    
+
     SortedKeyValueIterator<Key,Value> iter = IteratorUtil.loadIterators(IteratorScope.minc, source, new KeyExtent(new Text("tab"), null, null), conf,
         new DefaultIteratorEnvironment(conf));
     iter.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(1, 0)));
     assertTrue(iter.getTopValue().toString().equals("64"));
-    
+
     iter.next();
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(2, 0)));
     assertTrue(iter.getTopValue().toString().equals("81"));
-    
+
     iter.next();
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
-  
+
   @Test
   public void test5() throws IOException {
     ConfigurationCopy conf = new ConfigurationCopy();
-    
+
     // create an iterator that ages off
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".filter", "1," + AgeOffFilter.class.getName());
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".filter.opt.ttl", "100");
     conf.set(Property.TABLE_ITERATOR_PREFIX + IteratorScope.minc.name() + ".filter.opt.currentTime", "1000");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     MultiIteratorTest.nkv(tm, 1, 850, false, "1");
     MultiIteratorTest.nkv(tm, 2, 950, false, "2");
-    
+
     SortedMapIterator source = new SortedMapIterator(tm);
-    
+
     SortedKeyValueIterator<Key,Value> iter = IteratorUtil.loadIterators(IteratorScope.minc, source, new KeyExtent(new Text("tab"), null, null), conf,
         new DefaultIteratorEnvironment(conf));
     iter.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(iter.hasTop());
     assertTrue(iter.getTopKey().equals(MultiIteratorTest.nk(2, 950)));
     iter.next();
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
 
   @Test
   public void onlyReadsRelevantIteratorScopeConfigurations() throws Exception {
     Map<String,String> data = new HashMap<String,String>();
 
-    // Make some configuration items, one with a bogus scope 
+    // Make some configuration items, one with a bogus scope
     data.put(Property.TABLE_ITERATOR_SCAN_PREFIX + "foo", "50," + SummingCombiner.class.getName());
     data.put(Property.TABLE_ITERATOR_SCAN_PREFIX + "foo.opt." + SummingCombiner.ALL_OPTION, "true");
     data.put(Property.TABLE_ITERATOR_PREFIX + ".fakescope.bar", "50," + SummingCombiner.class.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/NumSummationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/NumSummationTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/NumSummationTest.java
index d08c31b..d8d3b47 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/NumSummationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/NumSummationTest.java
@@ -26,22 +26,22 @@ import org.apache.log4j.Logger;
  */
 @Deprecated
 public class NumSummationTest extends TestCase {
-  
+
   private static final Logger log = Logger.getLogger(NumSummationTest.class);
-  
+
   public byte[] init(int n) {
     byte[] b = new byte[n];
     for (int i = 0; i < b.length; i++)
       b[i] = 0;
     return b;
   }
-  
+
   public void test1() {
     try {
       long[] la = {1l, 2l, 3l};
       byte[] b = NumArraySummation.longArrayToBytes(la);
       long[] la2 = NumArraySummation.bytesToLongArray(b);
-      
+
       assertTrue(la.length == la2.length);
       for (int i = 0; i < la.length; i++) {
         assertTrue(i + ": " + la[i] + " does not equal " + la2[i], la[i] == la2[i]);
@@ -50,7 +50,7 @@ public class NumSummationTest extends TestCase {
       assertTrue(false);
     }
   }
-  
+
   public void test2() {
     try {
       NumArraySummation nas = new NumArraySummation();
@@ -72,7 +72,7 @@ public class NumSummationTest extends TestCase {
       assertTrue(false);
     }
   }
-  
+
   public void test3() {
     try {
       NumArraySummation nas = new NumArraySummation();
@@ -91,19 +91,19 @@ public class NumSummationTest extends TestCase {
       assertTrue(false);
     }
   }
-  
+
   public void test4() {
     try {
       long l = 5l;
       byte[] b = NumSummation.longToBytes(l);
       long l2 = NumSummation.bytesToLong(b);
-      
+
       assertTrue(l == l2);
     } catch (Exception e) {
       assertTrue(false);
     }
   }
-  
+
   public void test5() {
     try {
       NumSummation ns = new NumSummation();
@@ -112,23 +112,23 @@ public class NumSummationTest extends TestCase {
       }
       long l = NumSummation.bytesToLong(ns.aggregate().get());
       assertTrue("l was " + l, l == 13);
-      
+
       ns.collect(new Value(NumSummation.longToBytes(Long.MAX_VALUE)));
       l = NumSummation.bytesToLong(ns.aggregate().get());
       assertTrue("l was " + l, l == Long.MAX_VALUE);
-      
+
       ns.collect(new Value(NumSummation.longToBytes(Long.MIN_VALUE)));
       l = NumSummation.bytesToLong(ns.aggregate().get());
       assertTrue("l was " + l, l == -1);
-      
+
       ns.collect(new Value(NumSummation.longToBytes(Long.MIN_VALUE)));
       l = NumSummation.bytesToLong(ns.aggregate().get());
       assertTrue("l was " + l, l == Long.MIN_VALUE);
-      
+
       ns.collect(new Value(NumSummation.longToBytes(Long.MIN_VALUE)));
       l = NumSummation.bytesToLong(ns.aggregate().get());
       assertTrue("l was " + l, l == Long.MIN_VALUE);
-      
+
       ns.reset();
       l = NumSummation.bytesToLong(ns.aggregate().get());
       assertTrue("l was " + l, l == 0);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfigurationTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfigurationTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfigurationTest.java
index b17ff28..1a285bc 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfigurationTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfigurationTest.java
@@ -29,38 +29,38 @@ public class AggregatorConfigurationTest extends TestCase {
   public void testBinary() {
     Text colf = new Text();
     Text colq = new Text();
-    
+
     for (int i = 0; i < 256; i++) {
       colf.append(new byte[] {(byte) i}, 0, 1);
       colq.append(new byte[] {(byte) (255 - i)}, 0, 1);
     }
-    
+
     runTest(colf, colq);
     runTest(colf);
   }
-  
+
   public void testBasic() {
     runTest(new Text("colf1"), new Text("cq2"));
     runTest(new Text("colf1"));
   }
-  
+
   private void runTest(Text colf) {
     String encodedCols;
     PerColumnIteratorConfig ac3 = new PerColumnIteratorConfig(colf, "com.foo.SuperAgg");
     encodedCols = ac3.encodeColumns();
     PerColumnIteratorConfig ac4 = PerColumnIteratorConfig.decodeColumns(encodedCols, "com.foo.SuperAgg");
-    
+
     assertEquals(colf, ac4.getColumnFamily());
     assertNull(ac4.getColumnQualifier());
   }
-  
+
   private void runTest(Text colf, Text colq) {
     PerColumnIteratorConfig ac = new PerColumnIteratorConfig(colf, colq, "com.foo.SuperAgg");
     String encodedCols = ac.encodeColumns();
     PerColumnIteratorConfig ac2 = PerColumnIteratorConfig.decodeColumns(encodedCols, "com.foo.SuperAgg");
-    
+
     assertEquals(colf, ac2.getColumnFamily());
     assertEquals(colq, ac2.getColumnQualifier());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIteratorTest.java
index aae5e8a..fbe7fd5 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIteratorTest.java
@@ -31,36 +31,36 @@ import org.apache.accumulo.core.iterators.SortedMapIterator;
 import org.apache.hadoop.io.Text;
 
 public class ColumnFamilySkippingIteratorTest extends TestCase {
-  
+
   private static final Collection<ByteSequence> EMPTY_SET = new HashSet<ByteSequence>();
-  
+
   Key nk(String row, String cf, String cq, long time) {
     return new Key(new Text(row), new Text(cf), new Text(cq), time);
   }
-  
+
   Key nk(int row, int cf, int cq, long time) {
     return nk(String.format("%06d", row), String.format("%06d", cf), String.format("%06d", cq), time);
   }
-  
+
   void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, Value val) {
     tm.put(nk(row, cf, cq, time), val);
   }
-  
+
   void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, String val) {
     put(tm, row, cf, cq, time, new Value(val.getBytes()));
   }
-  
+
   void put(TreeMap<Key,Value> tm, int row, int cf, int cq, long time, int val) {
     tm.put(nk(row, cf, cq, time), new Value((val + "").getBytes()));
   }
-  
+
   private void aten(ColumnFamilySkippingIterator rdi, String row, String cf, String cq, long time, String val) throws Exception {
     assertTrue(rdi.hasTop());
     assertEquals(nk(row, cf, cq, time), rdi.getTopKey());
     assertEquals(val, rdi.getTopValue().toString());
     rdi.next();
   }
-  
+
   public void test1() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
@@ -69,12 +69,12 @@ public class ColumnFamilySkippingIteratorTest extends TestCase {
     put(tm1, "r2", "cf2", "cq4", 5, "v4");
     put(tm1, "r2", "cf2", "cq5", 5, "v5");
     put(tm1, "r3", "cf3", "cq6", 5, "v6");
-    
+
     ColumnFamilySkippingIterator cfi = new ColumnFamilySkippingIterator(new SortedMapIterator(tm1));
-    
+
     cfi.seek(new Range(), EMPTY_SET, true);
     assertFalse(cfi.hasTop());
-    
+
     cfi.seek(new Range(), EMPTY_SET, false);
     assertTrue(cfi.hasTop());
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
@@ -83,14 +83,14 @@ public class ColumnFamilySkippingIteratorTest extends TestCase {
       cfi.next();
     }
     assertEquals(tm1, tm2);
-    
+
     HashSet<ByteSequence> colfams = new HashSet<ByteSequence>();
     colfams.add(new ArrayByteSequence("cf2"));
     cfi.seek(new Range(), colfams, true);
     aten(cfi, "r2", "cf2", "cq4", 5, "v4");
     aten(cfi, "r2", "cf2", "cq5", 5, "v5");
     assertFalse(cfi.hasTop());
-    
+
     colfams.add(new ArrayByteSequence("cf3"));
     colfams.add(new ArrayByteSequence("cf4"));
     cfi.seek(new Range(), colfams, true);
@@ -98,18 +98,18 @@ public class ColumnFamilySkippingIteratorTest extends TestCase {
     aten(cfi, "r2", "cf2", "cq5", 5, "v5");
     aten(cfi, "r3", "cf3", "cq6", 5, "v6");
     assertFalse(cfi.hasTop());
-    
+
     cfi.seek(new Range(), colfams, false);
     aten(cfi, "r1", "cf1", "cq1", 5, "v1");
     aten(cfi, "r1", "cf1", "cq3", 5, "v2");
     aten(cfi, "r2", "cf1", "cq1", 5, "v3");
     assertFalse(cfi.hasTop());
-    
+
   }
-  
+
   public void test2() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     for (int r = 0; r < 10; r++) {
       for (int cf = 0; cf < 1000; cf++) {
         for (int cq = 0; cq < 3; cq++) {
@@ -117,80 +117,80 @@ public class ColumnFamilySkippingIteratorTest extends TestCase {
         }
       }
     }
-    
+
     HashSet<ByteSequence> allColfams = new HashSet<ByteSequence>();
     for (int cf = 0; cf < 1000; cf++) {
       allColfams.add(new ArrayByteSequence(String.format("%06d", cf)));
     }
-    
+
     ColumnFamilySkippingIterator cfi = new ColumnFamilySkippingIterator(new SortedMapIterator(tm1));
     HashSet<ByteSequence> colfams = new HashSet<ByteSequence>();
-    
+
     runTest(cfi, 30000, 0, allColfams, colfams);
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 60)));
     runTest(cfi, 30000, 30, allColfams, colfams);
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 602)));
     runTest(cfi, 30000, 60, allColfams, colfams);
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 0)));
     runTest(cfi, 30000, 90, allColfams, colfams);
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 999)));
     runTest(cfi, 30000, 120, allColfams, colfams);
-    
+
     colfams.remove(new ArrayByteSequence(String.format("%06d", 0)));
     runTest(cfi, 30000, 90, allColfams, colfams);
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 1000)));
     runTest(cfi, 30000, 90, allColfams, colfams);
-    
+
     colfams.remove(new ArrayByteSequence(String.format("%06d", 999)));
     runTest(cfi, 30000, 60, allColfams, colfams);
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 61)));
     runTest(cfi, 30000, 90, allColfams, colfams);
-    
+
     for (int i = 62; i < 100; i++)
       colfams.add(new ArrayByteSequence(String.format("%06d", i)));
-    
+
     runTest(cfi, 30000, 1230, allColfams, colfams);
-    
+
   }
-  
+
   private void runTest(ColumnFamilySkippingIterator cfi, int total, int expected, HashSet<ByteSequence> allColfams, HashSet<ByteSequence> colfams)
       throws Exception {
     cfi.seek(new Range(), colfams, true);
     HashSet<ByteSequence> excpected1 = new HashSet<ByteSequence>(colfams);
     excpected1.retainAll(allColfams);
     runTest(cfi, expected, excpected1);
-    
+
     HashSet<ByteSequence> excpected2 = new HashSet<ByteSequence>(allColfams);
     excpected2.removeAll(colfams);
     cfi.seek(new Range(), colfams, false);
     runTest(cfi, total - expected, excpected2);
   }
-  
+
   private void runTest(ColumnFamilySkippingIterator cfi, int expected, HashSet<ByteSequence> colfams) throws Exception {
     int count = 0;
-    
+
     HashSet<ByteSequence> ocf = new HashSet<ByteSequence>();
-    
+
     while (cfi.hasTop()) {
       count++;
       ocf.add(cfi.getTopKey().getColumnFamilyData());
       cfi.next();
     }
-    
+
     assertEquals(expected, count);
     assertEquals(colfams, ocf);
   }
-  
+
   public void test3() throws Exception {
     // construct test where ColumnFamilySkippingIterator might try to seek past the end of the user supplied range
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     for (int r = 0; r < 3; r++) {
       for (int cf = 4; cf < 1000; cf++) {
         for (int cq = 0; cq < 1; cq++) {
@@ -198,31 +198,31 @@ public class ColumnFamilySkippingIteratorTest extends TestCase {
         }
       }
     }
-    
+
     CountingIterator ci = new CountingIterator(new SortedMapIterator(tm1));
     ColumnFamilySkippingIterator cfi = new ColumnFamilySkippingIterator(ci);
     HashSet<ByteSequence> colfams = new HashSet<ByteSequence>();
     colfams.add(new ArrayByteSequence(String.format("%06d", 4)));
-    
+
     Range range = new Range(nk(0, 4, 0, 6), true, nk(0, 400, 0, 6), true);
     cfi.seek(range, colfams, true);
-    
+
     assertTrue(cfi.hasTop());
     assertEquals(nk(0, 4, 0, 6), cfi.getTopKey());
     cfi.next();
     assertFalse(cfi.hasTop());
-    
+
     colfams.add(new ArrayByteSequence(String.format("%06d", 500)));
     cfi.seek(range, colfams, true);
-    
+
     assertTrue(cfi.hasTop());
     assertEquals(nk(0, 4, 0, 6), cfi.getTopKey());
     cfi.next();
     assertFalse(cfi.hasTop());
-    
+
     range = new Range(nk(0, 4, 0, 6), true, nk(1, 400, 0, 6), true);
     cfi.seek(range, colfams, true);
-    
+
     assertTrue(cfi.hasTop());
     assertEquals(nk(0, 4, 0, 6), cfi.getTopKey());
     cfi.next();
@@ -233,7 +233,7 @@ public class ColumnFamilySkippingIteratorTest extends TestCase {
     assertEquals(nk(1, 4, 0, 6), cfi.getTopKey());
     cfi.next();
     assertFalse(cfi.hasTop());
-    
+
     // System.out.println(ci.getCount());
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFilterTest.java
index 58b3295..3fd66b4 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/ColumnFilterTest.java
@@ -26,51 +26,51 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.hadoop.io.Text;
 
 public class ColumnFilterTest extends TestCase {
-  
+
   Key nk(String row, String cf, String cq) {
     return new Key(new Text(row), new Text(cf), new Text(cq));
   }
-  
+
   Column nc(String cf) {
     return new Column(cf.getBytes(), null, null);
   }
-  
+
   Column nc(String cf, String cq) {
     return new Column(cf.getBytes(), cq.getBytes(), null);
   }
-  
+
   public void test1() {
     HashSet<Column> columns = new HashSet<Column>();
-    
+
     columns.add(nc("cf1"));
-    
+
     ColumnQualifierFilter cf = new ColumnQualifierFilter(null, columns);
-    
+
     assertTrue(cf.accept(nk("r1", "cf1", "cq1"), new Value(new byte[0])));
     assertTrue(cf.accept(nk("r1", "cf2", "cq1"), new Value(new byte[0])));
-    
+
   }
-  
+
   public void test2() {
     HashSet<Column> columns = new HashSet<Column>();
-    
+
     columns.add(nc("cf1"));
     columns.add(nc("cf2", "cq1"));
-    
+
     ColumnQualifierFilter cf = new ColumnQualifierFilter(null, columns);
-    
+
     assertTrue(cf.accept(nk("r1", "cf1", "cq1"), new Value(new byte[0])));
     assertTrue(cf.accept(nk("r1", "cf2", "cq1"), new Value(new byte[0])));
     assertFalse(cf.accept(nk("r1", "cf2", "cq2"), new Value(new byte[0])));
   }
-  
+
   public void test3() {
     HashSet<Column> columns = new HashSet<Column>();
-    
+
     columns.add(nc("cf2", "cq1"));
-    
+
     ColumnQualifierFilter cf = new ColumnQualifierFilter(null, columns);
-    
+
     assertFalse(cf.accept(nk("r1", "cf1", "cq1"), new Value(new byte[0])));
     assertTrue(cf.accept(nk("r1", "cf2", "cq1"), new Value(new byte[0])));
     assertFalse(cf.accept(nk("r1", "cf2", "cq2"), new Value(new byte[0])));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/DeletingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/DeletingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/DeletingIteratorTest.java
index f499680..4fd48d5 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/DeletingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/DeletingIteratorTest.java
@@ -32,19 +32,19 @@ import org.apache.accumulo.core.iterators.SortedMapIterator;
 import org.apache.hadoop.io.Text;
 
 public class DeletingIteratorTest extends TestCase {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-  
+
   public void test1() {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dvOld = new Value("old".getBytes());
     Value dvDel = new Value("old".getBytes());
     Value dvNew = new Value("new".getBytes());
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
     Key k;
-    
+
     for (int i = 0; i < 2; i++) {
       for (long j = 0; j < 5; j++) {
         k = new Key(new Text(String.format("%03d", i)), colf, colq, j);
@@ -61,12 +61,12 @@ public class DeletingIteratorTest extends TestCase {
       }
     }
     assertTrue("Initial size was " + tm.size(), tm.size() == 21);
-    
+
     Text checkRow = new Text("000");
     try {
       DeletingIterator it = new DeletingIterator(new SortedMapIterator(tm), false);
       it.seek(new Range(), EMPTY_COL_FAMS, false);
-      
+
       TreeMap<Key,Value> tmOut = new TreeMap<Key,Value>();
       while (it.hasTop()) {
         tmOut.put(it.getTopKey(), it.getTopValue());
@@ -84,7 +84,7 @@ public class DeletingIteratorTest extends TestCase {
     } catch (IOException e) {
       assertFalse(true);
     }
-    
+
     try {
       DeletingIterator it = new DeletingIterator(new SortedMapIterator(tm), true);
       it.seek(new Range(), EMPTY_COL_FAMS, false);
@@ -112,120 +112,120 @@ public class DeletingIteratorTest extends TestCase {
       assertFalse(true);
     }
   }
-  
+
   // seek test
   public void test2() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     nkv(tm, "r000", 4, false, "v4");
     nkv(tm, "r000", 3, false, "v3");
     nkv(tm, "r000", 2, true, "v2");
     nkv(tm, "r000", 1, false, "v1");
-    
+
     DeletingIterator it = new DeletingIterator(new SortedMapIterator(tm), false);
-    
+
     // SEEK two keys before delete
     it.seek(nr("r000", 4), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(it.hasTop());
     assertEquals(nk("r000", 4), it.getTopKey());
     assertEquals("v4", it.getTopValue().toString());
-    
+
     it.next();
-    
+
     assertTrue(it.hasTop());
     assertEquals(nk("r000", 3), it.getTopKey());
     assertEquals("v3", it.getTopValue().toString());
-    
+
     it.next();
-    
+
     assertFalse(it.hasTop());
-    
+
     // SEEK passed delete
     it.seek(nr("r000", 1), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(it.hasTop());
-    
+
     // SEEK to delete
     it.seek(nr("r000", 2), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(it.hasTop());
-    
+
     // SEEK right before delete
     it.seek(nr("r000", 3), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(it.hasTop());
     assertEquals(nk("r000", 3), it.getTopKey());
     assertEquals("v3", it.getTopValue().toString());
-    
+
     it.next();
-    
+
     assertFalse(it.hasTop());
   }
-  
+
   // test delete with same timestamp as existing key
   public void test3() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     nkv(tm, "r000", 3, false, "v3");
     nkv(tm, "r000", 2, false, "v2");
     nkv(tm, "r000", 2, true, "");
     nkv(tm, "r000", 1, false, "v1");
-    
+
     DeletingIterator it = new DeletingIterator(new SortedMapIterator(tm), false);
     it.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(it.hasTop());
     assertEquals(nk("r000", 3), it.getTopKey());
     assertEquals("v3", it.getTopValue().toString());
-    
+
     it.next();
-    
+
     assertFalse(it.hasTop());
-    
+
     it.seek(nr("r000", 2), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(it.hasTop());
   }
-  
+
   // test range inclusiveness
   public void test4() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     nkv(tm, "r000", 3, false, "v3");
     nkv(tm, "r000", 2, false, "v2");
     nkv(tm, "r000", 2, true, "");
     nkv(tm, "r000", 1, false, "v1");
-    
+
     DeletingIterator it = new DeletingIterator(new SortedMapIterator(tm), false);
-    
+
     it.seek(nr("r000", 3), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(it.hasTop());
     assertEquals(nk("r000", 3), it.getTopKey());
     assertEquals("v3", it.getTopValue().toString());
-    
+
     it.next();
-    
+
     assertFalse(it.hasTop());
-    
+
     it.seek(nr("r000", 3, false), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(it.hasTop());
   }
-  
+
   private Range nr(String row, long ts, boolean inclusive) {
     return new Range(nk(row, ts), inclusive, null, true);
   }
-  
+
   private Range nr(String row, long ts) {
     return nr(row, ts, true);
   }
-  
+
   private Key nk(String row, long ts) {
     return new Key(new Text(row), ts);
   }
-  
+
   private void nkv(TreeMap<Key,Value> tm, String row, long ts, boolean deleted, String val) {
     Key k = nk(row, ts);
     k.setDeleted(deleted);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/MultiIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/MultiIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/MultiIteratorTest.java
index 5fcef31..c8ef180 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/MultiIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/MultiIteratorTest.java
@@ -35,34 +35,34 @@ import org.apache.accumulo.core.util.LocalityGroupUtil;
 import org.apache.hadoop.io.Text;
 
 public class MultiIteratorTest extends TestCase {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-  
+
   public static Key nk(int row, long ts) {
     return new Key(nr(row), ts);
   }
-  
+
   public static Range nrng(int row, long ts) {
     return new Range(nk(row, ts), null);
   }
-  
+
   public static void nkv(TreeMap<Key,Value> tm, int row, long ts, boolean deleted, String val) {
     Key k = nk(row, ts);
     k.setDeleted(deleted);
     tm.put(k, new Value(val.getBytes()));
   }
-  
+
   public static Text nr(int row) {
     return new Text(String.format("r%03d", row));
   }
-  
+
   void verify(int start, int end, Key seekKey, Text endRow, Text prevEndRow, boolean init, boolean incrRow, List<TreeMap<Key,Value>> maps) throws IOException {
     List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(maps.size());
-    
+
     for (TreeMap<Key,Value> map : maps) {
       iters.add(new SortedMapIterator(map));
     }
-    
+
     MultiIterator mi;
     if (endRow == null && prevEndRow == null)
       mi = new MultiIterator(iters, init);
@@ -72,58 +72,58 @@ public class MultiIteratorTest extends TestCase {
         for (SortedKeyValueIterator<Key,Value> iter : iters)
           iter.seek(range, LocalityGroupUtil.EMPTY_CF_SET, false);
       mi = new MultiIterator(iters, range);
-      
+
       if (init)
         mi.seek(range, LocalityGroupUtil.EMPTY_CF_SET, false);
     }
-    
+
     if (seekKey != null)
       mi.seek(new Range(seekKey, null), EMPTY_COL_FAMS, false);
     else
       mi.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     int i = start;
     while (mi.hasTop()) {
       if (incrRow)
         assertEquals(nk(i, 0), mi.getTopKey());
       else
         assertEquals(nk(0, i), mi.getTopKey());
-      
+
       assertEquals("v" + i, mi.getTopValue().toString());
-      
+
       mi.next();
       if (incrRow)
         i++;
       else
         i--;
     }
-    
+
     assertEquals("start=" + start + " end=" + end + " seekKey=" + seekKey + " endRow=" + endRow + " prevEndRow=" + prevEndRow + " init=" + init + " incrRow="
         + incrRow + " maps=" + maps, end, i);
   }
-  
+
   void verify(int start, Key seekKey, List<TreeMap<Key,Value>> maps) throws IOException {
     if (seekKey != null) {
       verify(start, -1, seekKey, null, null, false, false, maps);
     }
-    
+
     verify(start, -1, seekKey, null, null, true, false, maps);
   }
-  
+
   void verify(int start, int end, Key seekKey, Text endRow, Text prevEndRow, List<TreeMap<Key,Value>> maps) throws IOException {
     if (seekKey != null) {
       verify(start, end, seekKey, endRow, prevEndRow, false, false, maps);
     }
-    
+
     verify(start, end, seekKey, endRow, prevEndRow, true, false, maps);
   }
-  
+
   public void test1() throws IOException {
     // TEST non overlapping inputs
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     List<TreeMap<Key,Value>> tmpList = new ArrayList<TreeMap<Key,Value>>(2);
-    
+
     for (int i = 0; i < 4; i++) {
       nkv(tm1, 0, i, false, "v" + i);
     }
@@ -140,14 +140,14 @@ public class MultiIteratorTest extends TestCase {
       verify(seek, nk(0, seek), tmpList);
     }
   }
-  
+
   public void test2() throws IOException {
     // TEST overlapping inputs
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     List<TreeMap<Key,Value>> tmpList = new ArrayList<TreeMap<Key,Value>>(2);
-    
+
     for (int i = 0; i < 8; i++) {
       if (i % 2 == 0)
         nkv(tm1, 0, i, false, "v" + i);
@@ -163,18 +163,18 @@ public class MultiIteratorTest extends TestCase {
       verify(seek, nk(0, seek), tmpList);
     }
   }
-  
+
   public void test3() throws IOException {
     // TEST single input
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     List<TreeMap<Key,Value>> tmpList = new ArrayList<TreeMap<Key,Value>>(2);
-    
+
     for (int i = 0; i < 8; i++) {
       nkv(tm1, 0, i, false, "v" + i);
     }
     tmpList.add(tm1);
-    
+
     for (int seek = -1; seek < 8; seek++) {
       if (seek == 7) {
         verify(seek, null, tmpList);
@@ -182,71 +182,71 @@ public class MultiIteratorTest extends TestCase {
       verify(seek, nk(0, seek), tmpList);
     }
   }
-  
+
   public void test4() throws IOException {
     // TEST empty input
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     List<SortedKeyValueIterator<Key,Value>> skvil = new ArrayList<SortedKeyValueIterator<Key,Value>>(1);
     skvil.add(new SortedMapIterator(tm1));
     MultiIterator mi = new MultiIterator(skvil, true);
-    
+
     assertFalse(mi.hasTop());
-    
+
     mi.seek(nrng(0, 6), EMPTY_COL_FAMS, false);
     assertFalse(mi.hasTop());
   }
-  
+
   public void test5() throws IOException {
     // TEST overlapping inputs AND prevRow AND endRow AND seek
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     List<TreeMap<Key,Value>> tmpList = new ArrayList<TreeMap<Key,Value>>(2);
-    
+
     for (int i = 0; i < 8; i++) {
       if (i % 2 == 0)
         nkv(tm1, i, 0, false, "v" + i);
       else
         nkv(tm2, i, 0, false, "v" + i);
     }
-    
+
     tmpList.add(tm1);
     tmpList.add(tm2);
     for (int seek = -1; seek < 9; seek++) {
       verify(Math.max(0, seek), 8, nk(seek, 0), null, null, true, true, tmpList);
       verify(Math.max(0, seek), 8, nk(seek, 0), null, null, false, true, tmpList);
-      
+
       for (int er = seek; er < 10; er++) {
-        
+
         int end = seek > er ? seek : Math.min(er + 1, 8);
-        
+
         int noSeekEnd = Math.min(er + 1, 8);
         if (er < 0) {
           noSeekEnd = 0;
         }
-        
+
         verify(0, noSeekEnd, null, nr(er), null, true, true, tmpList);
         verify(Math.max(0, seek), end, nk(seek, 0), nr(er), null, true, true, tmpList);
         verify(Math.max(0, seek), end, nk(seek, 0), nr(er), null, false, true, tmpList);
-        
+
         for (int per = -1; per < er; per++) {
-          
+
           int start = Math.max(per + 1, seek);
-          
+
           if (start > er)
             end = start;
-          
+
           if (per >= 8)
             end = start;
-          
+
           int noSeekStart = Math.max(0, per + 1);
-          
+
           if (er < 0 || per >= 7) {
             noSeekEnd = noSeekStart;
           }
-          
+
           verify(noSeekStart, noSeekEnd, null, nr(er), nr(per), true, true, tmpList);
           verify(Math.max(0, start), end, nk(seek, 0), nr(er), nr(per), true, true, tmpList);
           verify(Math.max(0, start), end, nk(seek, 0), nr(er), nr(per), false, true, tmpList);
@@ -254,80 +254,80 @@ public class MultiIteratorTest extends TestCase {
       }
     }
   }
-  
+
   public void test6() throws IOException {
     // TEst setting an endKey
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     nkv(tm1, 3, 0, false, "1");
     nkv(tm1, 4, 0, false, "2");
     nkv(tm1, 6, 0, false, "3");
-    
+
     List<SortedKeyValueIterator<Key,Value>> skvil = new ArrayList<SortedKeyValueIterator<Key,Value>>(1);
     skvil.add(new SortedMapIterator(tm1));
     MultiIterator mi = new MultiIterator(skvil, true);
     mi.seek(new Range(null, true, nk(5, 9), false), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(3, 0)));
     assertTrue(mi.getTopValue().toString().equals("1"));
     mi.next();
-    
+
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(4, 0)));
     assertTrue(mi.getTopValue().toString().equals("2"));
     mi.next();
-    
+
     assertFalse(mi.hasTop());
-    
+
     mi.seek(new Range(nk(4, 10), true, nk(5, 9), false), EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(4, 0)));
     assertTrue(mi.getTopValue().toString().equals("2"));
     mi.next();
-    
+
     assertFalse(mi.hasTop());
-    
+
     mi.seek(new Range(nk(4, 10), true, nk(6, 0), false), EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(4, 0)));
     assertTrue(mi.getTopValue().toString().equals("2"));
     mi.next();
-    
+
     assertFalse(mi.hasTop());
-    
+
     mi.seek(new Range(nk(4, 10), true, nk(6, 0), true), EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(4, 0)));
     assertTrue(mi.getTopValue().toString().equals("2"));
     mi.next();
-    
+
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(6, 0)));
     assertTrue(mi.getTopValue().toString().equals("3"));
     mi.next();
-    
+
     assertFalse(mi.hasTop());
-    
+
     mi.seek(new Range(nk(4, 0), true, nk(6, 0), false), EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(4, 0)));
     assertTrue(mi.getTopValue().toString().equals("2"));
     mi.next();
-    
+
     assertFalse(mi.hasTop());
-    
+
     mi.seek(new Range(nk(4, 0), false, nk(6, 0), false), EMPTY_COL_FAMS, false);
     assertFalse(mi.hasTop());
-    
+
     mi.seek(new Range(nk(4, 0), false, nk(6, 0), true), EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopKey().equals(nk(6, 0)));
     assertTrue(mi.getTopValue().toString().equals("3"));
     mi.next();
     assertFalse(mi.hasTop());
-    
+
   }
-  
+
   public void test7() throws IOException {
     // TEst setting an endKey
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
@@ -340,13 +340,13 @@ public class MultiIteratorTest extends TestCase {
     nkv(tm1, 1, 0, false, "7");
     nkv(tm1, 2, 1, false, "8");
     nkv(tm1, 2, 0, false, "9");
-    
+
     List<SortedKeyValueIterator<Key,Value>> skvil = new ArrayList<SortedKeyValueIterator<Key,Value>>(1);
     skvil.add(new SortedMapIterator(tm1));
-    
+
     KeyExtent extent = new KeyExtent(new Text("tablename"), nr(1), nr(0));
     MultiIterator mi = new MultiIterator(skvil, extent);
-    
+
     Range r1 = new Range((Text) null, (Text) null);
     mi.seek(r1, EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
@@ -359,7 +359,7 @@ public class MultiIteratorTest extends TestCase {
     assertTrue(mi.getTopValue().toString().equals("7"));
     mi.next();
     assertFalse(mi.hasTop());
-    
+
     Range r2 = new Range(nk(0, 0), true, nk(1, 1), true);
     mi.seek(r2, EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
@@ -369,32 +369,32 @@ public class MultiIteratorTest extends TestCase {
     assertTrue(mi.getTopValue().toString().equals("6"));
     mi.next();
     assertFalse(mi.hasTop());
-    
+
     Range r3 = new Range(nk(0, 0), false, nk(1, 1), false);
     mi.seek(r3, EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopValue().toString().equals("5"));
     mi.next();
     assertFalse(mi.hasTop());
-    
+
     Range r4 = new Range(nk(1, 2), true, nk(1, 1), false);
     mi.seek(r4, EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopValue().toString().equals("5"));
     mi.next();
     assertFalse(mi.hasTop());
-    
+
     Range r5 = new Range(nk(1, 2), false, nk(1, 1), true);
     mi.seek(r5, EMPTY_COL_FAMS, false);
     assertTrue(mi.hasTop());
     assertTrue(mi.getTopValue().toString().equals("6"));
     mi.next();
     assertFalse(mi.hasTop());
-    
+
     Range r6 = new Range(nk(2, 1), true, nk(2, 0), true);
     mi.seek(r6, EMPTY_COL_FAMS, false);
     assertFalse(mi.hasTop());
-    
+
     Range r7 = new Range(nk(0, 3), true, nk(0, 1), true);
     mi.seek(r7, EMPTY_COL_FAMS, false);
     assertFalse(mi.hasTop());


[38/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/KeyFunctor.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/KeyFunctor.java b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/KeyFunctor.java
index 69902a4..5c35f65 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/KeyFunctor.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/KeyFunctor.java
@@ -22,9 +22,9 @@ import org.apache.accumulo.core.data.Range;
 public interface KeyFunctor {
   /**
    * Implementations should return null if a range can not be converted to a bloom key.
-   * 
+   *
    */
   org.apache.hadoop.util.bloom.Key transform(Range range);
-  
+
   org.apache.hadoop.util.bloom.Key transform(Key key);
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/RowFunctor.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/RowFunctor.java b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/RowFunctor.java
index 20eb26c..b46f593 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/RowFunctor.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/RowFunctor.java
@@ -22,18 +22,18 @@ import org.apache.accumulo.core.data.Range;
 import org.apache.hadoop.util.bloom.Key;
 
 public class RowFunctor implements KeyFunctor {
-  
+
   @Override
   public Key transform(org.apache.accumulo.core.data.Key acuKey) {
     byte keyData[];
-    
+
     ByteSequence row = acuKey.getRowData();
     keyData = new byte[row.length()];
     System.arraycopy(row.getBackingArray(), 0, keyData, 0, row.length());
-    
+
     return new Key(keyData, 1.0);
   }
-  
+
   @Override
   public Key transform(Range range) {
     if (isRangeInBloomFilter(range, PartialKey.ROW)) {
@@ -41,16 +41,16 @@ public class RowFunctor implements KeyFunctor {
     }
     return null;
   }
-  
+
   static boolean isRangeInBloomFilter(Range range, PartialKey keyDepth) {
-    
+
     if (range.getStartKey() == null || range.getEndKey() == null) {
       return false;
     }
-    
+
     if (range.getStartKey().equals(range.getEndKey(), keyDepth))
       return true;
-    
+
     // include everything but the deleted flag in the comparison...
     return range.getStartKey().followingKey(keyDepth).equals(range.getEndKey(), PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME) && !range.isEndKeyInclusive();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/map/MapFileOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/map/MapFileOperations.java b/core/src/main/java/org/apache/accumulo/core/file/map/MapFileOperations.java
index 267f805..fb2762f 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/map/MapFileOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/map/MapFileOperations.java
@@ -43,66 +43,66 @@ import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.MapFile;
 
 public class MapFileOperations extends FileOperations {
-  
+
   public static class RangeIterator implements FileSKVIterator {
-    
+
     SortedKeyValueIterator<Key,Value> reader;
     private Range range;
     private boolean hasTop;
-    
+
     public RangeIterator(SortedKeyValueIterator<Key,Value> reader) {
       this.reader = reader;
     }
-    
+
     @Override
     public void close() throws IOException {
       ((FileSKVIterator) reader).close();
     }
-    
+
     @Override
     public Key getFirstKey() throws IOException {
       return ((FileSKVIterator) reader).getFirstKey();
     }
-    
+
     @Override
     public Key getLastKey() throws IOException {
       return ((FileSKVIterator) reader).getLastKey();
     }
-    
+
     @Override
     public DataInputStream getMetaStore(String name) throws IOException {
       return ((FileSKVIterator) reader).getMetaStore(name);
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
       return new RangeIterator(reader.deepCopy(env));
     }
-    
+
     @Override
     public Key getTopKey() {
       if (!hasTop)
         throw new IllegalStateException();
       return reader.getTopKey();
     }
-    
+
     @Override
     public Value getTopValue() {
       if (!hasTop)
         throw new IllegalStateException();
       return reader.getTopValue();
     }
-    
+
     @Override
     public boolean hasTop() {
       return hasTop;
     }
-    
+
     @Override
     public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public void next() throws IOException {
       if (!hasTop)
@@ -110,87 +110,87 @@ public class MapFileOperations extends FileOperations {
       reader.next();
       hasTop = reader.hasTop() && !range.afterEndKey(reader.getTopKey());
     }
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
       reader.seek(range, columnFamilies, inclusive);
       this.range = range;
-      
+
       hasTop = reader.hasTop() && !range.afterEndKey(reader.getTopKey());
-      
+
       while (hasTop() && range.beforeStartKey(getTopKey())) {
         next();
       }
     }
-    
+
     @Override
     public void closeDeepCopies() throws IOException {
       ((FileSKVIterator) reader).closeDeepCopies();
     }
-    
+
     @Override
     public void setInterruptFlag(AtomicBoolean flag) {
       ((FileSKVIterator) reader).setInterruptFlag(flag);
     }
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     FileSKVIterator iter = new RangeIterator(new MapFileIterator(acuconf, fs, file, conf));
-    
+
     if (seekToBeginning)
       iter.seek(new Range(new Key(), null), new ArrayList<ByteSequence>(), false);
-    
+
     return iter;
   }
-  
+
   @Override
   public FileSKVWriter openWriter(final String file, final FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
-    
+
     throw new UnsupportedOperationException();
 
   }
-  
+
   @Override
   public FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return new SequenceFileIterator(MapFileUtil.openIndex(conf, fs, new Path(file)), false);
   }
-  
+
   @Override
   public long getFileSize(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return fs.getFileStatus(new Path(file + "/" + MapFile.DATA_FILE_NAME)).getLen();
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf) throws IOException {
     MapFileIterator mfIter = new MapFileIterator(tableConf, fs, file, conf);
-    
+
     FileSKVIterator iter = new RangeIterator(mfIter);
-    
+
     iter.seek(range, columnFamilies, inclusive);
-    
+
     return iter;
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf, BlockCache dataCache, BlockCache indexCache) throws IOException {
-    
+
     return openReader(file, range, columnFamilies, inclusive, fs, conf, tableConf);
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf,
       BlockCache dataCache, BlockCache indexCache) throws IOException {
-    
+
     return openReader(file, seekToBeginning, fs, conf, acuconf);
   }
-  
+
   @Override
   public FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf, BlockCache dCache, BlockCache iCache)
       throws IOException {
-    
+
     return openIndex(file, fs, conf, acuconf);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/map/MapFileUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/map/MapFileUtil.java b/core/src/main/java/org/apache/accumulo/core/file/map/MapFileUtil.java
index 41b00d9..1373eac 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/map/MapFileUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/map/MapFileUtil.java
@@ -36,7 +36,7 @@ public class MapFileUtil {
       throw e;
     }
   }
-  
+
   @SuppressWarnings("deprecation")
   public static SequenceFile.Reader openIndex(Configuration conf, FileSystem fs, Path mapFile) throws IOException {
     Path indexPath = new Path(mapFile, MapFile.INDEX_FILE_NAME);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/BlockIndex.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/BlockIndex.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/BlockIndex.java
index 2156a67..1ed9aca 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/BlockIndex.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/BlockIndex.java
@@ -27,40 +27,40 @@ import org.apache.accumulo.core.file.blockfile.ABlockReader;
 import org.apache.accumulo.core.file.rfile.MultiLevelIndex.IndexEntry;
 
 /**
- * 
+ *
  */
 public class BlockIndex {
-  
+
   public static BlockIndex getIndex(ABlockReader cacheBlock, IndexEntry indexEntry) throws IOException {
-    
+
     BlockIndex blockIndex = cacheBlock.getIndex(BlockIndex.class);
-    
+
     int accessCount = blockIndex.accessCount.incrementAndGet();
-    
+
     // 1 is a power of two, but do not care about it
     if (accessCount >= 2 && isPowerOfTwo(accessCount)) {
       blockIndex.buildIndex(accessCount, cacheBlock, indexEntry);
     }
-    
+
     if (blockIndex.blockIndex != null)
       return blockIndex;
 
     return null;
   }
-  
+
   private static boolean isPowerOfTwo(int x) {
     return ((x > 0) && (x & (x - 1)) == 0);
   }
-  
+
   private AtomicInteger accessCount = new AtomicInteger(0);
   private volatile BlockIndexEntry[] blockIndex = null;
 
   public static class BlockIndexEntry implements Comparable<BlockIndexEntry> {
-    
+
     private Key prevKey;
     private int entriesLeft;
     private int pos;
-    
+
     public BlockIndexEntry(int pos, int entriesLeft, Key prevKey) {
       this.pos = pos;
       this.entriesLeft = entriesLeft;
@@ -70,7 +70,7 @@ public class BlockIndex {
     public BlockIndexEntry(Key key) {
       this.prevKey = key;
     }
-    
+
     public int getEntriesLeft() {
       return entriesLeft;
     }
@@ -79,39 +79,39 @@ public class BlockIndex {
     public int compareTo(BlockIndexEntry o) {
       return prevKey.compareTo(o.prevKey);
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof BlockIndexEntry)
         return compareTo((BlockIndexEntry) o) == 0;
       return false;
     }
-    
+
     @Override
     public String toString() {
       return prevKey + " " + entriesLeft + " " + pos;
     }
-    
+
     public Key getPrevKey() {
       return prevKey;
     }
-    
+
     @Override
     public int hashCode() {
       assert false : "hashCode not designed";
       return 42; // any arbitrary constant will do
     }
   }
-  
+
   public BlockIndexEntry seekBlock(Key startKey, ABlockReader cacheBlock) {
 
     // get a local ref to the index, another thread could change it
     BlockIndexEntry[] blockIndex = this.blockIndex;
-    
+
     int pos = Arrays.binarySearch(blockIndex, new BlockIndexEntry(startKey));
 
     int index;
-    
+
     if (pos < 0) {
       if (pos == -1)
         return null; // less than the first key in index, did not index the first key in block so just return null... code calling this will scan from beginning
@@ -127,7 +127,7 @@ public class BlockIndex {
           break;
       }
     }
-    
+
     // handle case where multiple keys in block are exactly the same, want to find the earliest key in the index
     while (index - 1 > 0) {
       if (blockIndex[index].getPrevKey().equals(blockIndex[index - 1].getPrevKey()))
@@ -136,7 +136,7 @@ public class BlockIndex {
         break;
 
     }
-    
+
     if (index == 0 && blockIndex[index].getPrevKey().equals(startKey))
       return null;
 
@@ -144,24 +144,24 @@ public class BlockIndex {
     cacheBlock.seek(bie.pos);
     return bie;
   }
-  
+
   private synchronized void buildIndex(int indexEntries, ABlockReader cacheBlock, IndexEntry indexEntry) throws IOException {
     cacheBlock.seek(0);
-    
+
     RelativeKey rk = new RelativeKey();
     Value val = new Value();
-    
+
     int interval = indexEntry.getNumEntries() / indexEntries;
-    
+
     if (interval <= 32)
       return;
-    
+
     // multiple threads could try to create the index with different sizes, do not replace a large index with a smaller one
     if (this.blockIndex != null && this.blockIndex.length > indexEntries - 1)
       return;
 
     int count = 0;
-    
+
     ArrayList<BlockIndexEntry> index = new ArrayList<BlockIndexEntry>(indexEntries - 1);
 
     while (count < (indexEntry.getNumEntries() - interval + 1)) {
@@ -174,7 +174,7 @@ public class BlockIndex {
       if (count > 0 && count % interval == 0) {
         index.add(new BlockIndexEntry(pos, indexEntry.getNumEntries() - count, myPrevKey));
       }
-      
+
       count++;
     }
 
@@ -182,7 +182,7 @@ public class BlockIndex {
 
     cacheBlock.seek(0);
   }
-  
+
   BlockIndexEntry[] getIndexEntries() {
     return blockIndex;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
index 9456331..cd6bff8 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
@@ -61,7 +61,9 @@ public class CreateEmpty {
   static class Opts extends Help {
     @Parameter(names = {"-c", "--codec"}, description = "the compression codec to use.", validateWith = IsSupportedCompressionAlgorithm.class)
     String codec = Compression.COMPRESSION_NONE;
-    @Parameter(description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.", required = true, validateWith = NamedLikeRFile.class)
+    @Parameter(
+        description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
+        required = true, validateWith = NamedLikeRFile.class)
     List<String> files = new ArrayList<String>();
   }
 
@@ -74,7 +76,8 @@ public class CreateEmpty {
     for (String arg : opts.files) {
       Path path = new Path(arg);
       log.info("Writing to file '" + path + "'");
-      FileSKVWriter writer = (new RFileOperations()).openWriter(arg, path.getFileSystem(conf), conf, DefaultConfiguration.getDefaultConfiguration(), opts.codec);
+      FileSKVWriter writer = (new RFileOperations())
+          .openWriter(arg, path.getFileSystem(conf), conf, DefaultConfiguration.getDefaultConfiguration(), opts.codec);
       writer.close();
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/IndexIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/IndexIterator.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/IndexIterator.java
index f9c8686..b1dab36 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/IndexIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/IndexIterator.java
@@ -30,10 +30,10 @@ import org.apache.accumulo.core.iterators.IteratorEnvironment;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 class IndexIterator implements SortedKeyValueIterator<Key,Value> {
-  
+
   private Key key;
   private Iterator<IndexEntry> indexIter;
-  
+
   IndexIterator(Iterator<IndexEntry> indexIter) {
     this.indexIter = indexIter;
     if (indexIter.hasNext())
@@ -41,32 +41,32 @@ class IndexIterator implements SortedKeyValueIterator<Key,Value> {
     else
       key = null;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public Key getTopKey() {
     return key;
   }
-  
+
   @Override
   public Value getTopValue() {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public boolean hasTop() {
     return key != null;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void next() throws IOException {
     if (indexIter.hasNext())
@@ -74,10 +74,10 @@ class IndexIterator implements SortedKeyValueIterator<Key,Value> {
     else
       key = null;
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiIndexIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiIndexIterator.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiIndexIterator.java
index 5dade97..f220a58 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiIndexIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiIndexIterator.java
@@ -35,62 +35,62 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.accumulo.core.iterators.system.HeapIterator;
 
 class MultiIndexIterator extends HeapIterator implements FileSKVIterator {
-  
+
   private RFile.Reader source;
-  
+
   MultiIndexIterator(RFile.Reader source, List<Iterator<IndexEntry>> indexes) {
     super(indexes.size());
-    
+
     this.source = source;
-    
+
     for (Iterator<IndexEntry> index : indexes) {
       addSource(new IndexIterator(index));
     }
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void close() throws IOException {
     source.close();
   }
-  
+
   @Override
   public void closeDeepCopies() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public Key getFirstKey() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public Key getLastKey() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public DataInputStream getMetaStore(String name) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiLevelIndex.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiLevelIndex.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiLevelIndex.java
index 632968e..2109478 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiLevelIndex.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/MultiLevelIndex.java
@@ -41,7 +41,7 @@ import org.apache.accumulo.core.file.rfile.bcfile.Utils;
 import org.apache.hadoop.io.WritableComparable;
 
 public class MultiLevelIndex {
-  
+
   public static class IndexEntry implements WritableComparable<IndexEntry> {
     private Key key;
     private int entries;
@@ -49,7 +49,7 @@ public class MultiLevelIndex {
     private long compressedSize;
     private long rawSize;
     private boolean newFormat;
-    
+
     IndexEntry(Key k, int e, long offset, long compressedSize, long rawSize) {
       this.key = k;
       this.entries = e;
@@ -58,11 +58,11 @@ public class MultiLevelIndex {
       this.rawSize = rawSize;
       newFormat = true;
     }
-    
+
     public IndexEntry(boolean newFormat) {
       this.newFormat = newFormat;
     }
-    
+
     @Override
     public void readFields(DataInput in) throws IOException {
       key = new Key();
@@ -78,7 +78,7 @@ public class MultiLevelIndex {
         rawSize = -1;
       }
     }
-    
+
     @Override
     public void write(DataOutput out) throws IOException {
       key.write(out);
@@ -89,59 +89,59 @@ public class MultiLevelIndex {
         Utils.writeVLong(out, rawSize);
       }
     }
-    
+
     public Key getKey() {
       return key;
     }
-    
+
     public int getNumEntries() {
       return entries;
     }
-    
+
     public long getOffset() {
       return offset;
     }
-    
+
     public long getCompressedSize() {
       return compressedSize;
     }
-    
+
     public long getRawSize() {
       return rawSize;
     }
-    
+
     @Override
     public int compareTo(IndexEntry o) {
       return key.compareTo(o.key);
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof IndexEntry)
-        return compareTo((IndexEntry)o) == 0;
+        return compareTo((IndexEntry) o) == 0;
       return false;
     }
-    
+
     @Override
     public int hashCode() {
       assert false : "hashCode not designed";
       return 42; // any arbitrary constant will do
     }
   }
-  
+
   // a list that deserializes index entries on demand
   private static class SerializedIndex extends AbstractList<IndexEntry> implements List<IndexEntry>, RandomAccess {
-    
+
     private int[] offsets;
     private byte[] data;
     private boolean newFormat;
-    
+
     SerializedIndex(int[] offsets, byte[] data, boolean newFormat) {
       this.offsets = offsets;
       this.data = data;
       this.newFormat = newFormat;
     }
-    
+
     @Override
     public IndexEntry get(int index) {
       int len;
@@ -149,41 +149,41 @@ public class MultiLevelIndex {
         len = data.length - offsets[index];
       else
         len = offsets[index + 1] - offsets[index];
-      
+
       ByteArrayInputStream bais = new ByteArrayInputStream(data, offsets[index], len);
       DataInputStream dis = new DataInputStream(bais);
-      
+
       IndexEntry ie = new IndexEntry(newFormat);
       try {
         ie.readFields(dis);
       } catch (IOException e) {
         throw new RuntimeException(e);
       }
-      
+
       return ie;
     }
-    
+
     @Override
     public int size() {
       return offsets.length;
     }
-    
+
     public long sizeInBytes() {
       return data.length + 4 * offsets.length;
     }
-    
+
   }
-  
+
   private static class KeyIndex extends AbstractList<Key> implements List<Key>, RandomAccess {
-    
+
     private int[] offsets;
     private byte[] data;
-    
+
     KeyIndex(int[] offsets, byte[] data) {
       this.offsets = offsets;
       this.data = data;
     }
-    
+
     @Override
     public Key get(int index) {
       int len;
@@ -191,122 +191,122 @@ public class MultiLevelIndex {
         len = data.length - offsets[index];
       else
         len = offsets[index + 1] - offsets[index];
-      
+
       ByteArrayInputStream bais = new ByteArrayInputStream(data, offsets[index], len);
       DataInputStream dis = new DataInputStream(bais);
-      
+
       Key key = new Key();
       try {
         key.readFields(dis);
       } catch (IOException e) {
         throw new RuntimeException(e);
       }
-      
+
       return key;
     }
-    
+
     @Override
     public int size() {
       return offsets.length;
     }
   }
-  
+
   static class IndexBlock {
-    
+
     private ByteArrayOutputStream indexBytes;
     private DataOutputStream indexOut;
-    
+
     private ArrayList<Integer> offsets;
     private int level;
     private int offset;
-    
+
     SerializedIndex index;
     KeyIndex keyIndex;
     private boolean hasNext;
-    
+
     public IndexBlock(int level, int totalAdded) {
       // System.out.println("IndexBlock("+level+","+levelCount+","+totalAdded+")");
-      
+
       this.level = level;
       this.offset = totalAdded;
-      
+
       indexBytes = new ByteArrayOutputStream();
       indexOut = new DataOutputStream(indexBytes);
       offsets = new ArrayList<Integer>();
     }
-    
+
     public IndexBlock() {}
-    
+
     public void add(Key key, int value, long offset, long compressedSize, long rawSize) throws IOException {
       offsets.add(indexOut.size());
       new IndexEntry(key, value, offset, compressedSize, rawSize).write(indexOut);
     }
-    
+
     int getSize() {
       return indexOut.size() + 4 * offsets.size();
     }
-    
+
     public void write(DataOutput out) throws IOException {
       out.writeInt(level);
       out.writeInt(offset);
       out.writeBoolean(hasNext);
-      
+
       out.writeInt(offsets.size());
       for (Integer offset : offsets) {
         out.writeInt(offset);
       }
-      
+
       indexOut.close();
       byte[] indexData = indexBytes.toByteArray();
-      
+
       out.writeInt(indexData.length);
       out.write(indexData);
     }
-    
+
     public void readFields(DataInput in, int version) throws IOException {
-      
+
       if (version == RFile.RINDEX_VER_6 || version == RFile.RINDEX_VER_7) {
         level = in.readInt();
         offset = in.readInt();
         hasNext = in.readBoolean();
-        
+
         int numOffsets = in.readInt();
         int[] offsets = new int[numOffsets];
-        
+
         for (int i = 0; i < numOffsets; i++)
           offsets[i] = in.readInt();
-        
+
         int indexSize = in.readInt();
         byte[] serializedIndex = new byte[indexSize];
         in.readFully(serializedIndex);
-        
+
         index = new SerializedIndex(offsets, serializedIndex, true);
         keyIndex = new KeyIndex(offsets, serializedIndex);
       } else if (version == RFile.RINDEX_VER_3) {
         level = 0;
         offset = 0;
         hasNext = false;
-        
+
         int size = in.readInt();
-        
+
         ByteArrayOutputStream baos = new ByteArrayOutputStream();
         DataOutputStream dos = new DataOutputStream(baos);
         ArrayList<Integer> oal = new ArrayList<Integer>();
-        
+
         for (int i = 0; i < size; i++) {
           IndexEntry ie = new IndexEntry(false);
           oal.add(dos.size());
           ie.readFields(in);
           ie.write(dos);
         }
-        
+
         dos.close();
-        
+
         int[] oia = new int[oal.size()];
         for (int i = 0; i < oal.size(); i++) {
           oia[i] = oal.get(i);
         }
-        
+
         byte[] serializedIndex = baos.toByteArray();
         index = new SerializedIndex(oia, serializedIndex, false);
         keyIndex = new KeyIndex(oia, serializedIndex);
@@ -314,100 +314,100 @@ public class MultiLevelIndex {
         level = 0;
         offset = 0;
         hasNext = false;
-        
+
         int numIndexEntries = in.readInt();
         int offsets[] = new int[numIndexEntries];
         for (int i = 0; i < numIndexEntries; i++) {
           offsets[i] = in.readInt();
         }
-        
+
         int size = in.readInt();
         byte[] indexData = new byte[size];
         in.readFully(indexData);
-        
+
         index = new SerializedIndex(offsets, indexData, false);
         keyIndex = new KeyIndex(offsets, indexData);
       } else {
         throw new RuntimeException("Unexpected version " + version);
       }
-      
+
     }
-    
+
     List<IndexEntry> getIndex() {
       return index;
     }
-    
+
     public List<Key> getKeyIndex() {
       return keyIndex;
     }
-    
+
     int getLevel() {
       return level;
     }
-    
+
     int getOffset() {
       return offset;
     }
-    
+
     boolean hasNext() {
       return hasNext;
     }
-    
+
     void setHasNext(boolean b) {
       this.hasNext = b;
     }
-    
+
   }
-  
+
   /**
    * this class buffers writes to the index so that chunks of index blocks are contiguous in the file instead of having index blocks sprinkled throughout the
    * file making scans of the entire index slow.
    */
   public static class BufferedWriter {
-    
+
     private Writer writer;
     private DataOutputStream buffer;
     private int buffered;
     private ByteArrayOutputStream baos;
-    
+
     public BufferedWriter(Writer writer) {
       this.writer = writer;
       baos = new ByteArrayOutputStream(1 << 20);
       buffer = new DataOutputStream(baos);
       buffered = 0;
     }
-    
+
     private void flush() throws IOException {
       buffer.close();
-      
+
       DataInputStream dis = new DataInputStream(new ByteArrayInputStream(baos.toByteArray()));
-      
+
       IndexEntry ie = new IndexEntry(true);
       for (int i = 0; i < buffered; i++) {
         ie.readFields(dis);
         writer.add(ie.getKey(), ie.getNumEntries(), ie.getOffset(), ie.getCompressedSize(), ie.getRawSize());
       }
-      
+
       buffered = 0;
       baos = new ByteArrayOutputStream(1 << 20);
       buffer = new DataOutputStream(baos);
-      
+
     }
-    
+
     public void add(Key key, int data, long offset, long compressedSize, long rawSize) throws IOException {
       if (buffer.size() > (10 * 1 << 20)) {
         flush();
       }
-      
+
       new IndexEntry(key, data, offset, compressedSize, rawSize).write(buffer);
       buffered++;
     }
-    
+
     public void addLast(Key key, int data, long offset, long compressedSize, long rawSize) throws IOException {
       flush();
       writer.addLast(key, data, offset, compressedSize, rawSize);
     }
-    
+
     public void close(DataOutput out) throws IOException {
       writer.close(out);
     }
@@ -415,74 +415,74 @@ public class MultiLevelIndex {
 
   public static class Writer {
     private int threshold;
-    
+
     private ArrayList<IndexBlock> levels;
-    
+
     private int totalAdded;
-    
+
     private boolean addedLast = false;
-    
+
     private BlockFileWriter blockFileWriter;
-    
+
     Writer(BlockFileWriter blockFileWriter, int maxBlockSize) {
       this.blockFileWriter = blockFileWriter;
       this.threshold = maxBlockSize;
       levels = new ArrayList<IndexBlock>();
     }
-    
+
     private void add(int level, Key key, int data, long offset, long compressedSize, long rawSize) throws IOException {
       if (level == levels.size()) {
         levels.add(new IndexBlock(level, 0));
       }
-      
+
       IndexBlock iblock = levels.get(level);
-      
+
       iblock.add(key, data, offset, compressedSize, rawSize);
     }
-    
+
     private void flush(int level, Key lastKey, boolean last) throws IOException {
-      
+
       if (last && level == levels.size() - 1)
         return;
-      
+
       IndexBlock iblock = levels.get(level);
       if ((iblock.getSize() > threshold && iblock.offsets.size() > 1) || last) {
         ABlockWriter out = blockFileWriter.prepareDataBlock();
         iblock.setHasNext(!last);
         iblock.write(out);
         out.close();
-        
+
         add(level + 1, lastKey, 0, out.getStartPos(), out.getCompressedSize(), out.getRawSize());
         flush(level + 1, lastKey, last);
-        
+
         if (last)
           levels.set(level, null);
         else
           levels.set(level, new IndexBlock(level, totalAdded));
       }
     }
-    
+
     public void add(Key key, int data, long offset, long compressedSize, long rawSize) throws IOException {
       totalAdded++;
       add(0, key, data, offset, compressedSize, rawSize);
       flush(0, key, false);
     }
-    
+
     public void addLast(Key key, int data, long offset, long compressedSize, long rawSize) throws IOException {
       if (addedLast)
         throw new IllegalStateException("already added last");
-      
+
       totalAdded++;
       add(0, key, data, offset, compressedSize, rawSize);
       flush(0, key, true);
       addedLast = true;
-      
+
     }
-    
+
     public void close(DataOutput out) throws IOException {
       if (totalAdded > 0 && !addedLast)
         throw new IllegalStateException("did not call addLast");
-      
+
       out.writeInt(totalAdded);
       // save root node
       if (levels.size() > 0) {
@@ -490,32 +490,32 @@ public class MultiLevelIndex {
       } else {
         new IndexBlock(0, 0).write(out);
       }
-      
+
     }
   }
-  
+
   public static class Reader {
     private IndexBlock rootBlock;
     private BlockFileReader blockStore;
     private int version;
     private int size;
-    
+
     public class Node {
-      
+
       private Node parent;
       private IndexBlock indexBlock;
       private int currentPos;
-      
+
       Node(Node parent, IndexBlock iBlock) {
         this.parent = parent;
         this.indexBlock = iBlock;
       }
-      
+
       Node(IndexBlock rootInfo) {
         this.parent = null;
         this.indexBlock = rootInfo;
       }
-      
+
       private Node lookup(Key key) throws IOException {
         int pos = Collections.binarySearch(indexBlock.getKeyIndex(), key, new Comparator<Key>() {
           @Override
@@ -523,86 +523,86 @@ public class MultiLevelIndex {
             return o1.compareTo(o2);
           }
         });
-        
+
         if (pos < 0)
           pos = (pos * -1) - 1;
-        
+
         if (pos == indexBlock.getIndex().size()) {
           if (parent != null)
             throw new IllegalStateException();
           this.currentPos = pos;
           return this;
         }
-        
+
         this.currentPos = pos;
-        
+
         if (indexBlock.getLevel() == 0) {
           return this;
         }
-        
+
         IndexEntry ie = indexBlock.getIndex().get(pos);
         Node child = new Node(this, getIndexBlock(ie));
         return child.lookup(key);
       }
-      
+
       private Node getLast() throws IOException {
         currentPos = indexBlock.getIndex().size() - 1;
         if (indexBlock.getLevel() == 0)
           return this;
-        
+
         IndexEntry ie = indexBlock.getIndex().get(currentPos);
         Node child = new Node(this, getIndexBlock(ie));
         return child.getLast();
       }
-      
+
       private Node getFirst() throws IOException {
         currentPos = 0;
         if (indexBlock.getLevel() == 0)
           return this;
-        
+
         IndexEntry ie = indexBlock.getIndex().get(currentPos);
         Node child = new Node(this, getIndexBlock(ie));
         return child.getFirst();
       }
-      
+
       private Node getPrevious() throws IOException {
         if (currentPos == 0)
           return parent.getPrevious();
-        
+
         currentPos--;
-        
+
         IndexEntry ie = indexBlock.getIndex().get(currentPos);
         Node child = new Node(this, getIndexBlock(ie));
         return child.getLast();
-        
+
       }
-      
+
       private Node getNext() throws IOException {
         if (currentPos == indexBlock.getIndex().size() - 1)
           return parent.getNext();
-        
+
         currentPos++;
-        
+
         IndexEntry ie = indexBlock.getIndex().get(currentPos);
         Node child = new Node(this, getIndexBlock(ie));
         return child.getFirst();
-        
+
       }
-      
+
       Node getNextNode() throws IOException {
         return parent.getNext();
       }
-      
+
       Node getPreviousNode() throws IOException {
         return parent.getPrevious();
       }
     }
-    
+
     static public class IndexIterator implements ListIterator<IndexEntry> {
-      
+
       private Node node;
       private ListIterator<IndexEntry> liter;
-      
+
       private Node getPrevNode() {
         try {
           return node.getPreviousNode();
@@ -610,7 +610,7 @@ public class MultiLevelIndex {
           throw new RuntimeException(e);
         }
       }
-      
+
       private Node getNextNode() {
         try {
           return node.getNextNode();
@@ -618,155 +618,155 @@ public class MultiLevelIndex {
           throw new RuntimeException(e);
         }
       }
-      
+
       public IndexIterator() {
         node = null;
       }
-      
+
       public IndexIterator(Node node) {
         this.node = node;
         liter = node.indexBlock.getIndex().listIterator(node.currentPos);
       }
-      
+
       @Override
       public boolean hasNext() {
         if (node == null)
           return false;
-        
+
         if (!liter.hasNext()) {
           return node.indexBlock.hasNext();
         } else {
           return true;
         }
-        
+
       }
-      
+
       public IndexEntry peekPrevious() {
         IndexEntry ret = previous();
         next();
         return ret;
       }
-      
+
       public IndexEntry peek() {
         IndexEntry ret = next();
         previous();
         return ret;
       }
-      
+
       @Override
       public IndexEntry next() {
         if (!liter.hasNext()) {
           node = getNextNode();
           liter = node.indexBlock.getIndex().listIterator();
         }
-        
+
         return liter.next();
       }
-      
+
       @Override
       public boolean hasPrevious() {
         if (node == null)
           return false;
-        
+
         if (!liter.hasPrevious()) {
           return node.indexBlock.getOffset() > 0;
         } else {
           return true;
         }
       }
-      
+
       @Override
       public IndexEntry previous() {
         if (!liter.hasPrevious()) {
           node = getPrevNode();
           liter = node.indexBlock.getIndex().listIterator(node.indexBlock.getIndex().size());
         }
-        
+
         return liter.previous();
       }
-      
+
       @Override
       public int nextIndex() {
         return node.indexBlock.getOffset() + liter.nextIndex();
       }
-      
+
       @Override
       public int previousIndex() {
         return node.indexBlock.getOffset() + liter.previousIndex();
       }
-      
+
       @Override
       public void remove() {
         throw new UnsupportedOperationException();
       }
-      
+
       @Override
       public void set(IndexEntry e) {
         throw new UnsupportedOperationException();
-        
+
       }
-      
+
       @Override
       public void add(IndexEntry e) {
         throw new UnsupportedOperationException();
       }
-      
+
     }
-    
+
     public Reader(BlockFileReader blockStore, int version) {
       this.version = version;
       this.blockStore = blockStore;
     }
-    
+
     private IndexBlock getIndexBlock(IndexEntry ie) throws IOException {
       IndexBlock iblock = new IndexBlock();
       ABlockReader in = blockStore.getMetaBlock(ie.getOffset(), ie.getCompressedSize(), ie.getRawSize());
       iblock.readFields(in, version);
       in.close();
-      
+
       return iblock;
     }
-    
+
     public IndexIterator lookup(Key key) throws IOException {
       Node node = new Node(rootBlock);
       return new IndexIterator(node.lookup(key));
     }
-    
+
     public void readFields(DataInput in) throws IOException {
-      
+
       size = 0;
-      
+
       if (version == RFile.RINDEX_VER_6 || version == RFile.RINDEX_VER_7) {
         size = in.readInt();
       }
-      
+
       rootBlock = new IndexBlock();
       rootBlock.readFields(in, version);
-      
+
       if (version == RFile.RINDEX_VER_3 || version == RFile.RINDEX_VER_4) {
         size = rootBlock.getIndex().size();
       }
     }
-    
+
     public int size() {
       return size;
     }
-    
+
     private void getIndexInfo(IndexBlock ib, Map<Integer,Long> sizesByLevel, Map<Integer,Long> countsByLevel) throws IOException {
       Long size = sizesByLevel.get(ib.getLevel());
       if (size == null)
         size = 0l;
-      
+
       Long count = countsByLevel.get(ib.getLevel());
       if (count == null)
         count = 0l;
-      
+
       size += ib.index.sizeInBytes();
       count++;
-      
+
       sizesByLevel.put(ib.getLevel(), size);
       countsByLevel.put(ib.getLevel(), count);
-      
+
       if (ib.getLevel() > 0) {
         for (IndexEntry ie : ib.index) {
           IndexBlock cib = getIndexBlock(ie);
@@ -774,14 +774,14 @@ public class MultiLevelIndex {
         }
       }
     }
-    
+
     public void getIndexInfo(Map<Integer,Long> sizes, Map<Integer,Long> counts) throws IOException {
       getIndexInfo(rootBlock, sizes, counts);
     }
-    
+
     public Key getLastKey() {
       return rootBlock.getIndex().get(rootBlock.getIndex().size() - 1).getKey();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/PrintInfo.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/PrintInfo.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/PrintInfo.java
index 43586dd..f29efcc 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/PrintInfo.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/PrintInfo.java
@@ -39,7 +39,7 @@ import com.beust.jcommander.Parameter;
 
 public class PrintInfo {
   private static final Logger log = Logger.getLogger(PrintInfo.class);
-  
+
   static class Opts extends Help {
     @Parameter(names = {"-d", "--dump"}, description = "dump the key/value pairs")
     boolean dump = false;
@@ -48,29 +48,29 @@ public class PrintInfo {
     @Parameter(description = " <file> { <file> ... }")
     List<String> files = new ArrayList<String>();
   }
-  
+
   public static void main(String[] args) throws Exception {
     Configuration conf = new Configuration();
 
     AccumuloConfiguration aconf = SiteConfiguration.getInstance(DefaultConfiguration.getInstance());
     // TODO ACCUMULO-2462 This will only work for RFiles (path only, not URI) in HDFS when the correct filesystem for the given file
-    // is on Property.INSTANCE_DFS_DIR or, when INSTANCE_DFS_DIR is not defined, is on the default filesystem 
+    // is on Property.INSTANCE_DFS_DIR or, when INSTANCE_DFS_DIR is not defined, is on the default filesystem
     // defined in the Hadoop's core-site.xml
     //
     // A workaround is to always provide a URI to this class
     FileSystem hadoopFs = VolumeConfiguration.getDefaultVolume(conf, aconf).getFileSystem();
-    FileSystem localFs  = FileSystem.getLocal(conf);
+    FileSystem localFs = FileSystem.getLocal(conf);
     Opts opts = new Opts();
     opts.parseArgs(PrintInfo.class.getName(), args);
     if (opts.files.isEmpty()) {
       System.err.println("No files were given");
       System.exit(-1);
     }
-    
+
     long countBuckets[] = new long[11];
     long sizeBuckets[] = new long[countBuckets.length];
     long totalSize = 0;
-    
+
     for (String arg : opts.files) {
       Path path = new Path(arg);
       FileSystem fs;
@@ -81,14 +81,14 @@ public class PrintInfo {
         log.warn("Attempting to find file across filesystems. Consider providing URI instead of path");
         fs = hadoopFs.exists(path) ? hadoopFs : localFs; // fall back to local
       }
-      
+
       CachableBlockFile.Reader _rdr = new CachableBlockFile.Reader(fs, path, conf, null, null, aconf);
       Reader iter = new RFile.Reader(_rdr);
-      
+
       iter.printInfo();
       System.out.println();
       org.apache.accumulo.core.file.rfile.bcfile.PrintInfo.main(new String[] {arg});
-      
+
       if (opts.histogram || opts.dump) {
         iter.seek(new Range((Key) null, (Key) null), new ArrayList<ByteSequence>(), false);
         while (iter.hasTop()) {
@@ -113,7 +113,7 @@ public class PrintInfo {
           System.out.println(String.format("%11.0f : %10d %6.2f%%", Math.pow(10, i), countBuckets[i], sizeBuckets[i] * 100. / totalSize));
         }
       }
-      
+
       // If the output stream has closed, there is no reason to keep going.
       if (System.out.checkError())
         return;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/RFile.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/RFile.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/RFile.java
index 9dcb3a5..0b464d8 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/RFile.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/RFile.java
@@ -68,47 +68,47 @@ import org.apache.hadoop.io.Writable;
 import org.apache.log4j.Logger;
 
 public class RFile {
-  
+
   public static final String EXTENSION = "rf";
-  
+
   private static final Logger log = Logger.getLogger(RFile.class);
-  
+
   private RFile() {}
-  
+
   private static final int RINDEX_MAGIC = 0x20637474;
   static final int RINDEX_VER_7 = 7;
   static final int RINDEX_VER_6 = 6;
   // static final int RINDEX_VER_5 = 5; // unreleased
   static final int RINDEX_VER_4 = 4;
   static final int RINDEX_VER_3 = 3;
-    
+
   private static class LocalityGroupMetadata implements Writable {
-    
+
     private int startBlock;
     private Key firstKey;
     private Map<ByteSequence,MutableLong> columnFamilies;
-    
+
     private boolean isDefaultLG = false;
     private String name;
     private Set<ByteSequence> previousColumnFamilies;
-    
+
     private MultiLevelIndex.BufferedWriter indexWriter;
     private MultiLevelIndex.Reader indexReader;
-    
+
     public LocalityGroupMetadata(int version, BlockFileReader br) {
       columnFamilies = new HashMap<ByteSequence,MutableLong>();
       indexReader = new MultiLevelIndex.Reader(br, version);
     }
-    
+
     public LocalityGroupMetadata(int nextBlock, Set<ByteSequence> pcf, int indexBlockSize, BlockFileWriter bfw) {
       this.startBlock = nextBlock;
       isDefaultLG = true;
       columnFamilies = new HashMap<ByteSequence,MutableLong>();
       previousColumnFamilies = pcf;
-      
+
       indexWriter = new MultiLevelIndex.BufferedWriter(new MultiLevelIndex.Writer(bfw, indexBlockSize));
     }
-    
+
     public LocalityGroupMetadata(String name, Set<ByteSequence> cfset, int nextBlock, int indexBlockSize, BlockFileWriter bfw) {
       this.startBlock = nextBlock;
       this.name = name;
@@ -117,22 +117,22 @@ public class RFile {
       for (ByteSequence cf : cfset) {
         columnFamilies.put(cf, new MutableLong(0));
       }
-      
+
       indexWriter = new MultiLevelIndex.BufferedWriter(new MultiLevelIndex.Writer(bfw, indexBlockSize));
     }
-    
+
     private Key getFirstKey() {
       return firstKey;
     }
-    
+
     private void setFirstKey(Key key) {
       if (firstKey != null)
         throw new IllegalStateException();
       this.firstKey = new Key(key);
     }
-    
+
     public void updateColumnCount(Key key) {
-      
+
       if (isDefaultLG && columnFamilies == null) {
         if (previousColumnFamilies.size() > 0) {
           // only do this check when there are previous column families
@@ -141,23 +141,23 @@ public class RFile {
             throw new IllegalArgumentException("Added column family \"" + cf + "\" to default locality group that was in previous locality group");
           }
         }
-        
+
         // no longer keeping track of column families, so return
         return;
       }
-      
+
       ByteSequence cf = key.getColumnFamilyData();
       MutableLong count = columnFamilies.get(cf);
-      
+
       if (count == null) {
         if (!isDefaultLG) {
           throw new IllegalArgumentException("invalid column family : " + cf);
         }
-        
+
         if (previousColumnFamilies.contains(cf)) {
           throw new IllegalArgumentException("Added column family \"" + cf + "\" to default locality group that was in previous locality group");
         }
-        
+
         if (columnFamilies.size() > Writer.MAX_CF_IN_DLG) {
           // stop keeping track, there are too many
           columnFamilies = null;
@@ -165,86 +165,86 @@ public class RFile {
         }
         count = new MutableLong(0);
         columnFamilies.put(new ArrayByteSequence(cf.getBackingArray(), cf.offset(), cf.length()), count);
-        
+
       }
-      
+
       count.increment();
-      
+
     }
-    
+
     @Override
     public void readFields(DataInput in) throws IOException {
-      
+
       isDefaultLG = in.readBoolean();
       if (!isDefaultLG) {
         name = in.readUTF();
       }
-      
+
       startBlock = in.readInt();
-      
+
       int size = in.readInt();
-      
+
       if (size == -1) {
         if (!isDefaultLG)
           throw new IllegalStateException("Non default LG " + name + " does not have column families");
-        
+
         columnFamilies = null;
       } else {
         if (columnFamilies == null)
           columnFamilies = new HashMap<ByteSequence,MutableLong>();
         else
           columnFamilies.clear();
-        
+
         for (int i = 0; i < size; i++) {
           int len = in.readInt();
           byte cf[] = new byte[len];
           in.readFully(cf);
           long count = in.readLong();
-          
+
           columnFamilies.put(new ArrayByteSequence(cf), new MutableLong(count));
         }
       }
-      
+
       if (in.readBoolean()) {
         firstKey = new Key();
         firstKey.readFields(in);
       } else {
         firstKey = null;
       }
-      
+
       indexReader.readFields(in);
     }
-    
+
     @Override
     public void write(DataOutput out) throws IOException {
-      
+
       out.writeBoolean(isDefaultLG);
       if (!isDefaultLG) {
         out.writeUTF(name);
       }
-      
+
       out.writeInt(startBlock);
-      
+
       if (isDefaultLG && columnFamilies == null) {
         // only expect null when default LG, otherwise let a NPE occur
         out.writeInt(-1);
       } else {
         out.writeInt(columnFamilies.size());
-        
+
         for (Entry<ByteSequence,MutableLong> entry : columnFamilies.entrySet()) {
           out.writeInt(entry.getKey().length());
           out.write(entry.getKey().getBackingArray(), entry.getKey().offset(), entry.getKey().length());
           out.writeLong(entry.getValue().longValue());
         }
       }
-      
+
       out.writeBoolean(firstKey != null);
       if (firstKey != null)
         firstKey.write(out);
-      
+
       indexWriter.close(out);
     }
-    
+
     public void printInfo() throws IOException {
       PrintStream out = System.out;
       out.println("Locality group         : " + (isDefaultLG ? "<DEFAULT>" : name));
@@ -258,55 +258,55 @@ public class RFile {
             + String.format("%,d bytes  %,d blocks", entry.getValue(), countsByLevel.get(entry.getKey())));
       }
       out.println("\tFirst key            : " + firstKey);
-      
+
       Key lastKey = null;
       if (indexReader.size() > 0) {
         lastKey = indexReader.getLastKey();
       }
-      
+
       out.println("\tLast key             : " + lastKey);
-      
+
       long numKeys = 0;
       IndexIterator countIter = indexReader.lookup(new Key());
       while (countIter.hasNext()) {
         numKeys += countIter.next().getNumEntries();
       }
-      
+
       out.println("\tNum entries          : " + String.format("%,d", numKeys));
       out.println("\tColumn families      : " + (isDefaultLG && columnFamilies == null ? "<UNKNOWN>" : columnFamilies.keySet()));
     }
-    
+
   }
-  
+
   public static class Writer implements FileSKVWriter {
-    
+
     public static final int MAX_CF_IN_DLG = 1000;
-    
+
     private BlockFileWriter fileWriter;
     private ABlockWriter blockWriter;
-    
+
     // private BlockAppender blockAppender;
     private long blockSize = 100000;
     private int indexBlockSize;
     private int entries = 0;
-    
+
     private ArrayList<LocalityGroupMetadata> localityGroups = new ArrayList<LocalityGroupMetadata>();
     private LocalityGroupMetadata currentLocalityGroup = null;
     private int nextBlock = 0;
-    
+
     private Key lastKeyInBlock = null;
-    
+
     private boolean dataClosed = false;
     private boolean closed = false;
     private Key prevKey = new Key();
     private boolean startedDefaultLocalityGroup = false;
-    
+
     private HashSet<ByteSequence> previousColumnFamilies;
-    
+
     public Writer(BlockFileWriter bfw, int blockSize) throws IOException {
       this(bfw, blockSize, (int) AccumuloConfiguration.getDefaultConfiguration().getMemoryInBytes(Property.TABLE_FILE_COMPRESSED_BLOCK_SIZE_INDEX));
     }
-    
+
     public Writer(BlockFileWriter bfw, int blockSize, int indexBlockSize) throws IOException {
       this.blockSize = blockSize;
       this.indexBlockSize = indexBlockSize;
@@ -314,123 +314,123 @@ public class RFile {
       this.blockWriter = null;
       previousColumnFamilies = new HashSet<ByteSequence>();
     }
-    
+
     @Override
     public synchronized void close() throws IOException {
-      
+
       if (closed) {
         return;
       }
-      
+
       closeData();
-      
+
       ABlockWriter mba = fileWriter.prepareMetaBlock("RFile.index");
-      
+
       mba.writeInt(RINDEX_MAGIC);
       mba.writeInt(RINDEX_VER_7);
-      
+
       if (currentLocalityGroup != null)
         localityGroups.add(currentLocalityGroup);
-      
+
       mba.writeInt(localityGroups.size());
-      
+
       for (LocalityGroupMetadata lc : localityGroups) {
         lc.write(mba);
       }
-      
+
       mba.close();
-      
+
       fileWriter.close();
-      
+
       closed = true;
     }
-    
+
     private void closeData() throws IOException {
-      
+
       if (dataClosed) {
         return;
       }
-      
+
       dataClosed = true;
-      
+
       if (blockWriter != null) {
         closeBlock(lastKeyInBlock, true);
       }
     }
-    
+
     @Override
     public void append(Key key, Value value) throws IOException {
-      
+
       if (dataClosed) {
         throw new IllegalStateException("Cannont append, data closed");
       }
-      
+
       if (key.compareTo(prevKey) < 0) {
         throw new IllegalStateException("Keys appended out-of-order.  New key " + key + ", previous key " + prevKey);
       }
-      
+
       currentLocalityGroup.updateColumnCount(key);
-      
+
       if (currentLocalityGroup.getFirstKey() == null) {
         currentLocalityGroup.setFirstKey(key);
       }
-      
+
       if (blockWriter == null) {
         blockWriter = fileWriter.prepareDataBlock();
       } else if (blockWriter.getRawSize() > blockSize) {
         closeBlock(prevKey, false);
         blockWriter = fileWriter.prepareDataBlock();
       }
-      
+
       RelativeKey rk = new RelativeKey(lastKeyInBlock, key);
-      
+
       rk.write(blockWriter);
       value.write(blockWriter);
       entries++;
-      
+
       prevKey = new Key(key);
       lastKeyInBlock = prevKey;
-      
+
     }
-    
+
     private void closeBlock(Key key, boolean lastBlock) throws IOException {
       blockWriter.close();
-      
+
       if (lastBlock)
         currentLocalityGroup.indexWriter.addLast(key, entries, blockWriter.getStartPos(), blockWriter.getCompressedSize(), blockWriter.getRawSize());
       else
         currentLocalityGroup.indexWriter.add(key, entries, blockWriter.getStartPos(), blockWriter.getCompressedSize(), blockWriter.getRawSize());
-      
+
       blockWriter = null;
       lastKeyInBlock = null;
       entries = 0;
       nextBlock++;
     }
-    
+
     @Override
     public DataOutputStream createMetaStore(String name) throws IOException {
       closeData();
-      
+
       return (DataOutputStream) fileWriter.prepareMetaBlock(name);
     }
-    
+
     private void _startNewLocalityGroup(String name, Set<ByteSequence> columnFamilies) throws IOException {
       if (dataClosed) {
         throw new IllegalStateException("data closed");
       }
-      
+
       if (startedDefaultLocalityGroup) {
         throw new IllegalStateException("Can not start anymore new locality groups after default locality group started");
       }
-      
+
       if (blockWriter != null) {
         closeBlock(lastKeyInBlock, true);
       }
-      
+
       if (currentLocalityGroup != null) {
         localityGroups.add(currentLocalityGroup);
       }
-      
+
       if (columnFamilies == null) {
         startedDefaultLocalityGroup = true;
         currentLocalityGroup = new LocalityGroupMetadata(nextBlock, previousColumnFamilies, indexBlockSize, fileWriter);
@@ -443,31 +443,31 @@ public class RFile {
         currentLocalityGroup = new LocalityGroupMetadata(name, columnFamilies, nextBlock, indexBlockSize, fileWriter);
         previousColumnFamilies.addAll(columnFamilies);
       }
-      
+
       prevKey = new Key();
     }
-    
+
     @Override
     public void startNewLocalityGroup(String name, Set<ByteSequence> columnFamilies) throws IOException {
       if (columnFamilies == null)
         throw new NullPointerException();
-      
+
       _startNewLocalityGroup(name, columnFamilies);
     }
-    
+
     @Override
     public void startDefaultLocalityGroup() throws IOException {
       _startNewLocalityGroup(null, null);
     }
-    
+
     @Override
     public boolean supportsLocalityGroups() {
       return true;
     }
   }
-  
+
   private static class LocalityGroupReader extends LocalityGroup implements FileSKVIterator {
-    
+
     private BlockFileReader reader;
     private MultiLevelIndex.Reader index;
     private int blockCount;
@@ -476,7 +476,7 @@ public class RFile {
     private boolean closed = false;
     private int version;
     private boolean checkRange = true;
-    
+
     private LocalityGroupReader(BlockFileReader reader, LocalityGroupMetadata lgm, int version) throws IOException {
       super(lgm.columnFamilies, lgm.isDefaultLG);
       this.firstKey = lgm.firstKey;
@@ -484,11 +484,11 @@ public class RFile {
       this.startBlock = lgm.startBlock;
       blockCount = index.size();
       this.version = version;
-      
+
       this.reader = reader;
-      
+
     }
-    
+
     public LocalityGroupReader(LocalityGroupReader lgr) {
       super(lgr.columnFamilies, lgr.isDefaultLocalityGroup);
       this.firstKey = lgr.firstKey;
@@ -498,20 +498,20 @@ public class RFile {
       this.reader = lgr.reader;
       this.version = lgr.version;
     }
-    
+
     Iterator<IndexEntry> getIndex() throws IOException {
       return index.lookup(new Key());
     }
-    
+
     @Override
     public void close() throws IOException {
       closed = true;
       hasTop = false;
       if (currBlock != null)
         currBlock.close();
-      
+
     }
-    
+
     private IndexIterator iiter;
     private int entriesLeft;
     private ABlockReader currBlock;
@@ -521,22 +521,22 @@ public class RFile {
     private Range range = null;
     private boolean hasTop = false;
     private AtomicBoolean interruptFlag;
-    
+
     @Override
     public Key getTopKey() {
       return rk.getKey();
     }
-    
+
     @Override
     public Value getTopValue() {
       return val;
     }
-    
+
     @Override
     public boolean hasTop() {
       return hasTop;
     }
-    
+
     @Override
     public void next() throws IOException {
       try {
@@ -546,20 +546,20 @@ public class RFile {
         throw ioe;
       }
     }
-    
+
     private void _next() throws IOException {
-      
+
       if (!hasTop)
         throw new IllegalStateException();
-      
+
       if (entriesLeft == 0) {
         currBlock.close();
-        
+
         if (iiter.hasNext()) {
           IndexEntry indexEntry = iiter.next();
           entriesLeft = indexEntry.getNumEntries();
           currBlock = getDataBlock(indexEntry);
-          
+
           checkRange = range.afterEndKey(indexEntry.getKey());
           if (!checkRange)
             hasTop = true;
@@ -571,7 +571,7 @@ public class RFile {
           return;
         }
       }
-      
+
       prevKey = rk.getKey();
       rk.readFields(currBlock);
       val.readFields(currBlock);
@@ -579,30 +579,30 @@ public class RFile {
       if (checkRange)
         hasTop = !range.afterEndKey(rk.getKey());
     }
-    
+
     private ABlockReader getDataBlock(IndexEntry indexEntry) throws IOException {
       if (interruptFlag != null && interruptFlag.get())
         throw new IterationInterruptedException();
-      
+
       if (version == RINDEX_VER_3 || version == RINDEX_VER_4)
         return reader.getDataBlock(startBlock + iiter.previousIndex());
       else
         return reader.getDataBlock(indexEntry.getOffset(), indexEntry.getCompressedSize(), indexEntry.getRawSize());
-      
+
     }
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-      
+
       if (closed)
         throw new IllegalStateException("Locality group reader closed");
-      
+
       if (columnFamilies.size() != 0 || inclusive)
         throw new IllegalArgumentException("I do not know how to filter column families");
-      
+
       if (interruptFlag != null && interruptFlag.get())
         throw new IterationInterruptedException();
-      
+
       try {
         _seek(range);
       } catch (IOException ioe) {
@@ -610,7 +610,7 @@ public class RFile {
         throw ioe;
       }
     }
-    
+
     private void reset() {
       rk = null;
       hasTop = false;
@@ -626,45 +626,45 @@ public class RFile {
         }
       }
     }
-    
+
     private void _seek(Range range) throws IOException {
-      
+
       this.range = range;
       this.checkRange = true;
-      
+
       if (blockCount == 0) {
         // its an empty file
         rk = null;
         return;
       }
-      
+
       Key startKey = range.getStartKey();
       if (startKey == null)
         startKey = new Key();
-      
+
       boolean reseek = true;
-      
+
       if (range.afterEndKey(firstKey)) {
         // range is before first key in rfile, so there is nothing to do
         reset();
         reseek = false;
       }
-      
+
       if (rk != null) {
         if (range.beforeStartKey(prevKey) && range.afterEndKey(getTopKey())) {
           // range is between the two keys in the file where the last range seeked to stopped, so there is
           // nothing to do
           reseek = false;
         }
-        
+
         if (startKey.compareTo(getTopKey()) <= 0 && startKey.compareTo(prevKey) > 0) {
           // current location in file can satisfy this request, no need to seek
           reseek = false;
         }
-        
+
         if (startKey.compareTo(getTopKey()) >= 0 && startKey.compareTo(iiter.peekPrevious().getKey()) <= 0) {
           // start key is within the unconsumed portion of the current block
-          
+
           // this code intentionally does not use the index associated with a cached block
           // because if only forward seeks are being done, then there is no benefit to building
           // and index for the block... could consider using the index if it exist but not
@@ -679,37 +679,37 @@ public class RFile {
             prevKey = skippr.prevKey;
             rk = skippr.rk;
           }
-          
+
           reseek = false;
         }
-        
+
         if (iiter.previousIndex() == 0 && getTopKey().equals(firstKey) && startKey.compareTo(firstKey) <= 0) {
           // seeking before the beginning of the file, and already positioned at the first key in the file
           // so there is nothing to do
           reseek = false;
         }
       }
-      
+
       if (reseek) {
         iiter = index.lookup(startKey);
-        
+
         reset();
-        
+
         if (!iiter.hasNext()) {
           // past the last key
         } else {
-          
+
           // if the index contains the same key multiple times, then go to the
           // earliest index entry containing the key
           while (iiter.hasPrevious() && iiter.peekPrevious().getKey().equals(iiter.peek().getKey())) {
             iiter.previous();
           }
-          
+
           if (iiter.hasPrevious())
             prevKey = new Key(iiter.peekPrevious().getKey()); // initially prevKey is the last key of the prev block
           else
             prevKey = new Key(); // first block in the file, so set prev key to minimal key
-            
+
           IndexEntry indexEntry = iiter.next();
           entriesLeft = indexEntry.getNumEntries();
           currBlock = getDataBlock(indexEntry);
@@ -736,7 +736,7 @@ public class RFile {
 
                 val.readFields(currBlock);
                 valbs = new MutableByteSequence(val.get(), 0, val.getSize());
-                
+
                 // just consumed one key from the input stream, so subtract one from entries left
                 entriesLeft = bie.getEntriesLeft() - 1;
                 prevKey = new Key(bie.getPrevKey());
@@ -754,76 +754,76 @@ public class RFile {
           rk = skippr.rk;
         }
       }
-      
+
       hasTop = rk != null && !range.afterEndKey(rk.getKey());
-      
+
       while (hasTop() && range.beforeStartKey(getTopKey())) {
         next();
       }
     }
-    
+
     @Override
     public Key getFirstKey() throws IOException {
       return firstKey;
     }
-    
+
     @Override
     public Key getLastKey() throws IOException {
       if (index.size() == 0)
         return null;
       return index.getLastKey();
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public void closeDeepCopies() throws IOException {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public DataInputStream getMetaStore(String name) throws IOException {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public void setInterruptFlag(AtomicBoolean flag) {
       this.interruptFlag = flag;
     }
-    
+
     @Override
     public InterruptibleIterator getIterator() {
       return this;
     }
   }
-  
+
   public static class Reader extends HeapIterator implements FileSKVIterator {
 
     private BlockFileReader reader;
-    
+
     private ArrayList<LocalityGroupMetadata> localityGroups = new ArrayList<LocalityGroupMetadata>();
-    
+
     private LocalityGroupReader lgReaders[];
     private HashSet<ByteSequence> nonDefaultColumnFamilies;
-    
+
     private List<Reader> deepCopies;
     private boolean deepCopy = false;
-    
+
     private AtomicBoolean interruptFlag;
-    
+
     public Reader(BlockFileReader rdr) throws IOException {
       this.reader = rdr;
-      
+
       ABlockReader mb = reader.getMetaBlock("RFile.index");
-      try{
+      try {
         int magic = mb.readInt();
         int ver = mb.readInt();
 
@@ -847,16 +847,16 @@ public class RFile {
       } finally {
         mb.close();
       }
-      
+
       nonDefaultColumnFamilies = new HashSet<ByteSequence>();
       for (LocalityGroupMetadata lgm : localityGroups) {
         if (!lgm.isDefaultLG)
           nonDefaultColumnFamilies.addAll(lgm.columnFamilies.keySet());
       }
-      
+
       createHeap(lgReaders.length);
     }
-    
+
     private Reader(Reader r) {
       super(r.lgReaders.length);
       this.reader = r.reader;
@@ -869,7 +869,7 @@ public class RFile {
         this.lgReaders[i].setInterruptFlag(r.interruptFlag);
       }
     }
-    
+
     private void closeLocalityGroupReaders() {
       for (LocalityGroupReader lgr : lgReaders) {
         try {
@@ -879,26 +879,26 @@ public class RFile {
         }
       }
     }
-    
+
     @Override
     public void closeDeepCopies() {
       if (deepCopy)
         throw new RuntimeException("Calling closeDeepCopies on a deep copy is not supported");
-      
+
       for (Reader deepCopy : deepCopies)
         deepCopy.closeLocalityGroupReaders();
-      
+
       deepCopies.clear();
     }
-    
+
     @Override
     public void close() throws IOException {
       if (deepCopy)
         throw new RuntimeException("Calling close on a deep copy is not supported");
-      
+
       closeDeepCopies();
       closeLocalityGroupReaders();
-      
+
       try {
         reader.close();
       } finally {
@@ -907,15 +907,15 @@ public class RFile {
          */
       }
     }
-    
+
     @Override
     public Key getFirstKey() throws IOException {
       if (lgReaders.length == 0) {
         return null;
       }
-      
+
       Key minKey = null;
-      
+
       for (int i = 0; i < lgReaders.length; i++) {
         if (minKey == null) {
           minKey = lgReaders[i].getFirstKey();
@@ -925,18 +925,18 @@ public class RFile {
             minKey = firstKey;
         }
       }
-      
+
       return minKey;
     }
-    
+
     @Override
     public Key getLastKey() throws IOException {
       if (lgReaders.length == 0) {
         return null;
       }
-      
+
       Key maxKey = null;
-      
+
       for (int i = 0; i < lgReaders.length; i++) {
         if (maxKey == null) {
           maxKey = lgReaders[i].getLastKey();
@@ -946,10 +946,10 @@ public class RFile {
             maxKey = lastKey;
         }
       }
-      
+
       return maxKey;
     }
-    
+
     @Override
     public DataInputStream getMetaStore(String name) throws IOException, NoSuchMetaStoreException {
       try {
@@ -958,7 +958,7 @@ public class RFile {
         throw new NoSuchMetaStoreException("name = " + name, e);
       }
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
       Reader copy = new Reader(this);
@@ -966,53 +966,53 @@ public class RFile {
       deepCopies.add(copy);
       return copy;
     }
-    
+
     @Override
     public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
       throw new UnsupportedOperationException();
-      
+
     }
-    
+
     private int numLGSeeked = 0;
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
       numLGSeeked = LocalityGroupIterator.seek(this, lgReaders, nonDefaultColumnFamilies, range, columnFamilies, inclusive);
     }
-    
+
     int getNumLocalityGroupsSeeked() {
       return numLGSeeked;
     }
-    
+
     public FileSKVIterator getIndex() throws IOException {
-      
+
       ArrayList<Iterator<IndexEntry>> indexes = new ArrayList<Iterator<IndexEntry>>();
-      
+
       for (LocalityGroupReader lgr : lgReaders) {
         indexes.add(lgr.getIndex());
       }
-      
+
       return new MultiIndexIterator(this, indexes);
     }
-    
+
     public void printInfo() throws IOException {
       for (LocalityGroupMetadata lgm : localityGroups) {
         lgm.printInfo();
       }
-      
+
     }
-    
+
     @Override
     public void setInterruptFlag(AtomicBoolean flag) {
       if (deepCopy)
         throw new RuntimeException("Calling setInterruptFlag on a deep copy is not supported");
-      
+
       if (deepCopies.size() != 0)
         throw new RuntimeException("Setting interrupt flag after calling deep copy not supported");
-      
+
       setInterruptFlagInternal(flag);
     }
-    
+
     private void setInterruptFlagInternal(AtomicBoolean flag) {
       this.interruptFlag = flag;
       for (LocalityGroupReader lgr : lgReaders) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/rfile/RFileOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/RFileOperations.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/RFileOperations.java
index 9fabe42..088abfe 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/RFileOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/RFileOperations.java
@@ -38,20 +38,20 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
 public class RFileOperations extends FileOperations {
-  
+
   private static final Collection<ByteSequence> EMPTY_CF_SET = Collections.emptySet();
-  
+
   @Override
   public long getFileSize(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return fs.getFileStatus(new Path(file)).getLen();
   }
-  
+
   @Override
   public FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
-    
+
     return openIndex(file, fs, conf, acuconf, null, null);
   }
-  
+
   @Override
   public FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf, BlockCache dataCache, BlockCache indexCache)
       throws IOException {
@@ -61,30 +61,30 @@ public class RFileOperations extends FileOperations {
     // Reader reader = new RFile.Reader(in, len , conf);
     CachableBlockFile.Reader _cbr = new CachableBlockFile.Reader(fs, path, conf, dataCache, indexCache, acuconf);
     final Reader reader = new RFile.Reader(_cbr);
-    
+
     return reader.getIndex();
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return openReader(file, seekToBeginning, fs, conf, acuconf, null, null);
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf,
       BlockCache dataCache, BlockCache indexCache) throws IOException {
     Path path = new Path(file);
-    
+
     CachableBlockFile.Reader _cbr = new CachableBlockFile.Reader(fs, path, conf, dataCache, indexCache, acuconf);
     Reader iter = new RFile.Reader(_cbr);
-    
+
     if (seekToBeginning) {
       iter.seek(new Range((Key) null, null), EMPTY_CF_SET, false);
     }
-    
+
     return iter;
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf) throws IOException {
@@ -92,7 +92,7 @@ public class RFileOperations extends FileOperations {
     iter.seek(range, columnFamilies, inclusive);
     return iter;
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf, BlockCache dataCache, BlockCache indexCache) throws IOException {
@@ -100,7 +100,7 @@ public class RFileOperations extends FileOperations {
     iter.seek(range, columnFamilies, inclusive);
     return iter;
   }
-  
+
   @Override
   public FileSKVWriter openWriter(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return openWriter(file, fs, conf, acuconf, acuconf.get(Property.TABLE_FILE_COMPRESSION_TYPE));
@@ -119,10 +119,10 @@ public class RFileOperations extends FileOperations {
     if (tblock > 0)
       block = tblock;
     int bufferSize = conf.getInt("io.file.buffer.size", 4096);
-    
+
     long blockSize = acuconf.getMemoryInBytes(Property.TABLE_FILE_COMPRESSED_BLOCK_SIZE);
     long indexBlockSize = acuconf.getMemoryInBytes(Property.TABLE_FILE_COMPRESSED_BLOCK_SIZE_INDEX);
-    
+
     CachableBlockFile.Writer _cbw = new CachableBlockFile.Writer(fs.create(new Path(file), false, bufferSize, (short) rep, block), compression, conf, acuconf);
     Writer writer = new RFile.Writer(_cbw, (int) blockSize, (int) indexBlockSize);
     return writer;


[05/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StartAll.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StartAll.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StartAll.java
index d164695..8504fd1 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StartAll.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StartAll.java
@@ -33,7 +33,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class StartAll extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     log.info("Starting all servers");
@@ -52,5 +52,5 @@ public class StartAll extends Test {
       }
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StopTabletServer.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StopTabletServer.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StopTabletServer.java
index c16160e..995a72e 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StopTabletServer.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/concurrent/StopTabletServer.java
@@ -38,7 +38,7 @@ import org.apache.zookeeper.KeeperException;
 import org.apache.zookeeper.data.Stat;
 
 public class StopTabletServer extends Test {
-  
+
   Set<TServerInstance> getTServers(Instance instance) throws KeeperException, InterruptedException {
     Set<TServerInstance> result = new HashSet<TServerInstance>();
     ZooReader rdr = new ZooReader(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
@@ -60,12 +60,12 @@ public class StopTabletServer extends Test {
     }
     return result;
   }
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     Instance instance = env.getInstance();
-    
+
     List<TServerInstance> currentServers = new ArrayList<TServerInstance>(getTServers(instance));
     Collections.shuffle(currentServers);
     Runtime runtime = Runtime.getRuntime();
@@ -80,5 +80,5 @@ public class StopTabletServer extends Test {
         throw new RuntimeException("Failed to stop " + victim);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Compact.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Compact.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Compact.java
index df9803f..1b0e03b 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Compact.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Compact.java
@@ -26,7 +26,7 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class Compact extends Test {
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Flush.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Flush.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Flush.java
index 37c8c91..cb5c5f1 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Flush.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Flush.java
@@ -26,7 +26,7 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class Flush extends Test {
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Init.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Init.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Init.java
index fe47813..ebe12ef 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Init.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Init.java
@@ -32,7 +32,7 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class Init extends Test {
 
@@ -68,21 +68,21 @@ public class Init extends Test {
 
         if (j % 1000 == 0 && j > 0) {
           Status status = cw.write(m).getStatus();
-          
-          while(status == Status.UNKNOWN)
+
+          while (status == Status.UNKNOWN)
             status = cw.write(m).getStatus();
-          
+
           if (status == Status.ACCEPTED)
             acceptedCount++;
           m = new ConditionalMutation(Utils.getBank(i));
         }
 
       }
-      if (m.getConditions().size() > 0){
+      if (m.getConditions().size() > 0) {
         Status status = cw.write(m).getStatus();
-        while(status == Status.UNKNOWN)
+        while (status == Status.UNKNOWN)
           status = cw.write(m).getStatus();
-        
+
         if (status == Status.ACCEPTED)
           acceptedCount++;
       }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Merge.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Merge.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Merge.java
index a6318d5..2124518 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Merge.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Merge.java
@@ -26,7 +26,7 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class Merge extends Test {
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Setup.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Setup.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Setup.java
index 90c2978..c22d5b8 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Setup.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Setup.java
@@ -33,7 +33,7 @@ public class Setup extends Test {
   public void visit(State state, Environment env, Properties props) throws Exception {
     Random rand = new Random();
     state.set("rand", rand);
-    
+
     int numBanks = Integer.parseInt(props.getProperty("numBanks", "1000"));
     log.debug("numBanks = " + numBanks);
     state.set("numBanks", numBanks);
@@ -53,9 +53,7 @@ public class Setup extends Test {
       log.debug("set " + Property.TABLE_BLOCKCACHE_ENABLED.getKey() + " " + blockCache);
     } catch (TableExistsException tee) {}
 
-
-    ConditionalWriter cw = env.getConnector()
-        .createConditionalWriter(tableName, new ConditionalWriterConfig().setMaxWriteThreads(1));
+    ConditionalWriter cw = env.getConnector().createConditionalWriter(tableName, new ConditionalWriterConfig().setMaxWriteThreads(1));
     state.set("cw", cw);
 
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Split.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Split.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Split.java
index c6665b9..a1ca830 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Split.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Split.java
@@ -28,7 +28,7 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class Split extends Test {
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/TearDown.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/TearDown.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/TearDown.java
index 3f326c0..0c59eae 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/TearDown.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/TearDown.java
@@ -24,7 +24,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 /**
- * 
+ *
  */
 public class TearDown extends Test {
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Transfer.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Transfer.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Transfer.java
index f9bfaaa..ec4edaf 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Transfer.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Transfer.java
@@ -38,7 +38,7 @@ import org.apache.commons.math.distribution.ZipfDistributionImpl;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class Transfer extends Test {
 
@@ -108,7 +108,6 @@ public class Transfer extends Test {
         throw new Exception("Unexpected column qual: " + cq);
     }
 
-
     int amt = rand.nextInt(50);
 
     log.debug("transfer req " + bank + " " + amt + " " + acct1 + " " + a1 + " " + acct2 + " " + a2);
@@ -123,14 +122,13 @@ public class Transfer extends Test {
 
       ConditionalWriter cw = (ConditionalWriter) state.get("cw");
       Status status = cw.write(cm).getStatus();
-      while(status == Status.UNKNOWN){
-        log.debug("retrying transfer "+status);
+      while (status == Status.UNKNOWN) {
+        log.debug("retrying transfer " + status);
         status = cw.write(cm).getStatus();
       }
       log.debug("transfer result " + bank + " " + status + " " + a1 + " " + a2);
     }
 
-
   }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Utils.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Utils.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Utils.java
index 043ea71..1c83686 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Utils.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Utils.java
@@ -16,9 +16,8 @@
  */
 package org.apache.accumulo.test.randomwalk.conditional;
 
-
 /**
- * 
+ *
  */
 public class Utils {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Verify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Verify.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Verify.java
index 8acc3ed..2690ffc 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Verify.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/conditional/Verify.java
@@ -34,7 +34,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 /**
- * 
+ *
  */
 public class Verify extends Test {
 
@@ -80,7 +80,6 @@ public class Verify extends Test {
       throw new Exception("Sum is off " + sum);
     }
 
-
     log.debug("Verified " + row + " count = " + count + " sum = " + sum + " min = " + min + " max = " + max);
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Commit.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Commit.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Commit.java
index 26b69cd..4468591 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Commit.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Commit.java
@@ -23,13 +23,13 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Commit extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     env.getMultiTableBatchWriter().flush();
-    
+
     log.debug("Committed " + state.getLong("numWrites") + " writes.  Total writes: " + state.getLong("totalWrites"));
     state.set("numWrites", Long.valueOf(0));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ImageFixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ImageFixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ImageFixture.java
index bcf7cad..3bcc41c 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ImageFixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ImageFixture.java
@@ -31,36 +31,36 @@ import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.impl.Tables;
-import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.Environment;
+import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.State;
 import org.apache.hadoop.io.Text;
 
 public class ImageFixture extends Fixture {
-  
+
   String imageTableName;
   String indexTableName;
-  
+
   @Override
   public void setUp(State state, Environment env) throws Exception {
-    
+
     Connector conn = env.getConnector();
     Instance instance = env.getInstance();
-    
+
     SortedSet<Text> splits = new TreeSet<Text>();
     for (int i = 1; i < 256; i++) {
       splits.add(new Text(String.format("%04x", i << 8)));
     }
-    
+
     String hostname = InetAddress.getLocalHost().getHostName().replaceAll("[-.]", "_");
     String pid = env.getPid();
-    
+
     imageTableName = String.format("img_%s_%s_%d", hostname, pid, System.currentTimeMillis());
     state.set("imageTableName", imageTableName);
-    
+
     indexTableName = String.format("img_ndx_%s_%s_%d", hostname, pid, System.currentTimeMillis());
     state.set("indexTableName", indexTableName);
-    
+
     try {
       conn.tableOperations().create(imageTableName);
       conn.tableOperations().addSplits(imageTableName, splits);
@@ -69,7 +69,7 @@ public class ImageFixture extends Fixture {
       log.error("Table " + imageTableName + " already exists.");
       throw e;
     }
-    
+
     try {
       conn.tableOperations().create(indexTableName);
       log.debug("Created table " + indexTableName + " (id:" + Tables.getNameToIdMap(instance).get(indexTableName) + ")");
@@ -77,35 +77,35 @@ public class ImageFixture extends Fixture {
       log.error("Table " + imageTableName + " already exists.");
       throw e;
     }
-    
+
     Random rand = new Random();
     if (rand.nextInt(10) < 5) {
       // setup locality groups
       Map<String,Set<Text>> groups = getLocalityGroups();
-      
+
       conn.tableOperations().setLocalityGroups(imageTableName, groups);
       log.debug("Configured locality groups for " + imageTableName + " groups = " + groups);
     }
-    
+
     state.set("numWrites", Long.valueOf(0));
     state.set("totalWrites", Long.valueOf(0));
     state.set("verified", Integer.valueOf(0));
     state.set("lastIndexRow", new Text(""));
   }
-  
+
   static Map<String,Set<Text>> getLocalityGroups() {
     Map<String,Set<Text>> groups = new HashMap<String,Set<Text>>();
-    
+
     HashSet<Text> lg1 = new HashSet<Text>();
     lg1.add(Write.CONTENT_COLUMN_FAMILY);
     groups.put("lg1", lg1);
-    
+
     HashSet<Text> lg2 = new HashSet<Text>();
     lg2.add(Write.META_COLUMN_FAMILY);
     groups.put("lg2", lg2);
     return groups;
   }
-  
+
   @Override
   public void tearDown(State state, Environment env) throws Exception {
     // We have resources we need to clean up
@@ -116,19 +116,19 @@ public class ImageFixture extends Fixture {
       } catch (MutationsRejectedException e) {
         log.error("Ignoring mutations that weren't flushed", e);
       }
-      
+
       // Reset the MTBW on the state to null
       env.resetMultiTableBatchWriter();
     }
-    
+
     // Now we can safely delete the tables
     log.debug("Dropping tables: " + imageTableName + " " + indexTableName);
-    
+
     Connector conn = env.getConnector();
-    
+
     conn.tableOperations().delete(imageTableName);
     conn.tableOperations().delete(indexTableName);
-    
+
     log.debug("Final total of writes: " + state.getLong("totalWrites"));
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ScanMeta.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ScanMeta.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ScanMeta.java
index 8a2ca3d..4b801c2 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ScanMeta.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/ScanMeta.java
@@ -16,13 +16,13 @@
  */
 package org.apache.accumulo.test.randomwalk.image;
 
+import java.util.ArrayList;
 import java.util.HashMap;
 import java.util.Iterator;
 import java.util.Map;
-import java.util.Random;
 import java.util.Map.Entry;
-import java.util.ArrayList;
 import java.util.Properties;
+import java.util.Random;
 import java.util.UUID;
 
 import org.apache.accumulo.core.client.BatchScanner;
@@ -38,74 +38,74 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class ScanMeta extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     // scan just the metadata of the images table to find N hashes... use the batch scanner to lookup those N hashes in the index table
     // this scan will test locality groups....
-    
+
     String indexTableName = state.getString("indexTableName");
     String imageTableName = state.getString("imageTableName");
-    
+
     String uuid = UUID.randomUUID().toString();
-    
+
     Connector conn = env.getConnector();
-    
+
     Scanner imageScanner = conn.createScanner(imageTableName, new Authorizations());
-    
+
     imageScanner.setRange(new Range(new Text(uuid), null));
     imageScanner.fetchColumn(Write.META_COLUMN_FAMILY, Write.SHA1_COLUMN_QUALIFIER);
-    
+
     int minScan = Integer.parseInt(props.getProperty("minScan"));
     int maxScan = Integer.parseInt(props.getProperty("maxScan"));
-    
+
     Random rand = new Random();
     int numToScan = rand.nextInt(maxScan - minScan) + minScan;
-    
+
     Map<Text,Text> hashes = new HashMap<Text,Text>();
-    
+
     Iterator<Entry<Key,Value>> iter = imageScanner.iterator();
-    
+
     while (iter.hasNext() && numToScan > 0) {
-      
+
       Entry<Key,Value> entry = iter.next();
-      
+
       hashes.put(new Text(entry.getValue().get()), entry.getKey().getRow());
-      
+
       numToScan--;
     }
-    
+
     log.debug("Found " + hashes.size() + " hashes starting at " + uuid);
 
     if (hashes.isEmpty()) {
       return;
     }
-    
+
     // use batch scanner to verify all of these exist in index
     BatchScanner indexScanner = conn.createBatchScanner(indexTableName, Authorizations.EMPTY, 3);
     ArrayList<Range> ranges = new ArrayList<Range>();
     for (Text row : hashes.keySet()) {
       ranges.add(new Range(row));
     }
-    
+
     indexScanner.setRanges(ranges);
-    
+
     Map<Text,Text> hashes2 = new HashMap<Text,Text>();
-    
+
     for (Entry<Key,Value> entry : indexScanner)
       hashes2.put(entry.getKey().getRow(), new Text(entry.getValue().get()));
-    
+
     log.debug("Looked up " + ranges.size() + " ranges, found " + hashes2.size());
-    
+
     if (!hashes.equals(hashes2)) {
       log.error("uuids from doc table : " + hashes.values());
       log.error("uuids from index     : " + hashes2.values());
       throw new Exception("Mismatch between document table and index " + indexTableName + " " + imageTableName);
     }
-    
+
     indexScanner.close();
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/image/TableOp.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/TableOp.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/TableOp.java
index bf3ff17..b62ec34 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/TableOp.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/TableOp.java
@@ -30,10 +30,10 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class TableOp extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     // choose a table
     Random rand = new Random();
     String tableName;
@@ -42,7 +42,7 @@ public class TableOp extends Test {
     } else {
       tableName = state.getString("indexTableName");
     }
-    
+
     // check if chosen table exists
     Connector conn = env.getConnector();
     TableOperations tableOps = conn.tableOperations();
@@ -50,7 +50,7 @@ public class TableOp extends Test {
       log.error("Table " + tableName + " does not exist!");
       return;
     }
-    
+
     // choose a random action
     int num = rand.nextInt(10);
     if (num > 6) {
@@ -63,10 +63,10 @@ public class TableOp extends Test {
       log.debug("Clearing locator cache for " + tableName);
       tableOps.clearLocatorCache(tableName);
     }
-    
+
     if (rand.nextInt(10) < 3) {
       Map<String,Set<Text>> groups = tableOps.getLocalityGroups(state.getString("imageTableName"));
-      
+
       if (groups.size() == 0) {
         log.debug("Adding locality groups to " + state.getString("imageTableName"));
         groups = ImageFixture.getLocalityGroups();
@@ -74,7 +74,7 @@ public class TableOp extends Test {
         log.debug("Removing locality groups from " + state.getString("imageTableName"));
         groups = new HashMap<String,Set<Text>>();
       }
-      
+
       tableOps.setLocalityGroups(state.getString("imageTableName"), groups);
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Verify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Verify.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Verify.java
index 6ca524a..cfac23f 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Verify.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Verify.java
@@ -37,91 +37,91 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Verify extends Test {
-  
+
   String indexTableName;
   String imageTableName;
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     Random rand = new Random();
-    
+
     int maxVerify = Integer.parseInt(props.getProperty("maxVerify"));
     int numVerifications = rand.nextInt(maxVerify - 1) + 1;
-    
+
     indexTableName = state.getString("indexTableName");
     imageTableName = state.getString("imageTableName");
-    
+
     Connector conn = env.getConnector();
-    
+
     Scanner indexScanner = conn.createScanner(indexTableName, new Authorizations());
     Scanner imageScanner = conn.createScanner(imageTableName, new Authorizations());
-    
+
     String uuid = UUID.randomUUID().toString();
-    
+
     MessageDigest alg = MessageDigest.getInstance("SHA-1");
     alg.update(uuid.getBytes(UTF_8));
-    
+
     indexScanner.setRange(new Range(new Text(alg.digest()), null));
     indexScanner.setBatchSize(numVerifications);
-    
+
     Text curRow = null;
     int count = 0;
     for (Entry<Key,Value> entry : indexScanner) {
-      
+
       curRow = entry.getKey().getRow();
       String rowToVerify = entry.getValue().toString();
-      
+
       verifyRow(imageScanner, rowToVerify);
-      
+
       count++;
       if (count == numVerifications) {
         break;
       }
     }
-    
+
     if (count != numVerifications && curRow != null) {
       Text lastRow = (Text) state.get("lastIndexRow");
       if (lastRow.compareTo(curRow) != 0) {
         log.error("Verified only " + count + " of " + numVerifications + " - curRow " + curRow + " lastKey " + lastRow);
       }
     }
-    
+
     int verified = ((Integer) state.get("verified")).intValue() + numVerifications;
     log.debug("Verified " + numVerifications + " - Total " + verified);
     state.set("verified", Integer.valueOf(verified));
   }
-  
+
   public void verifyRow(Scanner scanner, String row) throws Exception {
-    
+
     scanner.setRange(new Range(new Text(row)));
     scanner.clearColumns();
     scanner.fetchColumnFamily(Write.CONTENT_COLUMN_FAMILY);
     scanner.fetchColumn(Write.META_COLUMN_FAMILY, Write.SHA1_COLUMN_QUALIFIER);
-    
+
     Iterator<Entry<Key,Value>> scanIter = scanner.iterator();
-    
+
     if (scanIter.hasNext() == false) {
       log.error("Found row(" + row + ") in " + indexTableName + " but not " + imageTableName);
       return;
     }
-    
+
     // get image
     Entry<Key,Value> entry = scanIter.next();
     byte[] imageBytes = entry.getValue().get();
-    
+
     MessageDigest alg = MessageDigest.getInstance("SHA-1");
     alg.update(imageBytes);
     byte[] localHash = alg.digest();
-    
+
     // get stored hash
     entry = scanIter.next();
     byte[] storedHash = entry.getValue().get();
-    
+
     if (localHash.length != storedHash.length) {
       throw new Exception("Hash lens do not match for " + row);
     }
-    
+
     for (int i = 0; i < localHash.length; i++) {
       if (localHash[i] != storedHash[i]) {
         throw new Exception("Hashes do not match for " + row);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Write.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Write.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Write.java
index d239495..fc5d73e 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Write.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/image/Write.java
@@ -33,62 +33,62 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Write extends Test {
-  
+
   static final Text UUID_COLUMN_QUALIFIER = new Text("uuid");
   static final Text COUNT_COLUMN_QUALIFIER = new Text("count");
   static final Text SHA1_COLUMN_QUALIFIER = new Text("sha1");
   static final Text IMAGE_COLUMN_QUALIFIER = new Text("image");
   static final Text META_COLUMN_FAMILY = new Text("meta");
   static final Text CONTENT_COLUMN_FAMILY = new Text("content");
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     MultiTableBatchWriter mtbw = env.getMultiTableBatchWriter();
-    
+
     BatchWriter imagesBW = mtbw.getBatchWriter(state.getString("imageTableName"));
     BatchWriter indexBW = mtbw.getBatchWriter(state.getString("indexTableName"));
-    
+
     String uuid = UUID.randomUUID().toString();
     Mutation m = new Mutation(new Text(uuid));
-    
+
     // create a fake image between 4KB and 1MB
     int maxSize = Integer.parseInt(props.getProperty("maxSize"));
     int minSize = Integer.parseInt(props.getProperty("minSize"));
-    
+
     Random rand = new Random();
     int numBytes = rand.nextInt(maxSize - minSize) + minSize;
     byte[] imageBytes = new byte[numBytes];
     rand.nextBytes(imageBytes);
     m.put(CONTENT_COLUMN_FAMILY, IMAGE_COLUMN_QUALIFIER, new Value(imageBytes));
-    
+
     // store size
     m.put(META_COLUMN_FAMILY, new Text("size"), new Value(String.format("%d", numBytes).getBytes(UTF_8)));
-    
+
     // store hash
     MessageDigest alg = MessageDigest.getInstance("SHA-1");
     alg.update(imageBytes);
     byte[] hash = alg.digest();
     m.put(META_COLUMN_FAMILY, SHA1_COLUMN_QUALIFIER, new Value(hash));
-    
+
     // update write counts
     state.set("numWrites", state.getLong("numWrites") + 1);
     Long totalWrites = state.getLong("totalWrites") + 1;
     state.set("totalWrites", totalWrites);
-    
+
     // set count
     m.put(META_COLUMN_FAMILY, COUNT_COLUMN_QUALIFIER, new Value(String.format("%d", totalWrites).getBytes(UTF_8)));
-    
+
     // add mutation
     imagesBW.addMutation(m);
-    
+
     // now add mutation to index
     Text row = new Text(hash);
     m = new Mutation(row);
     m.put(META_COLUMN_FAMILY, UUID_COLUMN_QUALIFIER, new Value(uuid.getBytes(UTF_8)));
-    
+
     indexBW.addMutation(m);
-    
+
     Text lastRow = (Text) state.get("lastIndexRow");
     if (lastRow.compareTo(row) < 0) {
       state.set("lastIndexRow", new Text(row));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Commit.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Commit.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Commit.java
index 106de52..a81e396 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Commit.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Commit.java
@@ -23,18 +23,18 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Commit extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     env.getMultiTableBatchWriter().flush();
-    
+
     Long numWrites = state.getLong("numWrites");
     Long totalWrites = state.getLong("totalWrites") + numWrites;
-    
+
     log.debug("Committed " + numWrites + " writes.  Total writes: " + totalWrites);
-    
+
     state.set("totalWrites", totalWrites);
     state.set("numWrites", Long.valueOf(0));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/CopyTool.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/CopyTool.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/CopyTool.java
index d92dea2..d39f8df 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/CopyTool.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/CopyTool.java
@@ -35,7 +35,7 @@ import org.apache.log4j.Logger;
 
 public class CopyTool extends Configured implements Tool {
   protected final Logger log = Logger.getLogger(this.getClass());
-  
+
   public static class SeqMapClass extends Mapper<Key,Value,Text,Mutation> {
     @Override
     public void map(Key key, Value val, Context output) throws IOException, InterruptedException {
@@ -44,18 +44,18 @@ public class CopyTool extends Configured implements Tool {
       output.write(null, m);
     }
   }
-  
+
   @Override
   public int run(String[] args) throws Exception {
     @SuppressWarnings("deprecation")
     Job job = new Job(getConf(), this.getClass().getSimpleName());
     job.setJarByClass(this.getClass());
-    
+
     if (job.getJar() == null) {
       log.error("M/R requires a jar file!  Run mvn package.");
       return 1;
     }
-    
+
     ClientConfiguration clientConf = new ClientConfiguration().withInstance(args[3]).withZkHosts(args[4]);
 
     job.setInputFormatClass(AccumuloInputFormat.class);
@@ -63,19 +63,19 @@ public class CopyTool extends Configured implements Tool {
     AccumuloInputFormat.setInputTableName(job, args[2]);
     AccumuloInputFormat.setScanAuthorizations(job, Authorizations.EMPTY);
     AccumuloInputFormat.setZooKeeperInstance(job, clientConf);
-    
+
     job.setMapperClass(SeqMapClass.class);
     job.setMapOutputKeyClass(Text.class);
     job.setMapOutputValueClass(Mutation.class);
-    
+
     job.setNumReduceTasks(0);
-    
+
     job.setOutputFormatClass(AccumuloOutputFormat.class);
     AccumuloOutputFormat.setConnectorInfo(job, args[0], new PasswordToken(args[1]));
     AccumuloOutputFormat.setCreateTables(job, true);
     AccumuloOutputFormat.setDefaultTableName(job, args[5]);
     AccumuloOutputFormat.setZooKeeperInstance(job, clientConf);
-    
+
     job.waitForCompletion(true);
     return job.isSuccessful() ? 0 : 1;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/DropTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/DropTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/DropTable.java
index 9a670dc..2bbd873 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/DropTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/DropTable.java
@@ -26,21 +26,21 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class DropTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     @SuppressWarnings("unchecked")
     ArrayList<String> tables = (ArrayList<String>) state.get("tableList");
-    
+
     // don't drop a table if we only have one table or less
     if (tables.size() <= 1) {
       return;
     }
-    
+
     Random rand = new Random();
     String tableName = tables.remove(rand.nextInt(tables.size()));
-    
+
     try {
       env.getConnector().tableOperations().delete(tableName);
       log.debug("Dropped " + tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/MultiTableFixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/MultiTableFixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/MultiTableFixture.java
index 536de08..4d46295 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/MultiTableFixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/MultiTableFixture.java
@@ -23,24 +23,24 @@ import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableNotFoundException;
-import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.Environment;
+import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.State;
 
 public class MultiTableFixture extends Fixture {
-  
+
   @Override
   public void setUp(State state, Environment env) throws Exception {
-    
+
     String hostname = InetAddress.getLocalHost().getHostName().replaceAll("[-.]", "_");
-    
+
     state.set("tableNamePrefix", String.format("multi_%s_%s_%d", hostname, env.getPid(), System.currentTimeMillis()));
     state.set("nextId", Integer.valueOf(0));
     state.set("numWrites", Long.valueOf(0));
     state.set("totalWrites", Long.valueOf(0));
     state.set("tableList", new ArrayList<String>());
   }
-  
+
   @Override
   public void tearDown(State state, Environment env) throws Exception {
     // We have resources we need to clean up
@@ -51,16 +51,16 @@ public class MultiTableFixture extends Fixture {
       } catch (MutationsRejectedException e) {
         log.error("Ignoring mutations that weren't flushed", e);
       }
-      
+
       // Reset the MTBW on the state to null
       env.resetMultiTableBatchWriter();
     }
-    
+
     Connector conn = env.getConnector();
-    
+
     @SuppressWarnings("unchecked")
     ArrayList<String> tables = (ArrayList<String>) state.get("tableList");
-    
+
     for (String tableName : tables) {
       try {
         conn.tableOperations().delete(tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/OfflineTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/OfflineTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/OfflineTable.java
index bc54ada..90c555b 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/OfflineTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/OfflineTable.java
@@ -25,20 +25,20 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class OfflineTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     @SuppressWarnings("unchecked")
     ArrayList<String> tables = (ArrayList<String>) state.get("tableList");
-    
+
     if (tables.size() <= 0) {
       return;
     }
-    
+
     Random rand = new Random();
     String tableName = tables.get(rand.nextInt(tables.size()));
-    
+
     env.getConnector().tableOperations().offline(tableName, rand.nextBoolean());
     log.debug("Table " + tableName + " offline ");
     env.getConnector().tableOperations().online(tableName, rand.nextBoolean());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Write.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Write.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Write.java
index c3c91c0..494b794 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Write.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/multitable/Write.java
@@ -35,10 +35,10 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Write extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     @SuppressWarnings("unchecked")
     ArrayList<String> tables = (ArrayList<String>) state.get("tableList");
 
@@ -46,10 +46,10 @@ public class Write extends Test {
       log.debug("No tables to ingest into");
       return;
     }
-    
+
     Random rand = new Random();
     String tableName = tables.get(rand.nextInt(tables.size()));
-    
+
     BatchWriter bw = null;
     try {
       bw = env.getMultiTableBatchWriter().getBatchWriter(tableName);
@@ -60,30 +60,30 @@ public class Write extends Test {
       log.error("Table " + tableName + " not found!");
       return;
     }
-    
+
     Text meta = new Text("meta");
     String uuid = UUID.randomUUID().toString();
-    
+
     Mutation m = new Mutation(new Text(uuid));
-    
+
     // create a fake payload between 4KB and 16KB
     int numBytes = rand.nextInt(12000) + 4000;
     byte[] payloadBytes = new byte[numBytes];
     rand.nextBytes(payloadBytes);
     m.put(meta, new Text("payload"), new Value(payloadBytes));
-    
+
     // store size
     m.put(meta, new Text("size"), new Value(String.format("%d", numBytes).getBytes(UTF_8)));
-    
+
     // store hash
     MessageDigest alg = MessageDigest.getInstance("SHA-1");
     alg.update(payloadBytes);
     m.put(meta, new Text("sha1"), new Value(alg.digest()));
-    
+
     // add mutation
     bw.addMutation(m);
-    
+
     state.set("numWrites", state.getLong("numWrites") + 1);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterSystemPerm.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterSystemPerm.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterSystemPerm.java
index 663f2c2..ba4d2e9 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterSystemPerm.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterSystemPerm.java
@@ -28,17 +28,17 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class AlterSystemPerm extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     Connector conn = env.getConnector();
-    WalkingSecurity ws = new WalkingSecurity(state,env);
-    
+    WalkingSecurity ws = new WalkingSecurity(state, env);
+
     String action = props.getProperty("task", "toggle");
     String perm = props.getProperty("perm", "random");
-    
-    String targetUser = WalkingSecurity.get(state,env).getSysUserName();
-    
+
+    String targetUser = WalkingSecurity.get(state, env).getSysUserName();
+
     SystemPermission sysPerm;
     if (perm.equals("random")) {
       Random r = new Random();
@@ -46,9 +46,9 @@ public class AlterSystemPerm extends Test {
       sysPerm = SystemPermission.values()[i];
     } else
       sysPerm = SystemPermission.valueOf(perm);
-    
+
     boolean hasPerm = ws.hasSystemPermission(targetUser, sysPerm);
-    
+
     // toggle
     if (!"take".equals(action) && !"give".equals(action)) {
       if (hasPerm != conn.securityOperations().hasSystemPermission(targetUser, sysPerm))
@@ -58,7 +58,7 @@ public class AlterSystemPerm extends Test {
       else
         action = "give";
     }
-    
+
     if ("take".equals(action)) {
       try {
         conn.securityOperations().revokeSystemPermission(targetUser, sysPerm);
@@ -96,5 +96,5 @@ public class AlterSystemPerm extends Test {
       ws.grantSystemPermission(targetUser, sysPerm);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTable.java
index 4f9a8b9..70c98e4 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTable.java
@@ -30,22 +30,22 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class AlterTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state,env).getSysUserName(), WalkingSecurity.get(state,env).getSysToken());
-    
-    String tableName = WalkingSecurity.get(state,env).getTableName();
-    String namespaceName = WalkingSecurity.get(state,env).getNamespaceName();
-    
-    boolean exists = WalkingSecurity.get(state,env).getTableExists();
-    boolean hasPermission = WalkingSecurity.get(state,env).canAlterTable(WalkingSecurity.get(state,env).getSysCredentials(), tableName, namespaceName);
+    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env).getSysToken());
+
+    String tableName = WalkingSecurity.get(state, env).getTableName();
+    String namespaceName = WalkingSecurity.get(state, env).getNamespaceName();
+
+    boolean exists = WalkingSecurity.get(state, env).getTableExists();
+    boolean hasPermission = WalkingSecurity.get(state, env).canAlterTable(WalkingSecurity.get(state, env).getSysCredentials(), tableName, namespaceName);
     String newTableName = String.format("security_%s_%s_%d", InetAddress.getLocalHost().getHostName().replaceAll("[-.]", "_"), env.getPid(),
         System.currentTimeMillis());
-    
+
     renameTable(conn, state, env, tableName, newTableName, hasPermission, exists);
   }
-  
+
   public static void renameTable(Connector conn, State state, Environment env, String oldName, String newName, boolean hasPermission, boolean tableExists)
       throws AccumuloSecurityException, AccumuloException, TableExistsException {
     try {
@@ -57,7 +57,7 @@ public class AlterTable extends Test {
         else
           return;
       } else if (ae.getSecurityErrorCode().equals(SecurityErrorCode.BAD_CREDENTIALS)) {
-        if (WalkingSecurity.get(state,env).userPassTransient(conn.whoami()))
+        if (WalkingSecurity.get(state, env).userPassTransient(conn.whoami()))
           return;
       }
       throw new AccumuloException("Got unexpected ae error code", ae);
@@ -67,7 +67,7 @@ public class AlterTable extends Test {
       else
         return;
     }
-    WalkingSecurity.get(state,env).setTableName(newName);
+    WalkingSecurity.get(state, env).setTableName(newName);
     if (!hasPermission)
       throw new AccumuloException("Didn't get Security Exception when we should have");
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTablePerm.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTablePerm.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTablePerm.java
index 42ac364..8befe8a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTablePerm.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/AlterTablePerm.java
@@ -30,28 +30,28 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class AlterTablePerm extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     alter(state, env, props);
   }
-  
+
   public static void alter(State state, Environment env, Properties props) throws Exception {
     String action = props.getProperty("task", "toggle");
     String perm = props.getProperty("perm", "random");
     String sourceUserProp = props.getProperty("source", "system");
     String targetUser = props.getProperty("target", "table");
-    boolean tabExists = WalkingSecurity.get(state,env).getTableExists();
-    
+    boolean tabExists = WalkingSecurity.get(state, env).getTableExists();
+
     String target;
     if ("table".equals(targetUser))
-      target = WalkingSecurity.get(state,env).getTabUserName();
+      target = WalkingSecurity.get(state, env).getTabUserName();
     else
-      target = WalkingSecurity.get(state,env).getSysUserName();
-    
-    boolean exists = WalkingSecurity.get(state,env).userExists(target);
-    boolean tableExists = WalkingSecurity.get(state,env).getTableExists();
-    
+      target = WalkingSecurity.get(state, env).getSysUserName();
+
+    boolean exists = WalkingSecurity.get(state, env).userExists(target);
+    boolean tableExists = WalkingSecurity.get(state, env).getTableExists();
+
     TablePermission tabPerm;
     if (perm.equals("random")) {
       Random r = new Random();
@@ -59,26 +59,26 @@ public class AlterTablePerm extends Test {
       tabPerm = TablePermission.values()[i];
     } else
       tabPerm = TablePermission.valueOf(perm);
-    String tableName = WalkingSecurity.get(state,env).getTableName();
-    boolean hasPerm = WalkingSecurity.get(state,env).hasTablePermission(target, tableName, tabPerm);
+    String tableName = WalkingSecurity.get(state, env).getTableName();
+    boolean hasPerm = WalkingSecurity.get(state, env).hasTablePermission(target, tableName, tabPerm);
     boolean canGive;
     String sourceUser;
     AuthenticationToken sourceToken;
     if ("system".equals(sourceUserProp)) {
-      sourceUser = WalkingSecurity.get(state,env).getSysUserName();
-      sourceToken = WalkingSecurity.get(state,env).getSysToken();
+      sourceUser = WalkingSecurity.get(state, env).getSysUserName();
+      sourceToken = WalkingSecurity.get(state, env).getSysToken();
     } else if ("table".equals(sourceUserProp)) {
-      sourceUser = WalkingSecurity.get(state,env).getTabUserName();
-      sourceToken = WalkingSecurity.get(state,env).getTabToken();
+      sourceUser = WalkingSecurity.get(state, env).getTabUserName();
+      sourceToken = WalkingSecurity.get(state, env).getTabToken();
     } else {
       sourceUser = env.getUserName();
       sourceToken = env.getToken();
     }
     Connector conn = env.getInstance().getConnector(sourceUser, sourceToken);
-    
-    canGive = WalkingSecurity.get(state,env).canGrantTable(new Credentials(sourceUser, sourceToken).toThrift(env.getInstance()), target,
-        WalkingSecurity.get(state,env).getTableName(), WalkingSecurity.get(state,env).getNamespaceName());
-    
+
+    canGive = WalkingSecurity.get(state, env).canGrantTable(new Credentials(sourceUser, sourceToken).toThrift(env.getInstance()), target,
+        WalkingSecurity.get(state, env).getTableName(), WalkingSecurity.get(state, env).getNamespaceName());
+
     // toggle
     if (!"take".equals(action) && !"give".equals(action)) {
       try {
@@ -86,7 +86,7 @@ public class AlterTablePerm extends Test {
         if (hasPerm != (res = env.getConnector().securityOperations().hasTablePermission(target, tableName, tabPerm)))
           throw new AccumuloException("Test framework and accumulo are out of sync for user " + conn.whoami() + " for perm " + tabPerm.name()
               + " with local vs. accumulo being " + hasPerm + " " + res);
-        
+
         if (hasPerm)
           action = "take";
         else
@@ -108,8 +108,8 @@ public class AlterTablePerm extends Test {
         }
       }
     }
-    
-    boolean trans = WalkingSecurity.get(state,env).userPassTransient(conn.whoami());
+
+    boolean trans = WalkingSecurity.get(state, env).userPassTransient(conn.whoami());
     if ("take".equals(action)) {
       try {
         conn.securityOperations().revokeTablePermission(target, tableName, tabPerm);
@@ -137,7 +137,7 @@ public class AlterTablePerm extends Test {
             throw new AccumuloException("Got unexpected exception", ae);
         }
       }
-      WalkingSecurity.get(state,env).revokeTablePermission(target, tableName, tabPerm);
+      WalkingSecurity.get(state, env).revokeTablePermission(target, tableName, tabPerm);
     } else if ("give".equals(action)) {
       try {
         conn.securityOperations().grantTablePermission(target, tableName, tabPerm);
@@ -165,16 +165,16 @@ public class AlterTablePerm extends Test {
             throw new AccumuloException("Got unexpected exception", ae);
         }
       }
-      WalkingSecurity.get(state,env).grantTablePermission(target, tableName, tabPerm);
+      WalkingSecurity.get(state, env).grantTablePermission(target, tableName, tabPerm);
     }
-    
+
     if (!exists)
       throw new AccumuloException("User shouldn't have existed, but apparantly does");
     if (!tableExists)
       throw new AccumuloException("Table shouldn't have existed, but apparantly does");
     if (!canGive)
       throw new AccumuloException(conn.whoami() + " shouldn't have been able to grant privilege");
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Authenticate.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Authenticate.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Authenticate.java
index 1b4b15c..bceeaeb 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Authenticate.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Authenticate.java
@@ -30,36 +30,36 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Authenticate extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    authenticate(WalkingSecurity.get(state,env).getSysUserName(), WalkingSecurity.get(state,env).getSysToken(), state, env, props);
+    authenticate(WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env).getSysToken(), state, env, props);
   }
-  
+
   public static void authenticate(String principal, AuthenticationToken token, State state, Environment env, Properties props) throws Exception {
     String targetProp = props.getProperty("target");
     boolean success = Boolean.parseBoolean(props.getProperty("valid"));
-    
+
     Connector conn = env.getInstance().getConnector(principal, token);
-    
+
     String target;
-    
+
     if (targetProp.equals("table")) {
-      target = WalkingSecurity.get(state,env).getTabUserName();
+      target = WalkingSecurity.get(state, env).getTabUserName();
     } else {
-      target = WalkingSecurity.get(state,env).getSysUserName();
+      target = WalkingSecurity.get(state, env).getSysUserName();
     }
-    boolean exists = WalkingSecurity.get(state,env).userExists(target);
+    boolean exists = WalkingSecurity.get(state, env).userExists(target);
     // Copy so if failed it doesn't mess with the password stored in state
-    byte[] password = Arrays.copyOf(WalkingSecurity.get(state,env).getUserPassword(target), WalkingSecurity.get(state,env).getUserPassword(target).length);
-    boolean hasPermission = WalkingSecurity.get(state,env).canAskAboutUser(new Credentials(principal, token).toThrift(env.getInstance()), target);
-    
+    byte[] password = Arrays.copyOf(WalkingSecurity.get(state, env).getUserPassword(target), WalkingSecurity.get(state, env).getUserPassword(target).length);
+    boolean hasPermission = WalkingSecurity.get(state, env).canAskAboutUser(new Credentials(principal, token).toThrift(env.getInstance()), target);
+
     if (!success)
       for (int i = 0; i < password.length; i++)
         password[i]++;
-    
+
     boolean result;
-    
+
     try {
       result = conn.securityOperations().authenticateUser(target, new PasswordToken(password));
     } catch (AccumuloSecurityException ae) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/ChangePass.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/ChangePass.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/ChangePass.java
index 724ec98..28414c2 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/ChangePass.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/ChangePass.java
@@ -30,40 +30,40 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class ChangePass extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String target = props.getProperty("target");
     String source = props.getProperty("source");
-    
+
     String principal;
     AuthenticationToken token;
     if (source.equals("system")) {
-      principal = WalkingSecurity.get(state,env).getSysUserName();
-      token = WalkingSecurity.get(state,env).getSysToken();
+      principal = WalkingSecurity.get(state, env).getSysUserName();
+      token = WalkingSecurity.get(state, env).getSysToken();
     } else {
-      principal = WalkingSecurity.get(state,env).getTabUserName();
-      token = WalkingSecurity.get(state,env).getTabToken();
+      principal = WalkingSecurity.get(state, env).getTabUserName();
+      token = WalkingSecurity.get(state, env).getTabToken();
     }
     Connector conn = env.getInstance().getConnector(principal, token);
-    
+
     boolean hasPerm;
     boolean targetExists;
     if (target.equals("table")) {
-      target = WalkingSecurity.get(state,env).getTabUserName();
+      target = WalkingSecurity.get(state, env).getTabUserName();
     } else
-      target = WalkingSecurity.get(state,env).getSysUserName();
-    
-    targetExists = WalkingSecurity.get(state,env).userExists(target);
-    
-    hasPerm = WalkingSecurity.get(state,env).canChangePassword(new Credentials(principal, token).toThrift(env.getInstance()), target);
-    
+      target = WalkingSecurity.get(state, env).getSysUserName();
+
+    targetExists = WalkingSecurity.get(state, env).userExists(target);
+
+    hasPerm = WalkingSecurity.get(state, env).canChangePassword(new Credentials(principal, token).toThrift(env.getInstance()), target);
+
     Random r = new Random();
-    
+
     byte[] newPassw = new byte[r.nextInt(50) + 1];
     for (int i = 0; i < newPassw.length; i++)
       newPassw[i] = (byte) ((r.nextInt(26) + 65) & 0xFF);
-    
+
     PasswordToken newPass = new PasswordToken(newPassw);
     try {
       conn.securityOperations().changeLocalUserPassword(target, newPass);
@@ -78,14 +78,14 @@ public class ChangePass extends Test {
             throw new AccumuloException("User " + target + " doesn't exist and they SHOULD.", ae);
           return;
         case BAD_CREDENTIALS:
-          if (!WalkingSecurity.get(state,env).userPassTransient(conn.whoami()))
+          if (!WalkingSecurity.get(state, env).userPassTransient(conn.whoami()))
             throw new AccumuloException("Bad credentials for user " + conn.whoami());
           return;
         default:
           throw new AccumuloException("Got unexpected exception", ae);
       }
     }
-    WalkingSecurity.get(state,env).changePassword(target, newPass);
+    WalkingSecurity.get(state, env).changePassword(target, newPass);
     // Waiting 1 second for password to propogate through Zk
     Thread.sleep(1000);
     if (!hasPerm)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
index 0866923..fb4f309 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
@@ -29,16 +29,16 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class CreateTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state,env).getSysUserName(), WalkingSecurity.get(state,env).getSysToken());
-    
-    String tableName = WalkingSecurity.get(state,env).getTableName();
-    
-    boolean exists = WalkingSecurity.get(state,env).getTableExists();
-    boolean hasPermission = WalkingSecurity.get(state,env).canCreateTable(WalkingSecurity.get(state,env).getSysCredentials(), null, null);
-    
+    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env).getSysToken());
+
+    String tableName = WalkingSecurity.get(state, env).getTableName();
+
+    boolean exists = WalkingSecurity.get(state, env).getTableExists();
+    boolean hasPermission = WalkingSecurity.get(state, env).canCreateTable(WalkingSecurity.get(state, env).getSysCredentials(), null, null);
+
     try {
       conn.tableOperations().create(tableName);
     } catch (AccumuloSecurityException ae) {
@@ -50,7 +50,7 @@ public class CreateTable extends Test {
         {
           try {
             env.getConnector().tableOperations().create(tableName);
-            WalkingSecurity.get(state,env).initTable(tableName);
+            WalkingSecurity.get(state, env).initTable(tableName);
           } catch (TableExistsException tee) {
             if (exists)
               return;
@@ -67,9 +67,9 @@ public class CreateTable extends Test {
       else
         return;
     }
-    WalkingSecurity.get(state,env).initTable(tableName);
+    WalkingSecurity.get(state, env).initTable(tableName);
     for (TablePermission tp : TablePermission.values())
-      WalkingSecurity.get(state,env).grantTablePermission(conn.whoami(), tableName, tp);
+      WalkingSecurity.get(state, env).grantTablePermission(conn.whoami(), tableName, tp);
     if (!hasPermission)
       throw new AccumuloException("Didn't get Security Exception when we should have");
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
index 336b4e4..eb07c43 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
@@ -27,15 +27,15 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class CreateUser extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state,env).getSysUserName(), WalkingSecurity.get(state,env).getSysToken());
-    
-    String tableUserName = WalkingSecurity.get(state,env).getTabUserName();
-    
-    boolean exists = WalkingSecurity.get(state,env).userExists(tableUserName);
-    boolean hasPermission = WalkingSecurity.get(state,env).canCreateUser(WalkingSecurity.get(state,env).getSysCredentials(), tableUserName);
+    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env).getSysToken());
+
+    String tableUserName = WalkingSecurity.get(state, env).getTabUserName();
+
+    boolean exists = WalkingSecurity.get(state, env).userExists(tableUserName);
+    boolean hasPermission = WalkingSecurity.get(state, env).canCreateUser(WalkingSecurity.get(state, env).getSysCredentials(), tableUserName);
     PasswordToken tabUserPass = new PasswordToken("Super Sekret Table User Password");
     try {
       conn.securityOperations().createLocalUser(tableUserName, tabUserPass);
@@ -49,7 +49,7 @@ public class CreateUser extends Test {
           {
             if (!exists) {
               env.getConnector().securityOperations().createLocalUser(tableUserName, tabUserPass);
-              WalkingSecurity.get(state,env).createUser(tableUserName, tabUserPass);
+              WalkingSecurity.get(state, env).createUser(tableUserName, tabUserPass);
               Thread.sleep(1000);
             }
             return;
@@ -63,7 +63,7 @@ public class CreateUser extends Test {
           throw new AccumuloException("Got unexpected exception", ae);
       }
     }
-    WalkingSecurity.get(state,env).createUser(tableUserName, tabUserPass);
+    WalkingSecurity.get(state, env).createUser(tableUserName, tabUserPass);
     Thread.sleep(1000);
     if (!hasPermission)
       throw new AccumuloException("Didn't get Security Exception when we should have");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropTable.java
index 06aa037..a0bff7a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropTable.java
@@ -31,31 +31,32 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class DropTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     dropTable(state, env, props);
   }
-  
+
   public static void dropTable(State state, Environment env, Properties props) throws Exception {
     String sourceUser = props.getProperty("source", "system");
     String principal;
     AuthenticationToken token;
     if (sourceUser.equals("table")) {
-      principal = WalkingSecurity.get(state,env).getTabUserName();
-      token = WalkingSecurity.get(state,env).getTabToken();
+      principal = WalkingSecurity.get(state, env).getTabUserName();
+      token = WalkingSecurity.get(state, env).getTabToken();
     } else {
-      principal = WalkingSecurity.get(state,env).getSysUserName();
-      token = WalkingSecurity.get(state,env).getSysToken();
+      principal = WalkingSecurity.get(state, env).getSysUserName();
+      token = WalkingSecurity.get(state, env).getSysToken();
     }
     Connector conn = env.getInstance().getConnector(principal, token);
-    
-    String tableName = WalkingSecurity.get(state,env).getTableName();
-    String namespaceName = WalkingSecurity.get(state,env).getNamespaceName();
-    
-    boolean exists = WalkingSecurity.get(state,env).getTableExists();
-    boolean hasPermission = WalkingSecurity.get(state,env).canDeleteTable(new Credentials(principal, token).toThrift(env.getInstance()), tableName, namespaceName);
-    
+
+    String tableName = WalkingSecurity.get(state, env).getTableName();
+    String namespaceName = WalkingSecurity.get(state, env).getNamespaceName();
+
+    boolean exists = WalkingSecurity.get(state, env).getTableExists();
+    boolean hasPermission = WalkingSecurity.get(state, env).canDeleteTable(new Credentials(principal, token).toThrift(env.getInstance()), tableName,
+        namespaceName);
+
     try {
       conn.tableOperations().delete(tableName);
     } catch (AccumuloSecurityException ae) {
@@ -65,11 +66,11 @@ public class DropTable extends Test {
         else {
           // Drop anyway for sake of state
           env.getConnector().tableOperations().delete(tableName);
-          WalkingSecurity.get(state,env).cleanTablePermissions(tableName);
+          WalkingSecurity.get(state, env).cleanTablePermissions(tableName);
           return;
         }
       } else if (ae.getSecurityErrorCode().equals(SecurityErrorCode.BAD_CREDENTIALS)) {
-        if (WalkingSecurity.get(state,env).userPassTransient(conn.whoami()))
+        if (WalkingSecurity.get(state, env).userPassTransient(conn.whoami()))
           return;
       }
       throw new AccumuloException("Got unexpected ae error code", ae);
@@ -79,7 +80,7 @@ public class DropTable extends Test {
       else
         return;
     }
-    WalkingSecurity.get(state,env).cleanTablePermissions(tableName);
+    WalkingSecurity.get(state, env).cleanTablePermissions(tableName);
     if (!hasPermission)
       throw new AccumuloException("Didn't get Security Exception when we should have");
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropUser.java
index 40bde3b..e0ca60f 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/DropUser.java
@@ -26,16 +26,16 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class DropUser extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state,env).getSysUserName(), WalkingSecurity.get(state,env).getSysToken());
-    
-    String tableUserName = WalkingSecurity.get(state,env).getTabUserName();
-    
-    boolean exists = WalkingSecurity.get(state,env).userExists(tableUserName);
-    boolean hasPermission = WalkingSecurity.get(state,env).canDropUser(WalkingSecurity.get(state,env).getSysCredentials(), tableUserName);
-    
+    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env).getSysToken());
+
+    String tableUserName = WalkingSecurity.get(state, env).getTabUserName();
+
+    boolean exists = WalkingSecurity.get(state, env).userExists(tableUserName);
+    boolean hasPermission = WalkingSecurity.get(state, env).canDropUser(WalkingSecurity.get(state, env).getSysCredentials(), tableUserName);
+
     try {
       conn.securityOperations().dropLocalUser(tableUserName);
     } catch (AccumuloSecurityException ae) {
@@ -46,11 +46,11 @@ public class DropUser extends Test {
           else {
             if (exists) {
               env.getConnector().securityOperations().dropLocalUser(tableUserName);
-              WalkingSecurity.get(state,env).dropUser(tableUserName);
+              WalkingSecurity.get(state, env).dropUser(tableUserName);
             }
             return;
           }
-          
+
         case USER_DOESNT_EXIST:
           if (exists)
             throw new AccumuloException("Got user DNE exception when user should exists.", ae);
@@ -60,7 +60,7 @@ public class DropUser extends Test {
           throw new AccumuloException("Got unexpected exception", ae);
       }
     }
-    WalkingSecurity.get(state,env).dropUser(tableUserName);
+    WalkingSecurity.get(state, env).dropUser(tableUserName);
     Thread.sleep(1000);
     if (!hasPermission)
       throw new AccumuloException("Didn't get Security Exception when we should have");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityFixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityFixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityFixture.java
index 9ce5bfb..915eca0 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityFixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityFixture.java
@@ -24,24 +24,24 @@ import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
-import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.Environment;
+import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.State;
 
 public class SecurityFixture extends Fixture {
-  
+
   @Override
   public void setUp(State state, Environment env) throws Exception {
     String secTableName, systemUserName, tableUserName, secNamespaceName;
     Connector conn = env.getConnector();
-    
+
     String hostname = InetAddress.getLocalHost().getHostName().replaceAll("[-.]", "_");
-    
+
     systemUserName = String.format("system_%s", hostname);
     tableUserName = String.format("table_%s", hostname);
     secTableName = String.format("security_%s", hostname);
     secNamespaceName = String.format("securityNs_%s", hostname);
-    
+
     if (conn.tableOperations().exists(secTableName))
       conn.tableOperations().delete(secTableName);
     Set<String> users = conn.securityOperations().listLocalUsers();
@@ -49,63 +49,63 @@ public class SecurityFixture extends Fixture {
       conn.securityOperations().dropLocalUser(tableUserName);
     if (users.contains(systemUserName))
       conn.securityOperations().dropLocalUser(systemUserName);
-    
+
     PasswordToken sysUserPass = new PasswordToken("sysUser");
     conn.securityOperations().createLocalUser(systemUserName, sysUserPass);
-    
-    WalkingSecurity.get(state,env).setTableName(secTableName);
-    WalkingSecurity.get(state,env).setNamespaceName(secNamespaceName);
+
+    WalkingSecurity.get(state, env).setTableName(secTableName);
+    WalkingSecurity.get(state, env).setNamespaceName(secNamespaceName);
     state.set("rootUserPass", env.getToken());
-    
-    WalkingSecurity.get(state,env).setSysUserName(systemUserName);
-    WalkingSecurity.get(state,env).createUser(systemUserName, sysUserPass);
-    
-    WalkingSecurity.get(state,env).changePassword(tableUserName, new PasswordToken(new byte[0]));
-    
-    WalkingSecurity.get(state,env).setTabUserName(tableUserName);
-    
+
+    WalkingSecurity.get(state, env).setSysUserName(systemUserName);
+    WalkingSecurity.get(state, env).createUser(systemUserName, sysUserPass);
+
+    WalkingSecurity.get(state, env).changePassword(tableUserName, new PasswordToken(new byte[0]));
+
+    WalkingSecurity.get(state, env).setTabUserName(tableUserName);
+
     for (TablePermission tp : TablePermission.values()) {
-      WalkingSecurity.get(state,env).revokeTablePermission(systemUserName, secTableName, tp);
-      WalkingSecurity.get(state,env).revokeTablePermission(tableUserName, secTableName, tp);
+      WalkingSecurity.get(state, env).revokeTablePermission(systemUserName, secTableName, tp);
+      WalkingSecurity.get(state, env).revokeTablePermission(tableUserName, secTableName, tp);
     }
     for (SystemPermission sp : SystemPermission.values()) {
-      WalkingSecurity.get(state,env).revokeSystemPermission(systemUserName, sp);
-      WalkingSecurity.get(state,env).revokeSystemPermission(tableUserName, sp);
+      WalkingSecurity.get(state, env).revokeSystemPermission(systemUserName, sp);
+      WalkingSecurity.get(state, env).revokeSystemPermission(tableUserName, sp);
     }
-    WalkingSecurity.get(state,env).changeAuthorizations(tableUserName, new Authorizations());
+    WalkingSecurity.get(state, env).changeAuthorizations(tableUserName, new Authorizations());
   }
-  
+
   @Override
   public void tearDown(State state, Environment env) throws Exception {
     log.debug("One last validate");
     Validate.validate(state, env, log);
     Connector conn = env.getConnector();
-    
-    if (WalkingSecurity.get(state,env).getTableExists()) {
-      String secTableName = WalkingSecurity.get(state,env).getTableName();
+
+    if (WalkingSecurity.get(state, env).getTableExists()) {
+      String secTableName = WalkingSecurity.get(state, env).getTableName();
       log.debug("Dropping tables: " + secTableName);
-      
+
       conn.tableOperations().delete(secTableName);
     }
-    
-    if (WalkingSecurity.get(state,env).getNamespaceExists()) {
-      String secNamespaceName = WalkingSecurity.get(state,env).getNamespaceName();
+
+    if (WalkingSecurity.get(state, env).getNamespaceExists()) {
+      String secNamespaceName = WalkingSecurity.get(state, env).getNamespaceName();
       log.debug("Dropping namespace: " + secNamespaceName);
-      
+
       conn.namespaceOperations().delete(secNamespaceName);
     }
-    
-    if (WalkingSecurity.get(state,env).userExists(WalkingSecurity.get(state,env).getTabUserName())) {
-      String tableUserName = WalkingSecurity.get(state,env).getTabUserName();
+
+    if (WalkingSecurity.get(state, env).userExists(WalkingSecurity.get(state, env).getTabUserName())) {
+      String tableUserName = WalkingSecurity.get(state, env).getTabUserName();
       log.debug("Dropping user: " + tableUserName);
-      
+
       conn.securityOperations().dropLocalUser(tableUserName);
     }
-    String systemUserName = WalkingSecurity.get(state,env).getSysUserName();
+    String systemUserName = WalkingSecurity.get(state, env).getSysUserName();
     log.debug("Dropping user: " + systemUserName);
     conn.securityOperations().dropLocalUser(systemUserName);
     WalkingSecurity.clearInstance();
-    
+
     // Allow user drops to propagate, in case a new security test starts
     Thread.sleep(2000);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityHelper.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityHelper.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityHelper.java
index ba1893f..f106dd5 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityHelper.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SecurityHelper.java
@@ -32,150 +32,150 @@ import org.apache.log4j.Logger;
 
 public class SecurityHelper {
   private static final Logger log = Logger.getLogger(SecurityHelper.class);
-  
+
   private static final String tableName = "secTableName";
   private static final String masterName = "sysUserName";
   private static final String tUserName = "tabUserName";
-  
+
   private static final String masterPass = "sysUserPass";
   private static final String tUserPass = "tabUserPass";
-  
+
   private static final String tUserExists = "tabUserExists";
   private static final String tableExists = "secTableExists";
-  
+
   private static final String masterConn = "sysUserConn";
-  
+
   private static final String authsMap = "authorizationsCountMap";
   private static final String lastKey = "lastMutationKey";
   private static final String filesystem = "securityFileSystem";
-  
+
   public static String getTableName(State state) {
     return state.getString(tableName);
   }
-  
+
   public static void setTableName(State state, String tName) {
     state.set(tableName, tName);
   }
-  
+
   public static String getSysUserName(State state) {
     return state.getString(masterName);
   }
-  
+
   public static void setSysUserName(State state, String sysUserName) {
     state.set(masterName, sysUserName);
   }
-  
+
   public static String getTabUserName(State state) {
     return state.getString(tUserName);
   }
-  
+
   public static void setTabUserName(State state, String tabUserName) {
     state.set(tUserName, tabUserName);
   }
-  
+
   public static byte[] getSysUserPass(State state) {
     return (byte[]) state.get(masterPass);
   }
-  
+
   public static void setSysUserPass(State state, byte[] sysUserPass) {
     log.debug("Setting system user pass to " + new String(sysUserPass, UTF_8));
     state.set(masterPass, sysUserPass);
     state.set(masterPass + "time", System.currentTimeMillis());
-    
+
   }
-  
+
   public static boolean sysUserPassTransient(State state) {
     return System.currentTimeMillis() - state.getLong(masterPass + "time") < 1000;
   }
-  
+
   public static byte[] getTabUserPass(State state) {
     return (byte[]) state.get(tUserPass);
   }
-  
+
   public static void setTabUserPass(State state, byte[] tabUserPass) {
     log.debug("Setting table user pass to " + new String(tabUserPass, UTF_8));
     state.set(tUserPass, tabUserPass);
     state.set(tUserPass + "time", System.currentTimeMillis());
   }
-  
+
   public static boolean tabUserPassTransient(State state) {
     return System.currentTimeMillis() - state.getLong(tUserPass + "time") < 1000;
   }
-  
+
   public static boolean getTabUserExists(State state) {
     return Boolean.parseBoolean(state.getString(tUserExists));
   }
-  
+
   public static void setTabUserExists(State state, boolean exists) {
     state.set(tUserExists, Boolean.toString(exists));
   }
-  
+
   public static boolean getTableExists(State state) {
     return Boolean.parseBoolean(state.getString(tableExists));
   }
-  
+
   public static void setTableExists(State state, boolean exists) {
     state.set(tableExists, Boolean.toString(exists));
   }
-  
+
   public static Connector getSystemConnector(State state) {
     return (Connector) state.get(masterConn);
   }
-  
+
   public static void setSystemConnector(State state, Connector conn) {
     state.set(masterConn, conn);
   }
-  
+
   public static boolean getTabPerm(State state, String userName, TablePermission tp) {
     return Boolean.parseBoolean(state.getString("Tab" + userName + tp.name()));
   }
-  
+
   public static void setTabPerm(State state, String userName, TablePermission tp, boolean value) {
     log.debug((value ? "Gave" : "Took") + " the table permission " + tp.name() + (value ? " to" : " from") + " user " + userName);
     state.set("Tab" + userName + tp.name(), Boolean.toString(value));
     if (tp.equals(TablePermission.READ) || tp.equals(TablePermission.WRITE))
       state.set("Tab" + userName + tp.name() + "time", System.currentTimeMillis());
-    
+
   }
-  
+
   public static boolean getSysPerm(State state, String userName, SystemPermission tp) {
     return Boolean.parseBoolean(state.getString("Sys" + userName + tp.name()));
   }
-  
+
   public static void setSysPerm(State state, String userName, SystemPermission tp, boolean value) {
     log.debug((value ? "Gave" : "Took") + " the system permission " + tp.name() + (value ? " to" : " from") + " user " + userName);
     state.set("Sys" + userName + tp.name(), Boolean.toString(value));
   }
-  
+
   public static Authorizations getUserAuths(State state, String target) {
     return (Authorizations) state.get(target + "_auths");
   }
-  
+
   public static void setUserAuths(State state, String target, Authorizations auths) {
     state.set(target + "_auths", auths);
   }
-  
+
   @SuppressWarnings("unchecked")
   public static Map<String,Integer> getAuthsMap(State state) {
     return (Map<String,Integer>) state.get(authsMap);
   }
-  
+
   public static void setAuthsMap(State state, Map<String,Integer> map) {
     state.set(authsMap, map);
   }
-  
+
   public static String[] getAuthsArray() {
     return new String[] {"Fishsticks", "PotatoSkins", "Ribs", "Asparagus", "Paper", "Towels", "Lint", "Brush", "Celery"};
   }
-  
+
   public static String getLastKey(State state) {
     return state.getString(lastKey);
   }
-  
+
   public static void setLastKey(State state, String key) {
     state.set(lastKey, key);
   }
-  
+
   public static void increaseAuthMap(State state, String s, int increment) {
     Integer curVal = getAuthsMap(state).get(s);
     if (curVal == null) {
@@ -184,7 +184,7 @@ public class SecurityHelper {
     }
     curVal += increment;
   }
-  
+
   public static FileSystem getFs(State state) {
     FileSystem fs = null;
     try {
@@ -201,7 +201,7 @@ public class SecurityHelper {
     }
     return fs;
   }
-  
+
   public static boolean inAmbiguousZone(State state, String userName, TablePermission tp) {
     if (tp.equals(TablePermission.READ) || tp.equals(TablePermission.WRITE)) {
       Long setTime = (Long) state.get("Tab" + userName + tp.name() + "time");
@@ -210,5 +210,5 @@ public class SecurityHelper {
     }
     return false;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SetAuths.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SetAuths.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SetAuths.java
index e20367d..e80e475 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SetAuths.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/SetAuths.java
@@ -30,33 +30,34 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class SetAuths extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String authsString = props.getProperty("auths", "_random");
-    
+
     String targetUser = props.getProperty("system");
     String target;
     String authPrincipal;
     AuthenticationToken authToken;
     if ("table".equals(targetUser)) {
-      target = WalkingSecurity.get(state,env).getTabUserName();
-      authPrincipal = WalkingSecurity.get(state,env).getSysUserName();
-      authToken = WalkingSecurity.get(state,env).getSysToken();
+      target = WalkingSecurity.get(state, env).getTabUserName();
+      authPrincipal = WalkingSecurity.get(state, env).getSysUserName();
+      authToken = WalkingSecurity.get(state, env).getSysToken();
     } else {
-      target = WalkingSecurity.get(state,env).getSysUserName();
+      target = WalkingSecurity.get(state, env).getSysUserName();
       authPrincipal = env.getUserName();
       authToken = env.getToken();
     }
     Connector conn = env.getInstance().getConnector(authPrincipal, authToken);
-    
-    boolean exists = WalkingSecurity.get(state,env).userExists(target);
-    boolean hasPermission = WalkingSecurity.get(state,env).canChangeAuthorizations(new Credentials(authPrincipal, authToken).toThrift(env.getInstance()), target);
-    
+
+    boolean exists = WalkingSecurity.get(state, env).userExists(target);
+    boolean hasPermission = WalkingSecurity.get(state, env).canChangeAuthorizations(new Credentials(authPrincipal, authToken).toThrift(env.getInstance()),
+        target);
+
     Authorizations auths;
     if (authsString.equals("_random")) {
-      String[] possibleAuths = WalkingSecurity.get(state,env).getAuthsArray();
-      
+      String[] possibleAuths = WalkingSecurity.get(state, env).getAuthsArray();
+
       Random r = new Random();
       int i = r.nextInt(possibleAuths.length);
       String[] authSet = new String[i];
@@ -72,7 +73,7 @@ public class SetAuths extends Test {
     } else {
       auths = new Authorizations(authsString.split(","));
     }
-    
+
     try {
       conn.securityOperations().changeUserAuthorizations(target, auths);
     } catch (AccumuloSecurityException ae) {
@@ -91,9 +92,9 @@ public class SetAuths extends Test {
           throw new AccumuloException("Got unexpected exception", ae);
       }
     }
-    WalkingSecurity.get(state,env).changeAuthorizations(target, auths);
+    WalkingSecurity.get(state, env).changeAuthorizations(target, auths);
     if (!hasPermission)
       throw new AccumuloException("Didn't get Security Exception when we should have");
   }
-  
+
 }


[27/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/SourceSwitchingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/SourceSwitchingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/SourceSwitchingIteratorTest.java
index 23f08a8..7567871 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/SourceSwitchingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/SourceSwitchingIteratorTest.java
@@ -35,19 +35,19 @@ import org.apache.accumulo.core.iterators.system.SourceSwitchingIterator.DataSou
 import org.apache.hadoop.io.Text;
 
 public class SourceSwitchingIteratorTest extends TestCase {
-  
+
   Key nk(String row, String cf, String cq, long time) {
     return new Key(new Text(row), new Text(cf), new Text(cq), time);
   }
-  
+
   void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, Value val) {
     tm.put(nk(row, cf, cq, time), val);
   }
-  
+
   void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, String val) {
     put(tm, row, cf, cq, time, new Value(val.getBytes()));
   }
-  
+
   private void ane(SortedKeyValueIterator<Key,Value> rdi, String row, String cf, String cq, long time, String val, boolean callNext) throws Exception {
     assertTrue(rdi.hasTop());
     assertEquals(nk(row, cf, cq, time), rdi.getTopKey());
@@ -55,49 +55,49 @@ public class SourceSwitchingIteratorTest extends TestCase {
     if (callNext)
       rdi.next();
   }
-  
+
   class TestDataSource implements DataSource {
-    
+
     DataSource next;
     SortedKeyValueIterator<Key,Value> iter;
     List<TestDataSource> copies = new ArrayList<TestDataSource>();
     AtomicBoolean iflag;
-    
+
     TestDataSource(SortedKeyValueIterator<Key,Value> iter) {
       this(iter, new ArrayList<TestDataSource>());
     }
-    
+
     public TestDataSource(SortedKeyValueIterator<Key,Value> iter, List<TestDataSource> copies) {
       this.iter = iter;
       this.copies = copies;
       copies.add(this);
     }
-    
+
     @Override
     public DataSource getNewDataSource() {
       return next;
     }
-    
+
     @Override
     public boolean isCurrent() {
       return next == null;
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> iterator() {
       if (iflag != null)
         ((InterruptibleIterator) iter).setInterruptFlag(iflag);
       return iter;
     }
-    
+
     @Override
     public DataSource getDeepCopyDataSource(IteratorEnvironment env) {
       return new TestDataSource(iter.deepCopy(env), copies);
     }
-    
+
     void setNext(TestDataSource next) {
       this.next = next;
-      
+
       for (TestDataSource tds : copies) {
         if (tds != this)
           tds.next = new TestDataSource(next.iter.deepCopy(null), next.copies);
@@ -109,54 +109,54 @@ public class SourceSwitchingIteratorTest extends TestCase {
       this.iflag = flag;
     }
   }
-  
+
   public void test1() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq3", 5, "v2");
     put(tm1, "r2", "cf1", "cq1", 5, "v3");
-    
+
     SortedMapIterator smi = new SortedMapIterator(tm1);
     TestDataSource tds = new TestDataSource(smi);
     SourceSwitchingIterator ssi = new SourceSwitchingIterator(tds);
-    
+
     ssi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(ssi, "r1", "cf1", "cq1", 5, "v1", true);
     ane(ssi, "r1", "cf1", "cq3", 5, "v2", true);
     ane(ssi, "r2", "cf1", "cq1", 5, "v3", true);
     assertFalse(ssi.hasTop());
   }
-  
+
   public void test2() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq3", 5, "v2");
     put(tm1, "r2", "cf1", "cq1", 5, "v3");
-    
+
     SortedMapIterator smi = new SortedMapIterator(tm1);
     TestDataSource tds = new TestDataSource(smi);
     SourceSwitchingIterator ssi = new SourceSwitchingIterator(tds);
-    
+
     ssi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(ssi, "r1", "cf1", "cq1", 5, "v1", true);
-    
+
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     put(tm2, "r1", "cf1", "cq1", 5, "v4");
     put(tm2, "r1", "cf1", "cq3", 5, "v5");
     put(tm2, "r2", "cf1", "cq1", 5, "v6");
-    
+
     SortedMapIterator smi2 = new SortedMapIterator(tm2);
     TestDataSource tds2 = new TestDataSource(smi2);
     tds.next = tds2;
-    
+
     ane(ssi, "r1", "cf1", "cq3", 5, "v2", true);
     ane(ssi, "r2", "cf1", "cq1", 5, "v6", true);
     assertFalse(ssi.hasTop());
   }
-  
+
   public void test3() throws Exception {
     // test switching after a row
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq2", 5, "v2");
@@ -164,82 +164,82 @@ public class SourceSwitchingIteratorTest extends TestCase {
     put(tm1, "r1", "cf1", "cq4", 5, "v4");
     put(tm1, "r3", "cf1", "cq1", 5, "v5");
     put(tm1, "r3", "cf1", "cq2", 5, "v6");
-    
+
     SortedMapIterator smi = new SortedMapIterator(tm1);
     TestDataSource tds = new TestDataSource(smi);
     SourceSwitchingIterator ssi = new SourceSwitchingIterator(tds, true);
-    
+
     ssi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(ssi, "r1", "cf1", "cq1", 5, "v1", true);
-    
+
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>(tm1);
     put(tm2, "r1", "cf1", "cq5", 5, "v7"); // should not see this because it should not switch until the row is finished
     put(tm2, "r2", "cf1", "cq1", 5, "v8"); // should see this new row after it switches
-    
+
     // setup a new data source, but it should not switch until the current row is finished
     SortedMapIterator smi2 = new SortedMapIterator(tm2);
     TestDataSource tds2 = new TestDataSource(smi2);
     tds.next = tds2;
-    
+
     ane(ssi, "r1", "cf1", "cq2", 5, "v2", true);
     ane(ssi, "r1", "cf1", "cq3", 5, "v3", true);
     ane(ssi, "r1", "cf1", "cq4", 5, "v4", true);
     ane(ssi, "r2", "cf1", "cq1", 5, "v8", true);
     ane(ssi, "r3", "cf1", "cq1", 5, "v5", true);
     ane(ssi, "r3", "cf1", "cq2", 5, "v6", true);
-    
+
   }
-  
+
   public void test4() throws Exception {
     // ensure switch is done on initial seek
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq2", 5, "v2");
-    
+
     SortedMapIterator smi = new SortedMapIterator(tm1);
     TestDataSource tds = new TestDataSource(smi);
     SourceSwitchingIterator ssi = new SourceSwitchingIterator(tds, false);
-    
+
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     put(tm2, "r1", "cf1", "cq1", 6, "v3");
     put(tm2, "r1", "cf1", "cq2", 6, "v4");
-    
+
     SortedMapIterator smi2 = new SortedMapIterator(tm2);
     TestDataSource tds2 = new TestDataSource(smi2);
     tds.next = tds2;
-    
+
     ssi.seek(new Range(), new ArrayList<ByteSequence>(), false);
-    
+
     ane(ssi, "r1", "cf1", "cq1", 6, "v3", true);
     ane(ssi, "r1", "cf1", "cq2", 6, "v4", true);
-    
+
   }
-  
+
   public void test5() throws Exception {
     // esnure switchNow() works w/ deepCopy()
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq2", 5, "v2");
-    
+
     SortedMapIterator smi = new SortedMapIterator(tm1);
     TestDataSource tds = new TestDataSource(smi);
     SourceSwitchingIterator ssi = new SourceSwitchingIterator(tds, false);
-    
+
     SortedKeyValueIterator<Key,Value> dc1 = ssi.deepCopy(null);
-    
+
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     put(tm2, "r1", "cf1", "cq1", 6, "v3");
     put(tm2, "r2", "cf1", "cq2", 6, "v4");
-    
+
     SortedMapIterator smi2 = new SortedMapIterator(tm2);
     TestDataSource tds2 = new TestDataSource(smi2);
     tds.setNext(tds2);
-    
+
     ssi.switchNow();
-    
+
     ssi.seek(new Range("r1"), new ArrayList<ByteSequence>(), false);
     dc1.seek(new Range("r2"), new ArrayList<ByteSequence>(), false);
-    
+
     ane(ssi, "r1", "cf1", "cq1", 6, "v3", true);
     assertFalse(ssi.hasTop());
     ane(dc1, "r2", "cf1", "cq2", 6, "v4", true);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/TimeSettingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/TimeSettingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/TimeSettingIteratorTest.java
index 188ec47..783dbc0 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/TimeSettingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/TimeSettingIteratorTest.java
@@ -28,57 +28,57 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.SortedMapIterator;
 
 public class TimeSettingIteratorTest extends TestCase {
-  
+
   public void test1() throws Exception {
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     tm1.put(new Key("r0", "cf1", "cq1", 9l), new Value("v0".getBytes()));
     tm1.put(new Key("r1", "cf1", "cq1", Long.MAX_VALUE), new Value("v1".getBytes()));
     tm1.put(new Key("r1", "cf1", "cq1", 90l), new Value("v2".getBytes()));
     tm1.put(new Key("r1", "cf1", "cq1", 0l), new Value("v3".getBytes()));
     tm1.put(new Key("r2", "cf1", "cq1", 6l), new Value("v4".getBytes()));
-    
+
     TimeSettingIterator tsi = new TimeSettingIterator(new SortedMapIterator(tm1), 50);
-    
+
     tsi.seek(new Range(new Key("r1", "cf1", "cq1", 50l), true, new Key("r1", "cf1", "cq1", 50l), true), new HashSet<ByteSequence>(), false);
-    
+
     assertTrue(tsi.hasTop());
     assertEquals(new Key("r1", "cf1", "cq1", 50l), tsi.getTopKey());
     assertEquals("v1", tsi.getTopValue().toString());
     tsi.next();
-    
+
     assertTrue(tsi.hasTop());
     assertEquals(new Key("r1", "cf1", "cq1", 50l), tsi.getTopKey());
     assertEquals("v2", tsi.getTopValue().toString());
     tsi.next();
-    
+
     assertTrue(tsi.hasTop());
     assertEquals(new Key("r1", "cf1", "cq1", 50l), tsi.getTopKey());
     assertEquals("v3", tsi.getTopValue().toString());
     tsi.next();
-    
+
     assertFalse(tsi.hasTop());
-    
+
     tsi.seek(new Range(new Key("r1", "cf1", "cq1", 50l), false, null, true), new HashSet<ByteSequence>(), false);
-    
+
     assertTrue(tsi.hasTop());
     assertEquals(new Key("r2", "cf1", "cq1", 50l), tsi.getTopKey());
     assertEquals("v4", tsi.getTopValue().toString());
     tsi.next();
-    
+
     assertFalse(tsi.hasTop());
-    
+
     tsi.seek(new Range(null, true, new Key("r1", "cf1", "cq1", 50l), false), new HashSet<ByteSequence>(), false);
-    
+
     assertTrue(tsi.hasTop());
     assertEquals(new Key("r0", "cf1", "cq1", 50l), tsi.getTopKey());
     assertEquals("v0", tsi.getTopValue().toString());
     tsi.next();
-    
+
     assertFalse(tsi.hasTop());
-    
+
     tsi.seek(new Range(new Key("r1", "cf1", "cq1", 51l), true, new Key("r1", "cf1", "cq1", 50l), false), new HashSet<ByteSequence>(), false);
     assertFalse(tsi.hasTop());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/system/VisibilityFilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/system/VisibilityFilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/system/VisibilityFilterTest.java
index 6bc7bdc..667aa5f 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/system/VisibilityFilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/system/VisibilityFilterTest.java
@@ -32,21 +32,21 @@ import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 
 public class VisibilityFilterTest extends TestCase {
-  
+
   public void testBadVisibility() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     tm.put(new Key("r1", "cf1", "cq1", "A&"), new Value(new byte[0]));
     VisibilityFilter filter = new VisibilityFilter(new SortedMapIterator(tm), new Authorizations("A"), "".getBytes());
-    
+
     // suppress logging
     Level prevLevel = Logger.getLogger(VisibilityFilter.class).getLevel();
     Logger.getLogger(VisibilityFilter.class).setLevel(Level.FATAL);
 
     filter.seek(new Range(), new HashSet<ByteSequence>(), false);
     assertFalse(filter.hasTop());
-    
+
     Logger.getLogger(VisibilityFilter.class).setLevel(prevLevel);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/BigDecimalCombinerTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/BigDecimalCombinerTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/BigDecimalCombinerTest.java
index 9e3d975..c15fe55 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/BigDecimalCombinerTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/BigDecimalCombinerTest.java
@@ -50,7 +50,6 @@ public class BigDecimalCombinerTest {
   List<Column> columns;
   Combiner ai;
 
-
   @Before
   public void setup() {
     encoder = new BigDecimalCombiner.BigDecimalEncoder();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilterTest.java
index 44fe00f..f15bad4 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilterTest.java
@@ -40,238 +40,237 @@ import org.junit.Test;
 
 public class ColumnSliceFilterTest {
 
-    private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-
-    private static final SortedMap<Key,Value> TEST_DATA = new TreeMap<Key,Value>();
-    private static final Key KEY_1 = nkv(TEST_DATA, "boo1", "yup", "20080201", "dog");
-    private static final Key KEY_2 = nkv(TEST_DATA, "boo1", "yap", "20080202", "cat");
-    private static final Key KEY_3 = nkv(TEST_DATA, "boo2", "yap", "20080203", "hamster");
-    private static final Key KEY_4 = nkv(TEST_DATA, "boo2", "yop", "20080204", "lion");
-    private static final Key KEY_5 = nkv(TEST_DATA, "boo2", "yup", "20080206", "tiger");
-    private static final Key KEY_6 = nkv(TEST_DATA, "boo2", "yip", "20080203", "tiger");
-
-    private static IteratorEnvironment iteratorEnvironment;
-
-    private ColumnSliceFilter columnSliceFilter = new ColumnSliceFilter();
-    private IteratorSetting is;
-
-    private static Key nkv(SortedMap<Key,Value> tm, String row, String cf, String cq, String val) {
-        Key k = nk(row, cf, cq);
-        tm.put(k, new Value(val.getBytes()));
-        return k;
-    }
-
-    private static Key nk(String row, String cf, String cq) {
-        return new Key(new Text(row), new Text(cf), new Text(cq));
-    }
-
-    @Before
-    public void setUp() throws Exception {
-        columnSliceFilter.describeOptions();
-        iteratorEnvironment = new DefaultIteratorEnvironment();
-        is = new IteratorSetting(1, ColumnSliceFilter.class);
+  private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
+
+  private static final SortedMap<Key,Value> TEST_DATA = new TreeMap<Key,Value>();
+  private static final Key KEY_1 = nkv(TEST_DATA, "boo1", "yup", "20080201", "dog");
+  private static final Key KEY_2 = nkv(TEST_DATA, "boo1", "yap", "20080202", "cat");
+  private static final Key KEY_3 = nkv(TEST_DATA, "boo2", "yap", "20080203", "hamster");
+  private static final Key KEY_4 = nkv(TEST_DATA, "boo2", "yop", "20080204", "lion");
+  private static final Key KEY_5 = nkv(TEST_DATA, "boo2", "yup", "20080206", "tiger");
+  private static final Key KEY_6 = nkv(TEST_DATA, "boo2", "yip", "20080203", "tiger");
+
+  private static IteratorEnvironment iteratorEnvironment;
+
+  private ColumnSliceFilter columnSliceFilter = new ColumnSliceFilter();
+  private IteratorSetting is;
+
+  private static Key nkv(SortedMap<Key,Value> tm, String row, String cf, String cq, String val) {
+    Key k = nk(row, cf, cq);
+    tm.put(k, new Value(val.getBytes()));
+    return k;
+  }
+
+  private static Key nk(String row, String cf, String cq) {
+    return new Key(new Text(row), new Text(cf), new Text(cq));
+  }
+
+  @Before
+  public void setUp() throws Exception {
+    columnSliceFilter.describeOptions();
+    iteratorEnvironment = new DefaultIteratorEnvironment();
+    is = new IteratorSetting(1, ColumnSliceFilter.class);
+  }
+
+  @Test
+  public void testBasic() throws IOException {
+    ColumnSliceFilter.setSlice(is, "20080202", "20080204");
+
+    assertTrue(columnSliceFilter.validateOptions(is.getOptions()));
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, true);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testBothInclusive() throws IOException {
+    ColumnSliceFilter.setSlice(is, "20080202", true, "20080204", true);
+
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testBothExclusive() throws IOException {
+    ColumnSliceFilter.setSlice(is, "20080202", false, "20080204", false);
+
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testStartExclusiveEndInclusive() throws IOException {
+    ColumnSliceFilter.setSlice(is, "20080202", false, "20080204", true);
+
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testNullStart() throws IOException {
+    ColumnSliceFilter.setSlice(is, null, "20080204");
+
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_1));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testNullEnd() throws IOException {
+    ColumnSliceFilter.setSlice(is, "20080202", null);
+
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_5));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testBothNull() throws IOException {
+    ColumnSliceFilter.setSlice(is, null, null);
+
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
+
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_1));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
+    columnSliceFilter.next();
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_5));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testStartAfterEnd() throws IOException {
+    try {
+      ColumnSliceFilter.setSlice(is, "20080204", "20080202");
+      fail("IllegalArgumentException expected but not thrown");
+    } catch (IllegalArgumentException expectedException) {
+      // Exception successfully thrown
     }
-
-    @Test
-    public void testBasic() throws IOException {
-        ColumnSliceFilter.setSlice(is, "20080202", "20080204");
-
-        assertTrue(columnSliceFilter.validateOptions(is.getOptions()));
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, true);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testStartEqualToEndStartInclusiveEndExclusive() throws IOException {
+    try {
+      ColumnSliceFilter.setSlice(is, "20080202", "20080202");
+      fail("IllegalArgumentException expected but not thrown");
+    } catch (IllegalArgumentException expectedException) {
+      // Exception successfully thrown
     }
-
-    @Test
-    public void testBothInclusive() throws IOException {
-        ColumnSliceFilter.setSlice(is, "20080202", true, "20080204", true);
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
+  }
+
+  @Test
+  public void testStartEqualToEndStartExclusiveEndInclusive() throws IOException {
+    try {
+      ColumnSliceFilter.setSlice(is, "20080202", false, "20080202", true);
+      fail("IllegalArgumentException expected but not thrown");
+    } catch (IllegalArgumentException expectedException) {
+      // Exception successfully thrown
     }
+  }
 
-    @Test
-    public void testBothExclusive() throws IOException {
-        ColumnSliceFilter.setSlice(is, "20080202", false, "20080204", false);
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
-    }
+  @Test
+  public void testStartEqualToEndBothInclusive() throws IOException {
+    ColumnSliceFilter.setSlice(is, "20080202", true, "20080202", true);
 
-    @Test
-    public void testStartExclusiveEndInclusive() throws IOException {
-        ColumnSliceFilter.setSlice(is, "20080202", false, "20080204", true);
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
-    }
+    columnSliceFilter.validateOptions(is.getOptions());
+    columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
+    columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
 
-    @Test
-    public void testNullStart() throws IOException {
-        ColumnSliceFilter.setSlice(is, null, "20080204");
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_1));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
-    }
-    
-    @Test
-    public void testNullEnd() throws IOException {
-        ColumnSliceFilter.setSlice(is, "20080202", null);
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_5));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
-    }
-
-    @Test
-    public void testBothNull() throws IOException {
-        ColumnSliceFilter.setSlice(is, null, null);
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_1));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_3));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_6));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_4));
-        columnSliceFilter.next();
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_5));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
-    }
-
-    @Test
-    public void testStartAfterEnd() throws IOException {
-        try {
-            ColumnSliceFilter.setSlice(is, "20080204", "20080202");
-            fail("IllegalArgumentException expected but not thrown");
-        } catch(IllegalArgumentException expectedException) {
-            // Exception successfully thrown
-        }
-    }
-
-    @Test
-    public void testStartEqualToEndStartInclusiveEndExclusive() throws IOException {
-        try {
-            ColumnSliceFilter.setSlice(is, "20080202", "20080202");
-            fail("IllegalArgumentException expected but not thrown");
-        } catch(IllegalArgumentException expectedException) {
-            // Exception successfully thrown
-        }
-    }
-
-    @Test
-    public void testStartEqualToEndStartExclusiveEndInclusive() throws IOException {
-        try {
-            ColumnSliceFilter.setSlice(is, "20080202", false, "20080202", true);
-            fail("IllegalArgumentException expected but not thrown");
-        } catch(IllegalArgumentException expectedException) {
-            // Exception successfully thrown
-        }
-    }
-
-    @Test
-    public void testStartEqualToEndBothInclusive() throws IOException {
-        ColumnSliceFilter.setSlice(is, "20080202", true, "20080202", true);
-
-        columnSliceFilter.validateOptions(is.getOptions());
-        columnSliceFilter.init(new SortedMapIterator(TEST_DATA), is.getOptions(), iteratorEnvironment);
-        columnSliceFilter.seek(new Range(), EMPTY_COL_FAMS, false);
-
-        assertTrue(columnSliceFilter.hasTop());
-        assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
-        columnSliceFilter.next();
-        assertFalse(columnSliceFilter.hasTop());
-    }
+    assertTrue(columnSliceFilter.hasTop());
+    assertTrue(columnSliceFilter.getTopKey().equals(KEY_2));
+    columnSliceFilter.next();
+    assertFalse(columnSliceFilter.hasTop());
+  }
 }
-

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/CombinerTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/CombinerTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/CombinerTest.java
index 41d6425..cdac2fb 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/CombinerTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/CombinerTest.java
@@ -50,511 +50,511 @@ import org.junit.Assert;
 import org.junit.Test;
 
 public class CombinerTest {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-  
+
   static Key nk(int row, int colf, int colq, long ts, boolean deleted) {
     Key k = nk(row, colf, colq, ts);
     k.setDeleted(deleted);
     return k;
   }
-  
+
   static Key nk(int row, int colf, int colq, long ts) {
     return new Key(nr(row), new Text(String.format("cf%03d", colf)), new Text(String.format("cq%03d", colq)), ts);
   }
-  
+
   static Range nr(int row, int colf, int colq, long ts, boolean inclusive) {
     return new Range(nk(row, colf, colq, ts), inclusive, null, true);
   }
-  
+
   static Range nr(int row, int colf, int colq, long ts) {
     return nr(row, colf, colq, ts, true);
   }
-  
+
   static <V> void nkv(TreeMap<Key,Value> tm, int row, int colf, int colq, long ts, boolean deleted, V val, Encoder<V> encoder) {
     Key k = nk(row, colf, colq, ts);
     k.setDeleted(deleted);
     tm.put(k, new Value(encoder.encode(val)));
   }
-  
+
   static Text nr(int row) {
     return new Text(String.format("r%03d", row));
   }
-  
+
   @Test
   public void test1() throws IOException {
     Encoder<Long> encoder = LongCombiner.VAR_LEN_ENCODER;
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that do not aggregate
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, SummingCombiner.Type.VARLEN);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("2")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("4", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 1), ai.getTopKey());
     assertEquals("2", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking
-    
+
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 1), ai.getTopKey());
     assertEquals("2", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek after everything
     ai.seek(nr(1, 1, 1, 0), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
   }
-  
+
   @Test
   public void test2() throws IOException {
     Encoder<Long> encoder = LongCombiner.VAR_LEN_ENCODER;
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, VarLenEncoder.class);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking to the beginning of a key that aggregates
-    
+
     ai.seek(nr(1, 1, 1, 3), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking the middle of a key the aggregates
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking to the end of a key the aggregates
     ai.seek(nr(1, 1, 1, 1), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
     // try seeking before a key the aggregates
     ai.seek(nr(1, 1, 1, 4), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
   }
-  
+
   @Test
   public void test3() throws IOException {
     Encoder<Long> encoder = LongCombiner.FIXED_LEN_ENCODER;
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
-    
+
     // keys that do not aggregate
     nkv(tm1, 2, 2, 1, 1, false, 2l, encoder);
     nkv(tm1, 2, 2, 1, 2, false, 3l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, FixedLenEncoder.class.getName());
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 1), ai.getTopKey());
     assertEquals("2", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek after key that aggregates
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     // seek before key that aggregates
     ai.seek(nr(1, 1, 1, 4), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
   }
-  
+
   @Test
   public void testDeepCopy() throws IOException {
     Encoder<Long> encoder = LongCombiner.FIXED_LEN_ENCODER;
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
-    
+
     // keys that do not aggregate
     nkv(tm1, 2, 2, 1, 1, false, 2l, encoder);
     nkv(tm1, 2, 2, 1, 2, false, 3l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, FixedLenEncoder.class.getName());
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
-    
+
     SortedKeyValueIterator<Key,Value> ai2 = ai.deepCopy(null);
     SortedKeyValueIterator<Key,Value> ai3 = ai.deepCopy(null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 1), ai.getTopKey());
     assertEquals("2", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek after key that aggregates
     ai2.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai2.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai2.getTopKey());
     assertEquals("3", encoder.decode(ai2.getTopValue().get()).toString());
-    
+
     // seek before key that aggregates
     ai3.seek(nr(1, 1, 1, 4), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai3.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai3.getTopKey());
     assertEquals("9", encoder.decode(ai3.getTopValue().get()).toString());
-    
+
     ai3.next();
-    
+
     assertTrue(ai3.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai3.getTopKey());
     assertEquals("3", encoder.decode(ai3.getTopValue().get()).toString());
   }
-  
+
   @Test
   public void test4() throws IOException {
     Encoder<Long> encoder = LongCombiner.STRING_ENCODER;
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that do not aggregate
     nkv(tm1, 0, 0, 1, 1, false, 7l, encoder);
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
-    
+
     // keys that do not aggregate
     nkv(tm1, 2, 2, 1, 1, false, 2l, encoder);
     nkv(tm1, 2, 2, 1, 2, false, 3l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, SummingCombiner.Type.STRING);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(0, 0, 1, 1), ai.getTopKey());
     assertEquals("7", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 1), ai.getTopKey());
     assertEquals("2", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     // seek test
     ai.seek(nr(0, 0, 1, 0), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     // seek after key that aggregates
     ai.seek(nr(1, 1, 1, 2), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("3", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     // combine all columns
-    
+
     is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, SummingCombiner.Type.STRING);
     Combiner.setCombineAllColumns(is, true);
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(0, 0, 1, 1), ai.getTopKey());
     assertEquals("7", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(2, 2, 1, 2), ai.getTopKey());
     assertEquals("5", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
   }
-  
+
   @Test
   public void test5() throws IOException {
     Encoder<Long> encoder = LongCombiner.STRING_ENCODER;
     // try aggregating across multiple data sets that contain
     // the exact same keys w/ different values
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
-    
+
     TreeMap<Key,Value> tm2 = new TreeMap<Key,Value>();
     nkv(tm2, 1, 1, 1, 1, false, 3l, encoder);
-    
+
     TreeMap<Key,Value> tm3 = new TreeMap<Key,Value>();
     nkv(tm3, 1, 1, 1, 1, false, 4l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, StringEncoder.class);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     List<SortedKeyValueIterator<Key,Value>> sources = new ArrayList<SortedKeyValueIterator<Key,Value>>(3);
     sources.add(new SortedMapIterator(tm1));
     sources.add(new SortedMapIterator(tm2));
     sources.add(new SortedMapIterator(tm3));
-    
+
     MultiIterator mi = new MultiIterator(sources, true);
     ai.init(mi, is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 1), ai.getTopKey());
     assertEquals("9", encoder.decode(ai.getTopValue().get()).toString());
   }
-  
+
   @Test
   public void test6() throws IOException {
     Encoder<Long> encoder = LongCombiner.VAR_LEN_ENCODER;
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, 2l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, VarLenEncoder.class.getName());
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), new DefaultIteratorEnvironment());
-    
+
     // try seeking to the beginning of a key that aggregates
-    
+
     ai.seek(nr(1, 1, 1, 3, false), EMPTY_COL_FAMS, false);
-    
+
     assertFalse(ai.hasTop());
-    
+
   }
-  
+
   @Test
   public void test7() throws IOException {
     Encoder<Long> encoder = LongCombiner.FIXED_LEN_ENCODER;
-    
+
     // test that delete is not aggregated
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     nkv(tm1, 1, 1, 1, 2, true, 0l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 4l, encoder);
     nkv(tm1, 1, 1, 1, 4, false, 3l, encoder);
-    
+
     Combiner ai = new SummingCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, SummingCombiner.Type.FIXEDLEN);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), new DefaultIteratorEnvironment());
-    
+
     ai.seek(nr(1, 1, 1, 4, true), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 4), ai.getTopKey());
     assertEquals("7", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2, true), ai.getTopKey());
     assertEquals("0", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
     assertFalse(ai.hasTop());
-    
+
     tm1 = new TreeMap<Key,Value>();
     nkv(tm1, 1, 1, 1, 2, true, 0l, encoder);
     ai = new SummingCombiner();
     ai.init(new SortedMapIterator(tm1), is.getOptions(), new DefaultIteratorEnvironment());
-    
+
     ai.seek(nr(1, 1, 1, 4, true), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 2, true), ai.getTopKey());
     assertEquals("0", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
     assertFalse(ai.hasTop());
   }
-  
+
   @Test
   public void valueIteratorTest() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
@@ -567,7 +567,7 @@ public class CombinerTest {
     assertEquals(iter.next().toString(), "1");
     assertFalse(iter.hasNext());
   }
-  
+
   @Test
   public void sumAllColumns() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
@@ -585,11 +585,11 @@ public class CombinerTest {
     iter.init(smi, s.getOptions(), new DefaultIteratorEnvironment());
     Combiner iter2 = new SummingCombiner();
     IteratorSetting s2 = new IteratorSetting(10, "s2", SummingCombiner.class);
-    SummingCombiner.setColumns(s2, Collections.singletonList(new IteratorSetting.Column("count","a")));
+    SummingCombiner.setColumns(s2, Collections.singletonList(new IteratorSetting.Column("count", "a")));
     SummingCombiner.setEncodingType(s2, LongCombiner.StringEncoder.class);
     iter2.init(iter, s.getOptions(), new DefaultIteratorEnvironment());
     iter2.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(iter2.hasTop());
     assertEquals("2", iter2.getTopValue().toString());
     iter2.next();
@@ -601,51 +601,49 @@ public class CombinerTest {
     iter2.next();
     assertFalse(iter2.hasTop());
   }
-  
 
-  
   @Test
   public void maxMinTest() throws IOException {
     Encoder<Long> encoder = LongCombiner.VAR_LEN_ENCODER;
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, 4l, encoder);
     nkv(tm1, 1, 1, 1, 2, false, 3l, encoder);
     nkv(tm1, 1, 1, 1, 3, false, 2l, encoder);
-    
+
     Combiner ai = new MaxCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingCombiner.class);
     LongCombiner.setEncodingType(is, SummingCombiner.Type.VARLEN);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("4", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     ai = new MinCombiner();
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertEquals("2", encoder.decode(ai.getTopValue().get()).toString());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
   }
-  
+
   public static List<Long> nal(Long... longs) {
     List<Long> al = new ArrayList<Long>(longs.length);
     for (Long l : longs) {
@@ -653,110 +651,110 @@ public class CombinerTest {
     }
     return al;
   }
-  
+
   public static void assertBytesEqual(byte[] a, byte[] b) {
     assertEquals(a.length, b.length);
     for (int i = 0; i < a.length; i++)
       assertEquals(a[i], b[i]);
   }
-  
+
   public static void sumArray(Class<? extends Encoder<List<Long>>> encoderClass, SummingArrayCombiner.Type type) throws IOException, InstantiationException,
       IllegalAccessException {
     Encoder<List<Long>> encoder = encoderClass.newInstance();
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
-    
+
     // keys that aggregate
     nkv(tm1, 1, 1, 1, 1, false, nal(1l, 2l), encoder);
     nkv(tm1, 1, 1, 1, 2, false, nal(3l, 4l, 5l), encoder);
     nkv(tm1, 1, 1, 1, 3, false, nal(), encoder);
-    
+
     Combiner ai = new SummingArrayCombiner();
-    
+
     IteratorSetting is = new IteratorSetting(1, SummingArrayCombiner.class);
     SummingArrayCombiner.setEncodingType(is, type);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertBytesEqual(encoder.encode(nal(4l, 6l, 5l)), ai.getTopValue().get());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     is.clearOptions();
     SummingArrayCombiner.setEncodingType(is, encoderClass);
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertBytesEqual(encoder.encode(nal(4l, 6l, 5l)), ai.getTopValue().get());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     is.clearOptions();
     SummingArrayCombiner.setEncodingType(is, encoderClass.getName());
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
     ai.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(ai.hasTop());
     assertEquals(nk(1, 1, 1, 3), ai.getTopKey());
     assertBytesEqual(encoder.encode(nal(4l, 6l, 5l)), ai.getTopValue().get());
-    
+
     ai.next();
-    
+
     assertFalse(ai.hasTop());
-    
+
     is.clearOptions();
     SummingArrayCombiner.setEncodingType(is, SummingCombiner.VAR_LEN_ENCODER.getClass().getName());
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     try {
       ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
       Assert.fail();
     } catch (IllegalArgumentException e) {}
-    
+
     is.clearOptions();
     SummingArrayCombiner.setEncodingType(is, BadEncoder.class.getName());
     Combiner.setColumns(is, Collections.singletonList(new IteratorSetting.Column("cf001")));
-    
+
     try {
       ai.init(new SortedMapIterator(tm1), is.getOptions(), null);
       Assert.fail();
     } catch (IllegalArgumentException e) {}
   }
-  
+
   public static class BadEncoder implements Encoder<List<Long>> {
     @Override
     public byte[] encode(List<Long> v) {
       return new byte[0];
     }
-    
+
     @Override
     public List<Long> decode(byte[] b) {
       return new ArrayList<Long>();
     }
-    
+
   }
-  
+
   @Test
   public void sumArrayTest() throws IOException, InstantiationException, IllegalAccessException {
     sumArray(SummingArrayCombiner.VarLongArrayEncoder.class, SummingArrayCombiner.Type.VARLEN);
     sumArray(SummingArrayCombiner.FixedLongArrayEncoder.class, SummingArrayCombiner.Type.FIXEDLEN);
     sumArray(SummingArrayCombiner.StringArrayEncoder.class, SummingArrayCombiner.Type.STRING);
   }
-  
+
   @Test
   public void testEncoders() {
     TypedValueCombiner.testEncoder(SummingCombiner.FIXED_LEN_ENCODER, Long.MAX_VALUE);
@@ -774,12 +772,12 @@ public class CombinerTest {
     TypedValueCombiner.testEncoder(SummingCombiner.STRING_ENCODER, 42l);
     TypedValueCombiner.testEncoder(SummingCombiner.STRING_ENCODER, -42l);
     TypedValueCombiner.testEncoder(SummingCombiner.STRING_ENCODER, 0l);
-    
+
     TypedValueCombiner.testEncoder(SummingArrayCombiner.FIXED_LONG_ARRAY_ENCODER, Arrays.asList(0l, -1l, 10l, Long.MAX_VALUE, Long.MIN_VALUE));
     TypedValueCombiner.testEncoder(SummingArrayCombiner.VAR_LONG_ARRAY_ENCODER, Arrays.asList(0l, -1l, 10l, Long.MAX_VALUE, Long.MIN_VALUE));
     TypedValueCombiner.testEncoder(SummingArrayCombiner.STRING_ARRAY_ENCODER, Arrays.asList(0l, -1l, 10l, Long.MAX_VALUE, Long.MIN_VALUE));
   }
-  
+
   @Test
   public void testAdds() {
     assertEquals(LongCombiner.safeAdd(Long.MIN_VALUE + 5, -10), Long.MIN_VALUE);
@@ -787,5 +785,5 @@ public class CombinerTest {
     assertEquals(LongCombiner.safeAdd(Long.MIN_VALUE + 5, -5), Long.MIN_VALUE);
     assertEquals(LongCombiner.safeAdd(Long.MAX_VALUE - 5, 5), Long.MAX_VALUE);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/FilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/FilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/FilterTest.java
index b7a842f..3aefdf1 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/FilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/FilterTest.java
@@ -47,10 +47,10 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class FilterTest {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
   private static final Map<String,String> EMPTY_OPTS = new HashMap<String,String>();
-  
+
   public static class SimpleFilter extends Filter {
     public boolean accept(Key k, Value v) {
       // System.out.println(k.getRow());
@@ -59,7 +59,7 @@ public class FilterTest {
       return false;
     }
   }
-  
+
   public static class SimpleFilter2 extends Filter {
     public boolean accept(Key k, Value v) {
       if (k.getColumnFamily().toString().equals("a"))
@@ -67,7 +67,7 @@ public class FilterTest {
       return true;
     }
   }
-  
+
   private static int size(SortedKeyValueIterator<Key,Value> iterator) throws IOException {
     int size = 0;
     while (iterator.hasTop()) {
@@ -77,33 +77,33 @@ public class FilterTest {
     }
     return size;
   }
-  
+
   @Test
   public void test1() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     for (int i = 0; i < 1000; i++) {
       Key k = new Key(new Text(String.format("%03d", i)), colf, colq);
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     Filter filter1 = new SimpleFilter();
     filter1.init(new SortedMapIterator(tm), EMPTY_OPTS, null);
     filter1.seek(new Range(), EMPTY_COL_FAMS, false);
     int size = size(filter1);
     assertTrue("size = " + size, size == 100);
-    
+
     Filter fi = new SimpleFilter();
     fi.init(new SortedMapIterator(tm), EMPTY_OPTS, null);
     Key k = new Key(new Text("500"));
     fi.seek(new Range(k, null), EMPTY_COL_FAMS, false);
     size = size(fi);
     assertTrue("size = " + size, size == 50);
-    
+
     filter1 = new SimpleFilter();
     filter1.init(new SortedMapIterator(tm), EMPTY_OPTS, null);
     Filter filter2 = new SimpleFilter2();
@@ -112,36 +112,36 @@ public class FilterTest {
     size = size(filter2);
     assertTrue("size = " + size, size == 0);
   }
-  
+
   @Test
   public void test1neg() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     for (int i = 0; i < 1000; i++) {
       Key k = new Key(new Text(String.format("%03d", i)), colf, colq);
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     Filter filter = new SimpleFilter();
-    
+
     IteratorSetting is = new IteratorSetting(1, SimpleFilter.class);
     Filter.setNegate(is, true);
-    
+
     filter.init(new SortedMapIterator(tm), is.getOptions(), null);
     filter.seek(new Range(), EMPTY_COL_FAMS, false);
     int size = size(filter);
     assertTrue("size = " + size, size == 900);
-    
+
     filter.init(new SortedMapIterator(tm), is.getOptions(), null);
     Key k = new Key(new Text("500"));
     filter.seek(new Range(k, null), EMPTY_COL_FAMS, false);
     size = size(filter);
     assertTrue("size = " + size, size == 450);
-    
+
     filter.init(new SortedMapIterator(tm), EMPTY_OPTS, null);
     Filter filter2 = new SimpleFilter2();
     filter2.init(filter, is.getOptions(), null);
@@ -149,25 +149,25 @@ public class FilterTest {
     size = size(filter2);
     assertTrue("size = " + size, size == 100);
   }
-  
+
   @Test
   public void testDeepCopy() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     for (int i = 0; i < 1000; i++) {
       Key k = new Key(new Text(String.format("%03d", i)), colf, colq);
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     SimpleFilter filter = new SimpleFilter();
-    
+
     IteratorSetting is = new IteratorSetting(1, SimpleFilter.class);
     Filter.setNegate(is, true);
-    
+
     filter.init(new SortedMapIterator(tm), is.getOptions(), null);
     SortedKeyValueIterator<Key,Value> copy = filter.deepCopy(null);
     filter.seek(new Range(), EMPTY_COL_FAMS, false);
@@ -177,21 +177,21 @@ public class FilterTest {
     size = size(copy);
     assertTrue("size = " + size, size == 900);
   }
-  
+
   @Test
   public void test2() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     for (int i = 0; i < 1000; i++) {
       Key k = new Key(new Text(String.format("%03d", i)), colf, colq);
       k.setTimestamp(i);
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     SortedKeyValueIterator<Key,Value> a = new AgeOffFilter();
     IteratorSetting is = new IteratorSetting(1, AgeOffFilter.class);
     AgeOffFilter.setTTL(is, 101l);
@@ -210,7 +210,7 @@ public class FilterTest {
     copy.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(copy), 900);
   }
-  
+
   @Test
   public void test2a() throws IOException {
     Text colf = new Text("a");
@@ -220,26 +220,26 @@ public class FilterTest {
     IteratorSetting is = new IteratorSetting(1, ColumnAgeOffFilter.class);
     ColumnAgeOffFilter.addTTL(is, new IteratorSetting.Column("a"), 901l);
     long ts = System.currentTimeMillis();
-    
+
     for (long i = 0; i < 1000; i++) {
       Key k = new Key(new Text(String.format("%03d", i)), colf, colq, ts - i);
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     ColumnAgeOffFilter a = new ColumnAgeOffFilter();
     assertTrue(a.validateOptions(is.getOptions()));
     a.init(new SortedMapIterator(tm), is.getOptions(), new DefaultIteratorEnvironment());
     a.overrideCurrentTime(ts);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 902);
-    
+
     ColumnAgeOffFilter.addTTL(is, new IteratorSetting.Column("a", "b"), 101l);
     a.init(new SortedMapIterator(tm), is.getOptions(), new DefaultIteratorEnvironment());
     a.overrideCurrentTime(ts);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 102);
-    
+
     ColumnAgeOffFilter.removeTTL(is, new IteratorSetting.Column("a", "b"));
     a.init(new SortedMapIterator(tm), is.getOptions(), new DefaultIteratorEnvironment());
     a = (ColumnAgeOffFilter) a.deepCopy(null);
@@ -247,14 +247,14 @@ public class FilterTest {
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 902);
   }
-  
+
   @Test
   public void test3() throws IOException {
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
     HashSet<Column> hsc = new HashSet<Column>();
     hsc.add(new Column("c".getBytes(), null, null));
-    
+
     Text colf1 = new Text("a");
     Text colq1 = new Text("b");
     Text colf2 = new Text("c");
@@ -274,93 +274,93 @@ public class FilterTest {
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     ColumnQualifierFilter a = new ColumnQualifierFilter(new SortedMapIterator(tm), hsc);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 1000);
-    
+
     hsc = new HashSet<Column>();
     hsc.add(new Column("a".getBytes(), "b".getBytes(), null));
     a = new ColumnQualifierFilter(new SortedMapIterator(tm), hsc);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     int size = size(a);
     assertTrue("size was " + size, size == 500);
-    
+
     hsc = new HashSet<Column>();
     a = new ColumnQualifierFilter(new SortedMapIterator(tm), hsc);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     size = size(a);
     assertTrue("size was " + size, size == 1000);
   }
-  
+
   @Test
   public void test4() throws IOException {
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     ColumnVisibility le1 = new ColumnVisibility("L1");
     ColumnVisibility le2 = new ColumnVisibility("L0&OFFICIAL");
     ColumnVisibility le3 = new ColumnVisibility("L1&L2");
     ColumnVisibility le4 = new ColumnVisibility("L1&L2&G1");
     ColumnVisibility[] lea = {le1, le2, le3, le4};
     Authorizations auths = new Authorizations("L1", "L2", "L0", "OFFICIAL");
-    
+
     for (int i = 0; i < 1000; i++) {
       Key k = new Key(new Text(String.format("%03d", i)), new Text("a"), new Text("b"), new Text(lea[i % 4].getExpression()));
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 1000);
-    
+
     VisibilityFilter a = new VisibilityFilter(new SortedMapIterator(tm), auths, le2.getExpression());
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     int size = size(a);
     assertTrue("size was " + size, size == 750);
   }
-  
+
   private ColumnQualifierFilter ncqf(TreeMap<Key,Value> tm, Column... columns) throws IOException {
     HashSet<Column> hsc = new HashSet<Column>();
-    
+
     for (Column column : columns) {
       hsc.add(column);
     }
-    
+
     ColumnQualifierFilter a = new ColumnQualifierFilter(new SortedMapIterator(tm), hsc);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     return a;
   }
-  
+
   @Test
   public void test5() throws IOException {
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     tm.put(new Key(new Text(String.format("%03d", 1)), new Text("a"), new Text("x")), dv);
     tm.put(new Key(new Text(String.format("%03d", 2)), new Text("a"), new Text("y")), dv);
     tm.put(new Key(new Text(String.format("%03d", 3)), new Text("a"), new Text("z")), dv);
     tm.put(new Key(new Text(String.format("%03d", 4)), new Text("b"), new Text("x")), dv);
     tm.put(new Key(new Text(String.format("%03d", 5)), new Text("b"), new Text("y")), dv);
-    
+
     assertTrue(tm.size() == 5);
-    
+
     int size = size(ncqf(tm, new Column("c".getBytes(), null, null)));
     assertTrue(size == 5);
-    
+
     size = size(ncqf(tm, new Column("a".getBytes(), null, null)));
     assertTrue(size == 5);
-    
+
     size = size(ncqf(tm, new Column("a".getBytes(), "x".getBytes(), null)));
     assertTrue(size == 1);
-    
+
     size = size(ncqf(tm, new Column("a".getBytes(), "x".getBytes(), null), new Column("b".getBytes(), "x".getBytes(), null)));
     assertTrue(size == 2);
-    
+
     size = size(ncqf(tm, new Column("a".getBytes(), "x".getBytes(), null), new Column("b".getBytes(), "y".getBytes(), null)));
     assertTrue(size == 2);
-    
+
     size = size(ncqf(tm, new Column("a".getBytes(), "x".getBytes(), null), new Column("b".getBytes(), null, null)));
     assertTrue(size == 3);
   }
-  
+
   @Test
   public void testNoVisFilter() throws IOException {
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
@@ -370,28 +370,28 @@ public class FilterTest {
       tm.put(k, v);
     }
     assertTrue(tm.size() == 1000);
-    
+
     Filter filter = new ReqVisFilter();
     filter.init(new SortedMapIterator(tm), EMPTY_OPTS, null);
     filter.seek(new Range(), EMPTY_COL_FAMS, false);
     int size = size(filter);
     assertTrue("size = " + size, size == 100);
   }
-  
+
   @Test
   public void testTimestampFilter() throws IOException, ParseException {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     for (int i = 0; i < 100; i++) {
       Key k = new Key(new Text(String.format("%02d", i)), colf, colq);
       k.setTimestamp(i);
       tm.put(k, dv);
     }
     assertTrue(tm.size() == 100);
-    
+
     SimpleDateFormat dateParser = new SimpleDateFormat("yyyyMMddHHmmssz");
     long baseTime = dateParser.parse("19990101000000GMT").getTime();
     tm.clear();
@@ -412,76 +412,76 @@ public class FilterTest {
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 21);
-    
+
     TimestampFilter.setEnd(is, "19990101000031GMT", false);
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 20);
-    
+
     TimestampFilter.setStart(is, "19990101000011GMT", false);
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 19);
-    
+
     TimestampFilter.setEnd(is, "19990101000031GMT", true);
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 20);
-    
+
     is.clearOptions();
     TimestampFilter.setStart(is, "19990101000011GMT", true);
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 89);
-    
+
     TimestampFilter.setStart(is, "19990101000011GMT", false);
     assertTrue(a.validateOptions(is.getOptions()));
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 88);
-    
+
     is.clearOptions();
     TimestampFilter.setEnd(is, "19990101000031GMT", true);
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 32);
-    
+
     TimestampFilter.setEnd(is, "19990101000031GMT", false);
     assertTrue(a.validateOptions(is.getOptions()));
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 31);
-    
+
     TimestampFilter.setEnd(is, 253402300800001l, true);
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
-    
+
     is.clearOptions();
     is.addOption(TimestampFilter.START, "19990101000011GMT");
     assertTrue(a.validateOptions(is.getOptions()));
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 89);
-    
+
     is.clearOptions();
     is.addOption(TimestampFilter.END, "19990101000031GMT");
     assertTrue(a.validateOptions(is.getOptions()));
     a.init(new SortedMapIterator(tm), is.getOptions(), null);
     a.seek(new Range(), EMPTY_COL_FAMS, false);
     assertEquals(size(a), 32);
-    
+
     try {
       a.validateOptions(EMPTY_OPTS);
       assertTrue(false);
     } catch (IllegalArgumentException e) {}
   }
-  
+
   @Test
   public void testDeletes() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
     Value dv = new Value();
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     Key k = new Key(new Text("0"), colf, colq);
     tm.put(k, dv);
     k = new Key(new Text("1"), colf, colq, 10);
@@ -491,14 +491,14 @@ public class FilterTest {
     tm.put(k, dv);
     k = new Key(new Text("10"), colf, colq);
     tm.put(k, dv);
-    
+
     assertTrue(tm.size() == 4);
-    
+
     Filter filter = new SimpleFilter();
     filter.init(new SortedMapIterator(tm), EMPTY_OPTS, null);
     filter.seek(new Range(), EMPTY_COL_FAMS, false);
     int size = size(filter);
     assertTrue("size = " + size, size == 3);
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/GrepIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/GrepIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/GrepIteratorTest.java
index bfce3db..23af994 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/GrepIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/GrepIteratorTest.java
@@ -41,33 +41,33 @@ public class GrepIteratorTest {
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
   SortedMap<Key,Value> input;
   SortedMap<Key,Value> output;
-  
+
   @Before
   public void init() {
     input = new TreeMap<Key,Value>();
     output = new TreeMap<Key,Value>();
     input.put(new Key("abcdef", "xyz", "xyz", 0), new Value("xyz".getBytes()));
     output.put(new Key("abcdef", "xyz", "xyz", 0), new Value("xyz".getBytes()));
-    
+
     input.put(new Key("bdf", "ace", "xyz", 0), new Value("xyz".getBytes()));
     input.put(new Key("bdf", "abcdef", "xyz", 0), new Value("xyz".getBytes()));
     output.put(new Key("bdf", "abcdef", "xyz", 0), new Value("xyz".getBytes()));
     input.put(new Key("bdf", "xyz", "xyz", 0), new Value("xyz".getBytes()));
-    
+
     input.put(new Key("ceg", "xyz", "abcdef", 0), new Value("xyz".getBytes()));
     output.put(new Key("ceg", "xyz", "abcdef", 0), new Value("xyz".getBytes()));
     input.put(new Key("ceg", "xyz", "xyz", 0), new Value("xyz".getBytes()));
-    
+
     input.put(new Key("dfh", "xyz", "xyz", 0), new Value("abcdef".getBytes()));
     output.put(new Key("dfh", "xyz", "xyz", 0), new Value("abcdef".getBytes()));
     input.put(new Key("dfh", "xyz", "xyz", 1), new Value("xyz".getBytes()));
-    
+
     Key k = new Key("dfh", "xyz", "xyz", 1);
     k.setDeleted(true);
     input.put(k, new Value("xyz".getBytes()));
     output.put(k, new Value("xyz".getBytes()));
   }
-  
+
   public static void checkEntries(SortedKeyValueIterator<Key,Value> skvi, SortedMap<Key,Value> map) throws IOException {
     for (Entry<Key,Value> e : map.entrySet()) {
       assertTrue(skvi.hasTop());
@@ -77,7 +77,7 @@ public class GrepIteratorTest {
     }
     assertFalse(skvi.hasTop());
   }
-  
+
   @Test
   public void test() throws IOException {
     GrepIterator gi = new GrepIterator();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/IndexedDocIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/IndexedDocIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/IndexedDocIteratorTest.java
index ac0ab6b..117fcac 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/IndexedDocIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/IndexedDocIteratorTest.java
@@ -43,29 +43,29 @@ import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 
 public class IndexedDocIteratorTest extends TestCase {
-  
+
   private static final Logger log = Logger.getLogger(IndexedDocIteratorTest.class);
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
   private static final byte[] nullByte = {0};
-  
+
   private static IteratorEnvironment env = new DefaultIteratorEnvironment();
-  
+
   TreeMap<Key,Value> map;
   Text[] columnFamilies;
   Text[] otherColumnFamilies;
-  
+
   static int docid = 0;
   static String docColfPrefix = "doc";
   static Text indexColf = new Text("index");
   static Text docColf = new Text(docColfPrefix);
-  
+
   static {
     log.setLevel(Level.OFF);
     docColf.append(nullByte, 0, 1);
     docColf.append("type".getBytes(), 0, "type".getBytes().length);
   }
-  
+
   private TreeMap<Key,Value> createSortedMap(float hitRatio, int numRows, int numDocsPerRow, Text[] columnFamilies, Text[] otherColumnFamilies,
       HashSet<Text> docs, Text[] negatedColumns) {
     StringBuilder sb = new StringBuilder();
@@ -73,7 +73,7 @@ public class IndexedDocIteratorTest extends TestCase {
     Value v = new Value(new byte[0]);
     TreeMap<Key,Value> map = new TreeMap<Key,Value>();
     boolean[] negateMask = new boolean[columnFamilies.length];
-    
+
     for (int i = 0; i < columnFamilies.length; i++) {
       negateMask[i] = false;
       if (negatedColumns.length > 0)
@@ -131,20 +131,20 @@ public class IndexedDocIteratorTest extends TestCase {
     }
     return map;
   }
-  
+
   static TestRFile trf = new TestRFile(AccumuloConfiguration.getDefaultConfiguration());
-  
+
   private SortedKeyValueIterator<Key,Value> createIteratorStack(float hitRatio, int numRows, int numDocsPerRow, Text[] columnFamilies,
       Text[] otherColumnFamilies, HashSet<Text> docs) throws IOException {
     Text nullText[] = new Text[0];
     return createIteratorStack(hitRatio, numRows, numDocsPerRow, columnFamilies, otherColumnFamilies, docs, nullText);
   }
-  
+
   private SortedKeyValueIterator<Key,Value> createIteratorStack(float hitRatio, int numRows, int numDocsPerRow, Text[] columnFamilies,
       Text[] otherColumnFamilies, HashSet<Text> docs, Text[] negatedColumns) throws IOException {
     // write a map file
     trf.openWriter(false);
-    
+
     TreeMap<Key,Value> inMemoryMap = createSortedMap(hitRatio, numRows, numDocsPerRow, columnFamilies, otherColumnFamilies, docs, negatedColumns);
     trf.writer.startNewLocalityGroup("docs", RFileTest.ncfs(docColf.toString()));
     for (Entry<Key,Value> entry : inMemoryMap.entrySet()) {
@@ -156,28 +156,28 @@ public class IndexedDocIteratorTest extends TestCase {
       if (entry.getKey().getColumnFamily().equals(indexColf))
         trf.writer.append(entry.getKey(), entry.getValue());
     }
-    
+
     trf.closeWriter();
-    
+
     trf.openReader();
     return trf.reader;
   }
-  
+
   private synchronized static void cleanup() throws IOException {
     trf.closeReader();
     docid = 0;
   }
-  
+
   public void testNull() {}
-  
+
   @Override
   public void setUp() {
     Logger.getRootLogger().setLevel(Level.ERROR);
   }
-  
+
   private static final int NUM_ROWS = 5;
   private static final int NUM_DOCIDS = 200;
-  
+
   public void test1() throws IOException {
     columnFamilies = new Text[2];
     columnFamilies[0] = new Text("CC");
@@ -187,7 +187,7 @@ public class IndexedDocIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("B");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     HashSet<Text> docs = new HashSet<Text>();
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
@@ -204,17 +204,17 @@ public class IndexedDocIteratorTest extends TestCase {
       Value v = iter.getTopValue();
       // System.out.println(k.toString());
       // System.out.println(iter.getDocID(k));
-      
+
       Text d = IndexedDocIterator.parseDocID(k);
       assertTrue(docs.contains(d));
       assertTrue(new String(v.get()).endsWith(" docID=" + d));
-      
+
       iter.next();
     }
     assertEquals(hitCount, docs.size());
     cleanup();
   }
-  
+
   public void test2() throws IOException {
     columnFamilies = new Text[3];
     columnFamilies[0] = new Text("A");
@@ -225,7 +225,7 @@ public class IndexedDocIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("C");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     HashSet<Text> docs = new HashSet<Text>();
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
@@ -248,7 +248,7 @@ public class IndexedDocIteratorTest extends TestCase {
     assertEquals(hitCount, docs.size());
     cleanup();
   }
-  
+
   public void test3() throws IOException {
     columnFamilies = new Text[6];
     columnFamilies[0] = new Text("C");
@@ -262,7 +262,7 @@ public class IndexedDocIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("B");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     HashSet<Text> docs = new HashSet<Text>();
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
@@ -290,7 +290,7 @@ public class IndexedDocIteratorTest extends TestCase {
     assertEquals(hitCount, docs.size());
     cleanup();
   }
-  
+
   public void test4() throws IOException {
     columnFamilies = new Text[3];
     boolean[] notFlags = new boolean[3];
@@ -308,7 +308,7 @@ public class IndexedDocIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("C");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     HashSet<Text> docs = new HashSet<Text>();
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs, negatedColumns);


[58/66] [abbrv] accumulo git commit: ACCUMULO-3451 comply with checkstyle rules

Posted by ct...@apache.org.
ACCUMULO-3451 comply with checkstyle rules

  Adds javadoc and style changes to pass checkstyle enforcement rules.


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/901d60ef
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/901d60ef
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/901d60ef

Branch: refs/heads/master
Commit: 901d60ef1cf72c2d55c90746fce94e108a992d3b
Parents: d2c116f
Author: Christopher Tubbs <ct...@apache.org>
Authored: Wed Dec 24 15:22:30 2014 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:25:40 2015 -0500

----------------------------------------------------------------------
 .../core/client/admin/InstanceOperations.java        |  6 +++---
 .../java/org/apache/accumulo/core/data/Value.java    |  7 +++----
 .../apache/accumulo/core/file/rfile/CreateEmpty.java |  4 ++--
 .../accumulo/core/file/rfile/bcfile/TFile.java       |  4 ++--
 .../accumulo/core/file/rfile/bcfile/Utils.java       |  1 +
 .../org/apache/accumulo/core/iterators/Combiner.java |  2 +-
 .../apache/accumulo/core/iterators/LongCombiner.java |  4 ++--
 .../accumulo/core/iterators/OptionDescriber.java     |  4 ++--
 .../accumulo/core/iterators/TypedValueCombiner.java  | 10 +++++-----
 .../core/iterators/user/SummingArrayCombiner.java    |  5 +++--
 .../accumulo/core/security/ColumnVisibility.java     | 10 +++++-----
 .../crypto/DefaultSecretKeyEncryptionStrategy.java   |  3 ++-
 .../accumulo/core/util/format/BinaryFormatter.java   |  7 ++++---
 .../accumulo/core/util/shell/commands/DUCommand.java |  4 +++-
 .../core/util/shell/commands/EGrepCommand.java       |  3 ++-
 .../core/util/shell/commands/HistoryCommand.java     |  4 +---
 .../util/shell/commands/QuotedStringTokenizer.java   | 15 ++++++---------
 .../examples/simple/mapreduce/TableToFile.java       |  4 ++--
 .../java/org/apache/accumulo/server/Accumulo.java    |  4 ++--
 .../accumulo/server/tabletserver/Compactor.java      |  6 ++++--
 .../org/apache/accumulo/server/util/Initialize.java  |  3 ++-
 .../main/java/org/apache/accumulo/start/Main.java    |  4 ++--
 .../start/classloader/AccumuloClassLoader.java       |  4 ++--
 .../test/randomwalk/security/CreateTable.java        |  5 ++---
 .../test/randomwalk/security/CreateUser.java         |  5 ++---
 25 files changed, 65 insertions(+), 63 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
index 049dd85..8cb656c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
@@ -82,7 +82,7 @@ public interface InstanceOperations {
    * List the active scans on tablet server.
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @return A list of active scans on tablet server.
    */
 
@@ -92,7 +92,7 @@ public interface InstanceOperations {
    * List the active compaction running on a tablet server
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @return the list of active compactions
    * @since 1.5.0
    */
@@ -103,7 +103,7 @@ public interface InstanceOperations {
    * Throws an exception if a tablet server can not be contacted.
    *
    * @param tserver
-   *          The tablet server address should be of the form <ip address>:<port>
+   *          The tablet server address should be of the form {@code <ip address>:<port>}
    * @since 1.5.0
    */
   public void ping(String tserver) throws AccumuloException;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/data/Value.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Value.java b/core/src/main/java/org/apache/accumulo/core/data/Value.java
index ba89d6c..b937203 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Value.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Value.java
@@ -139,12 +139,13 @@ public class Value implements WritableComparable<Object> {
     return this.value.length;
   }
 
+  @Override
   public void readFields(final DataInput in) throws IOException {
     this.value = new byte[in.readInt()];
     in.readFully(this.value, 0, this.value.length);
   }
 
-  /** {@inheritDoc} */
+  @Override
   public void write(final DataOutput out) throws IOException {
     out.writeInt(this.value.length);
     out.write(this.value, 0, this.value.length);
@@ -152,7 +153,6 @@ public class Value implements WritableComparable<Object> {
 
   // Below methods copied from BytesWritable
 
-  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     return WritableComparator.hashBytes(value, this.value.length);
@@ -165,6 +165,7 @@ public class Value implements WritableComparable<Object> {
    *          The other bytes writable
    * @return Positive if left is bigger than right, 0 if they are equal, and negative if left is smaller than right.
    */
+  @Override
   public int compareTo(Object right_obj) {
     return compareTo(((Value) right_obj).get());
   }
@@ -179,7 +180,6 @@ public class Value implements WritableComparable<Object> {
     return (diff != 0) ? diff : WritableComparator.compareBytes(this.value, 0, this.value.length, that, 0, that.length);
   }
 
-  /** {@inheritDoc} */
   @Override
   public boolean equals(Object right_obj) {
     if (right_obj instanceof byte[]) {
@@ -207,7 +207,6 @@ public class Value implements WritableComparable<Object> {
       super(Value.class);
     }
 
-    /** {@inheritDoc} */
     @Override
     public int compare(byte[] b1, int s1, int l1, byte[] b2, int s2, int l2) {
       return comparator.compare(b1, s1, l1, b2, s2, l2);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
index bd9fa43..8d54088 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
@@ -59,8 +59,8 @@ public class CreateEmpty {
   static class Opts extends Help {
     @Parameter(names = {"-c", "--codec"}, description = "the compression codec to use.", validateWith = IsSupportedCompressionAlgorithm.class)
     String codec = TFile.COMPRESSION_NONE;
-    @Parameter(
-        description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
+    @Parameter(description = " <path> { <path> ... } Each path given is a URL. "
+        + "Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
         required = true, validateWith = NamedLikeRFile.class)
     List<String> files = new ArrayList<String>();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
index e21598a..400695b 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
@@ -1065,9 +1065,9 @@ public class TFile {
        * @param reader
        *          The TFile reader object.
        * @param beginKey
-       *          Begin key of the scan. If null, scan from the first <K,V> entry of the TFile.
+       *          Begin key of the scan. If null, scan from the first {@code <K,V>} entry of the TFile.
        * @param endKey
-       *          End key of the scan. If null, scan up to the last <K, V> entry of the TFile.
+       *          End key of the scan. If null, scan up to the last {@code <K, V>} entry of the TFile.
        */
       protected Scanner(Reader reader, RawComparable beginKey, RawComparable endKey) throws IOException {
         this(reader, (beginKey == null) ? reader.begin() : reader.getBlockContainsKey(beginKey, false), reader.end());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
index fb0277a..84b861b 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/Utils.java
@@ -164,6 +164,7 @@ public final class Utils {
    * <li>if (FB in [-104, -73]), return (FB+88)<<16 + (NB[0]&0xff)<<8 + NB[1]&0xff;
    * <li>if (FB in [-120, -105]), return (FB+112)<<24 + (NB[0]&0xff)<<16 + (NB[1]&0xff)<<8 + NB[2]&0xff;
    * <li>if (FB in [-128, -121]), return interpret NB[FB+129] as a signed big-endian integer.
+   * </ul>
    *
    * @param in
    *          input stream

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
index 41e4e1e..f75076d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
@@ -70,7 +70,7 @@ public abstract class Combiner extends WrappingIterator implements OptionDescrib
      * Constructs an iterator over Values whose Keys are versions of the current topKey of the source SortedKeyValueIterator.
      *
      * @param source
-     *          The SortedKeyValueIterator<Key,Value> from which to read data.
+     *          The {@code SortedKeyValueIterator<Key,Value>} from which to read data.
      */
     public ValueIterator(SortedKeyValueIterator<Key,Value> source) {
       this.source = source;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
index 0bffbf7..cfdfd6e 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.io.WritableUtils;
 /**
  * A TypedValueCombiner that translates each Value to a Long before reducing, then encodes the reduced Long back to a Value.
  *
- * Subclasses must implement a typedReduce method: public Long typedReduce(Key key, Iterator<Long> iter);
+ * Subclasses must implement a typedReduce method: {@code public Long typedReduce(Key key, Iterator<Long> iter);}
  *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Longs) for all non-deleted versions of that Key.
  *
@@ -226,7 +226,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
-   *          Class<? extends Encoder<Long>> specifying the encoding type.
+   *          {@code Class<? extends Encoder<Long>>} specifying the encoding type.
    */
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<Long>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
index 6158bfc..6593ecd 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
@@ -50,8 +50,8 @@ public interface OptionDescriber {
      * @param unnamedOptionDescriptions
      *          is a list of descriptions of additional options that don't have fixed names (null if unused). The descriptions are intended to describe a
      *          category, and the user will provide parameter names and values in that category; e.g., the FilteringIterator needs a list of Filters intended to
-     *          be named by their priority numbers, so its unnamedOptionDescriptions =
-     *          Collections.singletonList("<filterPriorityNumber> <ageoff|regex|filterClass>")
+     *          be named by their priority numbers, so its<br>
+     *          {@code unnamedOptionDescriptions = Collections.singletonList("<filterPriorityNumber> <ageoff|regex|filterClass>")}
      */
     public IteratorOptions(String name, String description, Map<String,String> namedOptions, List<String> unnamedOptionDescriptions) {
       this.name = name;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
index d1ae9f5..0f26fab 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
@@ -29,7 +29,7 @@ import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
 /**
  * A Combiner that decodes each Value to type V before reducing, then encodes the result of typedReduce back to Value.
  *
- * Subclasses must implement a typedReduce method: public V typedReduce(Key key, Iterator<V> iter);
+ * Subclasses must implement a typedReduce method: {@code public V typedReduce(Key key, Iterator<V> iter);}
  *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Vs) for all non-deleted versions of that Key.
  *
@@ -42,7 +42,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   protected static final String LOSSY = "lossy";
 
   /**
-   * A Java Iterator that translates an Iterator<Value> to an Iterator<V> using the decode method of an Encoder.
+   * A Java Iterator that translates an {@code Iterator<Value>} to an {@code Iterator<V>} using the decode method of an Encoder.
    */
   private static class VIterator<V> implements Iterator<V> {
     private Iterator<Value> source;
@@ -50,7 +50,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     private boolean lossy;
 
     /**
-     * Constructs an Iterator<V> from an Iterator<Value>
+     * Constructs an {@code Iterator<V>} from an {@code Iterator<Value>}
      *
      * @param iter
      *          The source iterator
@@ -114,14 +114,14 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   }
 
   /**
-   * Sets the Encoder<V> used to translate Values to V and back.
+   * Sets the {@code Encoder<V>} used to translate Values to V and back.
    */
   protected void setEncoder(Encoder<V> encoder) {
     this.encoder = encoder;
   }
 
   /**
-   * Instantiates and sets the Encoder<V> used to translate Values to V and back.
+   * Instantiates and sets the {@code Encoder<V>} used to translate Values to V and back.
    *
    * @throws IllegalArgumentException
    *           if ClassNotFoundException, InstantiationException, or IllegalAccessException occurs

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
index efced19..c2023c2 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
@@ -122,7 +122,8 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
     io.setName("sumarray");
-    io.setDescription("SummingArrayCombiner can interpret Values as arrays of Longs using a variety of encodings (arrays of variable length longs or fixed length longs, or comma-separated strings) before summing element-wise.");
+    io.setDescription("SummingArrayCombiner can interpret Values as arrays of Longs using a variety of encodings "
+        + "(arrays of variable length longs or fixed length longs, or comma-separated strings) before summing element-wise.");
     io.addNamedOption(TYPE, "<VARLEN|FIXEDLEN|STRING|fullClassName>");
     return io;
   }
@@ -248,7 +249,7 @@ public class SummingArrayCombiner extends TypedValueCombiner<List<Long>> {
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
-   *          Class<? extends Encoder<List<Long>>> specifying the encoding type.
+   *          {@code Class<? extends Encoder<List<Long>>>} specifying the encoding type.
    */
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<List<Long>>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
index e0c8e17..78ad4c9 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/ColumnVisibility.java
@@ -38,24 +38,24 @@ import org.apache.hadoop.io.WritableComparator;
  * definition of an expression.
  *
  * <P>
- * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&" and "|" indicating that both operands are
+ * The expression is a sequence of characters from the set [A-Za-z0-9_-.] along with the binary operators "&amp;" and "|" indicating that both operands are
  * necessary, or the either is necessary. The following are valid expressions for visibility:
  *
  * <pre>
  * A
  * A|B
- * (A|B)&(C|D)
- * orange|(red&yellow)
+ * (A|B)&amp;(C|D)
+ * orange|(red&amp;yellow)
  * </pre>
  *
  * <P>
  * The following are not valid expressions for visibility:
  *
  * <pre>
- * A|B&C
+ * A|B&amp;C
  * A=B
  * A|B|
- * A&|B
+ * A&amp;|B
  * ()
  * )
  * dog|!cat

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
index bb72ce5..4fd367c 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
@@ -204,7 +204,8 @@ public class DefaultSecretKeyEncryptionStrategy implements SecretKeyEncryptionSt
       if (!fs.exists(pathToKey)) {
 
         if (encryptionMode == Cipher.DECRYPT_MODE) {
-          log.error("There was a call to decrypt the session key but no key encryption key exists.  Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
+          log.error("There was a call to decrypt the session key but no key encryption key exists.  "
+              + "Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
           throw new RuntimeException("Could not find key encryption key file in configured location in HDFS (" + pathToKeyName + ")");
         } else {
           initializeKeyEncryptingKey(fs, pathToKey, context);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
index d60d076..89a380f 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
@@ -36,16 +36,19 @@ public class BinaryFormatter implements Formatter {
     doTimestamps = printTimestamps;
   }
 
+  @Override
   public boolean hasNext() {
     checkState(si, true);
     return si.hasNext();
   }
 
+  @Override
   public String next() {
     checkState(si, true);
     return formatEntry(si.next(), doTimestamps);
   }
 
+  @Override
   public void remove() {
     checkState(si, true);
     si.remove();
@@ -108,9 +111,7 @@ public class BinaryFormatter implements Formatter {
           sb.append("\\x").append(String.format("%02X", c));
       }
       return sb;
-    }
-
-    else {
+    } else {
       for (int i = 0; i < len; i++) {
 
         int c = 0xff & ba[offset + i];

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
index ca80e37..8215a5a 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
@@ -38,6 +38,7 @@ public class DUCommand extends Command {
 
   private Option optTablePattern, optHumanReadble;
 
+  @Override
   public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException, TableNotFoundException {
 
     final SortedSet<String> tablesToFlush = new TreeSet<String>(Arrays.asList(cl.getArgs()));
@@ -79,7 +80,8 @@ public class DUCommand extends Command {
 
   @Override
   public String description() {
-    return "prints how much space, in bytes, is used by files referenced by a table.  When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
+    return "prints how much space, in bytes, is used by files referenced by a table.  "
+        + "When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
index b2b2663..16cc5d1 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/EGrepCommand.java
@@ -41,7 +41,8 @@ public class EGrepCommand extends GrepCommand {
 
   @Override
   public String description() {
-    return "searches each row, column family, column qualifier and value, in parallel, on the server side (using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
+    return "searches each row, column family, column qualifier and value, in parallel, on the server side "
+        + "(using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
index 0ad94f1..145bb75 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
@@ -61,9 +61,7 @@ public class HistoryCommand extends Command {
           out.close();
         }
       }
-    }
-
-    else {
+    } else {
       BufferedReader in = null;
       try {
         in = new BufferedReader(new InputStreamReader(new FileInputStream(histDir + "/shell_history.txt"), UTF_8));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
index 18de460..f2b78cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/commands/QuotedStringTokenizer.java
@@ -73,9 +73,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
         } else {
           throw new BadArgumentException("can only escape single quotes, double quotes, the space character, the backslash, and hex input", input, i);
         }
-      }
-      // in a hex escape sequence
-      else if (hexChars != null) {
+      } else if (hexChars != null) {
+        // in a hex escape sequence
         final int digit = Character.digit(ch, 16);
         if (digit < 0) {
           throw new BadArgumentException("expected hex character", input, i);
@@ -93,9 +92,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
           token[tokenLength++] = b;
           hexChars = null;
         }
-      }
-      // in a quote, either end the quote, start escape, or continue a token
-      else if (inQuote) {
+      } else if (inQuote) {
+        // in a quote, either end the quote, start escape, or continue a token
         if (ch == inQuoteChar) {
           inQuote = false;
           tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
@@ -105,9 +103,8 @@ public class QuotedStringTokenizer implements Iterable<String> {
         } else {
           token[tokenLength++] = inputBytes[i];
         }
-      }
-      // not in a quote, either enter a quote, end a token, start escape, or continue a token
-      else {
+      } else {
+        // not in a quote, either enter a quote, end a token, start escape, or continue a token
         if (ch == '\'' || ch == '"') {
           if (tokenLength > 0) {
             tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
index 79aae12..a785727 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
@@ -41,8 +41,8 @@ import org.apache.hadoop.util.ToolRunner;
 import com.beust.jcommander.Parameter;
 
 /**
- * Takes a table and outputs the specified column to a set of part files on hdfs accumulo accumulo.examples.mapreduce.TableToFile <username> <password>
- * <tablename> <column> <hdfs-output-path>
+ * Takes a table and outputs the specified column to a set of part files on hdfs
+ * {@code accumulo accumulo.examples.mapreduce.TableToFile <username> <password> <tablename> <column> <hdfs-output-path>}
  */
 public class TableToFile extends Configured implements Tool {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/Accumulo.java b/server/src/main/java/org/apache/accumulo/server/Accumulo.java
index 0509601..1727bec 100644
--- a/server/src/main/java/org/apache/accumulo/server/Accumulo.java
+++ b/server/src/main/java/org/apache/accumulo/server/Accumulo.java
@@ -299,8 +299,8 @@ public class Accumulo {
       final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(
           ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE, ZooReaderWriter.getRetryingInstance()));
       if (!(fate.list().isEmpty())) {
-        throw new AccumuloException(
-            "Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
+        throw new AccumuloException("Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. "
+            + "Please see the README document for instructions on what to do under your previous version.");
       }
     } catch (Exception exception) {
       log.fatal("Problem verifying Fate readiness", exception);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java b/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
index c0c1e4b..9569e9a 100644
--- a/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
+++ b/server/src/main/java/org/apache/accumulo/server/tabletserver/Compactor.java
@@ -75,6 +75,7 @@ public class Compactor implements Callable<CompactionStats> {
 
     private long count;
 
+    @Override
     public CountingIterator deepCopy(IteratorEnvironment env) {
       return new CountingIterator(this, env);
     }
@@ -183,7 +184,7 @@ public class Compactor implements Callable<CompactionStats> {
 
       CompactionReason reason;
 
-      if (compactor.imm != null)
+      if (compactor.imm != null) {
         switch (compactor.mincReason) {
           case USER:
             reason = CompactionReason.USER;
@@ -196,7 +197,7 @@ public class Compactor implements Callable<CompactionStats> {
             reason = CompactionReason.SYSTEM;
             break;
         }
-      else
+      } else {
         switch (compactor.reason) {
           case USER:
             reason = CompactionReason.USER;
@@ -212,6 +213,7 @@ public class Compactor implements Callable<CompactionStats> {
             reason = CompactionReason.SYSTEM;
             break;
         }
+      }
 
       List<IterInfo> iiList = new ArrayList<IterInfo>();
       Map<String,Map<String,String>> iterOptions = new HashMap<String,Map<String,String>>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
----------------------------------------------------------------------
diff --git a/server/src/main/java/org/apache/accumulo/server/util/Initialize.java b/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
index e626fd8..31d53eb 100644
--- a/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
+++ b/server/src/main/java/org/apache/accumulo/server/util/Initialize.java
@@ -157,7 +157,8 @@ public class Initialize {
       c.printNewline();
       c.printString("   bin/accumulo " + org.apache.accumulo.server.util.ChangeSecret.class.getName() + " oldPassword newPassword.");
       c.printNewline();
-      c.printString("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. Without this accumulo will not operate correctly");
+      c.printString("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. "
+          + "Without this accumulo will not operate correctly");
       c.printNewline();
     }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/Main.java b/start/src/main/java/org/apache/accumulo/start/Main.java
index f9b1ab9..a80ebe6 100644
--- a/start/src/main/java/org/apache/accumulo/start/Main.java
+++ b/start/src/main/java/org/apache/accumulo/start/Main.java
@@ -141,7 +141,7 @@ public class Main {
   }
 
   private static void printUsage() {
-    System.out
-        .println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info | tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
+    System.out.println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info "
+        + "| tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
index 2673b5a..d9e2821 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
@@ -71,8 +71,8 @@ public class AccumuloClassLoader {
   }
 
   /**
-   * Parses and XML Document for a property node for a <name> with the value propertyName if it finds one the function return that property's value for its
-   * <value> node. If not found the function will return null
+   * Parses and XML Document for a property node for a &lt;name&gt; with the value propertyName if it finds one the function return that property's value for its
+   * &lt;value&gt; node. If not found the function will return null
    *
    * @param d
    *          XMLDocument to search through

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
index 90d7735..61b146a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
@@ -44,9 +44,8 @@ public class CreateTable extends Test {
       if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
         if (hasPermission)
           throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-        else
-        // create table anyway for sake of state
-        {
+        else {
+          // create table anyway for sake of state
           try {
             state.getConnector().tableOperations().create(tableName);
             WalkingSecurity.get(state).initTable(tableName);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/901d60ef/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
index 4ec9d22..1f539ff 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
@@ -43,9 +43,8 @@ public class CreateUser extends Test {
         case PERMISSION_DENIED:
           if (hasPermission)
             throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-          else
-          // create user anyway for sake of state
-          {
+          else {
+            // create user anyway for sake of state
             if (!exists) {
               state.getConnector().securityOperations().createLocalUser(tableUserName, tabUserPass);
               WalkingSecurity.get(state).createUser(tableUserName, tabUserPass);


[04/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/TableOp.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/TableOp.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/TableOp.java
index 2499af2..2612fc9 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/TableOp.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/TableOp.java
@@ -53,11 +53,11 @@ import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 
 public class TableOp extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state,env).getTabUserName(), WalkingSecurity.get(state,env).getTabToken());
-    
+    Connector conn = env.getInstance().getConnector(WalkingSecurity.get(state, env).getTabUserName(), WalkingSecurity.get(state, env).getTabToken());
+
     String action = props.getProperty("action", "_random");
     TablePermission tp;
     if ("_random".equalsIgnoreCase(action)) {
@@ -66,18 +66,18 @@ public class TableOp extends Test {
     } else {
       tp = TablePermission.valueOf(action);
     }
-    
-    boolean tableExists = WalkingSecurity.get(state,env).getTableExists();
-    String tableName = WalkingSecurity.get(state,env).getTableName();
-    String namespaceName = WalkingSecurity.get(state,env).getNamespaceName();
-    
+
+    boolean tableExists = WalkingSecurity.get(state, env).getTableExists();
+    String tableName = WalkingSecurity.get(state, env).getTableName();
+    String namespaceName = WalkingSecurity.get(state, env).getNamespaceName();
+
     switch (tp) {
       case READ: {
-        boolean canRead = WalkingSecurity.get(state,env).canScan(WalkingSecurity.get(state,env).getTabCredentials(), tableName, namespaceName);
-        Authorizations auths = WalkingSecurity.get(state,env).getUserAuthorizations(WalkingSecurity.get(state,env).getTabCredentials());
-        boolean ambiguousZone = WalkingSecurity.get(state,env).inAmbiguousZone(conn.whoami(), tp);
-        boolean ambiguousAuths = WalkingSecurity.get(state,env).ambiguousAuthorizations(conn.whoami());
-        
+        boolean canRead = WalkingSecurity.get(state, env).canScan(WalkingSecurity.get(state, env).getTabCredentials(), tableName, namespaceName);
+        Authorizations auths = WalkingSecurity.get(state, env).getUserAuthorizations(WalkingSecurity.get(state, env).getTabCredentials());
+        boolean ambiguousZone = WalkingSecurity.get(state, env).inAmbiguousZone(conn.whoami(), tp);
+        boolean ambiguousAuths = WalkingSecurity.get(state, env).ambiguousAuthorizations(conn.whoami());
+
         Scanner scan = null;
         try {
           scan = conn.createScanner(tableName, conn.securityOperations().getUserAuthorizations(conn.whoami()));
@@ -92,7 +92,7 @@ public class TableOp extends Test {
           }
           if (!canRead && !ambiguousZone)
             throw new AccumuloException("Was able to read when I shouldn't have had the perm with connection user " + conn.whoami() + " table " + tableName);
-          for (Entry<String,Integer> entry : WalkingSecurity.get(state,env).getAuthsMap().entrySet()) {
+          for (Entry<String,Integer> entry : WalkingSecurity.get(state, env).getAuthsMap().entrySet()) {
             if (auths.contains(entry.getKey().getBytes(UTF_8)))
               seen = seen - entry.getValue();
           }
@@ -131,25 +131,25 @@ public class TableOp extends Test {
             else
               throw new AccumuloException("Mismatched authorizations! ", re.getCause());
           }
-          
+
           throw new AccumuloException("Unexpected exception!", re);
         } finally {
           if (scan != null) {
             scan.close();
             scan = null;
           }
-          
+
         }
-        
+
         break;
       }
       case WRITE:
-        boolean canWrite = WalkingSecurity.get(state,env).canWrite(WalkingSecurity.get(state,env).getTabCredentials(), tableName, namespaceName);
-        boolean ambiguousZone = WalkingSecurity.get(state,env).inAmbiguousZone(conn.whoami(), tp);
-        
-        String key = WalkingSecurity.get(state,env).getLastKey() + "1";
+        boolean canWrite = WalkingSecurity.get(state, env).canWrite(WalkingSecurity.get(state, env).getTabCredentials(), tableName, namespaceName);
+        boolean ambiguousZone = WalkingSecurity.get(state, env).inAmbiguousZone(conn.whoami(), tp);
+
+        String key = WalkingSecurity.get(state, env).getLastKey() + "1";
         Mutation m = new Mutation(new Text(key));
-        for (String s : WalkingSecurity.get(state,env).getAuthsArray()) {
+        for (String s : WalkingSecurity.get(state, env).getAuthsArray()) {
           m.put(new Text(), new Text(), new ColumnVisibility(s), new Value("value".getBytes(UTF_8)));
         }
         BatchWriter writer = null;
@@ -170,7 +170,7 @@ public class TableOp extends Test {
             // For now, just wait a second and go again if they can write!
             if (!canWrite)
               return;
-            
+
             if (ambiguousZone) {
               Thread.sleep(1000);
               try {
@@ -184,8 +184,8 @@ public class TableOp extends Test {
             }
           }
           if (works)
-            for (String s : WalkingSecurity.get(state,env).getAuthsArray())
-              WalkingSecurity.get(state,env).increaseAuthMap(s, 1);
+            for (String s : WalkingSecurity.get(state, env).getAuthsArray())
+              WalkingSecurity.get(state, env).increaseAuthMap(s, 1);
         } finally {
           if (writer != null) {
             writer.close();
@@ -194,15 +194,15 @@ public class TableOp extends Test {
         }
         break;
       case BULK_IMPORT:
-        key = WalkingSecurity.get(state,env).getLastKey() + "1";
+        key = WalkingSecurity.get(state, env).getLastKey() + "1";
         SortedSet<Key> keys = new TreeSet<Key>();
-        for (String s : WalkingSecurity.get(state,env).getAuthsArray()) {
+        for (String s : WalkingSecurity.get(state, env).getAuthsArray()) {
           Key k = new Key(key, "", "", s);
           keys.add(k);
         }
         Path dir = new Path("/tmp", "bulk_" + UUID.randomUUID().toString());
         Path fail = new Path(dir.toString() + "_fail");
-        FileSystem fs = WalkingSecurity.get(state,env).getFs();
+        FileSystem fs = WalkingSecurity.get(state, env).getFs();
         FileSKVWriter f = FileOperations.getInstance().openWriter(dir + "/securityBulk." + RFile.EXTENSION, fs, fs.getConf(),
             AccumuloConfiguration.getDefaultConfiguration());
         f.startDefaultLocalityGroup();
@@ -218,28 +218,28 @@ public class TableOp extends Test {
           return;
         } catch (AccumuloSecurityException ae) {
           if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
-            if (WalkingSecurity.get(state,env).canBulkImport(WalkingSecurity.get(state,env).getTabCredentials(), tableName, namespaceName))
+            if (WalkingSecurity.get(state, env).canBulkImport(WalkingSecurity.get(state, env).getTabCredentials(), tableName, namespaceName))
               throw new AccumuloException("Bulk Import failed when it should have worked: " + tableName);
             return;
           } else if (ae.getSecurityErrorCode().equals(SecurityErrorCode.BAD_CREDENTIALS)) {
-            if (WalkingSecurity.get(state,env).userPassTransient(conn.whoami()))
+            if (WalkingSecurity.get(state, env).userPassTransient(conn.whoami()))
               return;
           }
           throw new AccumuloException("Unexpected exception!", ae);
         }
-        for (String s : WalkingSecurity.get(state,env).getAuthsArray())
-          WalkingSecurity.get(state,env).increaseAuthMap(s, 1);
+        for (String s : WalkingSecurity.get(state, env).getAuthsArray())
+          WalkingSecurity.get(state, env).increaseAuthMap(s, 1);
         fs.delete(dir, true);
         fs.delete(fail, true);
-        
-        if (!WalkingSecurity.get(state,env).canBulkImport(WalkingSecurity.get(state,env).getTabCredentials(), tableName, namespaceName))
+
+        if (!WalkingSecurity.get(state, env).canBulkImport(WalkingSecurity.get(state, env).getTabCredentials(), tableName, namespaceName))
           throw new AccumuloException("Bulk Import succeeded when it should have failed: " + dir + " table " + tableName);
         break;
       case ALTER_TABLE:
         AlterTable.renameTable(conn, state, env, tableName, tableName + "plus",
-            WalkingSecurity.get(state,env).canAlterTable(WalkingSecurity.get(state,env).getTabCredentials(), tableName, namespaceName), tableExists);
+            WalkingSecurity.get(state, env).canAlterTable(WalkingSecurity.get(state, env).getTabCredentials(), tableName, namespaceName), tableExists);
         break;
-      
+
       case GRANT:
         props.setProperty("task", "grant");
         props.setProperty("perm", "random");
@@ -247,7 +247,7 @@ public class TableOp extends Test {
         props.setProperty("target", "system");
         AlterTablePerm.alter(state, env, props);
         break;
-      
+
       case DROP_TABLE:
         props.setProperty("source", "table");
         DropTable.dropTable(state, env, props);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Validate.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Validate.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Validate.java
index 905327a..d2fc795 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Validate.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/Validate.java
@@ -21,8 +21,8 @@ import java.util.Properties;
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.Connector;
-import org.apache.accumulo.core.client.security.SecurityErrorCode;
 import org.apache.accumulo.core.client.impl.thrift.ThriftSecurityException;
+import org.apache.accumulo.core.client.security.SecurityErrorCode;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
@@ -32,34 +32,34 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.log4j.Logger;
 
 public class Validate extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     validate(state, env, log);
   }
-  
+
   public static void validate(State state, Environment env, Logger log) throws Exception {
     Connector conn = env.getConnector();
-    
-    boolean tableExists = WalkingSecurity.get(state,env).getTableExists();
-    boolean cloudTableExists = conn.tableOperations().list().contains(WalkingSecurity.get(state,env).getTableName());
+
+    boolean tableExists = WalkingSecurity.get(state, env).getTableExists();
+    boolean cloudTableExists = conn.tableOperations().list().contains(WalkingSecurity.get(state, env).getTableName());
     if (tableExists != cloudTableExists)
       throw new AccumuloException("Table existance out of sync");
-    
-    boolean tableUserExists = WalkingSecurity.get(state,env).userExists(WalkingSecurity.get(state,env).getTabUserName());
-    boolean cloudTableUserExists = conn.securityOperations().listLocalUsers().contains(WalkingSecurity.get(state,env).getTabUserName());
+
+    boolean tableUserExists = WalkingSecurity.get(state, env).userExists(WalkingSecurity.get(state, env).getTabUserName());
+    boolean cloudTableUserExists = conn.securityOperations().listLocalUsers().contains(WalkingSecurity.get(state, env).getTabUserName());
     if (tableUserExists != cloudTableUserExists)
       throw new AccumuloException("Table User existance out of sync");
-    
+
     Properties props = new Properties();
     props.setProperty("target", "system");
     Authenticate.authenticate(env.getUserName(), env.getToken(), state, env, props);
     props.setProperty("target", "table");
     Authenticate.authenticate(env.getUserName(), env.getToken(), state, env, props);
-    
-    for (String user : new String[] {WalkingSecurity.get(state,env).getSysUserName(), WalkingSecurity.get(state,env).getTabUserName()}) {
+
+    for (String user : new String[] {WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env).getTabUserName()}) {
       for (SystemPermission sp : SystemPermission.values()) {
-        boolean hasSp = WalkingSecurity.get(state,env).hasSystemPermission(user, sp);
+        boolean hasSp = WalkingSecurity.get(state, env).hasSystemPermission(user, sp);
         boolean accuHasSp;
         try {
           accuHasSp = conn.securityOperations().hasSystemPermission(user, sp);
@@ -76,12 +76,12 @@ public class Validate extends Test {
         if (hasSp != accuHasSp)
           throw new AccumuloException(user + " existance out of sync for system perm " + sp + " hasSp/CloudhasSP " + hasSp + " " + accuHasSp);
       }
-      
+
       for (TablePermission tp : TablePermission.values()) {
-        boolean hasTp = WalkingSecurity.get(state,env).hasTablePermission(user, WalkingSecurity.get(state,env).getTableName(), tp);
+        boolean hasTp = WalkingSecurity.get(state, env).hasTablePermission(user, WalkingSecurity.get(state, env).getTableName(), tp);
         boolean accuHasTp;
         try {
-          accuHasTp = conn.securityOperations().hasTablePermission(user, WalkingSecurity.get(state,env).getTableName(), tp);
+          accuHasTp = conn.securityOperations().hasTablePermission(user, WalkingSecurity.get(state, env).getTableName(), tp);
           log.debug("Just checked to see if user " + user + " has table perm " + tp.name() + " with answer " + accuHasTp);
         } catch (AccumuloSecurityException ae) {
           if (ae.getSecurityErrorCode().equals(SecurityErrorCode.USER_DOESNT_EXIST)) {
@@ -100,14 +100,14 @@ public class Validate extends Test {
         if (hasTp != accuHasTp)
           throw new AccumuloException(user + " existance out of sync for table perm " + tp + " hasTp/CloudhasTP " + hasTp + " " + accuHasTp);
       }
-      
+
     }
-    
+
     Authorizations accuAuths;
     Authorizations auths;
     try {
-      auths = WalkingSecurity.get(state,env).getUserAuthorizations(WalkingSecurity.get(state,env).getTabCredentials());
-      accuAuths = conn.securityOperations().getUserAuthorizations(WalkingSecurity.get(state,env).getTabUserName());
+      auths = WalkingSecurity.get(state, env).getUserAuthorizations(WalkingSecurity.get(state, env).getTabCredentials());
+      accuAuths = conn.securityOperations().getUserAuthorizations(WalkingSecurity.get(state, env).getTabUserName());
     } catch (ThriftSecurityException ae) {
       if (ae.getCode() == org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode.USER_DOESNT_EXIST) {
         if (tableUserExists)
@@ -120,5 +120,5 @@ public class Validate extends Test {
     if (!auths.equals(accuAuths))
       throw new AccumuloException("Table User authorizations out of sync");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/security/WalkingSecurity.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/WalkingSecurity.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/WalkingSecurity.java
index 0fb3a82..9d285e0 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/WalkingSecurity.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/WalkingSecurity.java
@@ -53,7 +53,7 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.log4j.Logger;
 
 /**
- * 
+ *
  */
 public class WalkingSecurity extends SecurityOperation implements Authorizor, Authenticator, PermissionHandler {
   State state = null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/BatchVerify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/BatchVerify.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/BatchVerify.java
index 6a725c7..8616655 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/BatchVerify.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/BatchVerify.java
@@ -20,9 +20,9 @@ import java.util.ArrayList;
 import java.util.Collections;
 import java.util.Iterator;
 import java.util.List;
+import java.util.Map.Entry;
 import java.util.Properties;
 import java.util.Random;
-import java.util.Map.Entry;
 
 import org.apache.accumulo.core.client.BatchScanner;
 import org.apache.accumulo.core.client.Connector;
@@ -36,23 +36,23 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class BatchVerify extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     Random rand = new Random();
-    
+
     long numWrites = state.getLong("numWrites");
     int maxVerify = Integer.parseInt(props.getProperty("maxVerify", "2000"));
     long numVerify = rand.nextInt(maxVerify - 1) + 1;
-    
+
     if (numVerify > (numWrites / 4)) {
       numVerify = numWrites / 4;
     }
-    
+
     Connector conn = env.getConnector();
     BatchScanner scanner = conn.createBatchScanner(state.getString("seqTableName"), new Authorizations(), 2);
-    
+
     try {
       int count = 0;
       List<Range> ranges = new ArrayList<Range>();
@@ -65,29 +65,29 @@ public class BatchVerify extends Test {
         count += rangeEnd - rangeStart + 1;
         ranges.add(new Range(new Text(String.format("%010d", rangeStart)), new Text(String.format("%010d", rangeEnd))));
       }
-      
+
       ranges = Range.mergeOverlapping(ranges);
       Collections.sort(ranges);
-      
+
       if (count == 0 || ranges.size() == 0)
         return;
-      
+
       log.debug(String.format("scanning %d rows in the following %d ranges:", count, ranges.size()));
       for (Range r : ranges) {
         log.debug(r);
       }
-      
+
       scanner.setRanges(ranges);
-      
+
       List<Key> keys = new ArrayList<Key>();
       for (Entry<Key,Value> entry : scanner) {
         keys.add(entry.getKey());
       }
-      
+
       log.debug("scan returned " + keys.size() + " rows. now verifying...");
-      
+
       Collections.sort(keys);
-      
+
       Iterator<Key> iterator = keys.iterator();
       int curKey = Integer.parseInt(iterator.next().getRow().toString());
       boolean done = false;
@@ -95,12 +95,12 @@ public class BatchVerify extends Test {
         int start = Integer.parseInt(r.getStartKey().getRow().toString());
         int end = Integer.parseInt(String.copyValueOf(r.getEndKey().getRow().toString().toCharArray(), 0, 10));
         for (int i = start; i <= end; i++) {
-          
+
           if (done) {
             log.error("missing key " + i);
             break;
           }
-          
+
           while (curKey < i) {
             log.error("extra key " + curKey);
             if (iterator.hasNext() == false) {
@@ -109,11 +109,11 @@ public class BatchVerify extends Test {
             }
             curKey = Integer.parseInt(iterator.next().getRow().toString());
           }
-          
+
           if (curKey > i) {
             log.error("missing key " + i);
           }
-          
+
           if (iterator.hasNext()) {
             curKey = Integer.parseInt(iterator.next().getRow().toString());
           } else {
@@ -121,7 +121,7 @@ public class BatchVerify extends Test {
           }
         }
       }
-      
+
       log.debug("verify is now complete");
     } finally {
       scanner.close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Commit.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Commit.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Commit.java
index 7d5102e..84adae1 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Commit.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Commit.java
@@ -23,14 +23,14 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Commit extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     env.getMultiTableBatchWriter().flush();
-    
+
     log.debug("Committed " + state.getLong("numWrites") + " writes.  Total writes: " + state.getLong("totalWrites"));
     state.set("numWrites", Long.valueOf(0));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerify.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerify.java
index 4337799..95c1d0b 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerify.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerify.java
@@ -16,8 +16,8 @@
  */
 package org.apache.accumulo.test.randomwalk.sequential;
 
-import java.util.Properties;
 import java.util.Map.Entry;
+import java.util.Properties;
 
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.Scanner;
@@ -32,10 +32,10 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.util.ToolRunner;
 
 public class MapRedVerify extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     String[] args = new String[8];
     args[0] = "-libjars";
     args[1] = getMapReduceJars();
@@ -45,15 +45,15 @@ public class MapRedVerify extends Test {
     args[5] = env.getInstance().getInstanceName();
     args[6] = env.getConfigProperty("ZOOKEEPERS");
     args[7] = args[4] + "_MR";
-    
+
     if (ToolRunner.run(CachedConfiguration.getInstance(), new MapRedVerifyTool(), args) != 0) {
       log.error("Failed to run map/red verify");
       return;
     }
-    
+
     Scanner outputScanner = env.getConnector().createScanner(args[7], Authorizations.EMPTY);
     outputScanner.setRange(new Range());
-    
+
     int count = 0;
     Key lastKey = null;
     for (Entry<Key,Value> entry : outputScanner) {
@@ -64,11 +64,11 @@ public class MapRedVerify extends Test {
       }
       lastKey = current;
     }
-    
+
     if (count > 1) {
       log.error("Gaps in output");
     }
-    
+
     log.debug("Dropping table: " + args[7]);
     Connector conn = env.getConnector();
     conn.tableOperations().delete(args[7]);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerifyTool.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerifyTool.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerifyTool.java
index 3a078ef..47b671c 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerifyTool.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/MapRedVerifyTool.java
@@ -38,7 +38,7 @@ import org.apache.log4j.Logger;
 
 public class MapRedVerifyTool extends Configured implements Tool {
   protected final Logger log = Logger.getLogger(this.getClass());
-  
+
   public static class SeqMapClass extends Mapper<Key,Value,NullWritable,IntWritable> {
     @Override
     public void map(Key row, Value data, Context output) throws IOException, InterruptedException {
@@ -46,16 +46,16 @@ public class MapRedVerifyTool extends Configured implements Tool {
       output.write(NullWritable.get(), new IntWritable(num.intValue()));
     }
   }
-  
+
   public static class SeqReduceClass extends Reducer<NullWritable,IntWritable,Text,Mutation> {
     @Override
     public void reduce(NullWritable ignore, Iterable<IntWritable> values, Context output) throws IOException, InterruptedException {
       Iterator<IntWritable> iterator = values.iterator();
-      
+
       if (iterator.hasNext() == false) {
         return;
       }
-      
+
       int start = iterator.next().get();
       int index = start;
       while (iterator.hasNext()) {
@@ -68,45 +68,45 @@ public class MapRedVerifyTool extends Configured implements Tool {
       }
       writeMutation(output, start, index);
     }
-    
+
     public void writeMutation(Context output, int start, int end) throws IOException, InterruptedException {
       Mutation m = new Mutation(new Text(String.format("%010d", start)));
       m.put(new Text(String.format("%010d", end)), new Text(""), new Value(new byte[0]));
       output.write(null, m);
     }
   }
-  
+
   @Override
   public int run(String[] args) throws Exception {
     @SuppressWarnings("deprecation")
     Job job = new Job(getConf(), this.getClass().getSimpleName());
     job.setJarByClass(this.getClass());
-    
+
     if (job.getJar() == null) {
       log.error("M/R requires a jar file!  Run mvn package.");
       return 1;
     }
-    
+
     ClientConfiguration clientConf = new ClientConfiguration().withInstance(args[3]).withZkHosts(args[4]);
 
     job.setInputFormatClass(AccumuloInputFormat.class);
     AccumuloInputFormat.setConnectorInfo(job, args[0], new PasswordToken(args[1]));
     AccumuloInputFormat.setInputTableName(job, args[2]);
     AccumuloInputFormat.setZooKeeperInstance(job, clientConf);
-    
+
     job.setMapperClass(SeqMapClass.class);
     job.setMapOutputKeyClass(NullWritable.class);
     job.setMapOutputValueClass(IntWritable.class);
-    
+
     job.setReducerClass(SeqReduceClass.class);
     job.setNumReduceTasks(1);
-    
+
     job.setOutputFormatClass(AccumuloOutputFormat.class);
     AccumuloOutputFormat.setConnectorInfo(job, args[0], new PasswordToken(args[1]));
     AccumuloOutputFormat.setCreateTables(job, true);
     AccumuloOutputFormat.setDefaultTableName(job, args[5]);
     AccumuloOutputFormat.setZooKeeperInstance(job, clientConf);
-    
+
     job.waitForCompletion(true);
     return job.isSuccessful() ? 0 : 1;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/SequentialFixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/SequentialFixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/SequentialFixture.java
index c2320a7..9565cc5 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/SequentialFixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/SequentialFixture.java
@@ -24,25 +24,25 @@ import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.impl.Tables;
-import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.Environment;
+import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.State;
 
 public class SequentialFixture extends Fixture {
-  
+
   String seqTableName;
-  
+
   @Override
   public void setUp(State state, Environment env) throws Exception {
-    
+
     Connector conn = env.getConnector();
     Instance instance = env.getInstance();
-    
+
     String hostname = InetAddress.getLocalHost().getHostName().replaceAll("[-.]", "_");
-    
+
     seqTableName = String.format("sequential_%s_%s_%d", hostname, env.getPid(), System.currentTimeMillis());
     state.set("seqTableName", seqTableName);
-    
+
     try {
       conn.tableOperations().create(seqTableName);
       log.debug("Created table " + seqTableName + " (id:" + Tables.getNameToIdMap(instance).get(seqTableName) + ")");
@@ -51,11 +51,11 @@ public class SequentialFixture extends Fixture {
       throw e;
     }
     conn.tableOperations().setProperty(seqTableName, "table.scan.max.memory", "1K");
-    
+
     state.set("numWrites", Long.valueOf(0));
     state.set("totalWrites", Long.valueOf(0));
   }
-  
+
   @Override
   public void tearDown(State state, Environment env) throws Exception {
     // We have resources we need to clean up
@@ -66,15 +66,15 @@ public class SequentialFixture extends Fixture {
       } catch (MutationsRejectedException e) {
         log.error("Ignoring mutations that weren't flushed", e);
       }
-      
+
       // Reset the MTBW on the state to null
       env.resetMultiTableBatchWriter();
     }
-    
+
     log.debug("Dropping tables: " + seqTableName);
-    
+
     Connector conn = env.getConnector();
-    
+
     conn.tableOperations().delete(seqTableName);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Write.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Write.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Write.java
index 1795434..5398d99 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Write.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/sequential/Write.java
@@ -29,20 +29,20 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Write extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     BatchWriter bw = env.getMultiTableBatchWriter().getBatchWriter(state.getString("seqTableName"));
-    
+
     state.set("numWrites", state.getLong("numWrites") + 1);
-    
+
     Long totalWrites = state.getLong("totalWrites") + 1;
     if (totalWrites % 10000 == 0) {
       log.debug("Total writes: " + totalWrites);
     }
     state.set("totalWrites", totalWrites);
-    
+
     Mutation m = new Mutation(new Text(String.format("%010d", totalWrites)));
     m.put(new Text("cf"), new Text("cq"), new Value("val".getBytes(UTF_8)));
     bw.addMutation(m);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/BulkInsert.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/BulkInsert.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/BulkInsert.java
index 5959105..1074e5a 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/BulkInsert.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/BulkInsert.java
@@ -49,23 +49,23 @@ import org.apache.hadoop.io.Text;
 import org.apache.hadoop.util.ToolRunner;
 
 public class BulkInsert extends Test {
-  
+
   class SeqfileBatchWriter implements BatchWriter {
-    
+
     SequenceFile.Writer writer;
-    
+
     @SuppressWarnings("deprecation")
     SeqfileBatchWriter(Configuration conf, FileSystem fs, String file) throws IOException {
       writer = new SequenceFile.Writer(fs, conf, new Path(file), Key.class, Value.class);
     }
-    
+
     @Override
     public void addMutation(Mutation m) throws MutationsRejectedException {
       List<ColumnUpdate> updates = m.getUpdates();
       for (ColumnUpdate cu : updates) {
         Key key = new Key(m.getRow(), cu.getColumnFamily(), cu.getColumnQualifier(), cu.getColumnVisibility(), Long.MAX_VALUE, false, false);
         Value val = new Value(cu.getValue(), false);
-        
+
         try {
           writer.append(key, val);
         } catch (IOException e) {
@@ -73,16 +73,16 @@ public class BulkInsert extends Test {
         }
       }
     }
-    
+
     @Override
     public void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException {
       for (Mutation mutation : iterable)
         addMutation(mutation);
     }
-    
+
     @Override
     public void flush() throws MutationsRejectedException {}
-    
+
     @Override
     public void close() throws MutationsRejectedException {
       try {
@@ -91,53 +91,53 @@ public class BulkInsert extends Test {
         throw new RuntimeException(e);
       }
     }
-    
+
   }
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     String indexTableName = (String) state.get("indexTableName");
     String dataTableName = (String) state.get("docTableName");
     int numPartitions = (Integer) state.get("numPartitions");
     Random rand = (Random) state.get("rand");
     long nextDocID = (Long) state.get("nextDocID");
-    
+
     int minInsert = Integer.parseInt(props.getProperty("minInsert"));
     int maxInsert = Integer.parseInt(props.getProperty("maxInsert"));
     int numToInsert = rand.nextInt(maxInsert - minInsert) + minInsert;
-    
+
     int maxSplits = Integer.parseInt(props.getProperty("maxSplits"));
-    
+
     Configuration conf = CachedConfiguration.getInstance();
     FileSystem fs = FileSystem.get(conf);
-    
+
     String rootDir = "/tmp/shard_bulk/" + dataTableName;
-    
+
     fs.mkdirs(new Path(rootDir));
-    
+
     BatchWriter dataWriter = new SeqfileBatchWriter(conf, fs, rootDir + "/data.seq");
     BatchWriter indexWriter = new SeqfileBatchWriter(conf, fs, rootDir + "/index.seq");
-    
+
     for (int i = 0; i < numToInsert; i++) {
       String docID = Insert.insertRandomDocument(nextDocID++, dataWriter, indexWriter, indexTableName, dataTableName, numPartitions, rand);
       log.debug("Bulk inserting document " + docID);
     }
-    
+
     state.set("nextDocID", Long.valueOf(nextDocID));
-    
+
     dataWriter.close();
     indexWriter.close();
-    
+
     sort(state, env, fs, dataTableName, rootDir + "/data.seq", rootDir + "/data_bulk", rootDir + "/data_work", maxSplits);
     sort(state, env, fs, indexTableName, rootDir + "/index.seq", rootDir + "/index_bulk", rootDir + "/index_work", maxSplits);
-    
+
     bulkImport(fs, state, env, dataTableName, rootDir, "data");
     bulkImport(fs, state, env, indexTableName, rootDir, "index");
-    
+
     fs.delete(new Path(rootDir), true);
   }
-  
+
   private void bulkImport(FileSystem fs, State state, Environment env, String tableName, String rootDir, String prefix) throws Exception {
     while (true) {
       String bulkDir = rootDir + "/" + prefix + "_bulk";
@@ -146,11 +146,11 @@ public class BulkInsert extends Test {
       fs.delete(failPath, true);
       fs.mkdirs(failPath);
       env.getConnector().tableOperations().importDirectory(tableName, bulkDir, failDir, true);
-      
+
       FileStatus[] failures = fs.listStatus(failPath);
       if (failures != null && failures.length > 0) {
         log.warn("Failed to bulk import some files, retrying ");
-        
+
         for (FileStatus failure : failures) {
           if (!failure.getPath().getName().endsWith(".seq"))
             fs.rename(failure.getPath(), new Path(new Path(bulkDir), failure.getPath().getName()));
@@ -162,28 +162,29 @@ public class BulkInsert extends Test {
         break;
     }
   }
-  
-  private void sort(State state, Environment env, FileSystem fs, String tableName, String seqFile, String outputDir, String workDir, int maxSplits) throws Exception {
-    
+
+  private void sort(State state, Environment env, FileSystem fs, String tableName, String seqFile, String outputDir, String workDir, int maxSplits)
+      throws Exception {
+
     PrintStream out = new PrintStream(new BufferedOutputStream(fs.create(new Path(workDir + "/splits.txt"))), false, UTF_8.name());
-    
+
     Connector conn = env.getConnector();
-    
+
     Collection<Text> splits = conn.tableOperations().listSplits(tableName, maxSplits);
     for (Text split : splits)
       out.println(Base64.encodeBase64String(TextUtil.getBytes(split)));
-    
+
     out.close();
-    
+
     SortTool sortTool = new SortTool(seqFile, outputDir, workDir + "/splits.txt", splits);
-    
+
     String[] args = new String[2];
     args[0] = "-libjars";
     args[1] = getMapReduceJars();
-    
+
     if (ToolRunner.run(CachedConfiguration.getInstance(), sortTool, args) != 0) {
       throw new Exception("Failed to run map/red verify");
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CloneIndex.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CloneIndex.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CloneIndex.java
index 77f34bc..0d3d38e 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CloneIndex.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CloneIndex.java
@@ -25,21 +25,21 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class CloneIndex extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     String indexTableName = (String) state.get("indexTableName");
     String tmpIndexTableName = indexTableName + "_tmp";
-    
+
     long t1 = System.currentTimeMillis();
     env.getConnector().tableOperations().flush(indexTableName, null, null, true);
     long t2 = System.currentTimeMillis();
     env.getConnector().tableOperations().clone(indexTableName, tmpIndexTableName, false, new HashMap<String,String>(), new HashSet<String>());
     long t3 = System.currentTimeMillis();
-    
+
     log.debug("Cloned " + tmpIndexTableName + " from " + indexTableName + " flush: " + (t2 - t1) + "ms clone: " + (t3 - t2) + "ms");
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Commit.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Commit.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Commit.java
index 8f928c7..d3f9ed7 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Commit.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Commit.java
@@ -23,11 +23,11 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Commit extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     env.getMultiTableBatchWriter().flush();
     log.debug("Committed inserts ");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CompactFilter.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CompactFilter.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CompactFilter.java
index 96c6c62..a73b311 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CompactFilter.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/CompactFilter.java
@@ -39,18 +39,18 @@ import org.apache.hadoop.io.Text;
  * Test deleting documents by using a compaction filter iterator
  */
 public class CompactFilter extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     String docTableName = (String) state.get("docTableName");
     Random rand = (Random) state.get("rand");
-    
+
     String deleteChar = Integer.toHexString(rand.nextInt(16)) + "";
     String regex = "^[0-9a-f][" + deleteChar + "].*";
 
     ArrayList<IteratorSetting> documentFilters = new ArrayList<IteratorSetting>();
-    
+
     IteratorSetting is = new IteratorSetting(21, "ii", RegExFilter.class);
     RegExFilter.setRegexs(is, regex, null, null, null, false);
     RegExFilter.setNegate(is, true);
@@ -60,35 +60,35 @@ public class CompactFilter extends Test {
     env.getConnector().tableOperations().compact(docTableName, null, null, documentFilters, true, true);
     long t2 = System.currentTimeMillis();
     long t3 = t2 - t1;
-    
+
     ArrayList<IteratorSetting> indexFilters = new ArrayList<IteratorSetting>();
-    
+
     is = new IteratorSetting(21, RegExFilter.class);
     RegExFilter.setRegexs(is, null, null, regex, null, false);
     RegExFilter.setNegate(is, true);
     indexFilters.add(is);
-    
+
     t1 = System.currentTimeMillis();
     env.getConnector().tableOperations().compact(indexTableName, null, null, indexFilters, true, true);
     t2 = System.currentTimeMillis();
-    
+
     log.debug("Filtered documents using compaction iterators " + regex + " " + (t3) + " " + (t2 - t1));
-    
+
     BatchScanner bscanner = env.getConnector().createBatchScanner(docTableName, new Authorizations(), 10);
-    
+
     List<Range> ranges = new ArrayList<Range>();
     for (int i = 0; i < 16; i++) {
       ranges.add(Range.prefix(new Text(Integer.toHexString(i) + "" + deleteChar)));
     }
-    
+
     bscanner.setRanges(ranges);
     Iterator<Entry<Key,Value>> iter = bscanner.iterator();
-    
+
     if (iter.hasNext()) {
       throw new Exception("Saw unexpected document " + iter.next().getKey());
     }
 
     bscanner.close();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Delete.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Delete.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Delete.java
index 4931af5..4cc0304 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Delete.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Delete.java
@@ -28,31 +28,31 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Delete extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     String dataTableName = (String) state.get("docTableName");
     int numPartitions = (Integer) state.get("numPartitions");
     Random rand = (Random) state.get("rand");
-    
+
     Entry<Key,Value> entry = Search.findRandomDocument(state, env, dataTableName, rand);
     if (entry == null)
       return;
-    
+
     String docID = entry.getKey().getRow().toString();
     String doc = entry.getValue().toString();
-    
+
     Insert.unindexDocument(env.getMultiTableBatchWriter().getBatchWriter(indexTableName), doc, docID, numPartitions);
-    
+
     Mutation m = new Mutation(docID);
     m.putDelete("doc", "");
-    
+
     env.getMultiTableBatchWriter().getBatchWriter(dataTableName).addMutation(m);
-    
+
     log.debug("Deleted document " + docID);
-    
+
     env.getMultiTableBatchWriter().flush();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteSomeDocs.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteSomeDocs.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteSomeDocs.java
index ed0a458..f7fb1df 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteSomeDocs.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteSomeDocs.java
@@ -33,48 +33,48 @@ import org.apache.accumulo.test.randomwalk.Test;
 
 //a test created to test the batch deleter
 public class DeleteSomeDocs extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     // delete documents that where the document id matches a given pattern from doc and index table
     // using the batch deleter
-    
+
     Random rand = (Random) state.get("rand");
     String indexTableName = (String) state.get("indexTableName");
     String dataTableName = (String) state.get("docTableName");
-    
+
     ArrayList<String> patterns = new ArrayList<String>();
-    
+
     for (Object key : props.keySet())
       if (key instanceof String && ((String) key).startsWith("pattern"))
         patterns.add(props.getProperty((String) key));
-    
+
     String pattern = patterns.get(rand.nextInt(patterns.size()));
     BatchWriterConfig bwc = new BatchWriterConfig();
     BatchDeleter ibd = env.getConnector().createBatchDeleter(indexTableName, Authorizations.EMPTY, 8, bwc);
     ibd.setRanges(Collections.singletonList(new Range()));
-    
+
     IteratorSetting iterSettings = new IteratorSetting(100, RegExFilter.class);
     RegExFilter.setRegexs(iterSettings, null, null, pattern, null, false);
-    
+
     ibd.addScanIterator(iterSettings);
-    
+
     ibd.delete();
-    
+
     ibd.close();
-    
+
     BatchDeleter dbd = env.getConnector().createBatchDeleter(dataTableName, Authorizations.EMPTY, 8, bwc);
     dbd.setRanges(Collections.singletonList(new Range()));
-    
+
     iterSettings = new IteratorSetting(100, RegExFilter.class);
     RegExFilter.setRegexs(iterSettings, pattern, null, null, null, false);
-    
+
     dbd.addScanIterator(iterSettings);
-    
+
     dbd.delete();
-    
+
     dbd.close();
-    
+
     log.debug("Deleted documents w/ id matching '" + pattern + "'");
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteWord.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteWord.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteWord.java
index 10bbff3..28b1899 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteWord.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/DeleteWord.java
@@ -17,8 +17,8 @@
 package org.apache.accumulo.test.randomwalk.shard;
 
 import java.util.ArrayList;
-import java.util.Properties;
 import java.util.Map.Entry;
+import java.util.Properties;
 import java.util.Random;
 
 import org.apache.accumulo.core.client.BatchScanner;
@@ -36,62 +36,62 @@ import org.apache.hadoop.io.Text;
 
 /**
  * Delete all documents containing a particular word.
- * 
+ *
  */
 
 public class DeleteWord extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     String docTableName = (String) state.get("docTableName");
     int numPartitions = (Integer) state.get("numPartitions");
     Random rand = (Random) state.get("rand");
-    
+
     String wordToDelete = Insert.generateRandomWord(rand);
-    
+
     // use index to find all documents containing word
     Scanner scanner = env.getConnector().createScanner(indexTableName, Authorizations.EMPTY);
     scanner.fetchColumnFamily(new Text(wordToDelete));
-    
+
     ArrayList<Range> documentsToDelete = new ArrayList<Range>();
-    
+
     for (Entry<Key,Value> entry : scanner)
       documentsToDelete.add(new Range(entry.getKey().getColumnQualifier()));
-    
+
     if (documentsToDelete.size() > 0) {
       // use a batch scanner to fetch all documents
       BatchScanner bscanner = env.getConnector().createBatchScanner(docTableName, Authorizations.EMPTY, 8);
       bscanner.setRanges(documentsToDelete);
-      
+
       BatchWriter ibw = env.getMultiTableBatchWriter().getBatchWriter(indexTableName);
       BatchWriter dbw = env.getMultiTableBatchWriter().getBatchWriter(docTableName);
-      
+
       int count = 0;
-      
+
       for (Entry<Key,Value> entry : bscanner) {
         String docID = entry.getKey().getRow().toString();
         String doc = entry.getValue().toString();
-        
+
         Insert.unindexDocument(ibw, doc, docID, numPartitions);
-        
+
         Mutation m = new Mutation(docID);
         m.putDelete("doc", "");
-        
+
         dbw.addMutation(m);
         count++;
       }
-      
+
       bscanner.close();
-      
+
       env.getMultiTableBatchWriter().flush();
-      
+
       if (count != documentsToDelete.size()) {
         throw new Exception("Batch scanner did not return expected number of docs " + count + " " + documentsToDelete.size());
       }
     }
-    
+
     log.debug("Deleted " + documentsToDelete.size() + " documents containing " + wordToDelete);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ExportIndex.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ExportIndex.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ExportIndex.java
index f5ede55..ca273ac 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ExportIndex.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ExportIndex.java
@@ -36,36 +36,36 @@ import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class ExportIndex extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     String indexTableName = (String) state.get("indexTableName");
     String tmpIndexTableName = indexTableName + "_tmp";
-    
+
     String exportDir = "/tmp/shard_export/" + indexTableName;
     String copyDir = "/tmp/shard_export/" + tmpIndexTableName;
-    
+
     FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
-    
+
     fs.delete(new Path("/tmp/shard_export/" + indexTableName), true);
     fs.delete(new Path("/tmp/shard_export/" + tmpIndexTableName), true);
-    
+
     // disable spits, so that splits can be compared later w/o worrying one table splitting and the other not
     env.getConnector().tableOperations().setProperty(indexTableName, Property.TABLE_SPLIT_THRESHOLD.getKey(), "20G");
 
     long t1 = System.currentTimeMillis();
-    
+
     env.getConnector().tableOperations().flush(indexTableName, null, null, true);
     env.getConnector().tableOperations().offline(indexTableName);
-    
+
     long t2 = System.currentTimeMillis();
 
     env.getConnector().tableOperations().exportTable(indexTableName, exportDir);
-    
+
     long t3 = System.currentTimeMillis();
 
     // copy files
@@ -78,31 +78,31 @@ public class ExportIndex extends Test {
     }
 
     reader.close();
-    
+
     long t4 = System.currentTimeMillis();
-    
+
     env.getConnector().tableOperations().online(indexTableName);
     env.getConnector().tableOperations().importTable(tmpIndexTableName, copyDir);
-    
+
     long t5 = System.currentTimeMillis();
 
     fs.delete(new Path(exportDir), true);
     fs.delete(new Path(copyDir), true);
-    
+
     HashSet<Text> splits1 = new HashSet<Text>(env.getConnector().tableOperations().listSplits(indexTableName));
     HashSet<Text> splits2 = new HashSet<Text>(env.getConnector().tableOperations().listSplits(tmpIndexTableName));
-    
+
     if (!splits1.equals(splits2))
       throw new Exception("Splits not equals " + indexTableName + " " + tmpIndexTableName);
-    
+
     HashMap<String,String> props1 = new HashMap<String,String>();
     for (Entry<String,String> entry : env.getConnector().tableOperations().getProperties(indexTableName))
       props1.put(entry.getKey(), entry.getValue());
-    
+
     HashMap<String,String> props2 = new HashMap<String,String>();
     for (Entry<String,String> entry : env.getConnector().tableOperations().getProperties(tmpIndexTableName))
       props2.put(entry.getKey(), entry.getValue());
-    
+
     if (!props1.equals(props2))
       throw new Exception("Props not equals " + indexTableName + " " + tmpIndexTableName);
 
@@ -112,7 +112,7 @@ public class ExportIndex extends Test {
 
     log.debug("Imported " + tmpIndexTableName + " from " + indexTableName + " flush: " + (t2 - t1) + "ms export: " + (t3 - t2) + "ms copy:" + (t4 - t3)
         + "ms import:" + (t5 - t4) + "ms");
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Flush.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Flush.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Flush.java
index 5bd4354..39d32c4 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Flush.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Flush.java
@@ -24,22 +24,22 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Flush extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     String dataTableName = (String) state.get("docTableName");
     Random rand = (Random) state.get("rand");
-    
+
     String table;
-    
+
     if (rand.nextDouble() < .5)
       table = indexTableName;
     else
       table = dataTableName;
-    
+
     env.getConnector().tableOperations().flush(table, null, null, true);
     log.debug("Flushed " + table);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Grep.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Grep.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Grep.java
index 66d53d8..d5c5e5d 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Grep.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Grep.java
@@ -37,61 +37,61 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Grep extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     // pick a few randoms words... grep for those words and search the index
     // ensure both return the same set of documents
-    
+
     String indexTableName = (String) state.get("indexTableName");
     String dataTableName = (String) state.get("docTableName");
     Random rand = (Random) state.get("rand");
-    
+
     Text words[] = new Text[rand.nextInt(4) + 2];
-    
+
     for (int i = 0; i < words.length; i++) {
       words[i] = new Text(Insert.generateRandomWord(rand));
     }
-    
+
     BatchScanner bs = env.getConnector().createBatchScanner(indexTableName, Authorizations.EMPTY, 16);
     IteratorSetting ii = new IteratorSetting(20, "ii", IntersectingIterator.class.getName());
     IntersectingIterator.setColumnFamilies(ii, words);
     bs.addScanIterator(ii);
     bs.setRanges(Collections.singleton(new Range()));
-    
+
     HashSet<Text> documentsFoundInIndex = new HashSet<Text>();
-    
+
     for (Entry<Key,Value> entry2 : bs) {
       documentsFoundInIndex.add(entry2.getKey().getColumnQualifier());
     }
-    
+
     bs.close();
-    
+
     bs = env.getConnector().createBatchScanner(dataTableName, Authorizations.EMPTY, 16);
-    
+
     for (int i = 0; i < words.length; i++) {
       IteratorSetting more = new IteratorSetting(20 + i, "ii" + i, RegExFilter.class);
       RegExFilter.setRegexs(more, null, null, null, "(^|(.*\\s))" + words[i] + "($|(\\s.*))", false);
       bs.addScanIterator(more);
     }
-    
+
     bs.setRanges(Collections.singleton(new Range()));
-    
+
     HashSet<Text> documentsFoundByGrep = new HashSet<Text>();
-    
+
     for (Entry<Key,Value> entry2 : bs) {
       documentsFoundByGrep.add(entry2.getKey().getRow());
     }
-    
+
     bs.close();
-    
+
     if (!documentsFoundInIndex.equals(documentsFoundByGrep)) {
       throw new Exception("Set of documents found not equal for words " + Arrays.asList(words).toString() + " " + documentsFoundInIndex + " "
           + documentsFoundByGrep);
     }
-    
+
     log.debug("Grep and index agree " + Arrays.asList(words).toString() + " " + documentsFoundInIndex.size());
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Insert.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Insert.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Insert.java
index 71d355b..4fdbbb9 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Insert.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Insert.java
@@ -31,11 +31,11 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Insert extends Test {
-  
+
   static final int NUM_WORDS = 100000;
   static final int MIN_WORDS_PER_DOC = 10;
   static final int MAX_WORDS_PER_DOC = 3000;
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
@@ -43,83 +43,83 @@ public class Insert extends Test {
     int numPartitions = (Integer) state.get("numPartitions");
     Random rand = (Random) state.get("rand");
     long nextDocID = (Long) state.get("nextDocID");
-    
+
     BatchWriter dataWriter = env.getMultiTableBatchWriter().getBatchWriter(dataTableName);
     BatchWriter indexWriter = env.getMultiTableBatchWriter().getBatchWriter(indexTableName);
-    
+
     String docID = insertRandomDocument(nextDocID++, dataWriter, indexWriter, indexTableName, dataTableName, numPartitions, rand);
-    
+
     log.debug("Inserted document " + docID);
-    
+
     state.set("nextDocID", Long.valueOf(nextDocID));
   }
-  
+
   static String insertRandomDocument(long did, BatchWriter dataWriter, BatchWriter indexWriter, String indexTableName, String dataTableName, int numPartitions,
       Random rand) throws TableNotFoundException, Exception, AccumuloException, AccumuloSecurityException {
     String doc = createDocument(rand);
-    
+
     String docID = new StringBuilder(String.format("%016x", did)).reverse().toString();
-    
+
     saveDocument(dataWriter, docID, doc);
     indexDocument(indexWriter, doc, docID, numPartitions);
-    
+
     return docID;
   }
-  
+
   static void saveDocument(BatchWriter bw, String docID, String doc) throws Exception {
-    
+
     Mutation m = new Mutation(docID);
     m.put("doc", "", doc);
-    
+
     bw.addMutation(m);
   }
-  
+
   static String createDocument(Random rand) {
     StringBuilder sb = new StringBuilder();
-    
+
     int numWords = rand.nextInt(MAX_WORDS_PER_DOC - MIN_WORDS_PER_DOC) + MIN_WORDS_PER_DOC;
-    
+
     for (int i = 0; i < numWords; i++) {
       String word = generateRandomWord(rand);
-      
+
       if (i > 0)
         sb.append(" ");
-      
+
       sb.append(word);
     }
-    
+
     return sb.toString();
   }
-  
+
   static String generateRandomWord(Random rand) {
     return Integer.toString(rand.nextInt(NUM_WORDS), Character.MAX_RADIX);
   }
-  
+
   static String genPartition(int partition) {
     return String.format("%06x", Math.abs(partition));
   }
-  
+
   static void indexDocument(BatchWriter bw, String doc, String docId, int numPartitions) throws Exception {
     indexDocument(bw, doc, docId, numPartitions, false);
   }
-  
+
   static void unindexDocument(BatchWriter bw, String doc, String docId, int numPartitions) throws Exception {
     indexDocument(bw, doc, docId, numPartitions, true);
   }
-  
+
   static void indexDocument(BatchWriter bw, String doc, String docId, int numPartitions, boolean delete) throws Exception {
-    
+
     String[] tokens = doc.split("\\W+");
-    
+
     String partition = genPartition(doc.hashCode() % numPartitions);
-    
+
     Mutation m = new Mutation(partition);
-    
+
     HashSet<String> tokensSeen = new HashSet<String>();
-    
+
     for (String token : tokens) {
       token = token.toLowerCase();
-      
+
       if (!tokensSeen.contains(token)) {
         tokensSeen.add(token);
         if (delete)
@@ -128,9 +128,9 @@ public class Insert extends Test {
           m.put(token, docId, new Value(new byte[0]));
       }
     }
-    
+
     if (m.size() > 0)
       bw.addMutation(m);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Merge.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Merge.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Merge.java
index 0c4fa3f..36a70f6 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Merge.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Merge.java
@@ -27,11 +27,11 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Merge extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
-    
+
     Collection<Text> splits = env.getConnector().tableOperations().listSplits(indexTableName);
     SortedSet<Text> splitSet = new TreeSet<Text>(splits);
     log.debug("merging " + indexTableName);
@@ -45,5 +45,5 @@ public class Merge extends Test {
       throw new Exception("There are more tablets after a merge: " + splits.size() + " was " + splitSet.size());
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Reindex.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Reindex.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Reindex.java
index de26b76..5aed6e3 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Reindex.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Reindex.java
@@ -31,36 +31,36 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Reindex extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     String tmpIndexTableName = indexTableName + "_tmp";
     String docTableName = (String) state.get("docTableName");
     int numPartitions = (Integer) state.get("numPartitions");
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     ShardFixture.createIndexTable(this.log, state, env, "_tmp", rand);
-    
+
     Scanner scanner = env.getConnector().createScanner(docTableName, Authorizations.EMPTY);
     BatchWriter tbw = env.getConnector().createBatchWriter(tmpIndexTableName, new BatchWriterConfig());
-    
+
     int count = 0;
-    
+
     for (Entry<Key,Value> entry : scanner) {
       String docID = entry.getKey().getRow().toString();
       String doc = entry.getValue().toString();
-      
+
       Insert.indexDocument(tbw, doc, docID, numPartitions);
-      
+
       count++;
     }
-    
+
     tbw.close();
-    
+
     log.debug("Reindexed " + count + " documents into " + tmpIndexTableName);
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Search.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Search.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Search.java
index c8888a1..899ec42 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Search.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Search.java
@@ -19,8 +19,8 @@ package org.apache.accumulo.test.randomwalk.shard;
 import java.util.Collections;
 import java.util.HashSet;
 import java.util.Iterator;
-import java.util.Properties;
 import java.util.Map.Entry;
+import java.util.Properties;
 import java.util.Random;
 
 import org.apache.accumulo.core.client.BatchScanner;
@@ -37,69 +37,69 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Search extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     String dataTableName = (String) state.get("docTableName");
-    
+
     Random rand = (Random) state.get("rand");
-    
+
     Entry<Key,Value> entry = findRandomDocument(state, env, dataTableName, rand);
     if (entry == null)
       return;
-    
+
     Text docID = entry.getKey().getRow();
     String doc = entry.getValue().toString();
-    
+
     String[] tokens = doc.split("\\W+");
     int numSearchTerms = rand.nextInt(6);
     if (numSearchTerms < 2)
       numSearchTerms = 2;
-    
+
     HashSet<String> searchTerms = new HashSet<String>();
     while (searchTerms.size() < numSearchTerms)
       searchTerms.add(tokens[rand.nextInt(tokens.length)]);
-    
+
     Text columns[] = new Text[searchTerms.size()];
     int index = 0;
     for (String term : searchTerms) {
       columns[index++] = new Text(term);
     }
-    
+
     log.debug("Looking up terms " + searchTerms + " expect to find " + docID);
-    
+
     BatchScanner bs = env.getConnector().createBatchScanner(indexTableName, Authorizations.EMPTY, 10);
     IteratorSetting ii = new IteratorSetting(20, "ii", IntersectingIterator.class);
     IntersectingIterator.setColumnFamilies(ii, columns);
     bs.addScanIterator(ii);
     bs.setRanges(Collections.singleton(new Range()));
-    
+
     boolean sawDocID = false;
-    
+
     for (Entry<Key,Value> entry2 : bs) {
       if (entry2.getKey().getColumnQualifier().equals(docID)) {
         sawDocID = true;
         break;
       }
     }
-    
+
     bs.close();
-    
+
     if (!sawDocID)
       throw new Exception("Did not see doc " + docID + " in index.  terms:" + searchTerms + " " + indexTableName + " " + dataTableName);
   }
-  
+
   static Entry<Key,Value> findRandomDocument(State state, Environment env, String dataTableName, Random rand) throws Exception {
     Scanner scanner = env.getConnector().createScanner(dataTableName, Authorizations.EMPTY);
     scanner.setBatchSize(1);
     scanner.setRange(new Range(Integer.toString(rand.nextInt(0xfffffff), 16), null));
-    
+
     Iterator<Entry<Key,Value>> iter = scanner.iterator();
     if (!iter.hasNext())
       return null;
-    
+
     return iter.next();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ShardFixture.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ShardFixture.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ShardFixture.java
index 36e9d5e..63500a0 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ShardFixture.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/ShardFixture.java
@@ -25,8 +25,8 @@ import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.conf.Property;
-import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.Environment;
+import org.apache.accumulo.test.randomwalk.Fixture;
 import org.apache.accumulo.test.randomwalk.State;
 import org.apache.hadoop.io.Text;
 import org.apache.log4j.Logger;
@@ -58,7 +58,7 @@ public class ShardFixture extends Fixture {
 
     String tableId = conn.tableOperations().tableIdMap().get(name);
     log.info("Created index table " + name + "(id:" + tableId + ")");
-    
+
     SortedSet<Text> splits = genSplits(numPartitions, rand.nextInt(numPartitions) + 1, "%06x");
     conn.tableOperations().addSplits(name, splits);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/SortTool.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/SortTool.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/SortTool.java
index 17af89e..1247bc2 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/SortTool.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/SortTool.java
@@ -36,38 +36,38 @@ public class SortTool extends Configured implements Tool {
   private String seqFile;
   private String splitFile;
   private Collection<Text> splits;
-  
+
   public SortTool(String seqFile, String outputDir, String splitFile, Collection<Text> splits) {
     this.outputDir = outputDir;
     this.seqFile = seqFile;
     this.splitFile = splitFile;
     this.splits = splits;
   }
-  
+
   public int run(String[] args) throws Exception {
     @SuppressWarnings("deprecation")
     Job job = new Job(getConf(), this.getClass().getSimpleName());
     job.setJarByClass(this.getClass());
-    
+
     if (job.getJar() == null) {
       log.error("M/R requires a jar file!  Run mvn package.");
       return 1;
     }
-    
+
     job.setInputFormatClass(SequenceFileInputFormat.class);
     SequenceFileInputFormat.setInputPaths(job, seqFile);
-    
+
     job.setPartitionerClass(KeyRangePartitioner.class);
     KeyRangePartitioner.setSplitFile(job, splitFile);
-    
+
     job.setMapOutputKeyClass(Key.class);
     job.setMapOutputValueClass(Value.class);
-    
+
     job.setNumReduceTasks(splits.size() + 1);
-    
+
     job.setOutputFormatClass(AccumuloFileOutputFormat.class);
     AccumuloFileOutputFormat.setOutputPath(job, new Path(outputDir));
-    
+
     job.waitForCompletion(true);
     return job.isSuccessful() ? 0 : 1;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Split.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Split.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Split.java
index 6c6ab00..0420fb8 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Split.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/Split.java
@@ -26,16 +26,16 @@ import org.apache.accumulo.test.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
 public class Split extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
     String indexTableName = (String) state.get("indexTableName");
     int numPartitions = (Integer) state.get("numPartitions");
     Random rand = (Random) state.get("rand");
-    
+
     SortedSet<Text> splitSet = ShardFixture.genSplits(numPartitions, rand.nextInt(numPartitions) + 1, "%06x");
     log.debug("adding splits " + indexTableName);
     env.getConnector().tableOperations().addSplits(indexTableName, splitSet);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/VerifyIndex.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/VerifyIndex.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/VerifyIndex.java
index c6f5703..70440d8 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/VerifyIndex.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/shard/VerifyIndex.java
@@ -30,42 +30,42 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class VerifyIndex extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {
-    
+
     String indexTableName = (String) state.get("indexTableName");
     String tmpIndexTableName = indexTableName + "_tmp";
-    
+
     // scan new and old index and verify identical
     Scanner indexScanner1 = env.getConnector().createScanner(tmpIndexTableName, Authorizations.EMPTY);
     Scanner indexScanner2 = env.getConnector().createScanner(indexTableName, Authorizations.EMPTY);
-    
+
     Iterator<Entry<Key,Value>> iter = indexScanner2.iterator();
-    
+
     int count = 0;
-    
+
     for (Entry<Key,Value> entry : indexScanner1) {
       if (!iter.hasNext())
         throw new Exception("index rebuild mismatch " + entry.getKey() + " " + indexTableName);
-      
+
       Key key1 = entry.getKey();
       Key key2 = iter.next().getKey();
-      
+
       if (!key1.equals(key2, PartialKey.ROW_COLFAM_COLQUAL))
         throw new Exception("index rebuild mismatch " + key1 + " " + key2 + " " + indexTableName + " " + tmpIndexTableName);
       count++;
       if (count % 1000 == 0)
         makingProgress();
     }
-    
+
     if (iter.hasNext())
       throw new Exception("index rebuild mismatch " + iter.next().getKey() + " " + tmpIndexTableName);
-    
+
     log.debug("Verified " + count + " index entries ");
-    
+
     env.getConnector().tableOperations().delete(indexTableName);
     env.getConnector().tableOperations().rename(tmpIndexTableName, indexTableName);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/CreateTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/CreateTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/CreateTable.java
index 1d1e926..2e14de7 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/CreateTable.java
@@ -24,7 +24,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class CreateTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/DeleteTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/DeleteTable.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/DeleteTable.java
index 379a17d..c0b7e65 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/DeleteTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/DeleteTable.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class DeleteTable extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Ingest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Ingest.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Ingest.java
index 461ef22..40f3555 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Ingest.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Ingest.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Ingest extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Scan.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Scan.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Scan.java
index bd82b39..5273ea8 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Scan.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Scan.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Scan extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Verify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Verify.java b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Verify.java
index 1b2f4cf..c41d633 100644
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Verify.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/unit/Verify.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.test.randomwalk.State;
 import org.apache.accumulo.test.randomwalk.Test;
 
 public class Verify extends Test {
-  
+
   @Override
   public void visit(State state, Environment env, Properties props) throws Exception {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/replication/ReplicationTablesPrinterThread.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/replication/ReplicationTablesPrinterThread.java b/test/src/main/java/org/apache/accumulo/test/replication/ReplicationTablesPrinterThread.java
index 750557c..0dec94c 100644
--- a/test/src/main/java/org/apache/accumulo/test/replication/ReplicationTablesPrinterThread.java
+++ b/test/src/main/java/org/apache/accumulo/test/replication/ReplicationTablesPrinterThread.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.core.replication.PrintReplicationRecords;
 import org.apache.accumulo.core.util.Daemon;
 
 /**
- * 
+ *
  */
 public class ReplicationTablesPrinterThread extends Daemon {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/replication/merkle/MerkleTreeNode.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/replication/merkle/MerkleTreeNode.java b/test/src/main/java/org/apache/accumulo/test/replication/merkle/MerkleTreeNode.java
index dfde41a..33ec056 100644
--- a/test/src/main/java/org/apache/accumulo/test/replication/merkle/MerkleTreeNode.java
+++ b/test/src/main/java/org/apache/accumulo/test/replication/merkle/MerkleTreeNode.java
@@ -32,7 +32,7 @@ import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
 /**
- * Encapsulates the level (height) within the tree, the ranges that it covers, and the new hash 
+ * Encapsulates the level (height) within the tree, the ranges that it covers, and the new hash
  */
 public class MerkleTreeNode {
   private static final Logger log = LoggerFactory.getLogger(MerkleTreeNode.class);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/replication/merkle/RangeSerialization.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/replication/merkle/RangeSerialization.java b/test/src/main/java/org/apache/accumulo/test/replication/merkle/RangeSerialization.java
index 62bc800..6b07f2f 100644
--- a/test/src/main/java/org/apache/accumulo/test/replication/merkle/RangeSerialization.java
+++ b/test/src/main/java/org/apache/accumulo/test/replication/merkle/RangeSerialization.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class RangeSerialization {
   private static final Text EMPTY = new Text(new byte[0]);
@@ -37,7 +37,7 @@ public class RangeSerialization {
     } else {
       startKey = new Key(holder);
     }
-    
+
     key.getColumnQualifier(holder);
     Key endKey;
     if (0 == holder.getLength()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/CompareTables.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/CompareTables.java b/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/CompareTables.java
index 44a5a3b..8e97beb 100644
--- a/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/CompareTables.java
+++ b/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/CompareTables.java
@@ -41,14 +41,14 @@ import com.beust.jcommander.Parameter;
 /**
  * Accepts a set of tables, computes the hashes for each, and prints the top-level hash for each table.
  * <p>
- * Will automatically create output tables for intermediate hashes instead of requiring their existence.
- * This will raise an exception when the table we want to use already exists.
+ * Will automatically create output tables for intermediate hashes instead of requiring their existence. This will raise an exception when the table we want to
+ * use already exists.
  */
 public class CompareTables {
   private static final Logger log = LoggerFactory.getLogger(CompareTables.class);
 
   public static class CompareTablesOpts extends ClientOpts {
-    @Parameter(names={"--tables"}, description = "Tables to compare", variableArity=true)
+    @Parameter(names = {"--tables"}, description = "Tables to compare", variableArity = true)
     public List<String> tables;
 
     @Parameter(names = {"-nt", "--numThreads"}, required = false, description = "number of concurrent threads calculating digests")

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/ComputeRootHash.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/ComputeRootHash.java b/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/ComputeRootHash.java
index ea241a6..cb2761b 100644
--- a/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/ComputeRootHash.java
+++ b/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/ComputeRootHash.java
@@ -74,7 +74,7 @@ public class ComputeRootHash {
   }
 
   protected ArrayList<MerkleTreeNode> getLeaves(Connector conn, String tableName) throws TableNotFoundException {
-    //TODO make this a bit more resilient to very large merkle trees by lazily reading more data from the table when necessary
+    // TODO make this a bit more resilient to very large merkle trees by lazily reading more data from the table when necessary
     final Scanner s = conn.createScanner(tableName, Authorizations.EMPTY);
     final ArrayList<MerkleTreeNode> leaves = new ArrayList<MerkleTreeNode>();
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/GenerateHashes.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/GenerateHashes.java b/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/GenerateHashes.java
index c2c9a5a..7e960b1 100644
--- a/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/GenerateHashes.java
+++ b/test/src/main/java/org/apache/accumulo/test/replication/merkle/cli/GenerateHashes.java
@@ -120,7 +120,8 @@ public class GenerateHashes {
     }
   }
 
-  public Collection<Range> getRanges(Connector conn, String tableName, String splitsFile) throws TableNotFoundException, AccumuloSecurityException, AccumuloException, FileNotFoundException {
+  public Collection<Range> getRanges(Connector conn, String tableName, String splitsFile) throws TableNotFoundException, AccumuloSecurityException,
+      AccumuloException, FileNotFoundException {
     if (null == splitsFile) {
       log.info("Using table split points");
       Collection<Text> endRows = conn.tableOperations().listSplits(tableName);


[23/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/AgeOffStore.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/AgeOffStore.java b/fate/src/main/java/org/apache/accumulo/fate/AgeOffStore.java
index f32fb70..3c88f48 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/AgeOffStore.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/AgeOffStore.java
@@ -28,49 +28,49 @@ import org.apache.log4j.Logger;
 
 /**
  * This store removes Repos, in the store it wraps, that are in a finished or new state for more than a configurable time period.
- * 
+ *
  * No external time source is used. It starts tracking idle time when its created.
- * 
+ *
  */
 public class AgeOffStore<T> implements TStore<T> {
-  
+
   public interface TimeSource {
     long currentTimeMillis();
   }
-  
+
   final private static Logger log = Logger.getLogger(AgeOffStore.class);
-  
+
   private TStore<T> store;
   private Map<Long,Long> candidates;
   private long ageOffTime;
   private long minTime;
   private TimeSource timeSource;
-  
+
   private synchronized void updateMinTime() {
     minTime = Long.MAX_VALUE;
-    
+
     for (Long time : candidates.values()) {
       if (time < minTime)
         minTime = time;
     }
   }
-  
+
   private synchronized void addCandidate(long txid) {
     long time = timeSource.currentTimeMillis();
     candidates.put(txid, time);
     if (time < minTime)
       minTime = time;
   }
-  
+
   private synchronized void removeCandidate(long txid) {
     Long time = candidates.remove(txid);
     if (time != null && time <= minTime)
       updateMinTime();
   }
-  
+
   public void ageOff() {
     HashSet<Long> oldTxs = new HashSet<Long>();
-    
+
     synchronized (this) {
       long time = timeSource.currentTimeMillis();
       if (minTime < time && time - minTime >= ageOffTime) {
@@ -79,12 +79,12 @@ public class AgeOffStore<T> implements TStore<T> {
             oldTxs.add(entry.getKey());
           }
         }
-        
+
         candidates.keySet().removeAll(oldTxs);
         updateMinTime();
       }
     }
-    
+
     for (Long txid : oldTxs) {
       try {
         store.reserve(txid);
@@ -99,7 +99,7 @@ public class AgeOffStore<T> implements TStore<T> {
             default:
               break;
           }
-          
+
         } finally {
           store.unreserve(txid, 0);
         }
@@ -108,15 +108,15 @@ public class AgeOffStore<T> implements TStore<T> {
       }
     }
   }
-  
+
   public AgeOffStore(TStore<T> store, long ageOffTime, TimeSource timeSource) {
     this.store = store;
     this.ageOffTime = ageOffTime;
     this.timeSource = timeSource;
     candidates = new HashMap<Long,Long>();
-    
+
     minTime = Long.MAX_VALUE;
-    
+
     List<Long> txids = store.list();
     for (Long txid : txids) {
       store.reserve(txid);
@@ -135,7 +135,7 @@ public class AgeOffStore<T> implements TStore<T> {
       }
     }
   }
-  
+
   public AgeOffStore(TStore<T> store, long ageOffTime) {
     this(store, ageOffTime, new TimeSource() {
       @Override
@@ -144,53 +144,53 @@ public class AgeOffStore<T> implements TStore<T> {
       }
     });
   }
-  
+
   @Override
   public long create() {
     long txid = store.create();
     addCandidate(txid);
     return txid;
   }
-  
+
   @Override
   public long reserve() {
     return store.reserve();
   }
-  
+
   @Override
   public void reserve(long tid) {
     store.reserve(tid);
   }
-  
+
   @Override
   public void unreserve(long tid, long deferTime) {
     store.unreserve(tid, deferTime);
   }
-  
+
   @Override
   public Repo<T> top(long tid) {
     return store.top(tid);
   }
-  
+
   @Override
   public void push(long tid, Repo<T> repo) throws StackOverflowException {
     store.push(tid, repo);
   }
-  
+
   @Override
   public void pop(long tid) {
     store.pop(tid);
   }
-  
+
   @Override
   public org.apache.accumulo.fate.TStore.TStatus getStatus(long tid) {
     return store.getStatus(tid);
   }
-  
+
   @Override
   public void setStatus(long tid, org.apache.accumulo.fate.TStore.TStatus status) {
     store.setStatus(tid, status);
-    
+
     switch (status) {
       case IN_PROGRESS:
       case FAILED_IN_PROGRESS:
@@ -204,28 +204,28 @@ public class AgeOffStore<T> implements TStore<T> {
         break;
     }
   }
-  
+
   @Override
   public org.apache.accumulo.fate.TStore.TStatus waitForStatusChange(long tid, EnumSet<org.apache.accumulo.fate.TStore.TStatus> expected) {
     return store.waitForStatusChange(tid, expected);
   }
-  
+
   @Override
   public void setProperty(long tid, String prop, Serializable val) {
     store.setProperty(tid, prop, val);
   }
-  
+
   @Override
   public Serializable getProperty(long tid, String prop) {
     return store.getProperty(tid, prop);
   }
-  
+
   @Override
   public void delete(long tid) {
     store.delete(tid);
     removeCandidate(tid);
   }
-  
+
   @Override
   public List<Long> list() {
     return store.list();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyRepo.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyRepo.java b/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyRepo.java
index 24d00d9..9ae7acb 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyRepo.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyRepo.java
@@ -19,8 +19,8 @@ package org.apache.accumulo.fate;
 /**
  * Read only access to a repeatable persisted operation.
  *
- * By definition, these methods are safe to call without impacting the state of FATE. They should also be
- * safe to call without impacting the state of system components.
+ * By definition, these methods are safe to call without impacting the state of FATE. They should also be safe to call without impacting the state of system
+ * components.
  *
  */
 public interface ReadOnlyRepo<T> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyStore.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyStore.java b/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyStore.java
index ad5e7e1..1f01090 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyStore.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyStore.java
@@ -24,9 +24,9 @@ import com.google.common.base.Preconditions;
 
 /**
  * This store decorates a TStore to make sure it can not be modified.
- * 
+ *
  * Unlike relying directly on the ReadOnlyTStore interface, this class will not allow subsequent users to cast back to a mutable TStore successfully.
- * 
+ *
  */
 public class ReadOnlyStore<T> implements ReadOnlyTStore<T> {
 
@@ -58,7 +58,7 @@ public class ReadOnlyStore<T> implements ReadOnlyTStore<T> {
 
   /**
    * Decorates a Repo to make sure it is treated as a ReadOnlyRepo.
-   * 
+   *
    * Similar to ReadOnlyStore, won't allow subsequent user to cast a ReadOnlyRepo back to a mutable Repo.
    */
   protected static class ReadOnlyRepoWrapper<X> implements ReadOnlyRepo<X> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyTStore.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyTStore.java b/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyTStore.java
index d390139..5c1344a 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyTStore.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/ReadOnlyTStore.java
@@ -23,8 +23,8 @@ import java.util.List;
 /**
  * Read only access to a Transaction Store.
  *
- * A transaction consists of a number of operations. Instances of this class may check on the queue of outstanding
- * transactions but may neither modify them nor create new ones.
+ * A transaction consists of a number of operations. Instances of this class may check on the queue of outstanding transactions but may neither modify them nor
+ * create new ones.
  */
 public interface ReadOnlyTStore<T> {
 
@@ -49,8 +49,7 @@ public interface ReadOnlyTStore<T> {
   /**
    * Reserve a transaction that is IN_PROGRESS or FAILED_IN_PROGRESS.
    *
-   * Reserving a transaction id ensures that nothing else in-process interacting via the same instance
-   * will be operating on that transaction id.
+   * Reserving a transaction id ensures that nothing else in-process interacting via the same instance will be operating on that transaction id.
    *
    * @return a transaction id that is safe to interact with, chosen by the store.
    */
@@ -59,8 +58,7 @@ public interface ReadOnlyTStore<T> {
   /**
    * Reserve the specific tid.
    *
-   * Reserving a transaction id ensures that nothing else in-process interacting via the same instance
-   * will be operating on that transaction id.
+   * Reserving a transaction id ensures that nothing else in-process interacting via the same instance will be operating on that transaction id.
    *
    */
   void reserve(long tid);
@@ -70,18 +68,20 @@ public interface ReadOnlyTStore<T> {
    *
    * upon successful return the store now controls the referenced transaction id. caller should no longer interact with it.
    *
-   * @param tid transaction id, previously reserved.
-   * @param deferTime time in millis to keep this transaction out of the pool used in the {@link #reserve() reserve} method. must be non-negative.
+   * @param tid
+   *          transaction id, previously reserved.
+   * @param deferTime
+   *          time in millis to keep this transaction out of the pool used in the {@link #reserve() reserve} method. must be non-negative.
    */
   void unreserve(long tid, long deferTime);
 
-
   /**
    * Get the current operation for the given transaction id.
    *
    * Caller must have already reserved tid.
    *
-   * @param tid transaction id, previously reserved.
+   * @param tid
+   *          transaction id, previously reserved.
    * @return a read-only view of the operation
    */
   ReadOnlyRepo<T> top(long tid);
@@ -91,7 +91,8 @@ public interface ReadOnlyTStore<T> {
    *
    * Caller must have already reserved tid.
    *
-   * @param tid transaction id, previously reserved.
+   * @param tid
+   *          transaction id, previously reserved.
    * @return execution status
    */
   TStatus getStatus(long tid);
@@ -99,8 +100,10 @@ public interface ReadOnlyTStore<T> {
   /**
    * Wait for the satus of a transaction to change
    *
-   * @param tid transaction id, need not have been reserved.
-   * @param expected a set of possible statuses we are interested in being notified about. may not be null.
+   * @param tid
+   *          transaction id, need not have been reserved.
+   * @param expected
+   *          a set of possible statuses we are interested in being notified about. may not be null.
    * @return execution status.
    */
   TStatus waitForStatusChange(long tid, EnumSet<TStatus> expected);
@@ -110,8 +113,10 @@ public interface ReadOnlyTStore<T> {
    *
    * Caller must have already reserved tid.
    *
-   * @param tid transaction id, previously reserved.
-   * @param prop name of property to retrieve.
+   * @param tid
+   *          transaction id, previously reserved.
+   * @param prop
+   *          name of property to retrieve.
    */
   Serializable getProperty(long tid, String prop);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/Repo.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/Repo.java b/fate/src/main/java/org/apache/accumulo/fate/Repo.java
index b0ebd1a..0dcfd7f 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/Repo.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/Repo.java
@@ -20,14 +20,14 @@ import java.io.Serializable;
 
 /**
  * Repeatable persisted operation
- * 
+ *
  */
 public interface Repo<T> extends ReadOnlyRepo<T>, Serializable {
-  
+
   Repo<T> call(long tid, T environment) throws Exception;
-  
+
   void undo(long tid, T environment) throws Exception;
-  
+
   // this allows the last fate op to return something to the user
   String getReturn();
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/StackOverflowException.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/StackOverflowException.java b/fate/src/main/java/org/apache/accumulo/fate/StackOverflowException.java
index 6e38f1b..0f385d4 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/StackOverflowException.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/StackOverflowException.java
@@ -17,11 +17,11 @@
 package org.apache.accumulo.fate;
 
 public class StackOverflowException extends Exception {
-  
+
   public StackOverflowException(String msg) {
     super(msg);
   }
-  
+
   private static final long serialVersionUID = 1L;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/TStore.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/TStore.java b/fate/src/main/java/org/apache/accumulo/fate/TStore.java
index 3adb493..bdcfba3 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/TStore.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/TStore.java
@@ -20,7 +20,7 @@ import java.io.Serializable;
 
 /**
  * Transaction Store: a place to save transactions
- * 
+ *
  * A transaction consists of a number of operations. To use, first create a transaction id, and then seed the transaction with an initial operation. An executor
  * service can then execute the transaction's operation, possibly pushing more operations onto the transaction as each step successfully completes. If a step
  * fails, the stack can be unwound, undoing each operation.
@@ -29,14 +29,14 @@ public interface TStore<T> extends ReadOnlyTStore<T> {
 
   /**
    * Create a new transaction id
-   * 
+   *
    * @return a transaction id
    */
   long create();
 
   /**
    * Get the current operation for the given transaction id.
-   * 
+   *
    * @param tid
    *          transaction id
    * @return the operation
@@ -46,7 +46,7 @@ public interface TStore<T> extends ReadOnlyTStore<T> {
 
   /**
    * Update the given transaction with the next operation
-   * 
+   *
    * @param tid
    *          the transaction id
    * @param repo
@@ -61,7 +61,7 @@ public interface TStore<T> extends ReadOnlyTStore<T> {
 
   /**
    * Update the state of a given transaction
-   * 
+   *
    * @param tid
    *          transaction id
    * @param status
@@ -73,7 +73,7 @@ public interface TStore<T> extends ReadOnlyTStore<T> {
 
   /**
    * Remove the transaction from the store.
-   * 
+   *
    * @param tid
    *          the transaction id
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/ZooStore.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/ZooStore.java b/fate/src/main/java/org/apache/accumulo/fate/ZooStore.java
index 0dc156e..6c8a7d6 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/ZooStore.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/ZooStore.java
@@ -46,7 +46,7 @@ import org.apache.zookeeper.KeeperException.NodeExistsException;
 //TODO document zookeeper layout - ACCUMULO-1298
 
 public class ZooStore<T> implements TStore<T> {
-  
+
   private String path;
   private IZooReaderWriter zk;
   private String lastReserved = "";
@@ -55,21 +55,21 @@ public class ZooStore<T> implements TStore<T> {
   private SecureRandom idgenerator;
   private long statusChangeEvents = 0;
   private int reservationsWaiting = 0;
-  
+
   private byte[] serialize(Object o) {
-    
+
     try {
       ByteArrayOutputStream baos = new ByteArrayOutputStream();
       ObjectOutputStream oos = new ObjectOutputStream(baos);
       oos.writeObject(o);
       oos.close();
-      
+
       return baos.toByteArray();
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   private Object deserialize(byte ser[]) {
     try {
       ByteArrayInputStream bais = new ByteArrayInputStream(ser);
@@ -79,26 +79,26 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e);
     }
   }
-  
+
   private String getTXPath(long tid) {
     return String.format("%s/tx_%016x", path, tid);
   }
-  
+
   private long parseTid(String txdir) {
     return Long.parseLong(txdir.split("_")[1], 16);
   }
-  
+
   public ZooStore(String path, IZooReaderWriter zk) throws KeeperException, InterruptedException {
-    
+
     this.path = path;
     this.zk = zk;
     this.reserved = new HashSet<Long>();
     this.defered = new HashMap<Long,Long>();
     this.idgenerator = new SecureRandom();
-    
+
     zk.putPersistentData(path, new byte[0], NodeExistsPolicy.SKIP);
   }
-  
+
   @Override
   public long create() {
     while (true) {
@@ -114,28 +114,28 @@ public class ZooStore<T> implements TStore<T> {
       }
     }
   }
-  
+
   @Override
   public long reserve() {
     try {
       while (true) {
-        
+
         long events;
         synchronized (this) {
           events = statusChangeEvents;
         }
-        
+
         List<String> txdirs = new ArrayList<String>(zk.getChildren(path));
         Collections.sort(txdirs);
-        
+
         synchronized (this) {
           if (txdirs.size() > 0 && txdirs.get(txdirs.size() - 1).compareTo(lastReserved) <= 0)
             lastReserved = "";
         }
-        
+
         for (String txdir : txdirs) {
           long tid = parseTid(txdir);
-          
+
           synchronized (this) {
             // this check makes reserve pick up where it left off, so that it cycles through all as it is repeatedly called.... failing to do so can lead to
             // starvation where fate ops that sort higher and hold a lock are never reserved.
@@ -151,13 +151,12 @@ public class ZooStore<T> implements TStore<T> {
             if (!reserved.contains(tid)) {
               reserved.add(tid);
               lastReserved = txdir;
-            }
-            else
+            } else
               continue;
           }
-          
+
           // have reserved id, status should not change
-          
+
           try {
             TStatus status = TStatus.valueOf(new String(zk.getData(path + "/" + txdir, null), UTF_8));
             if (status == TStatus.IN_PROGRESS || status == TStatus.FAILED_IN_PROGRESS) {
@@ -173,7 +172,7 @@ public class ZooStore<T> implements TStore<T> {
             throw e;
           }
         }
-        
+
         synchronized (this) {
           if (events == statusChangeEvents) {
             if (defered.size() > 0) {
@@ -190,7 +189,7 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e);
     }
   }
-  
+
   public void reserve(long tid) {
     synchronized (this) {
       reservationsWaiting++;
@@ -201,63 +200,63 @@ public class ZooStore<T> implements TStore<T> {
           } catch (InterruptedException e) {
             throw new RuntimeException(e);
           }
-        
+
         reserved.add(tid);
       } finally {
         reservationsWaiting--;
       }
     }
   }
-  
+
   private void unreserve(long tid) {
     synchronized (this) {
       if (!reserved.remove(tid))
         throw new IllegalStateException("Tried to unreserve id that was not reserved " + String.format("%016x", tid));
-      
+
       // do not want this unreserve to unesc wake up threads in reserve()... this leads to infinite loop when tx is stuck in NEW...
       // only do this when something external has called reserve(tid)...
       if (reservationsWaiting > 0)
         this.notifyAll();
     }
   }
-  
+
   @Override
   public void unreserve(long tid, long deferTime) {
-    
+
     if (deferTime < 0)
       throw new IllegalArgumentException("deferTime < 0 : " + deferTime);
-    
+
     synchronized (this) {
       if (!reserved.remove(tid))
         throw new IllegalStateException("Tried to unreserve id that was not reserved " + String.format("%016x", tid));
-      
+
       if (deferTime > 0)
         defered.put(tid, System.currentTimeMillis() + deferTime);
-      
+
       this.notifyAll();
     }
-    
+
   }
-  
+
   private void verifyReserved(long tid) {
     synchronized (this) {
       if (!reserved.contains(tid))
         throw new IllegalStateException("Tried to operate on unreserved transaction " + String.format("%016x", tid));
     }
   }
-  
+
   @SuppressWarnings("unchecked")
   @Override
   public Repo<T> top(long tid) {
     verifyReserved(tid);
-    
+
     while (true) {
       try {
         String txpath = getTXPath(tid);
         String top = findTop(txpath);
         if (top == null)
           return null;
-        
+
         byte[] ser = zk.getData(txpath + "/" + top, null);
         return (Repo<T>) deserialize(ser);
       } catch (KeeperException.NoNodeException ex) {
@@ -267,35 +266,35 @@ public class ZooStore<T> implements TStore<T> {
       }
     }
   }
-  
+
   private String findTop(String txpath) throws KeeperException, InterruptedException {
     List<String> ops = zk.getChildren(txpath);
-    
+
     ops = new ArrayList<String>(ops);
-    
+
     String max = "";
-    
+
     for (String child : ops)
       if (child.startsWith("repo_") && child.compareTo(max) > 0)
         max = child;
-    
+
     if (max.equals(""))
       return null;
-    
+
     return max;
   }
-  
+
   @Override
   public void push(long tid, Repo<T> repo) throws StackOverflowException {
     verifyReserved(tid);
-    
+
     String txpath = getTXPath(tid);
     try {
       String top = findTop(txpath);
       if (top != null && Long.parseLong(top.split("_")[1]) > 100) {
         throw new StackOverflowException("Repo stack size too large");
       }
-      
+
       zk.putPersistentSequential(txpath + "/repo_", serialize(repo));
     } catch (StackOverflowException soe) {
       throw soe;
@@ -303,11 +302,11 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public void pop(long tid) {
     verifyReserved(tid);
-    
+
     try {
       String txpath = getTXPath(tid);
       String top = findTop(txpath);
@@ -318,7 +317,7 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e);
     }
   }
-  
+
   private TStatus _getStatus(long tid) {
     try {
       return TStatus.valueOf(new String(zk.getData(getTXPath(tid), null), UTF_8));
@@ -328,13 +327,13 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public TStatus getStatus(long tid) {
     verifyReserved(tid);
     return _getStatus(tid);
   }
-  
+
   @Override
   public TStatus waitForStatusChange(long tid, EnumSet<TStatus> expected) {
     while (true) {
@@ -342,11 +341,11 @@ public class ZooStore<T> implements TStore<T> {
       synchronized (this) {
         events = statusChangeEvents;
       }
-      
+
       TStatus status = _getStatus(tid);
       if (expected.contains(status))
         return status;
-      
+
       synchronized (this) {
         if (events == statusChangeEvents) {
           try {
@@ -358,38 +357,38 @@ public class ZooStore<T> implements TStore<T> {
       }
     }
   }
-  
+
   @Override
   public void setStatus(long tid, TStatus status) {
     verifyReserved(tid);
-    
+
     try {
       zk.putPersistentData(getTXPath(tid), status.name().getBytes(UTF_8), NodeExistsPolicy.OVERWRITE);
     } catch (Exception e) {
       throw new RuntimeException(e);
     }
-    
+
     synchronized (this) {
       statusChangeEvents++;
     }
-    
+
   }
-  
+
   @Override
   public void delete(long tid) {
     verifyReserved(tid);
-    
+
     try {
       zk.recursiveDelete(getTXPath(tid), NodeMissingPolicy.SKIP);
     } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public void setProperty(long tid, String prop, Serializable so) {
     verifyReserved(tid);
-    
+
     try {
       if (so instanceof String) {
         zk.putPersistentData(getTXPath(tid) + "/prop_" + prop, ("S " + so).getBytes(UTF_8), NodeExistsPolicy.OVERWRITE);
@@ -405,14 +404,14 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e2);
     }
   }
-  
+
   @Override
   public Serializable getProperty(long tid, String prop) {
     verifyReserved(tid);
-    
+
     try {
       byte[] data = zk.getData(getTXPath(tid) + "/prop_" + prop, null);
-      
+
       if (data[0] == 'O') {
         byte[] sera = new byte[data.length - 2];
         System.arraycopy(data, 2, sera, 0, sera.length);
@@ -428,7 +427,7 @@ public class ZooStore<T> implements TStore<T> {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public List<Long> list() {
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/util/AddressUtil.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/util/AddressUtil.java b/fate/src/main/java/org/apache/accumulo/fate/util/AddressUtil.java
index 401bc1a..33e84cf 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/util/AddressUtil.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/util/AddressUtil.java
@@ -29,7 +29,7 @@ public class AddressUtil {
   /**
    * Fetch the security value that determines how long DNS failures are cached. Looks up the security property 'networkaddress.cache.negative.ttl'. Should that
    * fail returns the default value used in the Oracle JVM 1.4+, which is 10 seconds.
-   * 
+   *
    * @param originalException
    *          the host lookup that is the source of needing this lookup. maybe be null.
    * @return positive integer number of seconds

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/util/Daemon.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/util/Daemon.java b/fate/src/main/java/org/apache/accumulo/fate/util/Daemon.java
index 95fc6d3..da7c41c 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/util/Daemon.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/util/Daemon.java
@@ -17,44 +17,44 @@
 package org.apache.accumulo.fate.util;
 
 public class Daemon extends Thread {
-  
+
   public Daemon() {
     setDaemon(true);
   }
-  
+
   public Daemon(Runnable target) {
     super(target);
     setDaemon(true);
   }
-  
+
   public Daemon(String name) {
     super(name);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, Runnable target) {
     super(group, target);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, String name) {
     super(group, name);
     setDaemon(true);
   }
-  
+
   public Daemon(Runnable target, String name) {
     super(target, name);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, Runnable target, String name) {
     super(group, target, name);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, Runnable target, String name, long stackSize) {
     super(group, target, name, stackSize);
     setDaemon(true);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/util/LoggingRunnable.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/util/LoggingRunnable.java b/fate/src/main/java/org/apache/accumulo/fate/util/LoggingRunnable.java
index 32938e0..ba45488 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/util/LoggingRunnable.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/util/LoggingRunnable.java
@@ -23,12 +23,12 @@ import org.apache.log4j.Logger;
 public class LoggingRunnable implements Runnable {
   private Runnable runnable;
   private Logger log;
-  
+
   public LoggingRunnable(Logger log, Runnable r) {
     this.runnable = r;
     this.log = log;
   }
-  
+
   public void run() {
     try {
       runnable.run();
@@ -39,7 +39,7 @@ public class LoggingRunnable implements Runnable {
         // maybe the logging system is screwed up OR there is a bug in the exception, like t.getMessage() throws a NPE
         System.err.println("ERROR " + new Date() + " Failed to log message about thread death " + t2.getMessage());
         t2.printStackTrace();
-        
+
         // try to print original exception
         System.err.println("ERROR " + new Date() + " Exception that failed to log : " + t.getMessage());
         t.printStackTrace();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/util/UtilWaitThread.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/util/UtilWaitThread.java b/fate/src/main/java/org/apache/accumulo/fate/util/UtilWaitThread.java
index 32cdb5b..fb5701a 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/util/UtilWaitThread.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/util/UtilWaitThread.java
@@ -20,7 +20,7 @@ import org.apache.log4j.Logger;
 
 public class UtilWaitThread {
   private static final Logger log = Logger.getLogger(UtilWaitThread.class);
-  
+
   public static void sleep(long millis) {
     try {
       Thread.sleep(millis);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLock.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLock.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLock.java
index d10fcd2..fc94ff4 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLock.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLock.java
@@ -33,22 +33,22 @@ import org.apache.log4j.Logger;
 // A ReadWriteLock that can be implemented in ZooKeeper.  Features the ability to store data
 // with the lock, and recover the lock using that data to find the lock.
 public class DistributedReadWriteLock implements java.util.concurrent.locks.ReadWriteLock {
-  
+
   static enum LockType {
     READ, WRITE,
   };
-  
+
   // serializer for lock type and user data
   static class ParsedLock {
     public ParsedLock(LockType type, byte[] userData) {
       this.type = type;
       this.userData = Arrays.copyOf(userData, userData.length);
     }
-    
+
     public ParsedLock(byte[] lockData) {
       if (lockData == null || lockData.length < 1)
         throw new IllegalArgumentException();
-      
+
       int split = -1;
       for (int i = 0; i < lockData.length; i++) {
         if (lockData[i] == ':') {
@@ -56,22 +56,22 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
           break;
         }
       }
-      
+
       if (split == -1)
         throw new IllegalArgumentException();
-      
+
       this.type = LockType.valueOf(new String(lockData, 0, split, UTF_8));
       this.userData = Arrays.copyOfRange(lockData, split + 1, lockData.length);
     }
-    
+
     public LockType getType() {
       return type;
     }
-    
+
     public byte[] getUserData() {
       return userData;
     }
-    
+
     public byte[] getLockData() {
       byte typeBytes[] = type.name().getBytes(UTF_8);
       byte[] result = new byte[userData.length + 1 + typeBytes.length];
@@ -80,46 +80,46 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
       System.arraycopy(userData, 0, result, typeBytes.length + 1, userData.length);
       return result;
     }
-    
+
     private LockType type;
     private byte[] userData;
   }
-  
+
   // This kind of lock can be easily implemented by ZooKeeper
   // You make an entry at the bottom of the queue, readers run when there are no writers ahead of them,
   // a writer only runs when they are at the top of the queue.
   public interface QueueLock {
     SortedMap<Long,byte[]> getEarlierEntries(long entry);
-    
+
     void removeEntry(long entry);
-    
+
     long addEntry(byte[] data);
   }
-  
+
   private static final Logger log = Logger.getLogger(DistributedReadWriteLock.class);
-  
+
   static class ReadLock implements Lock {
-    
+
     QueueLock qlock;
     byte[] userData;
     long entry = -1;
-    
+
     ReadLock(QueueLock qlock, byte[] userData) {
       this.qlock = qlock;
       this.userData = userData;
     }
-    
+
     // for recovery
     ReadLock(QueueLock qlock, byte[] userData, long entry) {
       this.qlock = qlock;
       this.userData = userData;
       this.entry = entry;
     }
-    
+
     protected LockType lockType() {
       return LockType.READ;
     }
-    
+
     @Override
     public void lock() {
       while (true) {
@@ -131,7 +131,7 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
         }
       }
     }
-    
+
     @Override
     public void lockInterruptibly() throws InterruptedException {
       while (!Thread.currentThread().isInterrupted()) {
@@ -139,7 +139,7 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
           return;
       }
     }
-    
+
     @Override
     public boolean tryLock() {
       if (entry == -1) {
@@ -157,7 +157,7 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
       throw new IllegalStateException("Did not find our own lock in the queue: " + this.entry + " userData " + new String(this.userData, UTF_8) + " lockType "
           + lockType());
     }
-    
+
     @Override
     public boolean tryLock(long time, TimeUnit unit) throws InterruptedException {
       long now = System.currentTimeMillis();
@@ -171,7 +171,7 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
       }
       return false;
     }
-    
+
     @Override
     public void unlock() {
       if (entry == -1)
@@ -180,28 +180,28 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
       qlock.removeEntry(entry);
       entry = -1;
     }
-    
+
     @Override
     public Condition newCondition() {
       throw new NotImplementedException();
     }
   }
-  
+
   static class WriteLock extends ReadLock {
-    
+
     WriteLock(QueueLock qlock, byte[] userData) {
       super(qlock, userData);
     }
-    
+
     WriteLock(QueueLock qlock, byte[] userData, long entry) {
       super(qlock, userData, entry);
     }
-    
+
     @Override
     protected LockType lockType() {
       return LockType.WRITE;
     }
-    
+
     @Override
     public boolean tryLock() {
       if (entry == -1) {
@@ -211,22 +211,22 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
       SortedMap<Long,byte[]> entries = qlock.getEarlierEntries(entry);
       Iterator<Entry<Long,byte[]>> iterator = entries.entrySet().iterator();
       if (!iterator.hasNext())
-        throw new IllegalStateException("Did not find our own lock in the queue: " + this.entry + " userData " + new String(this.userData, UTF_8) + " lockType "
-            + lockType());
+        throw new IllegalStateException("Did not find our own lock in the queue: " + this.entry + " userData " + new String(this.userData, UTF_8)
+            + " lockType " + lockType());
       if (iterator.next().getKey().equals(entry))
         return true;
       return false;
     }
   }
-  
+
   private QueueLock qlock;
   private byte[] data;
-  
+
   public DistributedReadWriteLock(QueueLock qlock, byte[] data) {
     this.qlock = qlock;
     this.data = Arrays.copyOf(data, data.length);
   }
-  
+
   static public Lock recoverLock(QueueLock qlock, byte[] data) {
     SortedMap<Long,byte[]> entries = qlock.getEarlierEntries(Long.MAX_VALUE);
     for (Entry<Long,byte[]> entry : entries.entrySet()) {
@@ -242,12 +242,12 @@ public class DistributedReadWriteLock implements java.util.concurrent.locks.Read
     }
     return null;
   }
-  
+
   @Override
   public Lock readLock() {
     return new ReadLock(qlock, data);
   }
-  
+
   @Override
   public Lock writeLock() {
     return new WriteLock(qlock, data);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/IZooReaderWriter.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/IZooReaderWriter.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/IZooReaderWriter.java
index ad2a191..14d104b 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/IZooReaderWriter.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/IZooReaderWriter.java
@@ -40,7 +40,7 @@ public interface IZooReaderWriter extends IZooReader {
   boolean putPrivatePersistentData(String zPath, byte[] data, NodeExistsPolicy policy) throws KeeperException, InterruptedException;
 
   void putPersistentData(String zPath, byte[] data, int version, NodeExistsPolicy policy) throws KeeperException, InterruptedException;
-  
+
   boolean putPersistentData(String zPath, byte[] data, int version, NodeExistsPolicy policy, List<ACL> acls) throws KeeperException, InterruptedException;
 
   String putPersistentSequential(String zPath, byte[] data) throws KeeperException, InterruptedException;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/TransactionWatcher.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/TransactionWatcher.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/TransactionWatcher.java
index cc4bebf..bd27fb9 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/TransactionWatcher.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/TransactionWatcher.java
@@ -24,20 +24,21 @@ import java.util.concurrent.atomic.AtomicInteger;
 import org.apache.log4j.Logger;
 
 public class TransactionWatcher {
-  
+
   public interface Arbitrator {
     boolean transactionAlive(String type, long tid) throws Exception;
+
     boolean transactionComplete(String type, long tid) throws Exception;
   }
-  
+
   private static final Logger log = Logger.getLogger(TransactionWatcher.class);
   final private Map<Long,AtomicInteger> counts = new HashMap<Long,AtomicInteger>();
   final private Arbitrator arbitrator;
-  
+
   public TransactionWatcher(Arbitrator arbitrator) {
     this.arbitrator = arbitrator;
   }
-  
+
   public <T> T run(String ztxBulk, long tid, Callable<T> callable) throws Exception {
     synchronized (counts) {
       if (!arbitrator.transactionAlive(ztxBulk, tid)) {
@@ -62,7 +63,7 @@ public class TransactionWatcher {
       }
     }
   }
-  
+
   public boolean isActive(long tid) {
     synchronized (counts) {
       log.debug("Transactions in progress " + counts);
@@ -70,5 +71,5 @@ public class TransactionWatcher {
       return count != null && count.get() > 0;
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCache.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCache.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCache.java
index ee53ddd..5cfdbb8 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCache.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCache.java
@@ -40,8 +40,7 @@ import org.apache.zookeeper.data.Stat;
 import com.google.common.annotations.VisibleForTesting;
 
 /**
- * A cache for values stored in ZooKeeper. Values are kept up to date as they
- * change.
+ * A cache for values stored in ZooKeeper. Values are kept up to date as they change.
  */
 public class ZooCache {
   private static final Logger log = Logger.getLogger(ZooCache.class);
@@ -104,31 +103,36 @@ public class ZooCache {
   /**
    * Creates a new cache.
    *
-   * @param zooKeepers comma-separated list of ZooKeeper host[:port]s
-   * @param sessionTimeout ZooKeeper session timeout
+   * @param zooKeepers
+   *          comma-separated list of ZooKeeper host[:port]s
+   * @param sessionTimeout
+   *          ZooKeeper session timeout
    */
   public ZooCache(String zooKeepers, int sessionTimeout) {
     this(zooKeepers, sessionTimeout, null);
   }
 
   /**
-   * Creates a new cache. The given watcher is called whenever a watched node
-   * changes.
+   * Creates a new cache. The given watcher is called whenever a watched node changes.
    *
-   * @param zooKeepers comma-separated list of ZooKeeper host[:port]s
-   * @param sessionTimeout ZooKeeper session timeout
-   * @param watcher watcher object
+   * @param zooKeepers
+   *          comma-separated list of ZooKeeper host[:port]s
+   * @param sessionTimeout
+   *          ZooKeeper session timeout
+   * @param watcher
+   *          watcher object
    */
   public ZooCache(String zooKeepers, int sessionTimeout, Watcher watcher) {
     this(new ZooReader(zooKeepers, sessionTimeout), watcher);
   }
 
   /**
-   * Creates a new cache. The given watcher is called whenever a watched node
-   * changes.
+   * Creates a new cache. The given watcher is called whenever a watched node changes.
    *
-   * @param reader ZooKeeper reader
-   * @param watcher watcher object
+   * @param reader
+   *          ZooKeeper reader
+   * @param watcher
+   *          watcher object
    */
   public ZooCache(ZooReader reader, Watcher watcher) {
     this.zReader = reader;
@@ -187,7 +191,8 @@ public class ZooCache {
   /**
    * Gets the children of the given node. A watch is established by this call.
    *
-   * @param zPath path of node
+   * @param zPath
+   *          path of node
    * @return children list, or null if node has no children or does not exist
    */
   public synchronized List<String> getChildren(final String zPath) {
@@ -222,10 +227,10 @@ public class ZooCache {
   }
 
   /**
-   * Gets data at the given path. Status information is not returned. A watch is
-   * established by this call.
+   * Gets data at the given path. Status information is not returned. A watch is established by this call.
    *
-   * @param zPath path to get
+   * @param zPath
+   *          path to get
    * @return path data, or null if non-existent
    */
   public synchronized byte[] get(final String zPath) {
@@ -233,11 +238,12 @@ public class ZooCache {
   }
 
   /**
-   * Gets data at the given path, filling status information into the given
-   * <code>Stat</code> object. A watch is established by this call.
+   * Gets data at the given path, filling status information into the given <code>Stat</code> object. A watch is established by this call.
    *
-   * @param zPath path to get
-   * @param stat status object to populate
+   * @param zPath
+   *          path to get
+   * @param stat
+   *          status object to populate
    * @return path data, or null if non-existent
    */
   public synchronized byte[] get(final String zPath, Stat stat) {
@@ -332,17 +338,20 @@ public class ZooCache {
   /**
    * Checks if a data value (or lack of one) is cached.
    *
-   * @param zPath path of node
+   * @param zPath
+   *          path of node
    * @return true if data value is cached
    */
   @VisibleForTesting
   synchronized boolean dataCached(String zPath) {
     return cache.containsKey(zPath);
   }
+
   /**
    * Checks if children of a node (or lack of them) are cached.
    *
-   * @param zPath path of node
+   * @param zPath
+   *          path of node
    * @return true if children are cached
    */
   @VisibleForTesting
@@ -353,7 +362,8 @@ public class ZooCache {
   /**
    * Clears this cache of all information about nodes rooted at the given path.
    *
-   * @param zPath path of top node
+   * @param zPath
+   *          path of top node
    */
   public synchronized void clear(String zPath) {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactory.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactory.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactory.java
index 3c59a00..1475928 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactory.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactory.java
@@ -18,6 +18,7 @@ package org.apache.accumulo.fate.zookeeper;
 
 import java.util.HashMap;
 import java.util.Map;
+
 import org.apache.zookeeper.Watcher;
 
 /**
@@ -47,15 +48,17 @@ public class ZooCacheFactory {
       return zc;
     }
   }
+
   /**
-   * Gets a watched {@link ZooCache}. If the watcher is null, then the same (unwatched)
-   * object may be returned for multiple calls with the same remaining arguments.
+   * Gets a watched {@link ZooCache}. If the watcher is null, then the same (unwatched) object may be returned for multiple calls with the same remaining
+   * arguments.
    *
    * @param zooKeepers
    *          comma-seprated list of ZooKeeper host[:port]s
    * @param sessionTimeout
    *          session timeout
-   * @param watcher watcher (optional)
+   * @param watcher
+   *          watcher (optional)
    * @return cache object
    */
   public ZooCache getZooCache(String zooKeepers, int sessionTimeout, Watcher watcher) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooLock.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooLock.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooLock.java
index 1391d98..a0100b2 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooLock.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooLock.java
@@ -35,28 +35,28 @@ import org.apache.zookeeper.data.Stat;
 
 public class ZooLock implements Watcher {
   private static final Logger log = Logger.getLogger(ZooLock.class);
-  
+
   public static final String LOCK_PREFIX = "zlock-";
-  
+
   public enum LockLossReason {
     LOCK_DELETED, SESSION_EXPIRED
   }
-  
+
   public interface LockWatcher {
     void lostLock(LockLossReason reason);
-    
+
     /**
      * lost the ability to monitor the lock node, and its status is unknown
      */
     void unableToMonitorLockNode(Throwable e);
   }
-  
+
   public interface AsyncLockWatcher extends LockWatcher {
     void acquiredLock();
-    
+
     void failedToAcquireLock(Exception e);
   }
-  
+
   private boolean lockWasAcquired;
   final private String path;
   protected final IZooReaderWriter zooKeeper;
@@ -64,11 +64,11 @@ public class ZooLock implements Watcher {
   private LockWatcher lockWatcher;
   private boolean watchingParent = false;
   private String asyncLock;
-  
+
   public ZooLock(String zookeepers, int timeInMillis, String scheme, byte[] auth, String path) {
     this(new ZooCacheFactory().getZooCache(zookeepers, timeInMillis), ZooReaderWriter.getInstance(zookeepers, timeInMillis, scheme, auth), path);
   }
-  
+
   protected ZooLock(ZooCache zc, IZooReaderWriter zrw, String path) {
     getLockDataZooCache = zc;
     this.path = path;
@@ -81,66 +81,66 @@ public class ZooLock implements Watcher {
       throw new RuntimeException(ex);
     }
   }
-  
+
   private static class TryLockAsyncLockWatcher implements AsyncLockWatcher {
-    
+
     boolean acquiredLock = false;
     LockWatcher lw;
-    
+
     public TryLockAsyncLockWatcher(LockWatcher lw2) {
       this.lw = lw2;
     }
-    
+
     @Override
     public void acquiredLock() {
       acquiredLock = true;
     }
-    
+
     @Override
     public void failedToAcquireLock(Exception e) {}
-    
+
     @Override
     public void lostLock(LockLossReason reason) {
       lw.lostLock(reason);
     }
-    
+
     @Override
     public void unableToMonitorLockNode(Throwable e) {
       lw.unableToMonitorLockNode(e);
     }
-    
+
   }
-  
+
   public synchronized boolean tryLock(LockWatcher lw, byte data[]) throws KeeperException, InterruptedException {
-    
+
     TryLockAsyncLockWatcher tlalw = new TryLockAsyncLockWatcher(lw);
-    
+
     lockAsync(tlalw, data);
-    
+
     if (tlalw.acquiredLock) {
       return true;
     }
-    
+
     if (asyncLock != null) {
       zooKeeper.recursiveDelete(path + "/" + asyncLock, NodeMissingPolicy.SKIP);
       asyncLock = null;
     }
-    
+
     return false;
   }
-  
+
   private synchronized void lockAsync(final String myLock, final AsyncLockWatcher lw) throws KeeperException, InterruptedException {
-    
+
     if (asyncLock == null) {
       throw new IllegalStateException("Called lockAsync() when asyncLock == null");
     }
-    
+
     List<String> children = zooKeeper.getChildren(path);
-    
+
     if (!children.contains(myLock)) {
       throw new RuntimeException("Lock attempt ephemeral node no longer exist " + myLock);
     }
-    
+
     Collections.sort(children);
     if (log.isTraceEnabled()) {
       log.trace("Candidate lock nodes");
@@ -148,7 +148,7 @@ public class ZooLock implements Watcher {
         log.trace("- " + child);
       }
     }
-    
+
     if (children.get(0).equals(myLock)) {
       log.trace("First candidate is my lock, acquiring");
       if (!watchingParent) {
@@ -166,14 +166,14 @@ public class ZooLock implements Watcher {
       if (child.equals(myLock)) {
         break;
       }
-      
+
       prev = child;
     }
-    
+
     final String lockToWatch = path + "/" + prev;
     log.trace("Establishing watch on " + lockToWatch);
     Stat stat = zooKeeper.getStatus(lockToWatch, new Watcher() {
-      
+
       @Override
       public void process(WatchedEvent event) {
         if (log.isTraceEnabled()) {
@@ -224,39 +224,39 @@ public class ZooLock implements Watcher {
           }
         }
       }
-      
+
     });
-    
+
     if (stat == null)
       lockAsync(myLock, lw);
   }
-  
+
   private void lostLock(LockLossReason reason) {
     LockWatcher localLw = lockWatcher;
     lock = null;
     lockWatcher = null;
-    
+
     localLw.lostLock(reason);
   }
 
   public synchronized void lockAsync(final AsyncLockWatcher lw, byte data[]) {
-    
+
     if (lockWatcher != null || lock != null || asyncLock != null) {
       throw new IllegalStateException();
     }
-    
+
     lockWasAcquired = false;
-    
+
     try {
       final String asyncLockPath = zooKeeper.putEphemeralSequential(path + "/" + LOCK_PREFIX, data);
       log.trace("Ephemeral node " + asyncLockPath + " created");
       Stat stat = zooKeeper.getStatus(asyncLockPath, new Watcher() {
-        
-        private void failedToAcquireLock(){
+
+        private void failedToAcquireLock() {
           lw.failedToAcquireLock(new Exception("Lock deleted before acquired"));
           asyncLock = null;
         }
-        
+
         public void process(WatchedEvent event) {
           synchronized (ZooLock.this) {
             if (lock != null && event.getType() == EventType.NodeDeleted && event.getPath().equals(path + "/" + lock)) {
@@ -264,13 +264,13 @@ public class ZooLock implements Watcher {
             } else if (asyncLock != null && event.getType() == EventType.NodeDeleted && event.getPath().equals(path + "/" + asyncLock)) {
               failedToAcquireLock();
             } else if (event.getState() != KeeperState.Disconnected && event.getState() != KeeperState.Expired && (lock != null || asyncLock != null)) {
-              log.debug("Unexpected event watching lock node "+event+" "+asyncLockPath);
+              log.debug("Unexpected event watching lock node " + event + " " + asyncLockPath);
               try {
                 Stat stat2 = zooKeeper.getStatus(asyncLockPath, this);
-                if(stat2 == null){
-                  if(lock != null)
+                if (stat2 == null) {
+                  if (lock != null)
                     lostLock(LockLossReason.LOCK_DELETED);
-                  else if(asyncLock != null)
+                  else if (asyncLock != null)
                     failedToAcquireLock();
                 }
               } catch (Throwable e) {
@@ -278,105 +278,105 @@ public class ZooLock implements Watcher {
                 log.error("Failed to stat lock node " + asyncLockPath, e);
               }
             }
-           
+
           }
         }
       });
-      
+
       if (stat == null) {
         lw.failedToAcquireLock(new Exception("Lock does not exist after create"));
         return;
       }
-      
+
       asyncLock = asyncLockPath.substring(path.length() + 1);
-      
+
       lockAsync(asyncLock, lw);
-      
+
     } catch (KeeperException e) {
       lw.failedToAcquireLock(e);
     } catch (InterruptedException e) {
       lw.failedToAcquireLock(e);
     }
   }
-  
+
   public synchronized boolean tryToCancelAsyncLockOrUnlock() throws InterruptedException, KeeperException {
     boolean del = false;
-    
+
     if (asyncLock != null) {
       zooKeeper.recursiveDelete(path + "/" + asyncLock, NodeMissingPolicy.SKIP);
       del = true;
     }
-    
+
     if (lock != null) {
       unlock();
       del = true;
     }
-    
+
     return del;
   }
-  
+
   public synchronized void unlock() throws InterruptedException, KeeperException {
     if (lock == null) {
       throw new IllegalStateException();
     }
-    
+
     LockWatcher localLw = lockWatcher;
     String localLock = lock;
-    
+
     lock = null;
     lockWatcher = null;
-    
+
     zooKeeper.recursiveDelete(path + "/" + localLock, NodeMissingPolicy.SKIP);
-    
+
     localLw.lostLock(LockLossReason.LOCK_DELETED);
   }
-  
+
   public synchronized String getLockPath() {
     if (lock == null) {
       return null;
     }
     return path + "/" + lock;
   }
-  
+
   public synchronized String getLockName() {
     return lock;
   }
-  
+
   public synchronized LockID getLockID() {
     if (lock == null) {
       throw new IllegalStateException("Lock not held");
     }
     return new LockID(path, lock, zooKeeper.getZooKeeper().getSessionId());
   }
-  
+
   /**
    * indicates if the lock was acquired in the past.... helps discriminate between the case where the lock was never held, or held and lost....
-   * 
+   *
    * @return true if the lock was aquired, otherwise false.
    */
   public synchronized boolean wasLockAcquired() {
     return lockWasAcquired;
   }
-  
+
   public synchronized boolean isLocked() {
     return lock != null;
   }
-  
+
   public synchronized void replaceLockData(byte[] b) throws KeeperException, InterruptedException {
-    if (getLockPath()!=null)
+    if (getLockPath() != null)
       zooKeeper.getZooKeeper().setData(getLockPath(), b, -1);
   }
-  
+
   @Override
   public synchronized void process(WatchedEvent event) {
     log.debug("event " + event.getPath() + " " + event.getType() + " " + event.getState());
-    
+
     watchingParent = false;
 
     if (event.getState() == KeeperState.Expired && lock != null) {
       lostLock(LockLossReason.SESSION_EXPIRED);
     } else {
-      
+
       try { // set the watch on the parent node again
         zooKeeper.getStatus(path, this);
         watchingParent = true;
@@ -389,151 +389,151 @@ public class ZooLock implements Watcher {
           log.error("Error resetting watch on ZooLock " + lock == null ? asyncLock : lock + " " + event, ex);
         }
       }
-       
+
     }
 
   }
-  
+
   public static boolean isLockHeld(ZooKeeper zk, LockID lid) throws KeeperException, InterruptedException {
-    
+
     List<String> children = zk.getChildren(lid.path, false);
-    
+
     if (children == null || children.size() == 0) {
       return false;
     }
-    
+
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
     if (!lid.node.equals(lockNode))
       return false;
-    
+
     Stat stat = zk.exists(lid.path + "/" + lid.node, false);
     return stat != null && stat.getEphemeralOwner() == lid.eid;
   }
-  
+
   public static boolean isLockHeld(ZooCache zc, LockID lid) {
-    
+
     List<String> children = zc.getChildren(lid.path);
-    
+
     if (children == null || children.size() == 0) {
       return false;
     }
-    
+
     children = new ArrayList<String>(children);
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
     if (!lid.node.equals(lockNode))
       return false;
-    
+
     Stat stat = new Stat();
     return zc.get(lid.path + "/" + lid.node, stat) != null && stat.getEphemeralOwner() == lid.eid;
   }
-  
+
   public static byte[] getLockData(ZooKeeper zk, String path) throws KeeperException, InterruptedException {
     List<String> children = zk.getChildren(path, false);
-    
+
     if (children == null || children.size() == 0) {
       return null;
     }
-    
+
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
-    
+
     return zk.getData(path + "/" + lockNode, false, null);
   }
-  
+
   public static byte[] getLockData(org.apache.accumulo.fate.zookeeper.ZooCache zc, String path, Stat stat) {
-    
+
     List<String> children = zc.getChildren(path);
-    
+
     if (children == null || children.size() == 0) {
       return null;
     }
-    
+
     children = new ArrayList<String>(children);
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
-    
+
     if (!lockNode.startsWith(LOCK_PREFIX)) {
       throw new RuntimeException("Node " + lockNode + " at " + path + " is not a lock node");
     }
-    
+
     return zc.get(path + "/" + lockNode, stat);
   }
-  
+
   public static long getSessionId(ZooCache zc, String path) throws KeeperException, InterruptedException {
     List<String> children = zc.getChildren(path);
-    
+
     if (children == null || children.size() == 0) {
       return 0;
     }
-    
+
     children = new ArrayList<String>(children);
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
-    
+
     Stat stat = new Stat();
     if (zc.get(path + "/" + lockNode, stat) != null)
       return stat.getEphemeralOwner();
     return 0;
   }
-  
+
   private static ZooCache getLockDataZooCache;
-  
+
   public long getSessionId() throws KeeperException, InterruptedException {
     return getSessionId(getLockDataZooCache, path);
   }
-  
+
   public static void deleteLock(IZooReaderWriter zk, String path) throws InterruptedException, KeeperException {
     List<String> children;
-    
+
     children = zk.getChildren(path);
-    
+
     if (children == null || children.size() == 0) {
       throw new IllegalStateException("No lock is held at " + path);
     }
-    
+
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
-    
+
     if (!lockNode.startsWith(LOCK_PREFIX)) {
       throw new RuntimeException("Node " + lockNode + " at " + path + " is not a lock node");
     }
-    
+
     zk.recursiveDelete(path + "/" + lockNode, NodeMissingPolicy.SKIP);
-    
+
   }
-  
+
   public static boolean deleteLock(IZooReaderWriter zk, String path, String lockData) throws InterruptedException, KeeperException {
     List<String> children;
-    
+
     children = zk.getChildren(path);
-    
+
     if (children == null || children.size() == 0) {
       throw new IllegalStateException("No lock is held at " + path);
     }
-    
+
     Collections.sort(children);
-    
+
     String lockNode = children.get(0);
-    
+
     if (!lockNode.startsWith(LOCK_PREFIX)) {
       throw new RuntimeException("Node " + lockNode + " at " + path + " is not a lock node");
     }
-    
+
     byte[] data = zk.getData(path + "/" + lockNode, null);
-    
+
     if (lockData.equals(new String(data, UTF_8))) {
       zk.recursiveDelete(path + "/" + lockNode, NodeMissingPolicy.FAIL);
       return true;
     }
-    
+
     return false;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriter.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriter.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriter.java
index 38c7b64..2786c4f 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriter.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriter.java
@@ -209,5 +209,4 @@ public class ZooReaderWriter extends ZooReader implements IZooReaderWriter {
     putPersistentData(path, new byte[] {}, NodeExistsPolicy.SKIP);
   }
 
-
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReservation.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReservation.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReservation.java
index f1ba428..c5a0ce3 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReservation.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooReservation.java
@@ -27,11 +27,11 @@ import org.apache.zookeeper.KeeperException.NodeExistsException;
 import org.apache.zookeeper.data.Stat;
 
 public class ZooReservation {
-  
+
   public static boolean attempt(IZooReaderWriter zk, String path, String reservationID, String debugInfo) throws KeeperException, InterruptedException {
     if (reservationID.contains(":"))
       throw new IllegalArgumentException();
-    
+
     while (true) {
       try {
         zk.putPersistentData(path, (reservationID + ":" + debugInfo).getBytes(UTF_8), NodeExistsPolicy.FAIL);
@@ -44,18 +44,18 @@ public class ZooReservation {
         } catch (NoNodeException nne) {
           continue;
         }
-        
+
         String idInZoo = new String(zooData, UTF_8).split(":")[0];
-        
+
         return idInZoo.equals(reservationID);
       }
     }
-    
+
   }
-  
+
   public static void release(IZooReaderWriter zk, String path, String reservationID) throws KeeperException, InterruptedException {
     byte[] zooData;
-    
+
     try {
       zooData = zk.getData(path, null);
     } catch (NoNodeException e) {
@@ -63,15 +63,15 @@ public class ZooReservation {
       Logger.getLogger(ZooReservation.class).debug("Node does not exist " + path);
       return;
     }
-    
+
     String zooDataStr = new String(zooData, UTF_8);
     String idInZoo = zooDataStr.split(":")[0];
-    
+
     if (!idInZoo.equals(reservationID)) {
       throw new IllegalStateException("Tried to release reservation " + path + " with data mismatch " + reservationID + " " + zooDataStr);
     }
-    
+
     zk.recursiveDelete(path, NodeMissingPolicy.SKIP);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooSession.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooSession.java b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooSession.java
index 0059af7..6b5ec43 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooSession.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/zookeeper/ZooSession.java
@@ -68,11 +68,16 @@ public class ZooSession {
   }
 
   /**
-   * @param host comma separated list of zk servers
-   * @param timeout in milliseconds
-   * @param scheme authentication type, e.g. 'digest', may be null
-   * @param auth authentication-scheme-specific token, may be null
-   * @param watcher ZK notifications, may be null
+   * @param host
+   *          comma separated list of zk servers
+   * @param timeout
+   *          in milliseconds
+   * @param scheme
+   *          authentication type, e.g. 'digest', may be null
+   * @param auth
+   *          authentication-scheme-specific token, may be null
+   * @param watcher
+   *          ZK notifications, may be null
    */
   public static ZooKeeper connect(String host, int timeout, String scheme, byte[] auth, Watcher watcher) {
     final int TIME_BETWEEN_CONNECT_CHECKS_MS = 100;
@@ -99,7 +104,7 @@ public class ZooSession {
       } catch (IOException e) {
         if (e instanceof UnknownHostException) {
           /*
-             Make sure we wait atleast as long as the JVM TTL for negative DNS responses
+           * Make sure we wait atleast as long as the JVM TTL for negative DNS responses
            */
           sleepTime = Math.max(sleepTime, (AddressUtil.getAddressCacheNegativeTtl((UnknownHostException) e) + 1) * 1000);
         }
@@ -121,14 +126,13 @@ public class ZooSession {
       if (tryAgain) {
         if (startTime + 2 * timeout < System.currentTimeMillis() + sleepTime + connectTimeWait)
           sleepTime = startTime + 2 * timeout - System.currentTimeMillis() - connectTimeWait;
-        if (sleepTime < 0)
-        {
+        if (sleepTime < 0) {
           connectTimeWait -= sleepTime;
           sleepTime = 0;
         }
         UtilWaitThread.sleep(sleepTime);
         if (sleepTime < 10000)
-          sleepTime = sleepTime + (long)(sleepTime * Math.random());
+          sleepTime = sleepTime + (long) (sleepTime * Math.random());
       }
     }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/AgeOffStoreTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/AgeOffStoreTest.java b/fate/src/test/java/org/apache/accumulo/fate/AgeOffStoreTest.java
index 4f5b112..518ab81 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/AgeOffStoreTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/AgeOffStoreTest.java
@@ -25,17 +25,17 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class AgeOffStoreTest {
-  
+
   private static class TestTimeSource implements TimeSource {
     long time = 0;
-    
+
     public long currentTimeMillis() {
       return time;
     }
-    
+
   }
 
   @Test
@@ -44,14 +44,14 @@ public class AgeOffStoreTest {
     TestTimeSource tts = new TestTimeSource();
     SimpleStore<String> sstore = new SimpleStore<String>();
     AgeOffStore<String> aoStore = new AgeOffStore<String>(sstore, 10, tts);
-    
+
     aoStore.ageOff();
 
     Long txid1 = aoStore.create();
     aoStore.reserve(txid1);
     aoStore.setStatus(txid1, TStatus.IN_PROGRESS);
     aoStore.unreserve(txid1, 0);
-    
+
     aoStore.ageOff();
 
     Long txid2 = aoStore.create();
@@ -59,7 +59,7 @@ public class AgeOffStoreTest {
     aoStore.setStatus(txid2, TStatus.IN_PROGRESS);
     aoStore.setStatus(txid2, TStatus.FAILED);
     aoStore.unreserve(txid2, 0);
-    
+
     tts.time = 6;
 
     Long txid3 = aoStore.create();
@@ -67,21 +67,21 @@ public class AgeOffStoreTest {
     aoStore.setStatus(txid3, TStatus.IN_PROGRESS);
     aoStore.setStatus(txid3, TStatus.SUCCESSFUL);
     aoStore.unreserve(txid3, 0);
-    
+
     Long txid4 = aoStore.create();
-    
+
     aoStore.ageOff();
 
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1, txid2, txid3, txid4)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(4, new HashSet<Long>(aoStore.list()).size());
-    
+
     tts.time = 15;
-    
+
     aoStore.ageOff();
-    
+
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1, txid3, txid4)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(3, new HashSet<Long>(aoStore.list()).size());
-    
+
     tts.time = 30;
 
     aoStore.ageOff();
@@ -89,71 +89,71 @@ public class AgeOffStoreTest {
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(1, new HashSet<Long>(aoStore.list()).size());
   }
-  
+
   @Test
   public void testNonEmpty() {
     // test age off when source store starts off non empty
-    
+
     TestTimeSource tts = new TestTimeSource();
     SimpleStore<String> sstore = new SimpleStore<String>();
     Long txid1 = sstore.create();
     sstore.reserve(txid1);
     sstore.setStatus(txid1, TStatus.IN_PROGRESS);
     sstore.unreserve(txid1, 0);
-    
+
     Long txid2 = sstore.create();
     sstore.reserve(txid2);
     sstore.setStatus(txid2, TStatus.IN_PROGRESS);
     sstore.setStatus(txid2, TStatus.FAILED);
     sstore.unreserve(txid2, 0);
-    
+
     Long txid3 = sstore.create();
     sstore.reserve(txid3);
     sstore.setStatus(txid3, TStatus.IN_PROGRESS);
     sstore.setStatus(txid3, TStatus.SUCCESSFUL);
     sstore.unreserve(txid3, 0);
-    
+
     Long txid4 = sstore.create();
-    
+
     AgeOffStore<String> aoStore = new AgeOffStore<String>(sstore, 10, tts);
-    
+
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1, txid2, txid3, txid4)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(4, new HashSet<Long>(aoStore.list()).size());
-    
+
     aoStore.ageOff();
-    
+
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1, txid2, txid3, txid4)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(4, new HashSet<Long>(aoStore.list()).size());
-    
+
     tts.time = 15;
 
     aoStore.ageOff();
-    
+
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(1, new HashSet<Long>(aoStore.list()).size());
-    
+
     aoStore.reserve(txid1);
     aoStore.setStatus(txid1, TStatus.FAILED_IN_PROGRESS);
     aoStore.unreserve(txid1, 0);
-    
+
     tts.time = 30;
-    
+
     aoStore.ageOff();
-    
+
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(1, new HashSet<Long>(aoStore.list()).size());
-    
+
     aoStore.reserve(txid1);
     aoStore.setStatus(txid1, TStatus.FAILED);
     aoStore.unreserve(txid1, 0);
-    
+
     aoStore.ageOff();
-    
+
     Assert.assertEquals(new HashSet<Long>(Arrays.asList(txid1)), new HashSet<Long>(aoStore.list()));
     Assert.assertEquals(1, new HashSet<Long>(aoStore.list()).size());
-    
+
     tts.time = 42;
-    
+
     aoStore.ageOff();
 
     Assert.assertEquals(0, new HashSet<Long>(aoStore.list()).size());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/ReadOnlyStoreTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/ReadOnlyStoreTest.java b/fate/src/test/java/org/apache/accumulo/fate/ReadOnlyStoreTest.java
index c2d5f92..eea5f1b 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/ReadOnlyStoreTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/ReadOnlyStoreTest.java
@@ -20,7 +20,6 @@ import java.util.Collections;
 import java.util.EnumSet;
 
 import org.apache.accumulo.fate.ReadOnlyTStore.TStatus;
-
 import org.easymock.EasyMock;
 import org.junit.Assert;
 import org.junit.Test;
@@ -47,7 +46,7 @@ public class ReadOnlyStoreTest {
 
     EasyMock.expect(mock.waitForStatusChange(0xdeadbeefl, EnumSet.allOf(TStatus.class))).andReturn(TStatus.UNKNOWN);
     EasyMock.expect(mock.getProperty(0xdeadbeefl, "com.example.anyproperty")).andReturn("property");
-    EasyMock.expect(mock.list()).andReturn(Collections.<Long>emptyList());
+    EasyMock.expect(mock.list()).andReturn(Collections.<Long> emptyList());
 
     EasyMock.replay(repo);
     EasyMock.replay(mock);
@@ -64,7 +63,7 @@ public class ReadOnlyStoreTest {
 
     Assert.assertEquals(TStatus.UNKNOWN, store.waitForStatusChange(0xdeadbeefl, EnumSet.allOf(TStatus.class)));
     Assert.assertEquals("property", store.getProperty(0xdeadbeefl, "com.example.anyproperty"));
-    Assert.assertEquals(Collections.<Long>emptyList(), store.list());
+    Assert.assertEquals(Collections.<Long> emptyList(), store.list());
 
     EasyMock.verify(repo);
     EasyMock.verify(mock);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/SimpleStore.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/SimpleStore.java b/fate/src/test/java/org/apache/accumulo/fate/SimpleStore.java
index 60eabfb..f0bac88 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/SimpleStore.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/SimpleStore.java
@@ -31,62 +31,62 @@ import org.apache.commons.lang.NotImplementedException;
  * Transient in memory store for transactions.
  */
 public class SimpleStore<T> implements TStore<T> {
-  
+
   private long nextId = 1;
   private Map<Long,TStatus> statuses = new HashMap<Long,TStore.TStatus>();
   private Set<Long> reserved = new HashSet<Long>();
-  
+
   @Override
   public long create() {
     statuses.put(nextId, TStatus.NEW);
     return nextId++;
   }
-  
+
   @Override
   public long reserve() {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public void reserve(long tid) {
     if (reserved.contains(tid))
       throw new IllegalStateException(); // zoo store would wait, but do not expect test to reserve twice... if test change, then change this
     reserved.add(tid);
   }
-  
+
   @Override
   public void unreserve(long tid, long deferTime) {
     if (!reserved.remove(tid)) {
       throw new IllegalStateException();
     }
   }
-  
+
   @Override
   public Repo<T> top(long tid) {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public void push(long tid, Repo<T> repo) throws StackOverflowException {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public void pop(long tid) {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public org.apache.accumulo.fate.TStore.TStatus getStatus(long tid) {
     if (!reserved.contains(tid))
       throw new IllegalStateException();
-    
+
     TStatus status = statuses.get(tid);
     if (status == null)
       return TStatus.UNKNOWN;
     return status;
   }
-  
+
   @Override
   public void setStatus(long tid, org.apache.accumulo.fate.TStore.TStatus status) {
     if (!reserved.contains(tid))
@@ -95,32 +95,32 @@ public class SimpleStore<T> implements TStore<T> {
       throw new IllegalStateException();
     statuses.put(tid, status);
   }
-  
+
   @Override
   public org.apache.accumulo.fate.TStore.TStatus waitForStatusChange(long tid, EnumSet<org.apache.accumulo.fate.TStore.TStatus> expected) {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public void setProperty(long tid, String prop, Serializable val) {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public Serializable getProperty(long tid, String prop) {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public void delete(long tid) {
     if (!reserved.contains(tid))
       throw new IllegalStateException();
     statuses.remove(tid);
   }
-  
+
   @Override
   public List<Long> list() {
     return new ArrayList<Long>(statuses.keySet());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/util/AddressUtilTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/util/AddressUtilTest.java b/fate/src/test/java/org/apache/accumulo/fate/util/AddressUtilTest.java
index 6e6a3c3..ab9032d 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/util/AddressUtilTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/util/AddressUtilTest.java
@@ -24,7 +24,7 @@ import org.apache.log4j.Logger;
 
 /**
  * Test the AddressUtil class.
- * 
+ *
  */
 public class AddressUtilTest extends TestCase {
 
@@ -88,7 +88,7 @@ public class AddressUtilTest extends TestCase {
       log.info("AddressUtil is (hopefully) going to spit out an error about DNS lookups. you can ignore it.");
       AddressUtil.getAddressCacheNegativeTtl(null);
       fail("The JVM Security settings cache DNS failures forever, this should cause an exception.");
-    } catch(IllegalArgumentException exception) {
+    } catch (IllegalArgumentException exception) {
       assertTrue(true);
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLockTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLockTest.java b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLockTest.java
index a9a8e3c..9e540e4 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLockTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/DistributedReadWriteLockTest.java
@@ -16,31 +16,31 @@
  */
 package org.apache.accumulo.fate.zookeeper;
 
+import static org.junit.Assert.assertEquals;
+
 import java.util.SortedMap;
 import java.util.TreeMap;
 import java.util.concurrent.locks.Lock;
 import java.util.concurrent.locks.ReadWriteLock;
 
-import static org.junit.Assert.*;
-
 import org.apache.accumulo.fate.zookeeper.DistributedReadWriteLock.QueueLock;
 import org.junit.Test;
 
 public class DistributedReadWriteLockTest {
-  
+
   // Non-zookeeper version of QueueLock
   public static class MockQueueLock implements QueueLock {
-    
+
     long next = 0L;
     final SortedMap<Long,byte[]> locks = new TreeMap<Long,byte[]>();
-    
+
     @Override
     synchronized public SortedMap<Long,byte[]> getEarlierEntries(long entry) {
       SortedMap<Long,byte[]> result = new TreeMap<Long,byte[]>();
       result.putAll(locks.headMap(entry + 1));
       return result;
     }
-    
+
     @Override
     synchronized public void removeEntry(long entry) {
       synchronized (locks) {
@@ -48,7 +48,7 @@ public class DistributedReadWriteLockTest {
         locks.notifyAll();
       }
     }
-    
+
     @Override
     synchronized public long addEntry(byte[] data) {
       long result;
@@ -59,31 +59,31 @@ public class DistributedReadWriteLockTest {
       return result;
     }
   }
-  
+
   // some data that is probably not going to update atomically
   static class SomeData {
     volatile int[] data = new int[100];
     volatile int counter;
-    
+
     void read() {
       for (int i = 0; i < data.length; i++)
         assertEquals(counter, data[i]);
     }
-    
+
     void write() {
       ++counter;
       for (int i = data.length - 1; i >= 0; i--)
         data[i] = counter;
     }
   }
-  
+
   @Test
   public void testLock() throws Exception {
     final SomeData data = new SomeData();
     data.write();
     data.read();
     QueueLock qlock = new MockQueueLock();
-    
+
     final ReadWriteLock locker = new DistributedReadWriteLock(qlock, "locker1".getBytes());
     final Lock readLock = locker.readLock();
     final Lock writeLock = locker.writeLock();
@@ -93,7 +93,7 @@ public class DistributedReadWriteLockTest {
     writeLock.unlock();
     readLock.lock();
     readLock.unlock();
-    
+
     // do a bunch of reads/writes in separate threads, look for inconsistent updates
     Thread[] threads = new Thread[2];
     for (int i = 0; i < threads.length; i++) {
@@ -128,5 +128,5 @@ public class DistributedReadWriteLockTest {
       t.join();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/zookeeper/TransactionWatcherTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/TransactionWatcherTest.java b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/TransactionWatcherTest.java
index 840d33b..0e4e329 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/TransactionWatcherTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/TransactionWatcherTest.java
@@ -23,15 +23,14 @@ import java.util.Map;
 import java.util.concurrent.Callable;
 
 import org.junit.Assert;
-
 import org.junit.Test;
 
 public class TransactionWatcherTest {
-  
+
   static class SimpleArbitrator implements TransactionWatcher.Arbitrator {
     Map<String,List<Long>> started = new HashMap<String,List<Long>>();
     Map<String,List<Long>> cleanedUp = new HashMap<String,List<Long>>();
-    
+
     public synchronized void start(String txType, Long txid) throws Exception {
       List<Long> txids = started.get(txType);
       if (txids == null)
@@ -40,7 +39,7 @@ public class TransactionWatcherTest {
         throw new Exception("transaction already started");
       txids.add(txid);
       started.put(txType, txids);
-      
+
       txids = cleanedUp.get(txType);
       if (txids == null)
         txids = new ArrayList<Long>();
@@ -49,7 +48,7 @@ public class TransactionWatcherTest {
       txids.add(txid);
       cleanedUp.put(txType, txids);
     }
-    
+
     public synchronized void stop(String txType, Long txid) throws Exception {
       List<Long> txids = started.get(txType);
       if (txids != null && txids.contains(txid)) {
@@ -58,7 +57,7 @@ public class TransactionWatcherTest {
       }
       throw new Exception("transaction does not exist");
     }
-    
+
     public synchronized void cleanup(String txType, Long txid) throws Exception {
       List<Long> txids = cleanedUp.get(txType);
       if (txids != null && txids.contains(txid)) {
@@ -67,7 +66,7 @@ public class TransactionWatcherTest {
       }
       throw new Exception("transaction does not exist");
     }
-    
+
     @Override
     synchronized public boolean transactionAlive(String txType, long tid) throws Exception {
       List<Long> txids = started.get(txType);
@@ -83,9 +82,9 @@ public class TransactionWatcherTest {
         return true;
       return !txids.contains(tid);
     }
-    
+
   }
-  
+
   @Test
   public void testTransactionWatcher() throws Exception {
     final String txType = "someName";
@@ -109,12 +108,12 @@ public class TransactionWatcherTest {
       }
     });
     Assert.assertFalse(txw.isActive(txid));
-    Assert.assertFalse(sa.transactionComplete(txType,  txid));
+    Assert.assertFalse(sa.transactionComplete(txType, txid));
     sa.stop(txType, txid);
     Assert.assertFalse(sa.transactionAlive(txType, txid));
-    Assert.assertFalse(sa.transactionComplete(txType,  txid));
+    Assert.assertFalse(sa.transactionComplete(txType, txid));
     sa.cleanup(txType, txid);
-    Assert.assertTrue(sa.transactionComplete(txType,  txid));
+    Assert.assertTrue(sa.transactionComplete(txType, txid));
     try {
       txw.run(txType, txid, new Callable<Object>() {
         @Override
@@ -150,7 +149,7 @@ public class TransactionWatcherTest {
         return null;
       }
     });
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactoryTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactoryTest.java b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactoryTest.java
index e7dffc1..19d8f67 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactoryTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheFactoryTest.java
@@ -16,15 +16,16 @@
  */
 package org.apache.accumulo.fate.zookeeper;
 
-import org.apache.zookeeper.Watcher;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
 import static org.easymock.EasyMock.createMock;
 import static org.junit.Assert.assertNotNull;
 import static org.junit.Assert.assertNotSame;
 import static org.junit.Assert.assertSame;
 
+import org.apache.zookeeper.Watcher;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
+
 public class ZooCacheFactoryTest {
   private ZooCacheFactory zcf;
 
@@ -65,6 +66,7 @@ public class ZooCacheFactoryTest {
     ZooCache zc1 = zcf.getZooCache(zks1, timeout1, watcher);
     assertNotNull(zc1);
   }
+
   @Test
   public void testGetZooCacheWatcher_Null() {
     String zks1 = "zk1";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheTest.java b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheTest.java
index e3db785..5dd6f61 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooCacheTest.java
@@ -16,22 +16,6 @@
  */
 package org.apache.accumulo.fate.zookeeper;
 
-import java.util.List;
-import org.apache.zookeeper.KeeperException;
-import org.apache.zookeeper.WatchedEvent;
-import org.apache.zookeeper.Watcher;
-import org.apache.zookeeper.ZooKeeper;
-import org.apache.zookeeper.data.Stat;
-
-import org.junit.Before;
-import org.junit.Test;
-import static org.junit.Assert.assertArrayEquals;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertNull;
-import static org.junit.Assert.assertSame;
-import static org.junit.Assert.assertTrue;
-import org.easymock.Capture;
 import static org.easymock.EasyMock.anyObject;
 import static org.easymock.EasyMock.capture;
 import static org.easymock.EasyMock.createMock;
@@ -41,6 +25,23 @@ import static org.easymock.EasyMock.expect;
 import static org.easymock.EasyMock.expectLastCall;
 import static org.easymock.EasyMock.replay;
 import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertArrayEquals;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertNull;
+import static org.junit.Assert.assertSame;
+import static org.junit.Assert.assertTrue;
+
+import java.util.List;
+
+import org.apache.zookeeper.KeeperException;
+import org.apache.zookeeper.WatchedEvent;
+import org.apache.zookeeper.Watcher;
+import org.apache.zookeeper.ZooKeeper;
+import org.apache.zookeeper.data.Stat;
+import org.easymock.Capture;
+import org.junit.Before;
+import org.junit.Test;
 
 public class ZooCacheTest {
   private static final String ZPATH = "/some/path/in/zk";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriterTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriterTest.java b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriterTest.java
index 59fb498..9203b39 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriterTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooReaderWriterTest.java
@@ -96,7 +96,7 @@ public class ZooReaderWriterTest {
   @Test(expected = SessionExpiredException.class)
   public void testMutateNodeCreationFails() throws Exception {
     final String path = "/foo";
-    final byte[] value = new byte[]{0};
+    final byte[] value = new byte[] {0};
     final List<ACL> acls = Collections.<ACL> emptyList();
     Mutator mutator = new Mutator() {
       @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooSessionTest.java
----------------------------------------------------------------------
diff --git a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooSessionTest.java b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooSessionTest.java
index fae7e6e..82d3d1d 100644
--- a/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooSessionTest.java
+++ b/fate/src/test/java/org/apache/accumulo/fate/zookeeper/ZooSessionTest.java
@@ -21,10 +21,10 @@ import org.junit.Test;
 
 public class ZooSessionTest {
 
-  private static final int MINIMUM_TIMEOUT=10000;
+  private static final int MINIMUM_TIMEOUT = 10000;
   private static final String UNKNOWN_HOST = "hostname.that.should.not.exist.example.com:2181";
 
-  @Test(expected=RuntimeException.class, timeout=MINIMUM_TIMEOUT*4)
+  @Test(expected = RuntimeException.class, timeout = MINIMUM_TIMEOUT * 4)
   public void testUnknownHost() throws Exception {
     ZooKeeper session = ZooSession.connect(UNKNOWN_HOST, MINIMUM_TIMEOUT, null, null, null);
     session.close();


[08/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsRandomAccessContent.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsRandomAccessContent.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsRandomAccessContent.java
index aada2c7..a9bbe55 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsRandomAccessContent.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsRandomAccessContent.java
@@ -30,16 +30,16 @@ import org.apache.hadoop.fs.Path;
 /**
  * Provides random access to content in an HdfsFileObject. Currently this only supports read operations. All write operations throw an
  * {@link UnsupportedOperationException}.
- * 
+ *
  * @since 2.1
  */
 public class HdfsRandomAccessContent implements RandomAccessContent {
   private final FileSystem fs;
   private final Path path;
   private final FSDataInputStream fis;
-  
+
   /**
-   * 
+   *
    * @param path
    *          A Hadoop Path
    * @param fs
@@ -52,7 +52,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
     this.path = path;
     this.fis = this.fs.open(this.path);
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.RandomAccessContent#close()
    */
@@ -60,7 +60,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void close() throws IOException {
     this.fis.close();
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.RandomAccessContent#getFilePointer()
    */
@@ -68,7 +68,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public long getFilePointer() throws IOException {
     return this.fis.getPos();
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.RandomAccessContent#getInputStream()
    */
@@ -76,7 +76,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public InputStream getInputStream() throws IOException {
     return this.fis;
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.RandomAccessContent#length()
    */
@@ -84,7 +84,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public long length() throws IOException {
     return this.fs.getFileStatus(this.path).getLen();
   }
-  
+
   /**
    * @see java.io.DataInput#readBoolean()
    */
@@ -92,7 +92,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public boolean readBoolean() throws IOException {
     return this.fis.readBoolean();
   }
-  
+
   /**
    * @see java.io.DataInput#readByte()
    */
@@ -100,7 +100,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public byte readByte() throws IOException {
     return this.fis.readByte();
   }
-  
+
   /**
    * @see java.io.DataInput#readChar()
    */
@@ -108,7 +108,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public char readChar() throws IOException {
     return this.fis.readChar();
   }
-  
+
   /**
    * @see java.io.DataInput#readDouble()
    */
@@ -116,7 +116,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public double readDouble() throws IOException {
     return this.fis.readDouble();
   }
-  
+
   /**
    * @see java.io.DataInput#readFloat()
    */
@@ -124,7 +124,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public float readFloat() throws IOException {
     return this.fis.readFloat();
   }
-  
+
   /**
    * @see java.io.DataInput#readFully(byte[])
    */
@@ -132,7 +132,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void readFully(final byte[] b) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataInput#readFully(byte[], int, int)
    */
@@ -140,7 +140,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void readFully(final byte[] b, final int off, final int len) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataInput#readInt()
    */
@@ -148,7 +148,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public int readInt() throws IOException {
     return this.fis.readInt();
   }
-  
+
   /**
    * @see java.io.DataInput#readLine()
    */
@@ -157,7 +157,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
     BufferedReader d = new BufferedReader(new InputStreamReader(this.fis, Charset.forName("UTF-8")));
     return d.readLine();
   }
-  
+
   /**
    * @see java.io.DataInput#readLong()
    */
@@ -165,7 +165,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public long readLong() throws IOException {
     return this.fis.readLong();
   }
-  
+
   /**
    * @see java.io.DataInput#readShort()
    */
@@ -173,7 +173,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public short readShort() throws IOException {
     return this.fis.readShort();
   }
-  
+
   /**
    * @see java.io.DataInput#readUnsignedByte()
    */
@@ -181,7 +181,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public int readUnsignedByte() throws IOException {
     return this.fis.readUnsignedByte();
   }
-  
+
   /**
    * @see java.io.DataInput#readUnsignedShort()
    */
@@ -189,7 +189,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public int readUnsignedShort() throws IOException {
     return this.fis.readUnsignedShort();
   }
-  
+
   /**
    * @see java.io.DataInput#readUTF()
    */
@@ -197,7 +197,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public String readUTF() throws IOException {
     return this.fis.readUTF();
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.RandomAccessContent#seek(long)
    */
@@ -205,7 +205,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void seek(final long pos) throws IOException {
     this.fis.seek(pos);
   }
-  
+
   /**
    * @see java.io.DataInput#skipBytes(int)
    */
@@ -213,7 +213,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public int skipBytes(final int n) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#write(byte[])
    */
@@ -221,7 +221,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void write(final byte[] b) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#write(byte[], int, int)
    */
@@ -229,7 +229,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void write(final byte[] b, final int off, final int len) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#write(int)
    */
@@ -237,7 +237,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void write(final int b) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeBoolean(boolean)
    */
@@ -245,7 +245,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeBoolean(final boolean v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeByte(int)
    */
@@ -253,7 +253,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeByte(final int v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeBytes(java.lang.String)
    */
@@ -261,7 +261,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeBytes(final String s) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeChar(int)
    */
@@ -269,7 +269,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeChar(final int v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeChars(java.lang.String)
    */
@@ -277,7 +277,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeChars(final String s) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeDouble(double)
    */
@@ -285,7 +285,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeDouble(final double v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeFloat(float)
    */
@@ -293,7 +293,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeFloat(final float v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeInt(int)
    */
@@ -301,7 +301,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeInt(final int v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeLong(long)
    */
@@ -309,7 +309,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeLong(final long v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeShort(int)
    */
@@ -317,7 +317,7 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeShort(final int v) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   /**
    * @see java.io.DataOutput#writeUTF(java.lang.String)
    */
@@ -325,5 +325,5 @@ public class HdfsRandomAccessContent implements RandomAccessContent {
   public void writeUTF(final String s) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/test/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoaderTest.java
----------------------------------------------------------------------
diff --git a/start/src/test/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoaderTest.java b/start/src/test/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoaderTest.java
index 3d52832..5adab86 100644
--- a/start/src/test/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoaderTest.java
+++ b/start/src/test/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoaderTest.java
@@ -132,7 +132,8 @@ public class AccumuloReloadingVFSClassLoaderTest {
   //
   // This is caused by the filed being deleted and then readded in the same monitor tick. This causes the file to ultimately register the deletion over any
   // other events.
-  @Test @Ignore
+  @Test
+  @Ignore
   public void testFastDeleteAndReAdd() throws Exception {
     FileObject testDir = vfs.resolveFile(folder1.getRoot().toURI().toString());
     FileObject[] dirContents = testDir.getChildren();
@@ -161,10 +162,10 @@ public class AccumuloReloadingVFSClassLoaderTest {
     // Update the class
     FileUtils.copyURLToFile(this.getClass().getResource("/HelloWorld.jar"), folder1.newFile("HelloWorld.jar"));
 
-    //Wait for the monitor to notice
+    // Wait for the monitor to notice
     // VFS-487 significantly wait to avoid failure
     Thread.sleep(7000);
-    
+
     Class<?> clazz2 = arvcl.getClassLoader().loadClass("test.HelloWorld");
     Object o2 = clazz2.newInstance();
     Assert.assertEquals("Hello World!", o2.toString());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/test/java/org/apache/accumulo/start/classloader/vfs/ContextManagerTest.java
----------------------------------------------------------------------
diff --git a/start/src/test/java/org/apache/accumulo/start/classloader/vfs/ContextManagerTest.java b/start/src/test/java/org/apache/accumulo/start/classloader/vfs/ContextManagerTest.java
index 65f0292..d1e6813 100644
--- a/start/src/test/java/org/apache/accumulo/start/classloader/vfs/ContextManagerTest.java
+++ b/start/src/test/java/org/apache/accumulo/start/classloader/vfs/ContextManagerTest.java
@@ -60,7 +60,7 @@ public class ContextManagerTest {
     FileUtils.copyURLToFile(this.getClass().getResource("/HelloWorld.jar"), folder2.newFile("HelloWorld.jar"));
 
     uri1 = new File(folder1.getRoot(), "HelloWorld.jar").toURI().toString();
-    uri2 = folder2.getRoot().toURI().toString()+".*";
+    uri2 = folder2.getRoot().toURI().toString() + ".*";
 
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/test/java/org/apache/accumulo/start/classloader/vfs/providers/ReadOnlyHdfsFileProviderTest.java
----------------------------------------------------------------------
diff --git a/start/src/test/java/org/apache/accumulo/start/classloader/vfs/providers/ReadOnlyHdfsFileProviderTest.java b/start/src/test/java/org/apache/accumulo/start/classloader/vfs/providers/ReadOnlyHdfsFileProviderTest.java
index 8873d17..b5cec83 100644
--- a/start/src/test/java/org/apache/accumulo/start/classloader/vfs/providers/ReadOnlyHdfsFileProviderTest.java
+++ b/start/src/test/java/org/apache/accumulo/start/classloader/vfs/providers/ReadOnlyHdfsFileProviderTest.java
@@ -33,12 +33,12 @@ import org.junit.Before;
 import org.junit.Test;
 
 public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
-  
+
   private static final String TEST_DIR1 = getHdfsUri() + "/test-dir";
   private static final Path DIR1_PATH = new Path("/test-dir");
   private static final String TEST_FILE1 = TEST_DIR1 + "/accumulo-test-1.jar";
   private static final Path FILE1_PATH = new Path(DIR1_PATH, "accumulo-test-1.jar");
-  
+
   private DefaultFileSystemManager manager = null;
   private FileSystem hdfs = null;
 
@@ -49,16 +49,16 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     manager.init();
     this.hdfs = cluster.getFileSystem();
   }
-    
+
   private FileObject createTestFile(FileSystem hdfs) throws IOException {
-    //Create the directory
+    // Create the directory
     hdfs.mkdirs(DIR1_PATH);
     FileObject dir = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(dir);
     Assert.assertTrue(dir.exists());
     Assert.assertTrue(dir.getType().equals(FileType.FOLDER));
-    
-    //Create the file in the directory
+
+    // Create the file in the directory
     hdfs.create(FILE1_PATH).close();
     FileObject f = manager.resolveFile(TEST_FILE1);
     Assert.assertNotNull(f);
@@ -66,84 +66,84 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     Assert.assertTrue(f.getType().equals(FileType.FILE));
     return f;
   }
-  
+
   @Test
   public void testInit() throws Exception {
     FileObject fo = manager.resolveFile(TEST_FILE1);
     Assert.assertNotNull(fo);
   }
-  
+
   @Test
   public void testExistsFails() throws Exception {
     FileObject fo = manager.resolveFile(TEST_FILE1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
   }
-  
+
   @Test
   public void testExistsSucceeds() throws Exception {
     FileObject fo = manager.resolveFile(TEST_FILE1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
-    
-    //Create the file
+
+    // Create the file
     @SuppressWarnings("unused")
     FileObject f = createTestFile(hdfs);
-    
+
   }
 
-  @Test(expected=UnsupportedOperationException.class)
+  @Test(expected = UnsupportedOperationException.class)
   public void testCanRenameTo() throws Exception {
     FileObject fo = createTestFile(this.hdfs);
     Assert.assertNotNull(fo);
     fo.canRenameTo(fo);
   }
-  
+
   @Test
   public void testDoListChildren() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     FileObject dir = file.getParent();
-    
+
     FileObject[] children = dir.getChildren();
     Assert.assertTrue(children.length == 1);
     Assert.assertTrue(children[0].getName().equals(file.getName()));
-    
+
   }
-  
+
   @Test
   public void testGetContentSize() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     Assert.assertEquals(0, file.getContent().getSize());
   }
-  
+
   @Test
   public void testGetInputStream() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     file.getContent().getInputStream().close();
   }
-  
+
   @Test
   public void testIsHidden() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     Assert.assertFalse(file.isHidden());
   }
@@ -154,7 +154,7 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     Assert.assertTrue(file.isReadable());
   }
@@ -165,29 +165,29 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     Assert.assertFalse(file.isWriteable());
   }
-  
+
   @Test
   public void testLastModificationTime() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     Assert.assertFalse(-1 == file.getContent().getLastModifiedTime());
   }
-  
+
   @Test
   public void testGetAttributes() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     Map<String,Object> attributes = file.getContent().getAttributes();
     Assert.assertTrue(attributes.containsKey(HdfsFileAttributes.BLOCK_SIZE.toString()));
@@ -198,14 +198,14 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     Assert.assertTrue(attributes.containsKey(HdfsFileAttributes.OWNER.toString()));
     Assert.assertTrue(attributes.containsKey(HdfsFileAttributes.PERMISSIONS.toString()));
   }
-  
-  @Test(expected=FileSystemException.class)
+
+  @Test(expected = FileSystemException.class)
   public void testRandomAccessContent() throws Exception {
     FileObject fo = manager.resolveFile(TEST_DIR1);
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     file.getContent().getRandomAccessContent(RandomAccessMode.READWRITE).close();
   }
@@ -216,7 +216,7 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
     file.getContent().getRandomAccessContent(RandomAccessMode.READ).close();
   }
@@ -227,20 +227,20 @@ public class ReadOnlyHdfsFileProviderTest extends AccumuloDFSBase {
     Assert.assertNotNull(fo);
     Assert.assertFalse(fo.exists());
 
-    //Create the test file
+    // Create the test file
     FileObject file = createTestFile(hdfs);
-    //Get a handle to the same file
+    // Get a handle to the same file
     FileObject file2 = manager.resolveFile(TEST_FILE1);
     Assert.assertEquals(file, file2);
   }
-  
+
   @After
   public void tearDown() throws Exception {
     if (null != hdfs) {
       hdfs.delete(DIR1_PATH, true);
-      hdfs.close();  
+      hdfs.close();
     }
     manager.close();
   }
-    
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/test/java/org/apache/accumulo/test/AccumuloDFSBase.java
----------------------------------------------------------------------
diff --git a/start/src/test/java/org/apache/accumulo/test/AccumuloDFSBase.java b/start/src/test/java/org/apache/accumulo/test/AccumuloDFSBase.java
index 8e2d534..feab493 100644
--- a/start/src/test/java/org/apache/accumulo/test/AccumuloDFSBase.java
+++ b/start/src/test/java/org/apache/accumulo/test/AccumuloDFSBase.java
@@ -41,7 +41,7 @@ public class AccumuloDFSBase {
   protected static Configuration conf = null;
   protected static DefaultFileSystemManager vfs = null;
   protected static MiniDFSCluster cluster = null;
- 
+
   private static URI HDFS_URI;
 
   protected static URI getHdfsUri() {
@@ -53,7 +53,7 @@ public class AccumuloDFSBase {
     System.setProperty("java.io.tmpdir", System.getProperty("user.dir") + "/target");
     // System.setProperty("org.apache.commons.logging.Log", "org.apache.commons.logging.impl.NoOpLog");
     // Logger.getRootLogger().setLevel(Level.ERROR);
-    
+
     // Put the MiniDFSCluster directory in the target directory
     System.setProperty("test.build.data", "target/build/test/data");
 
@@ -63,7 +63,7 @@ public class AccumuloDFSBase {
 
     conf.set("dfs.datanode.data.dir.perm", MiniDFSUtil.computeDatanodeDirectoryPermission());
     conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 1024 * 1024); // 1M blocksize
-    
+
     try {
       cluster = new MiniDFSCluster(conf, 1, true, null);
       cluster.waitClusterUp();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/test/java/test/Test.java
----------------------------------------------------------------------
diff --git a/start/src/test/java/test/Test.java b/start/src/test/java/test/Test.java
index da8cd49..849199f 100644
--- a/start/src/test/java/test/Test.java
+++ b/start/src/test/java/test/Test.java
@@ -17,9 +17,9 @@
 package test;
 
 public interface Test {
-  
+
   public String hello();
-  
+
   public int add();
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/BulkImportDirectory.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/BulkImportDirectory.java b/test/src/main/java/org/apache/accumulo/test/BulkImportDirectory.java
index 8693d03..a0cc26e 100644
--- a/test/src/main/java/org/apache/accumulo/test/BulkImportDirectory.java
+++ b/test/src/main/java/org/apache/accumulo/test/BulkImportDirectory.java
@@ -36,15 +36,14 @@ import com.beust.jcommander.Parameter;
 
 public class BulkImportDirectory {
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(names={"-s","--source"}, description="directory to import from")
+    @Parameter(names = {"-s", "--source"}, description = "directory to import from")
     String source = null;
-    @Parameter(names={"-f","--failures"}, description="directory to copy failures into: will be deleted before the bulk import")
+    @Parameter(names = {"-f", "--failures"}, description = "directory to copy failures into: will be deleted before the bulk import")
     String failures = null;
-    @Parameter(description="<username> <password> <tablename> <sourcedir> <failuredir>")
+    @Parameter(description = "<username> <password> <tablename> <sourcedir> <failuredir>")
     List<String> args = new ArrayList<String>();
   }
-  
-  
+
   public static void main(String[] args) throws IOException, AccumuloException, AccumuloSecurityException, TableNotFoundException {
     final FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
     Opts opts = new Opts();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java b/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java
index c113b87..9b99f39 100644
--- a/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java
+++ b/test/src/main/java/org/apache/accumulo/test/CreateRFiles.java
@@ -21,68 +21,69 @@ import java.util.concurrent.Executors;
 import java.util.concurrent.TimeUnit;
 
 import org.apache.accumulo.core.cli.Help;
+import org.apache.log4j.Logger;
 
 import com.beust.jcommander.Parameter;
-import org.apache.log4j.Logger;
 
 public class CreateRFiles {
-  
+
   private static final Logger log = Logger.getLogger(CreateRFiles.class);
-  
+
   static class Opts extends Help {
-    
-    @Parameter(names="--output", description="the destiation directory")
+
+    @Parameter(names = "--output", description = "the destiation directory")
     String outputDirectory;
-    
-    @Parameter(names="--numThreads", description="number of threads to use when generating files")
+
+    @Parameter(names = "--numThreads", description = "number of threads to use when generating files")
     int numThreads = 4;
-    
-    @Parameter(names="--start", description="the start number for test data")
+
+    @Parameter(names = "--start", description = "the start number for test data")
     long start = 0;
-    
-    @Parameter(names="--end", description="the maximum number for test data")
-    long end = 10*1000*1000;
-    
-    @Parameter(names="--splits", description="the number of splits in the data")
+
+    @Parameter(names = "--end", description = "the maximum number for test data")
+    long end = 10 * 1000 * 1000;
+
+    @Parameter(names = "--splits", description = "the number of splits in the data")
     long numsplits = 4;
   }
-  
+
   public static void main(String[] args) {
     Opts opts = new Opts();
     opts.parseArgs(CreateRFiles.class.getName(), args);
-    
+
     long splitSize = Math.round((opts.end - opts.start) / (double) opts.numsplits);
-    
+
     long currStart = opts.start;
     long currEnd = opts.start + splitSize;
-    
+
     ExecutorService threadPool = Executors.newFixedThreadPool(opts.numThreads);
-    
+
     int count = 0;
     while (currEnd <= opts.end && currStart < currEnd) {
-      
-      final String tia = String.format("--rfile %s/mf%05d --timestamp 1 --size 50 --random 56 --rows %d --start %d --user root", opts.outputDirectory, count, currEnd - currStart, currStart);
-      
+
+      final String tia = String.format("--rfile %s/mf%05d --timestamp 1 --size 50 --random 56 --rows %d --start %d --user root", opts.outputDirectory, count,
+          currEnd - currStart, currStart);
+
       Runnable r = new Runnable() {
-        
+
         @Override
         public void run() {
           try {
             TestIngest.main(tia.split(" "));
           } catch (Exception e) {
-            log.error("Could not run "+TestIngest.class.getName()+".main using the input '"+tia+"'", e);
+            log.error("Could not run " + TestIngest.class.getName() + ".main using the input '" + tia + "'", e);
           }
         }
-        
+
       };
-      
+
       threadPool.execute(r);
-      
+
       count++;
       currStart = currEnd;
       currEnd = Math.min(opts.end, currStart + splitSize);
     }
-    
+
     threadPool.shutdown();
     while (!threadPool.isTerminated())
       try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/CreateRandomRFile.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/CreateRandomRFile.java b/test/src/main/java/org/apache/accumulo/test/CreateRandomRFile.java
index 5ef1681..ada8504 100644
--- a/test/src/main/java/org/apache/accumulo/test/CreateRandomRFile.java
+++ b/test/src/main/java/org/apache/accumulo/test/CreateRandomRFile.java
@@ -33,21 +33,21 @@ import org.apache.hadoop.io.Text;
 public class CreateRandomRFile {
   private static int num;
   private static String file;
-  
+
   public static byte[] createValue(long rowid, int dataSize) {
     Random r = new Random(rowid);
     byte value[] = new byte[dataSize];
-    
+
     r.nextBytes(value);
-    
+
     // transform to printable chars
     for (int j = 0; j < value.length; j++) {
       value[j] = (byte) (((0xff & value[j]) % 92) + ' ');
     }
-    
+
     return value;
   }
-  
+
   public static void main(String[] args) {
     if (args.length != 2) {
       System.err.println("Usage CreateRandomRFile <filename> <size>");
@@ -56,15 +56,15 @@ public class CreateRandomRFile {
     file = args[0];
     num = Integer.parseInt(args[1]);
     long rands[] = new long[num];
-    
+
     Random r = new Random();
-    
+
     for (int i = 0; i < rands.length; i++) {
       rands[i] = (r.nextLong() & 0x7fffffffffffffffl) % 10000000000l;
     }
-    
+
     Arrays.sort(rands);
-    
+
     Configuration conf = CachedConfiguration.getInstance();
     FileSKVWriter mfw;
     try {
@@ -73,25 +73,25 @@ public class CreateRandomRFile {
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
     for (int i = 0; i < rands.length; i++) {
       Text row = new Text(String.format("row_%010d", rands[i]));
       Key key = new Key(row);
-      
+
       Value dv = new Value(createValue(rands[i], 40));
-      
+
       try {
         mfw.append(key, dv);
       } catch (IOException e) {
         throw new RuntimeException(e);
       }
     }
-    
+
     try {
       mfw.close();
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/EstimateInMemMapOverhead.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/EstimateInMemMapOverhead.java b/test/src/main/java/org/apache/accumulo/test/EstimateInMemMapOverhead.java
index 5716311..1ab2a8a 100644
--- a/test/src/main/java/org/apache/accumulo/test/EstimateInMemMapOverhead.java
+++ b/test/src/main/java/org/apache/accumulo/test/EstimateInMemMapOverhead.java
@@ -27,17 +27,17 @@ import org.apache.hadoop.io.Text;
 
 abstract class MemoryUsageTest {
   abstract void addEntry(int i);
-  
+
   abstract int getEstimatedBytesPerEntry();
-  
+
   abstract void clear();
-  
+
   abstract int getNumPasses();
-  
+
   abstract String getName();
-  
+
   abstract void init();
-  
+
   public void run() {
     System.gc();
     long usedMem = Runtime.getRuntime().totalMemory() - Runtime.getRuntime().freeMemory();
@@ -47,53 +47,53 @@ abstract class MemoryUsageTest {
       usedMem = Runtime.getRuntime().totalMemory() - Runtime.getRuntime().freeMemory();
       count++;
     }
-    
+
     init();
-    
+
     for (int i = 0; i < getNumPasses(); i++) {
       addEntry(i);
     }
-    
+
     System.gc();
-    
+
     long memSize = (Runtime.getRuntime().totalMemory() - Runtime.getRuntime().freeMemory()) - usedMem;
-    
+
     double actualBytesPerEntry = memSize / (double) getNumPasses();
     double expectedBytesPerEntry = getEstimatedBytesPerEntry();
     double diff = actualBytesPerEntry - expectedBytesPerEntry;
     double ratio = actualBytesPerEntry / expectedBytesPerEntry * 100;
-    
+
     System.out.printf("%30s | %,10d | %6.2fGB | %6.2f | %6.2f | %6.2f | %6.2f%s%n", getName(), getNumPasses(), memSize / (1024 * 1024 * 1024.0),
         actualBytesPerEntry, expectedBytesPerEntry, diff, ratio, "%");
-    
+
     clear();
-    
+
   }
-  
+
 }
 
 class TextMemoryUsageTest extends MemoryUsageTest {
-  
+
   private int keyLen;
   private int colFamLen;
   private int colQualLen;
   private int dataLen;
   private TreeMap<Text,Value> map;
   private int passes;
-  
+
   TextMemoryUsageTest(int passes, int keyLen, int colFamLen, int colQualLen, int dataLen) {
     this.keyLen = keyLen;
     this.colFamLen = colFamLen;
     this.colQualLen = colQualLen;
     this.dataLen = dataLen;
     this.passes = passes;
-    
+
   }
-  
+
   void init() {
     map = new TreeMap<Text,Value>();
   }
-  
+
   public void addEntry(int i) {
     Text key = new Text(String.format("%0" + keyLen + "d:%0" + colFamLen + "d:%0" + colQualLen + "d", i, 0, 0).getBytes());
     //
@@ -102,45 +102,45 @@ class TextMemoryUsageTest extends MemoryUsageTest {
       data[j] = (byte) (j % 10 + 65);
     }
     Value value = new Value(data);
-    
+
     map.put(key, value);
-    
+
   }
-  
+
   public void clear() {
     map.clear();
     map = null;
   }
-  
+
   public int getEstimatedBytesPerEntry() {
     return keyLen + colFamLen + colQualLen + dataLen;
   }
-  
+
   int getNumPasses() {
     return passes;
   }
-  
+
   String getName() {
     return "Text " + keyLen + " " + colFamLen + " " + colQualLen + " " + dataLen;
   }
-  
+
 }
 
 class InMemoryMapMemoryUsageTest extends MemoryUsageTest {
-  
+
   private int keyLen;
   private int colFamLen;
   private int colQualLen;
   private int colVisLen;
   private int dataLen;
-  
+
   private InMemoryMap imm;
   private Text key;
   private Text colf;
   private Text colq;
   private ColumnVisibility colv;
   private int passes;
-  
+
   InMemoryMapMemoryUsageTest(int passes, int keyLen, int colFamLen, int colQualLen, int colVisLen, int dataLen) {
     this.keyLen = keyLen;
     this.colFamLen = colFamLen;
@@ -148,68 +148,68 @@ class InMemoryMapMemoryUsageTest extends MemoryUsageTest {
     this.dataLen = dataLen;
     this.passes = passes;
     this.colVisLen = colVisLen;
-    
+
   }
-  
+
   void init() {
     imm = new InMemoryMap(false, "/tmp");
     key = new Text();
-    
+
     colf = new Text(String.format("%0" + colFamLen + "d", 0));
     colq = new Text(String.format("%0" + colQualLen + "d", 0));
     colv = new ColumnVisibility(String.format("%0" + colVisLen + "d", 0));
   }
-  
+
   public void addEntry(int i) {
     key.set(String.format("%0" + keyLen + "d", i));
-    
+
     Mutation m = new Mutation(key);
-    
+
     byte data[] = new byte[dataLen];
     for (int j = 0; j < data.length; j++) {
       data[j] = (byte) (j % 10 + 65);
     }
     Value idata = new Value(data);
-    
+
     m.put(colf, colq, colv, idata);
-    
+
     imm.mutate(Collections.singletonList(m));
-    
+
   }
-  
+
   public int getEstimatedBytesPerEntry() {
     return keyLen + colFamLen + colQualLen + dataLen + 4 + colVisLen;
   }
-  
+
   public void clear() {
     imm = null;
     key = null;
     colf = null;
     colq = null;
   }
-  
+
   int getNumPasses() {
     return passes;
   }
-  
+
   String getName() {
     return "IMM " + keyLen + " " + colFamLen + " " + colQualLen + " " + dataLen;
   }
 }
 
 class MutationMemoryUsageTest extends MemoryUsageTest {
-  
+
   private int keyLen;
   private int colFamLen;
   private int colQualLen;
   private int dataLen;
-  
+
   private Mutation[] mutations;
   private Text key;
   private Text colf;
   private Text colq;
   private int passes;
-  
+
   MutationMemoryUsageTest(int passes, int keyLen, int colFamLen, int colQualLen, int dataLen) {
     this.keyLen = keyLen;
     this.colFamLen = colFamLen;
@@ -217,126 +217,126 @@ class MutationMemoryUsageTest extends MemoryUsageTest {
     this.dataLen = dataLen;
     this.passes = passes;
     mutations = new Mutation[passes];
-    
+
   }
-  
+
   void init() {
     key = new Text();
-    
+
     colf = new Text(String.format("%0" + colFamLen + "d", 0));
     colq = new Text(String.format("%0" + colQualLen + "d", 0));
-    
+
     byte data[] = new byte[dataLen];
     for (int i = 0; i < data.length; i++) {
       data[i] = (byte) (i % 10 + 65);
     }
   }
-  
+
   public void addEntry(int i) {
     key.set(String.format("%0" + keyLen + "d", i));
-    
+
     Mutation m = new Mutation(key);
-    
+
     byte data[] = new byte[dataLen];
     for (int j = 0; j < data.length; j++) {
       data[j] = (byte) (j % 10 + 65);
     }
     Value idata = new Value(data);
-    
+
     m.put(colf, colq, idata);
-    
+
     mutations[i] = m;
   }
-  
+
   public int getEstimatedBytesPerEntry() {
     return keyLen + colFamLen + colQualLen + dataLen;
   }
-  
+
   public void clear() {
     key = null;
     colf = null;
     colq = null;
     mutations = null;
   }
-  
+
   int getNumPasses() {
     return passes;
   }
-  
+
   String getName() {
     return "Mutation " + keyLen + " " + colFamLen + " " + colQualLen + " " + dataLen;
   }
 }
 
 class IntObjectMemoryUsageTest extends MemoryUsageTest {
-  
+
   private int passes;
   private Object data[];
-  
+
   static class SimpleObject {
     int d;
-    
+
     SimpleObject(int d) {
       this.d = d;
     }
   }
-  
+
   IntObjectMemoryUsageTest(int numPasses) {
     this.passes = numPasses;
   }
-  
+
   void init() {
     data = new Object[passes];
   }
-  
+
   void addEntry(int i) {
     data[i] = new SimpleObject(i);
-    
+
   }
-  
+
   void clear() {}
-  
+
   int getEstimatedBytesPerEntry() {
     return 4;
   }
-  
+
   String getName() {
     return "int obj";
   }
-  
+
   int getNumPasses() {
     return passes;
   }
-  
+
 }
 
 public class EstimateInMemMapOverhead {
-  
+
   private static void runTest(int numEntries, int keyLen, int colFamLen, int colQualLen, int colVisLen, int dataLen) {
     new IntObjectMemoryUsageTest(numEntries).run();
     new InMemoryMapMemoryUsageTest(numEntries, keyLen, colFamLen, colQualLen, colVisLen, dataLen).run();
     new TextMemoryUsageTest(numEntries, keyLen, colFamLen, colQualLen, dataLen).run();
     new MutationMemoryUsageTest(numEntries, keyLen, colFamLen, colQualLen, dataLen).run();
   }
-  
+
   public static void main(String[] args) {
     runTest(10000, 10, 4, 4, 4, 20);
     runTest(100000, 10, 4, 4, 4, 20);
     runTest(500000, 10, 4, 4, 4, 20);
     runTest(1000000, 10, 4, 4, 4, 20);
     runTest(2000000, 10, 4, 4, 4, 20);
-    
+
     runTest(10000, 20, 5, 5, 5, 500);
     runTest(100000, 20, 5, 5, 5, 500);
     runTest(500000, 20, 5, 5, 5, 500);
     runTest(1000000, 20, 5, 5, 5, 500);
     runTest(2000000, 20, 5, 5, 5, 500);
-    
+
     runTest(10000, 40, 10, 10, 10, 1000);
     runTest(100000, 40, 10, 10, 10, 1000);
     runTest(500000, 40, 10, 10, 10, 1000);
     runTest(1000000, 40, 10, 10, 10, 1000);
     runTest(2000000, 40, 10, 10, 10, 1000);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/FaultyConditionalWriter.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/FaultyConditionalWriter.java b/test/src/main/java/org/apache/accumulo/test/FaultyConditionalWriter.java
index 7e7480f..673e61d 100644
--- a/test/src/main/java/org/apache/accumulo/test/FaultyConditionalWriter.java
+++ b/test/src/main/java/org/apache/accumulo/test/FaultyConditionalWriter.java
@@ -24,17 +24,16 @@ import java.util.Random;
 import org.apache.accumulo.core.client.ConditionalWriter;
 import org.apache.accumulo.core.data.ConditionalMutation;
 
-
 /**
  * A writer that will sometimes return unknown. When it returns unknown the condition may or may not have been written.
  */
 public class FaultyConditionalWriter implements ConditionalWriter {
-  
+
   private ConditionalWriter cw;
   private double up;
   private Random rand;
   private double wp;
-  
+
   public FaultyConditionalWriter(ConditionalWriter cw, double unknownProbability, double writeProbability) {
     this.cw = cw;
     this.up = unknownProbability;
@@ -46,7 +45,7 @@ public class FaultyConditionalWriter implements ConditionalWriter {
   public Iterator<Result> write(Iterator<ConditionalMutation> mutations) {
     ArrayList<Result> resultList = new ArrayList<Result>();
     ArrayList<ConditionalMutation> writes = new ArrayList<ConditionalMutation>();
-    
+
     while (mutations.hasNext()) {
       ConditionalMutation cm = mutations.next();
       if (rand.nextDouble() <= up && rand.nextDouble() > wp)
@@ -54,13 +53,13 @@ public class FaultyConditionalWriter implements ConditionalWriter {
       else
         writes.add(cm);
     }
-    
+
     if (writes.size() > 0) {
       Iterator<Result> results = cw.write(writes.iterator());
-      
+
       while (results.hasNext()) {
         Result result = results.next();
-        
+
         if (rand.nextDouble() <= up && rand.nextDouble() <= wp)
           result = new Result(Status.UNKNOWN, result.getMutation(), result.getTabletServer());
         resultList.add(result);
@@ -68,14 +67,14 @@ public class FaultyConditionalWriter implements ConditionalWriter {
     }
     return resultList.iterator();
   }
-  
+
   public Result write(ConditionalMutation mutation) {
     return write(Collections.singleton(mutation).iterator()).next();
   }
-  
+
   @Override
   public void close() {
     cw.close();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/GetMasterStats.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/GetMasterStats.java b/test/src/main/java/org/apache/accumulo/test/GetMasterStats.java
index 5f9ea51..5c2cbf3 100644
--- a/test/src/main/java/org/apache/accumulo/test/GetMasterStats.java
+++ b/test/src/main/java/org/apache/accumulo/test/GetMasterStats.java
@@ -56,7 +56,7 @@ public class GetMasterStats {
     out(0, "Unassigned tablets: %d", stats.unassignedTablets);
     if (stats.badTServers != null && stats.badTServers.size() > 0) {
       out(0, "Bad servers");
-      
+
       for (Entry<String,Byte> entry : stats.badTServers.entrySet()) {
         out(1, "%s: %d", entry.getKey(), (int) entry.getValue());
       }
@@ -120,12 +120,12 @@ public class GetMasterStats {
       }
     }
   }
-  
+
   private static void out(int indent, String string, Object... args) {
     for (int i = 0; i < indent; i++) {
       System.out.print(" ");
     }
     System.out.println(String.format(string, args));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/IMMLGBenchmark.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/IMMLGBenchmark.java b/test/src/main/java/org/apache/accumulo/test/IMMLGBenchmark.java
index f993ba1..3d86eab 100644
--- a/test/src/main/java/org/apache/accumulo/test/IMMLGBenchmark.java
+++ b/test/src/main/java/org/apache/accumulo/test/IMMLGBenchmark.java
@@ -48,21 +48,21 @@ import org.apache.accumulo.core.util.UtilWaitThread;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class IMMLGBenchmark {
   public static void main(String[] args) throws Exception {
     ZooKeeperInstance zki = new ZooKeeperInstance(new ClientConfiguration().withInstance("test16").withZkHosts("localhost"));
     Connector conn = zki.getConnector("root", new PasswordToken("secret"));
-    
+
     int numlg = Integer.parseInt(args[0]);
-    
+
     ArrayList<byte[]> cfset = new ArrayList<byte[]>();
-    
+
     for (int i = 0; i < 32; i++) {
       cfset.add(String.format("%04x", i).getBytes());
     }
-    
+
     Map<String,Stat> stats = new TreeMap<String,Stat>();
 
     for (int i = 0; i < 5; i++) {
@@ -78,14 +78,14 @@ public class IMMLGBenchmark {
 
   private static void runTest(Connector conn, int numlg, ArrayList<byte[]> cfset, Map<String,Stat> stats) throws Exception {
     String table = "immlgb";
-    
+
     try {
       conn.tableOperations().delete(table);
     } catch (TableNotFoundException tnfe) {}
 
     conn.tableOperations().create(table);
     conn.tableOperations().setProperty(table, Property.TABLE_FILE_COMPRESSION_TYPE.getKey(), "snappy");
-    
+
     setupLocalityGroups(conn, numlg, cfset, table);
 
     addStat(stats, "write", write(conn, cfset, table));
@@ -96,13 +96,13 @@ public class IMMLGBenchmark {
     long t1 = System.currentTimeMillis();
     conn.tableOperations().flush(table, null, null, true);
     long t2 = System.currentTimeMillis();
-    
+
     addStat(stats, "flush", t2 - t1);
   }
-  
+
   private static void addStat(Map<String,Stat> stats, String s, long wt) {
     System.out.println(s + ":" + wt);
-    
+
     if (stats == null)
       return;
 
@@ -113,55 +113,56 @@ public class IMMLGBenchmark {
     }
     stat.addStat(wt);
   }
-  
+
   private static long scan(Connector conn, ArrayList<byte[]> cfset, String table, boolean cq) throws TableNotFoundException {
     Scanner scanner = conn.createScanner(table, Authorizations.EMPTY);
-    
+
     if (!cq)
       scanner.fetchColumnFamily(new Text(cfset.get(15)));
     else
       scanner.fetchColumn(new Text(cfset.get(15)), new Text(cfset.get(15)));
 
     long t1 = System.currentTimeMillis();
-    
+
     @SuppressWarnings("unused")
     int count = 0;
-    for (@SuppressWarnings("unused") Entry<Key,Value> entry : scanner) {
+    for (@SuppressWarnings("unused")
+    Entry<Key,Value> entry : scanner) {
       count++;
     }
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     return t2 - t1;
-    
+
   }
-  
+
   private static long write(Connector conn, ArrayList<byte[]> cfset, String table) throws TableNotFoundException, MutationsRejectedException {
     Random rand = new Random();
-    
+
     byte val[] = new byte[50];
-    
+
     BatchWriter bw = conn.createBatchWriter(table, new BatchWriterConfig());
-    
+
     long t1 = System.currentTimeMillis();
 
     for (int i = 0; i < 1 << 15; i++) {
       byte[] row = FastFormat.toZeroPaddedString(abs(rand.nextLong()), 16, 16, new byte[0]);
-      
+
       Mutation m = new Mutation(row);
       for (byte[] cf : cfset) {
         byte[] cq = FastFormat.toZeroPaddedString(rand.nextInt(1 << 16), 4, 16, new byte[0]);
         rand.nextBytes(val);
         m.put(cf, cq, val);
       }
-      
+
       bw.addMutation(m);
     }
-    
+
     bw.close();
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     return t2 - t1;
   }
 
@@ -170,7 +171,7 @@ public class IMMLGBenchmark {
     if (numlg > 1) {
       int numCF = cfset.size() / numlg;
       int gNum = 0;
-      
+
       Iterator<byte[]> cfiter = cfset.iterator();
       Map<String,Set<Text>> groups = new HashMap<String,Set<Text>>();
       while (cfiter.hasNext()) {
@@ -178,19 +179,19 @@ public class IMMLGBenchmark {
         for (int i = 0; i < numCF && cfiter.hasNext(); i++) {
           groupCols.add(new Text(cfiter.next()));
         }
-        
+
         groups.put("lg" + (gNum++), groupCols);
       }
-      
+
       conn.tableOperations().setLocalityGroups(table, groups);
       conn.tableOperations().offline(table);
       UtilWaitThread.sleep(1000);
       conn.tableOperations().online(table);
     }
   }
-  
+
   public static long abs(long l) {
-    l = Math.abs(l);  // abs(Long.MIN_VALUE) == Long.MIN_VALUE... 
+    l = Math.abs(l); // abs(Long.MIN_VALUE) == Long.MIN_VALUE...
     if (l < 0)
       return 0;
     return l;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/ListTables.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/ListTables.java b/test/src/main/java/org/apache/accumulo/test/ListTables.java
index 468b2d5..be8a7d3 100644
--- a/test/src/main/java/org/apache/accumulo/test/ListTables.java
+++ b/test/src/main/java/org/apache/accumulo/test/ListTables.java
@@ -18,8 +18,8 @@ package org.apache.accumulo.test;
 
 import java.util.Map.Entry;
 
-import org.apache.accumulo.server.cli.ClientOpts;
 import org.apache.accumulo.core.client.impl.Tables;
+import org.apache.accumulo.server.cli.ClientOpts;
 
 /**
  * This little program is used by the functional test to get a list of table ids.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/NativeMapConcurrencyTest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/NativeMapConcurrencyTest.java b/test/src/main/java/org/apache/accumulo/test/NativeMapConcurrencyTest.java
index c9d18e1..015cda4 100644
--- a/test/src/main/java/org/apache/accumulo/test/NativeMapConcurrencyTest.java
+++ b/test/src/main/java/org/apache/accumulo/test/NativeMapConcurrencyTest.java
@@ -27,38 +27,38 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.util.FastFormat;
 import org.apache.accumulo.tserver.NativeMap;
 import org.apache.hadoop.io.Text;
+import org.apache.log4j.Logger;
 
 import com.beust.jcommander.JCommander;
 import com.beust.jcommander.Parameter;
-import org.apache.log4j.Logger;
 
 public class NativeMapConcurrencyTest {
-  
+
   private static final Logger log = Logger.getLogger(NativeMapConcurrencyTest.class);
-  
+
   private static final byte ROW_PREFIX[] = new byte[] {'r'};
   private static final byte COL_PREFIX[] = new byte[] {'c'};
-  
+
   static Mutation nm(int r) {
     return new Mutation(new Text(FastFormat.toZeroPaddedString(r, 6, 10, ROW_PREFIX)));
   }
-  
+
   private static final Text ET = new Text();
-  
+
   private static void pc(Mutation m, int c, Value v) {
     m.put(new Text(FastFormat.toZeroPaddedString(c, 3, 10, COL_PREFIX)), ET, v);
   }
-  
+
   static NativeMap create(int numRows, int numCols) {
-    
+
     NativeMap nm = new NativeMap();
-    
+
     populate(0, numRows, numCols, nm);
-    
+
     return nm;
-    
+
   }
-  
+
   private static void populate(int start, int numRows, int numCols, NativeMap nm) {
     long t1 = System.currentTimeMillis();
     int mc = 1;
@@ -71,27 +71,27 @@ public class NativeMapConcurrencyTest {
       nm.mutate(m, mc++);
     }
     long t2 = System.currentTimeMillis();
-    
+
     System.out.printf("inserted %,d in %,d %,d %,6.2f%n", (numRows * numCols), (t2 - t1), nm.size(), rate((numRows * numCols), (t2 - t1)));
   }
-  
+
   private static double rate(int num, long ms) {
     return num / (ms / 1000.0);
   }
-  
+
   static class Opts {
-    @Parameter(names="--rows", description="rows", required = true)
+    @Parameter(names = "--rows", description = "rows", required = true)
     int rows = 0;
-    @Parameter(names="--cols", description="cols")
+    @Parameter(names = "--cols", description = "cols")
     int cols = 1;
-    @Parameter(names="--threads", description="threads")
+    @Parameter(names = "--threads", description = "threads")
     int threads = 1;
-    @Parameter(names="--writeThreads", description="write threads")
+    @Parameter(names = "--writeThreads", description = "write threads")
     int writeThreads = 1;
-    @Parameter(names="-help", help=true)
+    @Parameter(names = "-help", help = true)
     boolean help = false;
   }
-  
+
   public static void main(String[] args) {
     Opts opts = new Opts();
     JCommander jc = new JCommander(opts);
@@ -105,86 +105,86 @@ public class NativeMapConcurrencyTest {
     runTest(nm, opts.rows, opts.cols, opts.threads, opts.writeThreads);
     nm.delete();
   }
-  
+
   static class ScanTask implements Runnable {
-    
+
     private NativeMap nm;
-    
+
     ScanTask(NativeMap nm) {
       this.nm = nm;
     }
-    
+
     @Override
     public void run() {
-      
+
       for (int i = 0; i < 10; i++) {
-        
+
         Iterator<Entry<Key,Value>> iter = nm.iterator();
-        
+
         long t1 = System.currentTimeMillis();
-        
+
         int count = 0;
-        
+
         while (iter.hasNext()) {
           count++;
           iter.next();
         }
-        
+
         long t2 = System.currentTimeMillis();
-        
+
         System.out.printf("%d %,d %,d %,d %,d %,6.2f%n", Thread.currentThread().getId(), (t2 - t1), t1, t2, count, rate(count, (t2 - t1)));
       }
     }
-    
+
   }
-  
+
   static class WriteTask implements Runnable {
-    
+
     private int start;
     private int rows;
     private int cols;
     private NativeMap nm;
-    
+
     WriteTask(int start, int rows, int cols, NativeMap nm) {
       this.start = start;
       this.rows = rows;
       this.cols = cols;
       this.nm = nm;
     }
-    
+
     @Override
     public void run() {
       populate(start, rows, cols, nm);
     }
-    
+
   }
-  
+
   private static void runTest(NativeMap nm, int rows, int cols, int numReadThreads, int writeThreads) {
-    
+
     Thread threads[] = new Thread[numReadThreads + writeThreads];
-    
+
     for (int i = 0; i < numReadThreads; i++) {
       threads[i] = new Thread(new ScanTask(nm));
     }
-    
+
     int start = 0;
     for (int i = numReadThreads; i < writeThreads + numReadThreads; i++) {
       threads[i] = new Thread(new WriteTask(start, rows, cols, nm));
       // start += rows;
     }
-    
+
     for (Thread thread : threads) {
       thread.start();
     }
-    
+
     for (Thread thread : threads) {
       try {
         thread.join();
       } catch (InterruptedException e) {
-        log.error("Could not join thread '"+thread.getName()+"'", e);
+        log.error("Could not join thread '" + thread.getName() + "'", e);
       }
     }
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/NativeMapPerformanceTest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/NativeMapPerformanceTest.java b/test/src/main/java/org/apache/accumulo/test/NativeMapPerformanceTest.java
index 1e1006b..0285092 100644
--- a/test/src/main/java/org/apache/accumulo/test/NativeMapPerformanceTest.java
+++ b/test/src/main/java/org/apache/accumulo/test/NativeMapPerformanceTest.java
@@ -35,29 +35,29 @@ import org.apache.accumulo.tserver.NativeMap;
 import org.apache.hadoop.io.Text;
 
 public class NativeMapPerformanceTest {
-  
+
   private static final byte ROW_PREFIX[] = new byte[] {'r'};
   private static final byte COL_PREFIX[] = new byte[] {'c'};
-  
+
   static Key nk(int r, int c) {
     return new Key(new Text(FastFormat.toZeroPaddedString(r, 9, 10, ROW_PREFIX)), new Text(FastFormat.toZeroPaddedString(c, 6, 10, COL_PREFIX)));
   }
-  
+
   static Mutation nm(int r) {
     return new Mutation(new Text(FastFormat.toZeroPaddedString(r, 9, 10, ROW_PREFIX)));
   }
-  
+
   static Text ET = new Text();
-  
+
   private static void pc(Mutation m, int c, Value v) {
     m.put(new Text(FastFormat.toZeroPaddedString(c, 6, 10, COL_PREFIX)), ET, Long.MAX_VALUE, v);
   }
-  
+
   static void runPerformanceTest(int numRows, int numCols, int numLookups, String mapType) {
-    
+
     SortedMap<Key,Value> tm = null;
     NativeMap nm = null;
-    
+
     if (mapType.equals("SKIP_LIST"))
       tm = new ConcurrentSkipListMap<Key,Value>();
     else if (mapType.equals("TREE_MAP"))
@@ -66,12 +66,12 @@ public class NativeMapPerformanceTest {
       nm = new NativeMap();
     else
       throw new IllegalArgumentException(" map type must be SKIP_LIST, TREE_MAP, or NATIVE_MAP");
-    
+
     Random rand = new Random(19);
-    
+
     // puts
     long tps = System.currentTimeMillis();
-    
+
     if (nm != null) {
       for (int i = 0; i < numRows; i++) {
         int row = rand.nextInt(1000000000);
@@ -94,9 +94,9 @@ public class NativeMapPerformanceTest {
         }
       }
     }
-    
+
     long tpe = System.currentTimeMillis();
-    
+
     // Iteration
     Iterator<Entry<Key,Value>> iter;
     if (nm != null) {
@@ -104,15 +104,15 @@ public class NativeMapPerformanceTest {
     } else {
       iter = tm.entrySet().iterator();
     }
-    
+
     long tis = System.currentTimeMillis();
-    
+
     while (iter.hasNext()) {
       iter.next();
     }
-    
+
     long tie = System.currentTimeMillis();
-    
+
     rand = new Random(19);
     int rowsToLookup[] = new int[numLookups];
     int colsToLookup[] = new int[numLookups];
@@ -122,13 +122,13 @@ public class NativeMapPerformanceTest {
       for (int j = 0; j < numCols; j++) {
         col = rand.nextInt(1000000);
       }
-      
+
       rowsToLookup[i] = row;
       colsToLookup[i] = col;
     }
-    
+
     // get
-    
+
     long tgs = System.currentTimeMillis();
     if (nm != null) {
       for (int i = 0; i < numLookups; i++) {
@@ -146,51 +146,51 @@ public class NativeMapPerformanceTest {
       }
     }
     long tge = System.currentTimeMillis();
-    
+
     long memUsed = 0;
     if (nm != null) {
       memUsed = nm.getMemoryUsed();
     }
-    
+
     int size = (nm == null ? tm.size() : nm.size());
-    
+
     // delete
     long tds = System.currentTimeMillis();
-    
+
     if (nm != null)
       nm.delete();
-    
+
     long tde = System.currentTimeMillis();
-    
+
     if (tm != null)
       tm.clear();
-    
+
     System.gc();
     System.gc();
     System.gc();
     System.gc();
-    
+
     UtilWaitThread.sleep(3000);
-    
+
     System.out.printf("mapType:%10s   put rate:%,6.2f  scan rate:%,6.2f  get rate:%,6.2f  delete time : %6.2f  mem : %,d%n", "" + mapType, (numRows * numCols)
         / ((tpe - tps) / 1000.0), (size) / ((tie - tis) / 1000.0), numLookups / ((tge - tgs) / 1000.0), (tde - tds) / 1000.0, memUsed);
-    
+
   }
-  
+
   public static void main(String[] args) {
-    
+
     if (args.length != 3) {
       throw new IllegalArgumentException("Usage : " + NativeMapPerformanceTest.class.getName() + " <map type> <rows> <columns>");
     }
-    
+
     String mapType = args[0];
     int rows = Integer.parseInt(args[1]);
     int cols = Integer.parseInt(args[2]);
-    
+
     runPerformanceTest(rows, cols, 10000, mapType);
     runPerformanceTest(rows, cols, 10000, mapType);
     runPerformanceTest(rows, cols, 10000, mapType);
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/NativeMapStressTest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/NativeMapStressTest.java b/test/src/main/java/org/apache/accumulo/test/NativeMapStressTest.java
index c8e821b..72831d8 100644
--- a/test/src/main/java/org/apache/accumulo/test/NativeMapStressTest.java
+++ b/test/src/main/java/org/apache/accumulo/test/NativeMapStressTest.java
@@ -36,144 +36,144 @@ import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 
 public class NativeMapStressTest {
-  
+
   private static final Logger log = Logger.getLogger(NativeMapStressTest.class);
-  
+
   public static void main(String[] args) {
     testLotsOfMapDeletes(true);
     testLotsOfMapDeletes(false);
     testLotsOfOverwrites();
     testLotsOfGetsAndScans();
   }
-  
+
   private static void put(NativeMap nm, String row, String val, int mc) {
     Mutation m = new Mutation(new Text(row));
     m.put(new Text(), new Text(), Long.MAX_VALUE, new Value(val.getBytes(UTF_8)));
     nm.mutate(m, mc);
   }
-  
+
   private static void testLotsOfGetsAndScans() {
-    
+
     ArrayList<Thread> threads = new ArrayList<Thread>();
-    
+
     final int numThreads = 8;
     final int totalGets = 100000000;
     final int mapSizePerThread = (int) (4000000 / (double) numThreads);
     final int getsPerThread = (int) (totalGets / (double) numThreads);
-    
+
     for (int tCount = 0; tCount < numThreads; tCount++) {
       Runnable r = new Runnable() {
         @Override
         public void run() {
           NativeMap nm = new NativeMap();
-          
+
           Random r = new Random();
-          
+
           OpTimer opTimer = new OpTimer(log, Level.INFO);
-          
+
           opTimer.start("Creating map of size " + mapSizePerThread);
-          
+
           for (int i = 0; i < mapSizePerThread; i++) {
             String row = String.format("r%08d", i);
             String val = row + "v";
             put(nm, row, val, i);
           }
-          
+
           opTimer.stop("Created map of size " + nm.size() + " in %DURATION%");
-          
+
           opTimer.start("Doing " + getsPerThread + " gets()");
-          
+
           for (int i = 0; i < getsPerThread; i++) {
             String row = String.format("r%08d", r.nextInt(mapSizePerThread));
             String val = row + "v";
-            
+
             Value value = nm.get(new Key(new Text(row)));
             if (value == null || !value.toString().equals(val)) {
               log.error("nm.get(" + row + ") failed");
             }
           }
-          
+
           opTimer.stop("Finished " + getsPerThread + " gets in %DURATION%");
-          
+
           int scanned = 0;
-          
+
           opTimer.start("Doing " + getsPerThread + " random iterations");
-          
+
           for (int i = 0; i < getsPerThread; i++) {
             int startRow = r.nextInt(mapSizePerThread);
             String row = String.format("r%08d", startRow);
-            
+
             Iterator<Entry<Key,Value>> iter = nm.iterator(new Key(new Text(row)));
-            
+
             int count = 0;
-            
+
             while (iter.hasNext() && count < 10) {
               String row2 = String.format("r%08d", startRow + count);
               String val2 = row2 + "v";
-              
+
               Entry<Key,Value> entry = iter.next();
               if (!entry.getValue().toString().equals(val2) || !entry.getKey().equals(new Key(new Text(row2)))) {
                 log.error("nm.iter(" + row2 + ") failed row = " + row + " count = " + count + " row2 = " + row + " val2 = " + val2);
               }
-              
+
               count++;
             }
-            
+
             scanned += count;
           }
-          
+
           opTimer.stop("Finished " + getsPerThread + " random iterations (scanned = " + scanned + ") in %DURATION%");
-          
+
           nm.delete();
         }
       };
-      
+
       Thread t = new Thread(r);
       t.start();
-      
+
       threads.add(t);
     }
-    
+
     for (Thread thread : threads) {
       try {
         thread.join();
       } catch (InterruptedException e) {
-        log.error("Could not join thread '"+thread.getName()+"'.", e);
+        log.error("Could not join thread '" + thread.getName() + "'.", e);
         throw new RuntimeException(e);
       }
     }
   }
-  
+
   private static void testLotsOfMapDeletes(final boolean doRemoves) {
     final int numThreads = 8;
     final int rowRange = 10000;
     final int mapsPerThread = 50;
     final int totalInserts = 100000000;
     final int insertsPerMapPerThread = (int) (totalInserts / (double) numThreads / mapsPerThread);
-    
+
     System.out.println("insertsPerMapPerThread " + insertsPerMapPerThread);
-    
+
     ArrayList<Thread> threads = new ArrayList<Thread>();
-    
+
     for (int i = 0; i < numThreads; i++) {
       Runnable r = new Runnable() {
         @Override
         public void run() {
-          
+
           int inserts = 0;
           int removes = 0;
-          
+
           for (int i = 0; i < mapsPerThread; i++) {
-            
+
             NativeMap nm = new NativeMap();
-            
+
             for (int j = 0; j < insertsPerMapPerThread; j++) {
               String row = String.format("r%08d", j % rowRange);
               String val = row + "v";
               put(nm, row, val, j);
               inserts++;
             }
-            
+
             if (doRemoves) {
               Iterator<Entry<Key,Value>> iter = nm.iterator();
               while (iter.hasNext()) {
@@ -182,61 +182,61 @@ public class NativeMapStressTest {
                 removes++;
               }
             }
-            
+
             nm.delete();
           }
-          
+
           System.out.println("inserts " + inserts + " removes " + removes + " " + Thread.currentThread().getName());
         }
       };
-      
+
       Thread t = new Thread(r);
       t.start();
-      
+
       threads.add(t);
     }
-    
+
     for (Thread thread : threads) {
       try {
         thread.join();
       } catch (InterruptedException e) {
-        log.error("Could not join thread '"+thread.getName()+"'.", e);
+        log.error("Could not join thread '" + thread.getName() + "'.", e);
         throw new RuntimeException(e);
       }
     }
   }
-  
+
   private static void testLotsOfOverwrites() {
     final Map<Integer,NativeMap> nativeMaps = new HashMap<Integer,NativeMap>();
-    
+
     int numThreads = 8;
     final int insertsPerThread = (int) (100000000 / (double) numThreads);
     final int rowRange = 10000;
     final int numMaps = 50;
-    
+
     ArrayList<Thread> threads = new ArrayList<Thread>();
-    
+
     for (int i = 0; i < numThreads; i++) {
       Runnable r = new Runnable() {
         @Override
         public void run() {
           Random r = new Random();
           int inserts = 0;
-          
+
           for (int i = 0; i < insertsPerThread / 100.0; i++) {
             int map = r.nextInt(numMaps);
-            
+
             NativeMap nm;
-            
+
             synchronized (nativeMaps) {
               nm = nativeMaps.get(map);
               if (nm == null) {
                 nm = new NativeMap();
                 nativeMaps.put(map, nm);
-                
+
               }
             }
-            
+
             synchronized (nm) {
               for (int j = 0; j < 100; j++) {
                 String row = String.format("r%08d", r.nextInt(rowRange));
@@ -246,30 +246,30 @@ public class NativeMapStressTest {
               }
             }
           }
-          
+
           System.out.println("inserts " + inserts + " " + Thread.currentThread().getName());
         }
       };
-      
+
       Thread t = new Thread(r);
       t.start();
-      
+
       threads.add(t);
     }
-    
+
     for (Thread thread : threads) {
       try {
         thread.join();
       } catch (InterruptedException e) {
-        log.error("Could not join thread '"+thread.getName()+"'.", e);
+        log.error("Could not join thread '" + thread.getName() + "'.", e);
         throw new RuntimeException(e);
       }
     }
-    
+
     Set<Entry<Integer,NativeMap>> es = nativeMaps.entrySet();
     for (Entry<Integer,NativeMap> entry : es) {
       entry.getValue().delete();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/NullBatchWriter.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/NullBatchWriter.java b/test/src/main/java/org/apache/accumulo/test/NullBatchWriter.java
index d260c95..3bb2f7f 100644
--- a/test/src/main/java/org/apache/accumulo/test/NullBatchWriter.java
+++ b/test/src/main/java/org/apache/accumulo/test/NullBatchWriter.java
@@ -21,10 +21,10 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.data.Mutation;
 
 public class NullBatchWriter implements BatchWriter {
-  
+
   private int mutationsAdded;
   private long startTime;
-  
+
   @Override
   public void addMutation(Mutation m) throws MutationsRejectedException {
     if (mutationsAdded == 0) {
@@ -33,23 +33,23 @@ public class NullBatchWriter implements BatchWriter {
     mutationsAdded++;
     m.numBytes();
   }
-  
+
   @Override
   public void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException {
     for (Mutation mutation : iterable) {
       addMutation(mutation);
     }
   }
-  
+
   @Override
   public void close() throws MutationsRejectedException {
     flush();
   }
-  
+
   @Override
   public void flush() throws MutationsRejectedException {
     System.out.printf("Mutation add rate : %,6.2f mutations/sec%n", mutationsAdded / ((System.currentTimeMillis() - startTime) / 1000.0));
     mutationsAdded = 0;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/QueryMetadataTable.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/QueryMetadataTable.java b/test/src/main/java/org/apache/accumulo/test/QueryMetadataTable.java
index b6a9bda..713f3ae 100644
--- a/test/src/main/java/org/apache/accumulo/test/QueryMetadataTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/QueryMetadataTable.java
@@ -41,43 +41,43 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.server.cli.ClientOpts;
 import org.apache.accumulo.server.client.HdfsZooInstance;
 import org.apache.hadoop.io.Text;
+import org.apache.log4j.Logger;
 
 import com.beust.jcommander.Parameter;
-import org.apache.log4j.Logger;
 
 public class QueryMetadataTable {
   private static final Logger log = Logger.getLogger(QueryMetadataTable.class);
-  
+
   private static String principal;
   private static AuthenticationToken token;
-  
+
   static String location;
-  
+
   static class MDTQuery implements Runnable {
     private Text row;
-    
+
     MDTQuery(Text row) {
       this.row = row;
     }
-    
+
     @Override
     public void run() {
       try {
         KeyExtent extent = new KeyExtent(row, (Text) null);
-        
+
         Connector connector = HdfsZooInstance.getInstance().getConnector(principal, token);
         Scanner mdScanner = connector.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
         Text row = extent.getMetadataEntry();
-        
+
         mdScanner.setRange(new Range(row));
-        
+
         for (Entry<Key,Value> entry : mdScanner) {
           if (!entry.getKey().getRow().equals(row))
             break;
         }
-        
+
       } catch (TableNotFoundException e) {
-        log.error("Table '"+MetadataTable.NAME+"' not found.", e);
+        log.error("Table '" + MetadataTable.NAME + "' not found.", e);
         throw new RuntimeException(e);
       } catch (AccumuloException e) {
         log.error("AccumuloException encountered.", e);
@@ -88,28 +88,28 @@ public class QueryMetadataTable {
       }
     }
   }
-  
+
   static class Opts extends ClientOpts {
     @Parameter(names = "--numQueries", description = "number of queries to run")
     int numQueries = 1;
     @Parameter(names = "--numThreads", description = "number of threads used to run the queries")
     int numThreads = 1;
   }
-  
+
   public static void main(String[] args) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(QueryMetadataTable.class.getName(), args, scanOpts);
-    
+
     Connector connector = opts.getConnector();
     Scanner scanner = connector.createScanner(MetadataTable.NAME, opts.auths);
     scanner.setBatchSize(scanOpts.scanBatchSize);
     Text mdrow = new Text(KeyExtent.getMetadataEntry(new Text(MetadataTable.ID), null));
-    
+
     HashSet<Text> rowSet = new HashSet<Text>();
-    
+
     int count = 0;
-    
+
     for (Entry<Key,Value> entry : scanner) {
       System.out.print(".");
       if (count % 72 == 0) {
@@ -119,37 +119,37 @@ public class QueryMetadataTable {
         System.out.println(entry.getKey() + " " + entry.getValue());
         location = entry.getValue().toString();
       }
-      
+
       if (!entry.getKey().getRow().toString().startsWith(MetadataTable.ID))
         rowSet.add(entry.getKey().getRow());
       count++;
     }
-    
+
     System.out.printf(" %,d%n", count);
-    
+
     ArrayList<Text> rows = new ArrayList<Text>(rowSet);
-    
+
     Random r = new Random();
-    
+
     ExecutorService tp = Executors.newFixedThreadPool(opts.numThreads);
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     for (int i = 0; i < opts.numQueries; i++) {
       int index = r.nextInt(rows.size());
       MDTQuery mdtq = new MDTQuery(rows.get(index));
       tp.submit(mdtq);
     }
-    
+
     tp.shutdown();
-    
+
     try {
       tp.awaitTermination(1, TimeUnit.HOURS);
     } catch (InterruptedException e) {
       log.error("Failed while awaiting the ExcecutorService to terminate.", e);
       throw new RuntimeException(e);
     }
-    
+
     long t2 = System.currentTimeMillis();
     double delta = (t2 - t1) / 1000.0;
     System.out.println("time : " + delta + "  queries per sec : " + (opts.numQueries / delta));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/TestBinaryRows.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/TestBinaryRows.java b/test/src/main/java/org/apache/accumulo/test/TestBinaryRows.java
index 7b373c4..2615ae2 100644
--- a/test/src/main/java/org/apache/accumulo/test/TestBinaryRows.java
+++ b/test/src/main/java/org/apache/accumulo/test/TestBinaryRows.java
@@ -40,7 +40,7 @@ import com.beust.jcommander.Parameter;
 
 public class TestBinaryRows {
   private static final long byteOnes;
-  
+
   static {
     // safely build Byte.SIZE number of 1s as a long; not that I think Byte.SIZE will ever be anything but 8, but just for fun
     long b = 1;
@@ -48,51 +48,51 @@ public class TestBinaryRows {
       b |= (1L << i);
     byteOnes = b;
   }
-  
+
   static byte[] encodeLong(long l) {
     byte[] ba = new byte[Long.SIZE / Byte.SIZE];
-    
+
     // parse long into a sequence of bytes
     for (int i = 0; i < ba.length; ++i)
       ba[i] = (byte) (byteOnes & (l >>> (Byte.SIZE * (ba.length - i - 1))));
-    
+
     return ba;
   }
-  
+
   static long decodeLong(byte ba[]) {
     // validate byte array
     if (ba.length > Long.SIZE / Byte.SIZE)
       throw new IllegalArgumentException("Byte array of size " + ba.length + " is too big to hold a long");
-    
+
     // build the long from the bytes
     long l = 0;
     for (int i = 0; i < ba.length; ++i)
       l |= (byteOnes & ba[i]) << (Byte.SIZE * (ba.length - i - 1));
-    
+
     return l;
   }
-  
+
   public static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--mode", description="either 'ingest', 'delete', 'randomLookups', 'split', 'verify', 'verifyDeleted'", required=true)
+    @Parameter(names = "--mode", description = "either 'ingest', 'delete', 'randomLookups', 'split', 'verify', 'verifyDeleted'", required = true)
     public String mode;
-    @Parameter(names="--start", description="the lowest numbered row")
+    @Parameter(names = "--start", description = "the lowest numbered row")
     public long start = 0;
-    @Parameter(names="--count", description="number of rows to ingest", required=true)
+    @Parameter(names = "--count", description = "number of rows to ingest", required = true)
     public long num = 0;
   }
-  
+
   public static void runTest(Connector connector, Opts opts, BatchWriterOpts bwOpts, ScannerOpts scanOpts) throws Exception {
-    
+
     final Text CF = new Text("cf"), CQ = new Text("cq");
     final byte[] CF_BYTES = "cf".getBytes(UTF_8), CQ_BYTES = "cq".getBytes(UTF_8);
     if (opts.mode.equals("ingest") || opts.mode.equals("delete")) {
       BatchWriter bw = connector.createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig());
       boolean delete = opts.mode.equals("delete");
-      
+
       for (long i = 0; i < opts.num; i++) {
         byte[] row = encodeLong(i + opts.start);
         String value = "" + (i + opts.start);
-        
+
         Mutation m = new Mutation(new Text(row));
         if (delete) {
           m.putDelete(CF, CQ);
@@ -101,7 +101,7 @@ public class TestBinaryRows {
         }
         bw.addMutation(m);
       }
-      
+
       bw.close();
     } else if (opts.mode.equals("verifyDeleted")) {
       Scanner s = connector.createScanner(opts.getTableName(), opts.auths);
@@ -110,22 +110,22 @@ public class TestBinaryRows {
       Key stopKey = new Key(encodeLong(opts.start + opts.num - 1), CF_BYTES, CQ_BYTES, new byte[0], 0);
       s.setBatchSize(50000);
       s.setRange(new Range(startKey, stopKey));
-      
+
       for (Entry<Key,Value> entry : s) {
         throw new Exception("ERROR : saw entries in range that should be deleted ( first value : " + entry.getValue().toString() + ")");
       }
-      
+
     } else if (opts.mode.equals("verify")) {
       long t1 = System.currentTimeMillis();
-      
+
       Scanner s = connector.createScanner(opts.getTableName(), opts.auths);
       Key startKey = new Key(encodeLong(opts.start), CF_BYTES, CQ_BYTES, new byte[0], Long.MAX_VALUE);
       Key stopKey = new Key(encodeLong(opts.start + opts.num - 1), CF_BYTES, CQ_BYTES, new byte[0], 0);
       s.setBatchSize(scanOpts.scanBatchSize);
       s.setRange(new Range(startKey, stopKey));
-      
+
       long i = opts.start;
-      
+
       for (Entry<Key,Value> e : s) {
         Key k = e.getKey();
         Value v = e.getValue();
@@ -134,81 +134,81 @@ public class TestBinaryRows {
 
         i++;
       }
-      
+
       if (i != opts.start + opts.num) {
         throw new Exception("ERROR : did not see expected number of rows, saw " + (i - opts.start) + " expected " + opts.num);
       }
-      
+
       long t2 = System.currentTimeMillis();
-      
+
       System.out.printf("time : %9.2f secs%n", ((t2 - t1) / 1000.0));
       System.out.printf("rate : %9.2f entries/sec%n", opts.num / ((t2 - t1) / 1000.0));
-      
+
     } else if (opts.mode.equals("randomLookups")) {
       int numLookups = 1000;
-      
+
       Random r = new Random();
-      
+
       long t1 = System.currentTimeMillis();
-      
+
       for (int i = 0; i < numLookups; i++) {
         long row = ((r.nextLong() & 0x7fffffffffffffffl) % opts.num) + opts.start;
-        
+
         Scanner s = connector.createScanner(opts.getTableName(), opts.auths);
         s.setBatchSize(scanOpts.scanBatchSize);
         Key startKey = new Key(encodeLong(row), CF_BYTES, CQ_BYTES, new byte[0], Long.MAX_VALUE);
         Key stopKey = new Key(encodeLong(row), CF_BYTES, CQ_BYTES, new byte[0], 0);
         s.setRange(new Range(startKey, stopKey));
-        
+
         Iterator<Entry<Key,Value>> si = s.iterator();
-        
+
         if (si.hasNext()) {
           Entry<Key,Value> e = si.next();
           Key k = e.getKey();
           Value v = e.getValue();
-          
+
           checkKeyValue(row, k, v);
-          
+
           if (si.hasNext()) {
             throw new Exception("ERROR : lookup on " + row + " returned more than one result ");
           }
-          
+
         } else {
           throw new Exception("ERROR : lookup on " + row + " failed ");
         }
       }
-      
+
       long t2 = System.currentTimeMillis();
-      
+
       System.out.printf("time    : %9.2f secs%n", ((t2 - t1) / 1000.0));
       System.out.printf("lookups : %9d keys%n", numLookups);
       System.out.printf("rate    : %9.2f lookups/sec%n", numLookups / ((t2 - t1) / 1000.0));
-      
+
     } else if (opts.mode.equals("split")) {
       TreeSet<Text> splits = new TreeSet<Text>();
       int shift = (int) opts.start;
       int count = (int) opts.num;
-      
+
       for (long i = 0; i < count; i++) {
         long splitPoint = i << shift;
-        
+
         splits.add(new Text(encodeLong(splitPoint)));
         System.out.printf("added split point 0x%016x  %,12d%n", splitPoint, splitPoint);
       }
-      
+
       connector.tableOperations().create(opts.getTableName());
       connector.tableOperations().addSplits(opts.getTableName(), splits);
-      
+
     } else {
       throw new Exception("ERROR : " + opts.mode + " is not a valid operation.");
     }
   }
-  
+
   private static void checkKeyValue(long expected, Key k, Value v) throws Exception {
     if (expected != decodeLong(TextUtil.getBytes(k.getRow()))) {
       throw new Exception("ERROR : expected row " + expected + " saw " + decodeLong(TextUtil.getBytes(k.getRow())));
     }
-    
+
     if (!v.toString().equals("" + expected)) {
       throw new Exception("ERROR : expected value " + expected + " saw " + v.toString());
     }
@@ -219,7 +219,7 @@ public class TestBinaryRows {
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(TestBinaryRows.class.getName(), args, scanOpts, bwOpts);
-    
+
     try {
       runTest(opts.getConnector(), opts, bwOpts, scanOpts);
     } catch (Exception e) {


[13/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/InMemoryMap.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/InMemoryMap.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/InMemoryMap.java
index 7378348..47936b6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/InMemoryMap.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/InMemoryMap.java
@@ -71,13 +71,13 @@ import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
 class MemKeyComparator implements Comparator<Key>, Serializable {
-  
+
   private static final long serialVersionUID = 1L;
 
   @Override
   public int compare(Key k1, Key k2) {
     int cmp = k1.compareTo(k2);
-    
+
     if (cmp == 0) {
       if (k1 instanceof MemKey)
         if (k2 instanceof MemKey)
@@ -87,36 +87,36 @@ class MemKeyComparator implements Comparator<Key>, Serializable {
       else if (k2 instanceof MemKey)
         cmp = -1;
     }
-    
+
     return cmp;
   }
 }
 
 class PartialMutationSkippingIterator extends SkippingIterator implements InterruptibleIterator {
-  
+
   int kvCount;
-  
+
   public PartialMutationSkippingIterator(SortedKeyValueIterator<Key,Value> source, int maxKVCount) {
     setSource(source);
     this.kvCount = maxKVCount;
   }
-  
+
   @Override
   protected void consume() throws IOException {
     while (getSource().hasTop() && ((MemKey) getSource().getTopKey()).kvCount > kvCount)
       getSource().next();
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new PartialMutationSkippingIterator(getSource().deepCopy(env), kvCount);
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     ((InterruptibleIterator) getSource()).setInterruptFlag(flag);
   }
-  
+
 }
 
 class MemKeyConversionIterator extends WrappingIterator implements InterruptibleIterator {
@@ -132,17 +132,17 @@ class MemKeyConversionIterator extends WrappingIterator implements Interruptible
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new MemKeyConversionIterator(getSource().deepCopy(env));
   }
-  
+
   @Override
   public Key getTopKey() {
     return currKey;
   }
-  
+
   @Override
   public Value getTopValue() {
     return currVal;
   }
-  
+
   private void getTopKeyVal() {
     Key k = super.getTopKey();
     Value v = super.getTopValue();
@@ -156,7 +156,7 @@ class MemKeyConversionIterator extends WrappingIterator implements Interruptible
     currKey = new MemKey(k, mc);
 
   }
-  
+
   public void next() throws IOException {
     super.next();
     if (hasTop())
@@ -165,7 +165,7 @@ class MemKeyConversionIterator extends WrappingIterator implements Interruptible
 
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     super.seek(range, columnFamilies, inclusive);
-    
+
     if (hasTop())
       getTopKeyVal();
 
@@ -185,14 +185,14 @@ class MemKeyConversionIterator extends WrappingIterator implements Interruptible
 
 public class InMemoryMap {
   private SimpleMap map = null;
-  
+
   private static final Logger log = Logger.getLogger(InMemoryMap.class);
-  
+
   private volatile String memDumpFile = null;
   private final String memDumpDir;
 
   private Map<String,Set<ByteSequence>> lggroups;
-  
+
   public InMemoryMap(boolean useNativeMap, String memDumpDir) {
     this(new HashMap<String,Set<ByteSequence>>(), useNativeMap, memDumpDir);
   }
@@ -200,17 +200,17 @@ public class InMemoryMap {
   public InMemoryMap(Map<String,Set<ByteSequence>> lggroups, boolean useNativeMap, String memDumpDir) {
     this.memDumpDir = memDumpDir;
     this.lggroups = lggroups;
-    
+
     if (lggroups.size() == 0)
       map = newMap(useNativeMap);
     else
       map = new LocalityGroupMap(lggroups, useNativeMap);
   }
-  
+
   public InMemoryMap(AccumuloConfiguration config) throws LocalityGroupConfigurationError {
     this(LocalityGroupUtil.getLocalityGroups(config), config.getBoolean(Property.TSERV_NATIVEMAP_ENABLED), config.get(Property.TSERV_MEMDUMP_DIR));
   }
-  
+
   private static SimpleMap newMap(boolean useNativeMap) {
     if (useNativeMap && NativeMap.isLoaded()) {
       try {
@@ -219,43 +219,43 @@ public class InMemoryMap {
         log.error("Failed to create native map", t);
       }
     }
-    
+
     return new DefaultMap();
   }
-  
+
   private interface SimpleMap {
     Value get(Key key);
-    
+
     Iterator<Entry<Key,Value>> iterator(Key startKey);
-    
+
     int size();
-    
+
     InterruptibleIterator skvIterator();
-    
+
     void delete();
-    
+
     long getMemoryUsed();
-    
+
     void mutate(List<Mutation> mutations, int kvCount);
   }
-  
+
   private static class LocalityGroupMap implements SimpleMap {
-    
+
     private Map<ByteSequence,MutableLong> groupFams[];
-    
+
     // the last map in the array is the default locality group
     private SimpleMap maps[];
     private Partitioner partitioner;
     private List<Mutation>[] partitioned;
     private Set<ByteSequence> nonDefaultColumnFamilies;
-    
+
     @SuppressWarnings("unchecked")
     LocalityGroupMap(Map<String,Set<ByteSequence>> groups, boolean useNativeMap) {
       this.groupFams = new Map[groups.size()];
       this.maps = new SimpleMap[groups.size() + 1];
       this.partitioned = new List[groups.size() + 1];
       this.nonDefaultColumnFamilies = new HashSet<ByteSequence>();
-      
+
       for (int i = 0; i < maps.length; i++) {
         maps[i] = newMap(useNativeMap);
       }
@@ -268,9 +268,9 @@ public class InMemoryMap {
         this.groupFams[count++] = map;
         nonDefaultColumnFamilies.addAll(cfset);
       }
-      
+
       partitioner = new LocalityGroupUtil.Partitioner(this.groupFams);
-      
+
       for (int i = 0; i < partitioned.length; i++) {
         partitioned[i] = new ArrayList<Mutation>();
       }
@@ -280,12 +280,12 @@ public class InMemoryMap {
     public Value get(Key key) {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public Iterator<Entry<Key,Value>> iterator(Key startKey) {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public int size() {
       int sum = 0;
@@ -293,7 +293,7 @@ public class InMemoryMap {
         sum += map.size();
       return sum;
     }
-    
+
     @Override
     public InterruptibleIterator skvIterator() {
       LocalityGroup groups[] = new LocalityGroup[maps.length];
@@ -304,16 +304,15 @@ public class InMemoryMap {
           groups[i] = new LocalityGroup(maps[i].skvIterator(), null, true);
       }
 
-
       return new LocalityGroupIterator(groups, nonDefaultColumnFamilies);
     }
-    
+
     @Override
     public void delete() {
       for (SimpleMap map : maps)
         map.delete();
     }
-    
+
     @Override
     public long getMemoryUsed() {
       long sum = 0;
@@ -321,16 +320,16 @@ public class InMemoryMap {
         sum += map.getMemoryUsed();
       return sum;
     }
-    
+
     @Override
     public synchronized void mutate(List<Mutation> mutations, int kvCount) {
       // this method is synchronized because it reuses objects to avoid allocation,
       // currently, the method that calls this is synchronized so there is no
       // loss in parallelism.... synchronization was added here for future proofing
-      
-      try{
+
+      try {
         partitioner.partition(mutations, partitioned);
-        
+
         for (int i = 0; i < partitioned.length; i++) {
           if (partitioned[i].size() > 0) {
             maps[i].mutate(partitioned[i], kvCount);
@@ -345,14 +344,14 @@ public class InMemoryMap {
         }
       }
     }
-    
+
   }
 
   private static class DefaultMap implements SimpleMap {
     private ConcurrentSkipListMap<Key,Value> map = new ConcurrentSkipListMap<Key,Value>(new MemKeyComparator());
     private AtomicLong bytesInMemory = new AtomicLong();
     private AtomicInteger size = new AtomicInteger();
-    
+
     public void put(Key key, Value value) {
       // Always a MemKey, so account for the kvCount int
       bytesInMemory.addAndGet(key.getLength() + 4);
@@ -360,42 +359,42 @@ public class InMemoryMap {
       if (map.put(key, value) == null)
         size.incrementAndGet();
     }
-    
+
     public Value get(Key key) {
       return map.get(key);
     }
-    
+
     public Iterator<Entry<Key,Value>> iterator(Key startKey) {
       Key lk = new Key(startKey);
       SortedMap<Key,Value> tm = map.tailMap(lk);
       return tm.entrySet().iterator();
     }
-    
+
     public int size() {
       return size.get();
     }
-    
+
     public synchronized InterruptibleIterator skvIterator() {
       if (map == null)
         throw new IllegalStateException();
-      
+
       return new SortedMapIterator(map);
     }
-    
+
     public synchronized void delete() {
       map = null;
     }
-    
+
     public long getOverheadPerEntry() {
       // all of the java objects that are used to hold the
       // data and make it searchable have overhead... this
       // overhead is estimated using test.EstimateInMemMapOverhead
       // and is in bytes.. the estimates were obtained by running
       // java 6_16 in 64 bit server mode
-      
+
       return 200;
     }
-    
+
     @Override
     public void mutate(List<Mutation> mutations, int kvCount) {
       for (Mutation m : mutations) {
@@ -407,64 +406,64 @@ public class InMemoryMap {
         }
       }
     }
-    
+
     @Override
     public long getMemoryUsed() {
       return bytesInMemory.get() + (size() * getOverheadPerEntry());
     }
   }
-  
+
   private static class NativeMapWrapper implements SimpleMap {
     private NativeMap nativeMap;
-    
+
     NativeMapWrapper() {
       nativeMap = new NativeMap();
     }
-    
+
     public Value get(Key key) {
       return nativeMap.get(key);
     }
-    
+
     public Iterator<Entry<Key,Value>> iterator(Key startKey) {
       return nativeMap.iterator(startKey);
     }
-    
+
     public int size() {
       return nativeMap.size();
     }
-    
+
     public InterruptibleIterator skvIterator() {
       return (InterruptibleIterator) nativeMap.skvIterator();
     }
-    
+
     public void delete() {
       nativeMap.delete();
     }
-    
+
     public long getMemoryUsed() {
       return nativeMap.getMemoryUsed();
     }
-    
+
     @Override
     public void mutate(List<Mutation> mutations, int kvCount) {
       nativeMap.mutate(mutations, kvCount);
     }
   }
-  
+
   private AtomicInteger nextKVCount = new AtomicInteger(1);
   private AtomicInteger kvCount = new AtomicInteger(0);
 
   private Object writeSerializer = new Object();
-  
+
   /**
    * Applies changes to a row in the InMemoryMap
-   * 
+   *
    */
   public void mutate(List<Mutation> mutations) {
     int numKVs = 0;
     for (int i = 0; i < mutations.size(); i++)
       numKVs += mutations.get(i).size();
-    
+
     // Can not update mutationCount while writes that started before
     // are in progress, this would cause partial mutations to be seen.
     // Also, can not continue until mutation count is updated, because
@@ -472,7 +471,7 @@ public class InMemoryMap {
     // wait for writes that started before to finish.
     //
     // using separate lock from this map, to allow read/write in parallel
-    synchronized (writeSerializer ) {
+    synchronized (writeSerializer) {
       int kv = nextKVCount.getAndAdd(numKVs);
       try {
         map.mutate(mutations, kv);
@@ -481,51 +480,51 @@ public class InMemoryMap {
       }
     }
   }
-  
+
   /**
    * Returns a long representing the size of the InMemoryMap
-   * 
+   *
    * @return bytesInMemory
    */
   public synchronized long estimatedSizeInBytes() {
     if (map == null)
       return 0;
-    
+
     return map.getMemoryUsed();
   }
-  
+
   Iterator<Map.Entry<Key,Value>> iterator(Key startKey) {
     return map.iterator(startKey);
   }
-  
+
   public synchronized long getNumEntries() {
     if (map == null)
       return 0;
     return map.size();
   }
-  
+
   private final Set<MemoryIterator> activeIters = Collections.synchronizedSet(new HashSet<MemoryIterator>());
-  
+
   class MemoryDataSource implements DataSource {
-    
+
     boolean switched = false;
     private InterruptibleIterator iter;
     private FileSKVIterator reader;
     private MemoryDataSource parent;
     private IteratorEnvironment env;
     private AtomicBoolean iflag;
-    
+
     MemoryDataSource() {
       this(null, false, null, null);
     }
-    
+
     public MemoryDataSource(MemoryDataSource parent, boolean switched, IteratorEnvironment env, AtomicBoolean iflag) {
       this.parent = parent;
       this.switched = switched;
       this.env = env;
       this.iflag = iflag;
     }
-    
+
     @Override
     public boolean isCurrent() {
       if (switched)
@@ -533,12 +532,12 @@ public class InMemoryMap {
       else
         return memDumpFile == null;
     }
-    
+
     @Override
     public DataSource getNewDataSource() {
       if (switched)
         throw new IllegalStateException();
-      
+
       if (!isCurrent()) {
         switched = true;
         iter = null;
@@ -549,15 +548,15 @@ public class InMemoryMap {
           throw new RuntimeException();
         }
       }
-      
+
       return this;
     }
-    
+
     private synchronized FileSKVIterator getReader() throws IOException {
       if (reader == null) {
         Configuration conf = CachedConfiguration.getInstance();
         FileSystem fs = FileSystem.getLocal(conf);
-        
+
         reader = new RFileOperations().openReader(memDumpFile, true, fs, conf, SiteConfiguration.getInstance());
         if (iflag != null)
           reader.setInterruptFlag(iflag);
@@ -583,10 +582,10 @@ public class InMemoryMap {
               iter = new MemKeyConversionIterator(parent.getReader().deepCopy(env));
             }
         }
-      
+
       return iter;
     }
-    
+
     @Override
     public DataSource getDeepCopyDataSource(IteratorEnvironment env) {
       return new MemoryDataSource(parent == null ? this : parent, switched, env, iflag);
@@ -596,36 +595,36 @@ public class InMemoryMap {
     public void setInterruptFlag(AtomicBoolean flag) {
       this.iflag = flag;
     }
-    
+
   }
-  
+
   public class MemoryIterator extends WrappingIterator implements InterruptibleIterator {
-    
+
     private AtomicBoolean closed;
     private SourceSwitchingIterator ssi;
     private MemoryDataSource mds;
-    
+
     protected SortedKeyValueIterator<Key,Value> getSource() {
       if (closed.get())
         throw new IllegalStateException("Memory iterator is closed");
       return super.getSource();
     }
-    
+
     private MemoryIterator(InterruptibleIterator source) {
       this(source, new AtomicBoolean(false));
     }
-    
+
     private MemoryIterator(SortedKeyValueIterator<Key,Value> source, AtomicBoolean closed) {
       setSource(source);
       this.closed = closed;
     }
-    
+
     public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
       return new MemoryIterator(getSource().deepCopy(env), closed);
     }
-    
+
     public void close() {
-      
+
       synchronized (this) {
         if (closed.compareAndSet(false, true)) {
           try {
@@ -636,41 +635,41 @@ public class InMemoryMap {
           }
         }
       }
-      
+
       // remove outside of sync to avoid deadlock
       activeIters.remove(this);
     }
-    
+
     private synchronized boolean switchNow() throws IOException {
       if (closed.get())
         return false;
-      
+
       ssi.switchNow();
       return true;
     }
-    
+
     @Override
     public void setInterruptFlag(AtomicBoolean flag) {
       ((InterruptibleIterator) getSource()).setInterruptFlag(flag);
     }
-    
+
     private void setSSI(SourceSwitchingIterator ssi) {
       this.ssi = ssi;
     }
-    
+
     public void setMDS(MemoryDataSource mds) {
       this.mds = mds;
     }
-    
+
   }
-  
+
   public synchronized MemoryIterator skvIterator() {
     if (map == null)
       throw new NullPointerException();
-    
+
     if (deleted)
       throw new IllegalStateException("Can not obtain iterator after map deleted");
-    
+
     int mc = kvCount.get();
     MemoryDataSource mds = new MemoryDataSource();
     SourceSwitchingIterator ssi = new SourceSwitchingIterator(new MemoryDataSource());
@@ -680,94 +679,93 @@ public class InMemoryMap {
     activeIters.add(mi);
     return mi;
   }
-  
+
   public SortedKeyValueIterator<Key,Value> compactionIterator() {
-    
+
     if (nextKVCount.get() - 1 != kvCount.get())
-      throw new IllegalStateException("Memory map in unexpected state : nextKVCount = " + nextKVCount.get() + " kvCount = "
-          + kvCount.get());
-    
+      throw new IllegalStateException("Memory map in unexpected state : nextKVCount = " + nextKVCount.get() + " kvCount = " + kvCount.get());
+
     return map.skvIterator();
   }
-  
+
   private boolean deleted = false;
-  
+
   public void delete(long waitTime) {
-    
+
     synchronized (this) {
       if (deleted)
         throw new IllegalStateException("Double delete");
-      
+
       deleted = true;
     }
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     while (activeIters.size() > 0 && System.currentTimeMillis() - t1 < waitTime) {
       UtilWaitThread.sleep(50);
     }
-    
+
     if (activeIters.size() > 0) {
       // dump memmap exactly as is to a tmp file on disk, and switch scans to that temp file
       try {
         Configuration conf = CachedConfiguration.getInstance();
         FileSystem fs = FileSystem.getLocal(conf);
-        
+
         String tmpFile = memDumpDir + "/memDump" + UUID.randomUUID() + "." + RFile.EXTENSION;
-        
+
         Configuration newConf = new Configuration(conf);
         newConf.setInt("io.seqfile.compress.blocksize", 100000);
-        
+
         FileSKVWriter out = new RFileOperations().openWriter(tmpFile, fs, newConf, SiteConfiguration.getInstance());
-        
+
         InterruptibleIterator iter = map.skvIterator();
-       
-        HashSet<ByteSequence> allfams= new HashSet<ByteSequence>();
-        
-        for(Entry<String, Set<ByteSequence>> entry : lggroups.entrySet()){
+
+        HashSet<ByteSequence> allfams = new HashSet<ByteSequence>();
+
+        for (Entry<String,Set<ByteSequence>> entry : lggroups.entrySet()) {
           allfams.addAll(entry.getValue());
           out.startNewLocalityGroup(entry.getKey(), entry.getValue());
           iter.seek(new Range(), entry.getValue(), true);
           dumpLocalityGroup(out, iter);
         }
-        
+
         out.startDefaultLocalityGroup();
         iter.seek(new Range(), allfams, false);
-       
+
         dumpLocalityGroup(out, iter);
-        
+
         out.close();
-        
+
         log.debug("Created mem dump file " + tmpFile);
-        
+
         memDumpFile = tmpFile;
-        
+
         synchronized (activeIters) {
           for (MemoryIterator mi : activeIters) {
             mi.switchNow();
           }
         }
-        
+
         // rely on unix behavior that file will be deleted when last
         // reader closes it
         fs.delete(new Path(memDumpFile), true);
-        
+
       } catch (IOException ioe) {
         log.error("Failed to create mem dump file ", ioe);
-        
+
         while (activeIters.size() > 0) {
           UtilWaitThread.sleep(100);
         }
       }
-      
+
     }
-    
+
     SimpleMap tmpMap = map;
-    
+
     synchronized (this) {
       map = null;
     }
-    
+
     tmpMap.delete();
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/MemKey.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/MemKey.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/MemKey.java
index 4bc8891..443ffb2 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/MemKey.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/MemKey.java
@@ -23,55 +23,55 @@ import java.io.IOException;
 import org.apache.accumulo.core.data.Key;
 
 class MemKey extends Key {
-  
+
   int kvCount;
-  
+
   public MemKey(byte[] row, byte[] cf, byte[] cq, byte[] cv, long ts, boolean del, boolean copy, int mc) {
     super(row, cf, cq, cv, ts, del, copy);
     this.kvCount = mc;
   }
-  
+
   public MemKey() {
     super();
     this.kvCount = Integer.MAX_VALUE;
   }
-  
+
   public MemKey(Key key, int mc) {
     super(key);
     this.kvCount = mc;
   }
-  
+
   public String toString() {
     return super.toString() + " mc=" + kvCount;
   }
-  
+
   @Override
   public Object clone() throws CloneNotSupportedException {
     return super.clone();
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     super.write(out);
     out.writeInt(kvCount);
   }
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     super.readFields(in);
     kvCount = in.readInt();
   }
-  
+
   @Override
   public int compareTo(Key k) {
-    
+
     int cmp = super.compareTo(k);
-    
+
     if (cmp == 0 && k instanceof MemKey) {
       cmp = ((MemKey) k).kvCount - kvCount;
     }
-    
+
     return cmp;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/MemValue.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/MemValue.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/MemValue.java
index f1fdde4..0ce3b9e 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/MemValue.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/MemValue.java
@@ -22,12 +22,12 @@ import java.io.IOException;
 import org.apache.accumulo.core.data.Value;
 
 /**
- * 
+ *
  */
 public class MemValue extends Value {
   int kvCount;
   boolean merged = false;
-  
+
   /**
    * @param value
    *          Value
@@ -38,17 +38,17 @@ public class MemValue extends Value {
     super(value);
     this.kvCount = kv;
   }
-  
+
   public MemValue() {
     super();
     this.kvCount = Integer.MAX_VALUE;
   }
-  
+
   public MemValue(Value value, int kv) {
     super(value);
     this.kvCount = kv;
   }
-  
+
   // Override
   @Override
   public void write(final DataOutput out) throws IOException {
@@ -64,7 +64,7 @@ public class MemValue extends Value {
     }
     super.write(out);
   }
-  
+
   @Override
   public void set(final byte[] b) {
     super.set(b);
@@ -76,16 +76,16 @@ public class MemValue extends Value {
     super.copy(b);
     merged = false;
   }
-  
+
   /**
    * Takes a Value and will take out the embedded kvCount, and then return that value while replacing the Value with the original unembedded version
-   * 
+   *
    * @return The kvCount embedded in v.
    */
   public static int splitKVCount(Value v) {
     if (v instanceof MemValue)
       return ((MemValue) v).kvCount;
-    
+
     byte[] originalBytes = new byte[v.getSize() - 4];
     byte[] combined = v.get();
     System.arraycopy(combined, 4, originalBytes, 0, originalBytes.length);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/MinorCompactionReason.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/MinorCompactionReason.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/MinorCompactionReason.java
index 82c791c..62c6ec6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/MinorCompactionReason.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/MinorCompactionReason.java
@@ -18,4 +18,4 @@ package org.apache.accumulo.tserver;
 
 public enum MinorCompactionReason {
   USER, SYSTEM, CLOSE, RECOVERY
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/Mutations.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/Mutations.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/Mutations.java
index 5ee1952..76061e6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/Mutations.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/Mutations.java
@@ -18,8 +18,8 @@ package org.apache.accumulo.tserver;
 
 import java.util.List;
 
-import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.client.Durability;
+import org.apache.accumulo.core.data.Mutation;
 
 public class Mutations {
   private final Durability durability;
@@ -29,10 +29,12 @@ public class Mutations {
     this.durability = durability;
     this.mutations = mutations;
   }
+
   public Durability getDurability() {
     return durability;
   }
+
   public List<Mutation> getMutations() {
     return mutations;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/NativeMap.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/NativeMap.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/NativeMap.java
index e22a54f..b330d1e 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/NativeMap.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/NativeMap.java
@@ -48,11 +48,11 @@ import org.apache.log4j.Logger;
 
 /**
  * This class stores data in a C++ map. Doing this allows us to store more in memory and avoid pauses caused by Java GC.
- * 
+ *
  * The strategy for dealing with native memory allocated for the native map is that java code using the native map should call delete() as soon as it is
  * finished using the native map. When the NativeMap object is garbage collected its native resources will be released if needed. However waiting for java GC
  * would be a mistake for long lived NativeMaps. Long lived objects are not garbage collected quickly, therefore a process could easily use too much memory.
- * 
+ *
  */
 
 public class NativeMap implements Iterable<Map.Entry<Key,Value>> {
@@ -92,7 +92,7 @@ public class NativeMap implements Iterable<Map.Entry<Key,Value>> {
    * If native libraries are not loaded, the specified search path will be used to attempt to load them. Directories will be searched by using the
    * system-specific library naming conventions. A path directly to a file can also be provided. Loading will continue until the search path is exhausted, or
    * until the native libraries are found and successfully loaded, whichever occurs first.
-   * 
+   *
    * @param searchPath
    *          a list of files and directories to search
    */
@@ -116,7 +116,7 @@ public class NativeMap implements Iterable<Map.Entry<Key,Value>> {
 
   /**
    * Check if native libraries are loaded.
-   * 
+   *
    * @return true if they are loaded; false otherwise
    */
   public static boolean isLoaded() {
@@ -360,10 +360,10 @@ public class NativeMap implements Iterable<Map.Entry<Key,Value>> {
 
     /**
      * The strategy for dealing with native memory allocated for iterators is to simply delete that memory when this Java Object is garbage collected.
-     * 
+     *
      * These iterators are likely short lived object and therefore will be quickly garbage collected. Even if the objects are long lived and therefore more
      * slowly garbage collected they only hold a small amount of native memory.
-     * 
+     *
      */
 
     private long nmiPointer;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/RowLocks.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/RowLocks.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/RowLocks.java
index 1b22f05..babd629 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/RowLocks.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/RowLocks.java
@@ -30,60 +30,60 @@ import org.apache.accumulo.tserver.ConditionalMutationSet.DeferFilter;
 import org.apache.accumulo.tserver.data.ServerConditionalMutation;
 
 /**
- * 
+ *
  */
 class RowLocks {
-  
+
   private Map<ByteSequence,RowLock> rowLocks = new HashMap<ByteSequence,RowLock>();
-  
+
   static class RowLock {
     ReentrantLock rlock;
     int count;
     ByteSequence rowSeq;
-    
+
     RowLock(ReentrantLock rlock, ByteSequence rowSeq) {
       this.rlock = rlock;
       this.count = 0;
       this.rowSeq = rowSeq;
     }
-    
+
     public boolean tryLock() {
       return rlock.tryLock();
     }
-    
+
     public void lock() {
       rlock.lock();
     }
-    
+
     public void unlock() {
       rlock.unlock();
     }
   }
-  
+
   private RowLock getRowLock(ArrayByteSequence rowSeq) {
-      RowLock lock = rowLocks.get(rowSeq);
-      if (lock == null) {
-        lock = new RowLock(new ReentrantLock(), rowSeq);
-        rowLocks.put(rowSeq, lock);
-      }
-      
-      lock.count++;
-      return lock;
+    RowLock lock = rowLocks.get(rowSeq);
+    if (lock == null) {
+      lock = new RowLock(new ReentrantLock(), rowSeq);
+      rowLocks.put(rowSeq, lock);
+    }
+
+    lock.count++;
+    return lock;
   }
-  
+
   private void returnRowLock(RowLock lock) {
-      if (lock.count == 0)
-        throw new IllegalStateException();
-      lock.count--;
-      
-      if (lock.count == 0) {
-        rowLocks.remove(lock.rowSeq);
-      }
+    if (lock.count == 0)
+      throw new IllegalStateException();
+    lock.count--;
+
+    if (lock.count == 0) {
+      rowLocks.remove(lock.rowSeq);
+    }
   }
-  
+
   List<RowLock> acquireRowlocks(Map<KeyExtent,List<ServerConditionalMutation>> updates, Map<KeyExtent,List<ServerConditionalMutation>> deferred) {
     ArrayList<RowLock> locks = new ArrayList<RowLock>();
-    
+
     // assume that mutations are in sorted order to avoid deadlock
     synchronized (rowLocks) {
       for (List<ServerConditionalMutation> scml : updates.values()) {
@@ -92,7 +92,7 @@ class RowLocks {
         }
       }
     }
-    
+
     HashSet<ByteSequence> rowsNotLocked = null;
 
     // acquire as many locks as possible, not blocking on rows that are already locked
@@ -108,9 +108,9 @@ class RowLocks {
       // if there is only one lock, then wait for it
       locks.get(0).lock();
     }
-    
+
     if (rowsNotLocked != null) {
-      
+
       final HashSet<ByteSequence> rnlf = rowsNotLocked;
       // assume will get locks needed, do something expensive otherwise
       ConditionalMutationSet.defer(updates, deferred, new DeferFilter() {
@@ -121,11 +121,11 @@ class RowLocks {
               deferred.add(scm);
             else
               okMutations.add(scm);
-            
+
           }
         }
       });
-      
+
       ArrayList<RowLock> filteredLocks = new ArrayList<RowLock>();
       ArrayList<RowLock> locksToReturn = new ArrayList<RowLock>();
       for (RowLock rowLock : locks) {
@@ -135,7 +135,7 @@ class RowLocks {
           filteredLocks.add(rowLock);
         }
       }
-      
+
       synchronized (rowLocks) {
         for (RowLock rowLock : locksToReturn) {
           returnRowLock(rowLock);
@@ -146,12 +146,12 @@ class RowLocks {
     }
     return locks;
   }
-  
+
   void releaseRowLocks(List<RowLock> locks) {
     for (RowLock rowLock : locks) {
       rowLock.unlock();
     }
-    
+
     synchronized (rowLocks) {
       for (RowLock rowLock : locks) {
         returnRowLock(rowLock);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TConstraintViolationException.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TConstraintViolationException.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TConstraintViolationException.java
index 83fc43e..5f121cc 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TConstraintViolationException.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TConstraintViolationException.java
@@ -51,4 +51,4 @@ public class TConstraintViolationException extends Exception {
   CommitSession getCommitSession() {
     return commitSession;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TLevel.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TLevel.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TLevel.java
index 8bdf08b..5705c9e 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TLevel.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TLevel.java
@@ -19,18 +19,18 @@ package org.apache.accumulo.tserver;
 import org.apache.log4j.Level;
 
 public class TLevel extends Level {
-  
+
   private static final long serialVersionUID = 1L;
   public final static Level TABLET_HIST = new TLevel();
-  
+
   protected TLevel() {
     super(Level.DEBUG_INT + 100, "TABLET_HIST", Level.DEBUG_INT + 100);
   }
-  
+
   static public Level toLevel(int val) {
     if (val == Level.DEBUG_INT + 100)
       return Level.DEBUG;
     return Level.toLevel(val);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletIteratorEnvironment.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletIteratorEnvironment.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletIteratorEnvironment.java
index d1fece5..e7477b9 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletIteratorEnvironment.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletIteratorEnvironment.java
@@ -34,73 +34,73 @@ import org.apache.accumulo.tserver.FileManager.ScanFileManager;
 import org.apache.hadoop.fs.Path;
 
 public class TabletIteratorEnvironment implements IteratorEnvironment {
-  
+
   private final ScanFileManager trm;
   private final IteratorScope scope;
   private final boolean fullMajorCompaction;
   private final AccumuloConfiguration config;
   private final ArrayList<SortedKeyValueIterator<Key,Value>> topLevelIterators = new ArrayList<SortedKeyValueIterator<Key,Value>>();
   private Map<FileRef,DataFileValue> files;
-  
+
   public TabletIteratorEnvironment(IteratorScope scope, AccumuloConfiguration config) {
     if (scope == IteratorScope.majc)
       throw new IllegalArgumentException("must set if compaction is full");
-    
+
     this.scope = scope;
     this.trm = null;
     this.config = config;
     this.fullMajorCompaction = false;
   }
-  
+
   public TabletIteratorEnvironment(IteratorScope scope, AccumuloConfiguration config, ScanFileManager trm, Map<FileRef,DataFileValue> files) {
     if (scope == IteratorScope.majc)
       throw new IllegalArgumentException("must set if compaction is full");
-    
+
     this.scope = scope;
     this.trm = trm;
     this.config = config;
     this.fullMajorCompaction = false;
     this.files = files;
   }
-  
+
   public TabletIteratorEnvironment(IteratorScope scope, boolean fullMajC, AccumuloConfiguration config) {
     if (scope != IteratorScope.majc)
       throw new IllegalArgumentException("Tried to set maj compaction type when scope was " + scope);
-    
+
     this.scope = scope;
     this.trm = null;
     this.config = config;
     this.fullMajorCompaction = fullMajC;
   }
-  
+
   @Override
   public AccumuloConfiguration getConfig() {
     return config;
   }
-  
+
   @Override
   public IteratorScope getIteratorScope() {
     return scope;
   }
-  
+
   @Override
   public boolean isFullMajorCompaction() {
     if (scope != IteratorScope.majc)
       throw new IllegalStateException("Asked about major compaction type when scope is " + scope);
     return fullMajorCompaction;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
     FileRef ref = new FileRef(mapFileName, new Path(mapFileName));
     return trm.openFiles(Collections.singletonMap(ref, files.get(ref)), false).get(0);
   }
-  
+
   @Override
   public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
     topLevelIterators.add(iter);
   }
-  
+
   public SortedKeyValueIterator<Key,Value> getTopLevelIterator(SortedKeyValueIterator<Key,Value> iter) {
     if (topLevelIterators.isEmpty())
       return iter;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletMutations.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletMutations.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletMutations.java
index 9cc07dc..21734f9 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletMutations.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletMutations.java
@@ -18,8 +18,8 @@ package org.apache.accumulo.tserver;
 
 import java.util.List;
 
-import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.client.Durability;
+import org.apache.accumulo.core.data.Mutation;
 
 public class TabletMutations {
   private final int tid;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServer.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServer.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServer.java
index d5c1d2f..2bfa5a0 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServer.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServer.java
@@ -349,7 +349,6 @@ public class TabletServer extends AccumuloServerContext implements Runnable {
 
   private final SessionManager sessionManager;
 
-
   private final WriteTracker writeTracker = new WriteTracker();
 
   private final RowLocks rowLocks = new RowLocks();
@@ -2122,16 +2121,17 @@ public class TabletServer extends AccumuloServerContext implements Runnable {
         locationToOpen = VolumeUtil.switchRootTabletVolume(extent, locationToOpen);
 
         tablet = new Tablet(TabletServer.this, extent, locationToOpen, trm, tabletsKeyValues);
-        /* @formatter:off
-         * If a minor compaction starts after a tablet opens, this indicates a log recovery occurred. This recovered data must be minor compacted.
+        /*
+         * @formatter:off If a minor compaction starts after a tablet opens, this indicates a log recovery occurred. This recovered data must be minor
+         * compacted.
          *
          * There are three reasons to wait for this minor compaction to finish before placing the tablet in online tablets.
          *
-         * 1) The log recovery code does not handle data written to the tablet on multiple tablet servers.
-         * 2) The log recovery code does not block if memory is full. Therefore recovering lots of tablets that use a lot of memory could run out of memory.
-         * 3) The minor compaction finish event did not make it to the logs (the file will be in metadata, preventing replay of compacted data)...
-         * but do not want a majc to wipe the file out from metadata and then have another process failure...
-         * this could cause duplicate data to replay.
+         * 1) The log recovery code does not handle data written to the tablet on multiple tablet servers. 2) The log recovery code does not block if memory is
+         * full. Therefore recovering lots of tablets that use a lot of memory could run out of memory. 3) The minor compaction finish event did not make it to
+         * the logs (the file will be in metadata, preventing replay of compacted data)... but do not want a majc to wipe the file out from metadata and then
+         * have another process failure... this could cause duplicate data to replay.
+         *
          * @formatter:on
          */
         if (tablet.getNumEntriesInMemory() > 0 && !tablet.minorCompactNow(MinorCompactionReason.RECOVERY)) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServerResourceManager.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServerResourceManager.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServerResourceManager.java
index bb9d427..351d526 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServerResourceManager.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletServerResourceManager.java
@@ -420,7 +420,7 @@ public class TabletServerResourceManager {
               if (!tablet.initiateMinorCompaction(MinorCompactionReason.SYSTEM)) {
                 if (tablet.isClosed()) {
                   // attempt to remove it from the current reports if still there
-                  synchronized(tabletReports) {
+                  synchronized (tabletReports) {
                     TabletStateImpl latestReport = tabletReports.remove(keyExtent);
                     if (latestReport != null) {
                       if (latestReport.getTablet() != tablet) {
@@ -644,8 +644,6 @@ public class TabletServerResourceManager {
       }
     }
 
-
-
     // END methods that Tablets call to make decisions about major compaction
 
     // tablets call this method to run minor compactions,

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletStatsKeeper.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletStatsKeeper.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletStatsKeeper.java
index 40906df..1e2cdf4 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletStatsKeeper.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TabletStatsKeeper.java
@@ -21,18 +21,18 @@ import org.apache.accumulo.core.tabletserver.thrift.TabletStats;
 import org.apache.accumulo.server.util.ActionStatsUpdator;
 
 public class TabletStatsKeeper {
-  
+
   // suspect we need more synchronization in this class
   private ActionStats major = new ActionStats();
   private ActionStats minor = new ActionStats();
   private ActionStats split = new ActionStats();
-  
+
   public enum Operation {
     MAJOR, SPLIT, MINOR
   }
-  
+
   private ActionStats[] map = new ActionStats[] {major, split, minor};
-  
+
   public void updateTime(Operation operation, long queued, long start, long count, boolean failed) {
     try {
       ActionStats data = map[operation.ordinal()];
@@ -42,7 +42,7 @@ public class TabletStatsKeeper {
       } else {
         double t = (System.currentTimeMillis() - start) / 1000.0;
         double q = (start - queued) / 1000.0;
-        
+
         data.status--;
         data.count += count;
         data.num++;
@@ -56,9 +56,9 @@ public class TabletStatsKeeper {
     } catch (Exception E) {
       resetTimes();
     }
-    
+
   }
-  
+
   public void updateTime(Operation operation, long start, long count, boolean failed) {
     try {
       ActionStats data = map[operation.ordinal()];
@@ -67,52 +67,52 @@ public class TabletStatsKeeper {
         data.status--;
       } else {
         double t = (System.currentTimeMillis() - start) / 1000.0;
-        
+
         data.status--;
         data.num++;
         data.elapsed += t;
         data.sumDev += t * t;
-        
+
         if (data.elapsed < 0 || data.sumDev < 0 || data.queueSumDev < 0 || data.queueTime < 0)
           resetTimes();
       }
     } catch (Exception E) {
       resetTimes();
     }
-    
+
   }
-  
+
   public void saveMajorMinorTimes(TabletStats t) {
     ActionStatsUpdator.update(minor, t.minors);
     ActionStatsUpdator.update(major, t.majors);
   }
-  
+
   public void saveMinorTimes(TabletStatsKeeper t) {
     ActionStatsUpdator.update(minor, t.minor);
   }
-  
+
   public void saveMajorTimes(TabletStatsKeeper t) {
     ActionStatsUpdator.update(major, t.major);
   }
-  
+
   public void resetTimes() {
     major = new ActionStats();
     split = new ActionStats();
     minor = new ActionStats();
   }
-  
+
   public void incrementStatusMinor() {
     minor.status++;
   }
-  
+
   public void incrementStatusMajor() {
     major.status++;
   }
-  
+
   public void incrementStatusSplit() {
     split.status++;
   }
-  
+
   public TabletStats getTabletStats() {
     return new TabletStats(null, major, minor, split, 0, 0, 0, 0);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TooManyFilesException.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TooManyFilesException.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TooManyFilesException.java
index 98c7a02..026f7e2 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TooManyFilesException.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TooManyFilesException.java
@@ -19,9 +19,9 @@ package org.apache.accumulo.tserver;
 import java.io.IOException;
 
 public class TooManyFilesException extends IOException {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   public TooManyFilesException(String msg) {
     super(msg);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/TservConstraintEnv.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/TservConstraintEnv.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/TservConstraintEnv.java
index dbb67a9..edaca31 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/TservConstraintEnv.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/TservConstraintEnv.java
@@ -80,4 +80,4 @@ public class TservConstraintEnv implements Environment {
       }
     };
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/WriteTracker.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/WriteTracker.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/WriteTracker.java
index 84b5cd0..9bf80b4 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/WriteTracker.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/WriteTracker.java
@@ -30,13 +30,13 @@ import org.apache.accumulo.tserver.tablet.Tablet;
 import org.apache.log4j.Logger;
 
 /**
- * This little class keeps track of writes in progress and allows readers to wait for writes that started before the read. It assumes that the operation ids
- * are monotonically increasing.
+ * This little class keeps track of writes in progress and allows readers to wait for writes that started before the read. It assumes that the operation ids are
+ * monotonically increasing.
  *
  */
 class WriteTracker {
   private static final Logger log = Logger.getLogger(WriteTracker.class);
-  
+
   private static final AtomicLong operationCounter = new AtomicLong(1);
   private final Map<TabletType,TreeSet<Long>> inProgressWrites = new EnumMap<TabletType,TreeSet<Long>>(TabletType.class);
 
@@ -93,4 +93,4 @@ class WriteTracker {
 
     return startWrite(TabletType.type(extents));
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionPlan.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionPlan.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionPlan.java
index 75c6bd8..8f98761 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionPlan.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionPlan.java
@@ -22,9 +22,10 @@ import java.util.HashSet;
 import java.util.List;
 import java.util.Set;
 
-import com.google.common.collect.Sets;
 import org.apache.accumulo.server.fs.FileRef;
 
+import com.google.common.collect.Sets;
+
 /**
  * A plan for a compaction: the input files, the files that are *not* inputs to a compaction that should simply be deleted, and the optional parameters used to
  * create the resulting output file.
@@ -59,7 +60,7 @@ public class CompactionPlan {
 
   /**
    * Validate compaction plan.
-   * 
+   *
    * @param allFiles
    *          All possible files
    * @throws IllegalStateException

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionStrategy.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionStrategy.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionStrategy.java
index 2d94884..40cb604 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionStrategy.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/CompactionStrategy.java
@@ -42,10 +42,10 @@ public abstract class CompactionStrategy {
    * Determine if this tablet is eligible for a major compaction. It's ok if it later determines (through {@link #gatherInformation(MajorCompactionRequest)} and
    * {@link #getCompactionPlan(MajorCompactionRequest)}) that it does not need to. Any state stored during shouldCompact will no longer exist when
    * {@link #gatherInformation(MajorCompactionRequest)} and {@link #getCompactionPlan(MajorCompactionRequest)} are called.
-   * 
+   *
    * <P>
    * Called while holding the tablet lock, so it should not be doing any blocking.
-   * 
+   *
    * <P>
    * Since no blocking should be done in this method, then its unexpected that this method will throw IOException. However since its in the API, it can not be
    * easily removed.
@@ -55,7 +55,7 @@ public abstract class CompactionStrategy {
   /**
    * Called prior to obtaining the tablet lock, useful for examining metadata or indexes. State collected during this method will be available during the call
    * the {@link #getCompactionPlan(MajorCompactionRequest)}.
-   * 
+   *
    * @param request
    *          basic details about the tablet
    */
@@ -63,11 +63,11 @@ public abstract class CompactionStrategy {
 
   /**
    * Get the plan for compacting a tablets files. Called while holding the tablet lock, so it should not be doing any blocking.
-   * 
+   *
    * <P>
    * Since no blocking should be done in this method, then its unexpected that this method will throw IOException. However since its in the API, it can not be
    * easily removed.
-   * 
+   *
    * @param request
    *          basic details about the tablet
    * @return the plan for a major compaction, or null to cancel the compaction.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategy.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategy.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategy.java
index 6d4dc79..6cc9025 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategy.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategy.java
@@ -26,7 +26,7 @@ import org.apache.accumulo.core.metadata.schema.DataFileValue;
 import org.apache.accumulo.server.fs.FileRef;
 
 /**
- * 
+ *
  */
 public class SizeLimitCompactionStrategy extends DefaultCompactionStrategy {
   public static final String SIZE_LIMIT_OPT = "sizeLimit";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/ConstraintChecker.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/ConstraintChecker.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/ConstraintChecker.java
index 7ea1388..fd9e521 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/ConstraintChecker.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/ConstraintChecker.java
@@ -36,10 +36,10 @@ import org.apache.log4j.Logger;
 import com.google.common.annotations.VisibleForTesting;
 
 public class ConstraintChecker {
-  
+
   private ArrayList<Constraint> constrains;
   private static final Logger log = Logger.getLogger(ConstraintChecker.class);
-  
+
   private ClassLoader loader;
   private TableConfiguration conf;
 
@@ -47,7 +47,7 @@ public class ConstraintChecker {
 
   public ConstraintChecker(TableConfiguration conf) {
     constrains = new ArrayList<Constraint>();
-    
+
     this.conf = conf;
 
     try {
@@ -58,7 +58,7 @@ public class ConstraintChecker {
       } else {
         loader = AccumuloVFSClassLoader.getClassLoader();
       }
-      
+
       for (Entry<String,String> entry : conf) {
         if (entry.getKey().startsWith(Property.TABLE_CONSTRAINT_PREFIX.getKey())) {
           String className = entry.getValue();
@@ -67,7 +67,7 @@ public class ConstraintChecker {
           constrains.add(clazz.newInstance());
         }
       }
-      
+
       lastCheck.set(System.currentTimeMillis());
 
     } catch (Throwable e) {
@@ -84,21 +84,21 @@ public class ConstraintChecker {
   }
 
   public boolean classLoaderChanged() {
-    
+
     if (constrains.size() == 0)
       return false;
 
     try {
       String context = conf.get(Property.TABLE_CLASSPATH);
-      
+
       ClassLoader currentLoader;
-      
+
       if (context != null && !context.equals("")) {
         currentLoader = AccumuloVFSClassLoader.getContextManager().getClassLoader(context);
       } else {
         currentLoader = AccumuloVFSClassLoader.getClassLoader();
       }
-      
+
       return currentLoader != loader;
     } catch (Exception e) {
       log.debug("Failed to check " + e.getMessage());
@@ -117,10 +117,10 @@ public class ConstraintChecker {
   public Violations check(Environment env, Mutation m) {
     if (!env.getExtent().contains(new ComparableBytes(m.getRow()))) {
       Violations violations = new Violations();
-      
+
       ConstraintViolationSummary cvs = new ConstraintViolationSummary(SystemConstraint.class.getName(), (short) -1, "Mutation outside of tablet extent", 1);
       violations.add(cvs);
-      
+
       // do not bother with further checks since this mutation does not go with this tablet
       return violations;
     }
@@ -133,8 +133,7 @@ public class ConstraintChecker {
         if (violationCodes != null) {
           String className = constraint.getClass().getName();
           for (Short vcode : violationCodes) {
-            violations = addViolation(violations, new ConstraintViolationSummary(
-                className, vcode, constraint.getViolationDescription(vcode), 1));
+            violations = addViolation(violations, new ConstraintViolationSummary(className, vcode, constraint.getViolationDescription(vcode), 1));
           }
         }
       } catch (Throwable throwable) {
@@ -161,8 +160,7 @@ public class ConstraintChecker {
           msg = "threw some Exception";
         }
 
-        violations = addViolation(violations, new ConstraintViolationSummary(
-            constraint.getClass().getName(), vcode, "CONSTRAINT FAILED : " + msg, 1));
+        violations = addViolation(violations, new ConstraintViolationSummary(constraint.getClass().getName(), vcode, "CONSTRAINT FAILED : " + msg, 1));
       }
     }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/UnsatisfiableConstraint.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/UnsatisfiableConstraint.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/UnsatisfiableConstraint.java
index cf0c176..64bc2cd 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/UnsatisfiableConstraint.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/constraints/UnsatisfiableConstraint.java
@@ -23,21 +23,21 @@ import org.apache.accumulo.core.constraints.Constraint;
 import org.apache.accumulo.core.data.Mutation;
 
 public class UnsatisfiableConstraint implements Constraint {
-  
+
   private List<Short> violations;
   private String vDesc;
-  
+
   public UnsatisfiableConstraint(short vcode, String violationDescription) {
     this.violations = Collections.unmodifiableList(Collections.singletonList(vcode));
     this.vDesc = violationDescription;
   }
-  
+
   public List<Short> check(Environment env, Mutation mutation) {
     return violations;
   }
-  
+
   public String getViolationDescription(short violationCode) {
     return vDesc;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/data/ServerConditionalMutation.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/data/ServerConditionalMutation.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/data/ServerConditionalMutation.java
index 975300b..84137cc 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/data/ServerConditionalMutation.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/data/ServerConditionalMutation.java
@@ -24,22 +24,22 @@ import org.apache.accumulo.core.data.thrift.TConditionalMutation;
 import org.apache.accumulo.server.data.ServerMutation;
 
 /**
- * 
+ *
  */
 public class ServerConditionalMutation extends ServerMutation {
-  
+
   public static class TCMTranslator extends Translator<TConditionalMutation,ServerConditionalMutation> {
     @Override
     public ServerConditionalMutation translate(TConditionalMutation input) {
       return new ServerConditionalMutation(input);
     }
   }
-  
+
   public static final TCMTranslator TCMT = new TCMTranslator();
 
   private long cmid;
   private List<TCondition> conditions;
-  
+
   public ServerConditionalMutation(TConditionalMutation input) {
     super(input.mutation);
 
@@ -50,10 +50,9 @@ public class ServerConditionalMutation extends ServerMutation {
   public long getID() {
     return cmid;
   }
-  
+
   public List<TCondition> getConditions() {
     return conditions;
   }
-  
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/log/DfsLogger.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/DfsLogger.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/DfsLogger.java
index 18aa192..6f9be7d 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/DfsLogger.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/DfsLogger.java
@@ -152,7 +152,7 @@ public class DfsLogger {
   private static final LogFileValue EMPTY = new LogFileValue();
 
   private boolean closed = false;
-  
+
   private class LogSyncingTask implements Runnable {
 
     @Override
@@ -170,8 +170,7 @@ public class DfsLogger {
         workQueue.drainTo(work);
 
         Method durabilityMethod = null;
-        loop:
-        for (LogWork logWork : work) {
+        loop: for (LogWork logWork : work) {
           switch (logWork.durability) {
             case DEFAULT:
             case NONE:
@@ -287,7 +286,9 @@ public class DfsLogger {
 
   /**
    * Reference a pre-existing log file.
-   * @param meta the cq for the "log" entry in +r/!0
+   *
+   * @param meta
+   *          the cq for the "log" entry in +r/!0
    */
   public DfsLogger(ServerResources conf, String filename, String meta) throws IOException {
     this.conf = conf;
@@ -387,7 +388,8 @@ public class DfsLogger {
     log.debug("DfsLogger.open() begin");
     VolumeManager fs = conf.getFileSystem();
 
-    logPath = fs.choose(Optional.<String> absent(), ServerConstants.getBaseUris()) + Path.SEPARATOR + ServerConstants.WAL_DIR + Path.SEPARATOR + logger + Path.SEPARATOR + filename;
+    logPath = fs.choose(Optional.<String> absent(), ServerConstants.getBaseUris()) + Path.SEPARATOR + ServerConstants.WAL_DIR + Path.SEPARATOR + logger
+        + Path.SEPARATOR + filename;
 
     metaReference = toString();
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/log/LocalWALRecovery.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/LocalWALRecovery.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/LocalWALRecovery.java
index 10bc903..2658c1f 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/LocalWALRecovery.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/LocalWALRecovery.java
@@ -35,8 +35,8 @@ import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.SequenceFile;
-import org.apache.hadoop.io.WritableName;
 import org.apache.hadoop.io.SequenceFile.Reader;
+import org.apache.hadoop.io.WritableName;
 import org.apache.log4j.Logger;
 
 import com.beust.jcommander.JCommander;
@@ -50,10 +50,10 @@ import com.google.common.annotations.VisibleForTesting;
 @SuppressWarnings("deprecation")
 public class LocalWALRecovery implements Runnable {
   private static final Logger log = Logger.getLogger(LocalWALRecovery.class);
-  
-  static { 
-    WritableName.addName(LogFileKey.class,  org.apache.accumulo.server.logger.LogFileKey.class.getName());
-    WritableName.addName(LogFileValue.class,  org.apache.accumulo.server.logger.LogFileValue.class.getName());
+
+  static {
+    WritableName.addName(LogFileKey.class, org.apache.accumulo.server.logger.LogFileKey.class.getName());
+    WritableName.addName(LogFileValue.class, org.apache.accumulo.server.logger.LogFileValue.class.getName());
   }
 
   public static void main(String[] args) throws IOException {
@@ -150,7 +150,7 @@ public class LocalWALRecovery implements Runnable {
 
         Path localWal = new Path(file.toURI());
         FileSystem localFs = FileSystem.getLocal(fs.getConf());
-        
+
         Reader reader = new SequenceFile.Reader(localFs, localWal, localFs.getConf());
         // Reader reader = new SequenceFile.Reader(localFs.getConf(), SequenceFile.Reader.file(localWal));
         Path tmp = new Path(options.destination + "/" + name + ".copy");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/log/SortedLogRecovery.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/SortedLogRecovery.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/SortedLogRecovery.java
index 2c6d415..405ec70 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/SortedLogRecovery.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/SortedLogRecovery.java
@@ -39,21 +39,25 @@ import org.apache.log4j.Logger;
 
 /**
  * Extract Mutations for a tablet from a set of logs that have been sorted by operation and tablet.
- * 
+ *
  */
 public class SortedLogRecovery {
   private static final Logger log = Logger.getLogger(SortedLogRecovery.class);
-  
+
   static class EmptyMapFileException extends Exception {
     private static final long serialVersionUID = 1L;
 
-    public EmptyMapFileException() { super(); }
+    public EmptyMapFileException() {
+      super();
+    }
   }
 
   static class UnusedException extends Exception {
     private static final long serialVersionUID = 1L;
 
-    public UnusedException() { super(); }
+    public UnusedException() {
+      super();
+    }
   }
 
   private VolumeManager fs;
@@ -61,28 +65,28 @@ public class SortedLogRecovery {
   public SortedLogRecovery(VolumeManager fs) {
     this.fs = fs;
   }
-  
+
   private enum Status {
     INITIAL, LOOKING_FOR_FINISH, COMPLETE
   };
-  
+
   private static class LastStartToFinish {
     long lastStart = -1;
     long seq = -1;
     long lastFinish = -1;
     Status compactionStatus = Status.INITIAL;
     String tserverSession = "";
-    
+
     private void update(long newFinish) {
       this.seq = this.lastStart;
       if (newFinish != -1)
         lastFinish = newFinish;
     }
-    
+
     private void update(int newStartFile, long newStart) {
       this.lastStart = newStart;
     }
-    
+
     private void update(String newSession) {
       this.lastStart = -1;
       this.lastFinish = -1;
@@ -90,7 +94,7 @@ public class SortedLogRecovery {
       this.tserverSession = newSession;
     }
   }
-  
+
   public void recover(KeyExtent extent, List<Path> recoveryLogs, Set<String> tabletFiles, MutationReceiver mr) throws IOException {
     int[] tids = new int[recoveryLogs.size()];
     LastStartToFinish lastStartToFinish = new LastStartToFinish();
@@ -115,12 +119,12 @@ public class SortedLogRecovery {
           log.warn("Ignoring error closing file");
         }
       }
-      
+
     }
-    
+
     if (lastStartToFinish.compactionStatus == Status.LOOKING_FOR_FINISH)
       throw new RuntimeException("COMPACTION_FINISH (without preceding COMPACTION_START) not followed by successful minor compaction");
-    
+
     for (int i = 0; i < recoveryLogs.size(); i++) {
       Path logfile = recoveryLogs.get(i);
       MultiReader reader = new MultiReader(fs, logfile);
@@ -136,7 +140,7 @@ public class SortedLogRecovery {
       log.info("Recovery complete for " + extent + " using " + logfile);
     }
   }
-  
+
   private String getPathSuffix(String pathString) {
     Path path = new Path(pathString);
     if (path.depth() < 2)
@@ -144,7 +148,8 @@ public class SortedLogRecovery {
     return path.getParent().getName() + "/" + path.getName();
   }
 
-  int findLastStartToFinish(MultiReader reader, int fileno, KeyExtent extent, Set<String> tabletFiles, LastStartToFinish lastStartToFinish) throws IOException, EmptyMapFileException, UnusedException {
+  int findLastStartToFinish(MultiReader reader, int fileno, KeyExtent extent, Set<String> tabletFiles, LastStartToFinish lastStartToFinish) throws IOException,
+      EmptyMapFileException, UnusedException {
 
     HashSet<String> suffixes = new HashSet<String>();
     for (String path : tabletFiles)
@@ -158,7 +163,7 @@ public class SortedLogRecovery {
       throw new EmptyMapFileException();
     if (key.event != OPEN)
       throw new RuntimeException("First log entry value is not OPEN");
-    
+
     if (key.tserverSession.compareTo(lastStartToFinish.tserverSession) != 0) {
       if (lastStartToFinish.compactionStatus == Status.LOOKING_FOR_FINISH)
         throw new RuntimeException("COMPACTION_FINISH (without preceding COMPACTION_START) is not followed by a successful minor compaction.");
@@ -168,9 +173,9 @@ public class SortedLogRecovery {
     if (extent.isRootTablet()) {
       alternative = RootTable.OLD_EXTENT;
     }
-    
+
     LogFileKey defineKey = null;
-    
+
     // find the maximum tablet id... because a tablet may leave a tserver and then come back, in which case it would have a different tablet id
     // for the maximum tablet id, find the minimum sequence #... may be ok to find the max seq, but just want to make the code behave like it used to
     while (reader.next(key, value)) {
@@ -188,9 +193,9 @@ public class SortedLogRecovery {
     if (tid < 0) {
       throw new UnusedException();
     }
-    
+
     log.debug("Found tid, seq " + tid + " " + defineKey.seq);
-    
+
     // Scan start/stop events for this tablet
     key = defineKey;
     key.event = COMPACTION_START;
@@ -205,7 +210,7 @@ public class SortedLogRecovery {
         if (key.seq <= lastStartToFinish.lastStart)
           throw new RuntimeException("Sequence numbers are not increasing for start/stop events.");
         lastStartToFinish.update(fileno, key.seq);
-        
+
         // Tablet server finished the minor compaction, but didn't remove the entry from the METADATA table.
         log.debug("minor compaction into " + key.filename + " finished, but was still in the METADATA");
         if (suffixes.contains(getPathSuffix(key.filename)))
@@ -225,11 +230,11 @@ public class SortedLogRecovery {
     }
     return tid;
   }
-  
+
   private void playbackMutations(MultiReader reader, int tid, LastStartToFinish lastStartToFinish, MutationReceiver mr) throws IOException {
     LogFileKey key = new LogFileKey();
     LogFileValue value = new LogFileValue();
-    
+
     // Playback mutations after the last stop to finish
     log.info("Scanning for mutations starting at sequence number " + lastStartToFinish.seq + " for tid " + tid);
     key.event = MUTATION;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/log/TabletServerLogger.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/TabletServerLogger.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/TabletServerLogger.java
index ceb76da..5c3fc2d 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/log/TabletServerLogger.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/log/TabletServerLogger.java
@@ -90,7 +90,7 @@ public class TabletServerLogger {
 
   private final AtomicLong syncCounter;
   private final AtomicLong flushCounter;
-  
+
   private final static int HALT_AFTER_ERROR_COUNT = 5;
   // Die if we get 5 WAL creation errors in 10 seconds
   private final Cache<Long,Object> walErrors = CacheBuilder.newBuilder().maximumSize(HALT_AFTER_ERROR_COUNT).expireAfterWrite(10, TimeUnit.SECONDS).build();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileKey.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileKey.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileKey.java
index 3a20e8d..829cf2f 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileKey.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileKey.java
@@ -29,7 +29,7 @@ import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.hadoop.io.WritableComparable;
 
 public class LogFileKey implements WritableComparable<LogFileKey> {
-  
+
   public LogEvents event;
   public String filename = null;
   public KeyExtent tablet = null;
@@ -37,7 +37,7 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
   public int tid = -1;
   public static final int VERSION = 2;
   public String tserverSession;
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     int value = in.readByte();
@@ -79,9 +79,9 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
       default:
         throw new RuntimeException("Unknown log event type: " + event);
     }
-    
+
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     out.writeByte(event.ordinal());
@@ -119,7 +119,7 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
         throw new IllegalArgumentException("Bad value for LogFileEntry type");
     }
   }
-  
+
   static int eventType(LogEvents event) {
     // Order logs by START, TABLET_DEFINITIONS, COMPACTIONS and then MUTATIONS
     if (event == MUTATION || event == MANY_MUTATIONS) {
@@ -133,7 +133,7 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
     }
     return 2;
   }
-  
+
   private static int sign(long l) {
     if (l < 0)
       return -1;
@@ -141,7 +141,7 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
       return 1;
     return 0;
   }
-  
+
   @Override
   public int compareTo(LogFileKey o) {
     if (eventType(this.event) != eventType(o.event)) {
@@ -154,7 +154,7 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
     }
     return sign(this.seq - o.seq);
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     if (obj instanceof LogFileKey) {
@@ -162,16 +162,16 @@ public class LogFileKey implements WritableComparable<LogFileKey> {
     }
     return false;
   }
-  
+
   @Override
   public int hashCode() {
     return (int) seq;
   }
-  
+
   public static void printEntry(LogFileKey entry) {
     System.out.println(entry.toString());
   }
-  
+
   @Override
   public String toString() {
     switch (event) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileValue.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileValue.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileValue.java
index 81cd593..9ca0f38 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileValue.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/logger/LogFileValue.java
@@ -31,11 +31,11 @@ import org.apache.accumulo.server.data.ServerMutation;
 import org.apache.hadoop.io.Writable;
 
 public class LogFileValue implements Writable {
-  
+
   private static final List<Mutation> empty = Collections.emptyList();
-  
+
   public List<Mutation> mutations = empty;
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     int count = in.readInt();
@@ -46,7 +46,7 @@ public class LogFileValue implements Writable {
       mutations.add(mutation);
     }
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     out.writeInt(mutations.size());
@@ -54,18 +54,18 @@ public class LogFileValue implements Writable {
       m.write(out);
     }
   }
-  
+
   public static void print(LogFileValue value) {
     System.out.println(value.toString());
   }
-  
+
   private static String displayLabels(byte[] labels) {
     String s = new String(labels, UTF_8);
     s = s.replace("&", " & ");
     s = s.replace("|", " | ");
     return s;
   }
-  
+
   public static String format(LogFileValue lfv, int maxMutations) {
     if (lfv.mutations.size() == 0)
       return "";
@@ -80,18 +80,17 @@ public class LogFileValue implements Writable {
       builder.append("  ").append(new String(m.getRow(), UTF_8)).append("\n");
       for (ColumnUpdate update : m.getUpdates()) {
         String value = new String(update.getValue());
-        builder.append("      ").append(new String(update.getColumnFamily(), UTF_8)).append(":")
-                .append(new String(update.getColumnQualifier(), UTF_8)).append(" ").append(update.hasTimestamp() ? "[user]:" : "[system]:")
-                .append(update.getTimestamp()).append(" [").append(displayLabels(update.getColumnVisibility())).append("] ")
-                .append(update.isDeleted() ? "<deleted>" : value).append("\n");
+        builder.append("      ").append(new String(update.getColumnFamily(), UTF_8)).append(":").append(new String(update.getColumnQualifier(), UTF_8))
+            .append(" ").append(update.hasTimestamp() ? "[user]:" : "[system]:").append(update.getTimestamp()).append(" [")
+            .append(displayLabels(update.getColumnVisibility())).append("] ").append(update.isDeleted() ? "<deleted>" : value).append("\n");
       }
     }
     return builder.toString();
   }
-  
+
   @Override
   public String toString() {
     return format(this, 5);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/MasterMessage.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/MasterMessage.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/MasterMessage.java
index ecca3ce..ab71794 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/MasterMessage.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/MasterMessage.java
@@ -22,7 +22,7 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
 import org.apache.thrift.TException;
 
 public interface MasterMessage {
-  
+
   void send(TCredentials info, String serverName, MasterClientService.Iface client) throws TException, ThriftSecurityException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/SplitReportMessage.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/SplitReportMessage.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/SplitReportMessage.java
index 5d61c9c..0c93a86 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/SplitReportMessage.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/SplitReportMessage.java
@@ -20,8 +20,8 @@ import java.util.Map;
 import java.util.TreeMap;
 
 import org.apache.accumulo.core.client.impl.Translator;
-import org.apache.accumulo.core.client.impl.thrift.ThriftSecurityException;
 import org.apache.accumulo.core.client.impl.Translators;
+import org.apache.accumulo.core.client.impl.thrift.ThriftSecurityException;
 import org.apache.accumulo.core.data.KeyExtent;
 import org.apache.accumulo.core.master.thrift.MasterClientService;
 import org.apache.accumulo.core.master.thrift.TabletSplit;
@@ -33,24 +33,24 @@ import org.apache.thrift.TException;
 public class SplitReportMessage implements MasterMessage {
   Map<KeyExtent,Text> extents;
   KeyExtent old_extent;
-  
+
   public SplitReportMessage(KeyExtent old_extent, Map<KeyExtent,Text> newExtents) {
     this.old_extent = old_extent;
     extents = new TreeMap<KeyExtent,Text>(newExtents);
   }
-  
+
   public SplitReportMessage(KeyExtent old_extent, KeyExtent ne1, Text np1, KeyExtent ne2, Text np2) {
     this.old_extent = old_extent;
     extents = new TreeMap<KeyExtent,Text>();
     extents.put(ne1, np1);
     extents.put(ne2, np2);
   }
-  
+
   public void send(TCredentials credentials, String serverName, MasterClientService.Iface client) throws TException, ThriftSecurityException {
     TabletSplit split = new TabletSplit();
     split.oldTablet = old_extent.toThrift();
     split.newTablets = Translator.translate(extents.keySet(), Translators.KET);
     client.reportSplitExtent(Tracer.traceInfo(), credentials, serverName, split);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/TabletStatusMessage.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/TabletStatusMessage.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/TabletStatusMessage.java
index 655414d..25db744 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/TabletStatusMessage.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/mastermessage/TabletStatusMessage.java
@@ -18,22 +18,22 @@ package org.apache.accumulo.tserver.mastermessage;
 
 import org.apache.accumulo.core.client.impl.thrift.ThriftSecurityException;
 import org.apache.accumulo.core.data.KeyExtent;
-import org.apache.accumulo.core.master.thrift.TabletLoadState;
 import org.apache.accumulo.core.master.thrift.MasterClientService.Iface;
+import org.apache.accumulo.core.master.thrift.TabletLoadState;
 import org.apache.accumulo.core.security.thrift.TCredentials;
 import org.apache.accumulo.core.trace.Tracer;
 import org.apache.thrift.TException;
 
 public class TabletStatusMessage implements MasterMessage {
-  
+
   private KeyExtent extent;
   private TabletLoadState status;
-  
+
   public TabletStatusMessage(TabletLoadState status, KeyExtent extent) {
     this.extent = extent;
     this.status = status;
   }
-  
+
   public void send(TCredentials auth, String serverName, Iface client) throws TException, ThriftSecurityException {
     client.reportTabletStatus(Tracer.traceInfo(), auth, serverName, status, extent.toThrift());
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/metrics/TabletServerMBean.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/metrics/TabletServerMBean.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/metrics/TabletServerMBean.java
index 3b7a637..dc35c28 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/metrics/TabletServerMBean.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/metrics/TabletServerMBean.java
@@ -17,34 +17,34 @@
 package org.apache.accumulo.tserver.metrics;
 
 public interface TabletServerMBean {
-  
+
   int getOnlineCount();
-  
+
   int getOpeningCount();
-  
+
   int getUnopenedCount();
-  
+
   int getMajorCompactions();
-  
+
   int getMajorCompactionsQueued();
-  
+
   int getMinorCompactions();
-  
+
   int getMinorCompactionsQueued();
-  
+
   long getEntries();
-  
+
   long getEntriesInMemory();
-  
+
   long getQueries();
-  
+
   long getIngest();
-  
+
   long getTotalMinorCompactions();
-  
+
   double getHoldTime();
-  
+
   String getName();
-  
+
   double getAverageFilesPerTablet();
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystem.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystem.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystem.java
index 7d6c59e..a07f354 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystem.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystem.java
@@ -234,9 +234,9 @@ public class AccumuloReplicaSystem implements ReplicaSystem {
     }
   }
 
-  protected Status replicateRFiles(ClientContext peerContext, final HostAndPort peerTserver, final ReplicationTarget target,
-      final Path p, final Status status, final long sizeLimit, final String remoteTableId, final TCredentials tcreds, final ReplicaSystemHelper helper)
-      throws TTransportException, AccumuloException, AccumuloSecurityException {
+  protected Status replicateRFiles(ClientContext peerContext, final HostAndPort peerTserver, final ReplicationTarget target, final Path p, final Status status,
+      final long sizeLimit, final String remoteTableId, final TCredentials tcreds, final ReplicaSystemHelper helper) throws TTransportException,
+      AccumuloException, AccumuloSecurityException {
     DataInputStream input;
     try {
       input = getRFileInputStream(p);
@@ -280,9 +280,9 @@ public class AccumuloReplicaSystem implements ReplicaSystem {
     }
   }
 
-  protected Status replicateLogs(ClientContext peerContext, final HostAndPort peerTserver, final ReplicationTarget target,
-      final Path p, final Status status, final long sizeLimit, final String remoteTableId, final TCredentials tcreds, final ReplicaSystemHelper helper)
-      throws TTransportException, AccumuloException, AccumuloSecurityException {
+  protected Status replicateLogs(ClientContext peerContext, final HostAndPort peerTserver, final ReplicationTarget target, final Path p, final Status status,
+      final long sizeLimit, final String remoteTableId, final TCredentials tcreds, final ReplicaSystemHelper helper) throws TTransportException,
+      AccumuloException, AccumuloSecurityException {
 
     final Set<Integer> tids;
     final DataInputStream input;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationServicerHandler.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationServicerHandler.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationServicerHandler.java
index e2af4df..cc79f31 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationServicerHandler.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationServicerHandler.java
@@ -94,8 +94,7 @@ public class ReplicationServicerHandler implements Iface {
       replayer = clz.newInstance();
     } catch (InstantiationException | IllegalAccessException e1) {
       log.error("Could not instantiate replayer class {}", clz.getName());
-      throw new RemoteReplicationException(RemoteReplicationErrorCode.CANNOT_INSTANTIATE_REPLAYER, "Could not instantiate replayer class"
-          + clz.getName());
+      throw new RemoteReplicationException(RemoteReplicationErrorCode.CANNOT_INSTANTIATE_REPLAYER, "Could not instantiate replayer class" + clz.getName());
     }
 
     long entriesReplicated;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationWorker.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationWorker.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationWorker.java
index 1d20e2b..bd6bcd3 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationWorker.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/replication/ReplicationWorker.java
@@ -32,7 +32,7 @@ import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
 /**
- * 
+ *
  */
 public class ReplicationWorker implements Runnable {
   private static final Logger log = LoggerFactory.getLogger(ReplicationWorker.class);


[09/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/mock/MockShell.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/mock/MockShell.java b/shell/src/main/java/org/apache/accumulo/shell/mock/MockShell.java
index f862b68..ce70a02 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/mock/MockShell.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/mock/MockShell.java
@@ -22,6 +22,7 @@ import java.io.ByteArrayInputStream;
 import java.io.IOException;
 import java.io.InputStream;
 import java.io.OutputStream;
+
 import jline.console.ConsoleReader;
 
 import org.apache.accumulo.core.client.mock.MockInstance;
@@ -33,7 +34,7 @@ import org.apache.accumulo.shell.ShellOptionsJC;
  */
 public class MockShell extends Shell {
   private static final String NEWLINE = "\n";
-  
+
   protected InputStream in;
   protected OutputStream out;
 
@@ -45,7 +46,7 @@ public class MockShell extends Shell {
 
   public boolean config(String... args) {
     configError = super.config(args);
-    
+
     // Update the ConsoleReader with the input and output "redirected"
     try {
       this.reader = new ConsoleReader(in, out);
@@ -53,16 +54,16 @@ public class MockShell extends Shell {
       printException(e);
       configError = true;
     }
-    
+
     // Don't need this for testing purposes
     this.reader.setHistoryEnabled(false);
     this.reader.setPaginationEnabled(false);
-    
+
     // Make the parsing from the client easier;
     this.verbose = false;
     return configError;
   }
-  
+
   @Override
   protected void setInstance(ShellOptionsJC options) {
     // We always want a MockInstance for this test
@@ -72,11 +73,11 @@ public class MockShell extends Shell {
   public int start() throws IOException {
     if (configError)
       return 1;
-    
+
     String input;
     if (isVerbose())
       printInfo();
-    
+
     if (execFile != null) {
       java.util.Scanner scanner = new java.util.Scanner(execFile, UTF_8.name());
       try {
@@ -92,22 +93,22 @@ public class MockShell extends Shell {
       }
       return exitCode;
     }
-    
+
     while (true) {
       if (hasExited())
         return exitCode;
-      
+
       reader.setPrompt(getDefaultPrompt());
       input = reader.readLine();
       if (input == null) {
         reader.println();
         return exitCode;
       } // user canceled
-      
+
       execCommand(input, false, false);
     }
   }
-  
+
   /**
    * @param in
    *          the in to set
@@ -115,7 +116,7 @@ public class MockShell extends Shell {
   public void setConsoleInputStream(InputStream in) {
     this.in = in;
   }
-  
+
   /**
    * @param out
    *          the output stream to set
@@ -126,18 +127,18 @@ public class MockShell extends Shell {
 
   /**
    * Convenience method to create the byte-array to hand to the console
-   * 
+   *
    * @param commands
    *          An array of commands to run
    * @return A byte[] input stream which can be handed to the console.
    */
   public static ByteArrayInputStream makeCommands(String... commands) {
     StringBuilder sb = new StringBuilder(commands.length * 8);
-    
+
     for (String command : commands) {
       sb.append(command).append(NEWLINE);
     }
-    
+
     return new ByteArrayInputStream(sb.toString().getBytes(UTF_8));
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/PasswordConverterTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/PasswordConverterTest.java b/shell/src/test/java/org/apache/accumulo/shell/PasswordConverterTest.java
index 8e03830..75d7293 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/PasswordConverterTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/PasswordConverterTest.java
@@ -38,40 +38,40 @@ import com.beust.jcommander.Parameter;
 import com.beust.jcommander.ParameterException;
 
 public class PasswordConverterTest {
-  
+
   private class Password {
     @Parameter(names = "--password", converter = PasswordConverter.class)
     String password;
   }
-  
+
   private String[] argv;
   private Password password;
   private static InputStream realIn;
-  
+
   @BeforeClass
   public static void saveIn() {
     realIn = System.in;
   }
-  
+
   @Before
   public void setup() throws IOException {
     argv = new String[] {"--password", ""};
     password = new Password();
-    
+
     PipedInputStream in = new PipedInputStream();
     PipedOutputStream out = new PipedOutputStream(in);
     OutputStreamWriter osw = new OutputStreamWriter(out);
     osw.write("secret");
     osw.close();
-    
+
     System.setIn(in);
   }
-  
+
   @After
   public void teardown() {
     System.setIn(realIn);
   }
-  
+
   @Test
   public void testPass() {
     String expected = String.valueOf(Math.random());
@@ -79,7 +79,7 @@ public class PasswordConverterTest {
     new JCommander(password, argv);
     assertEquals(expected, password.password);
   }
-  
+
   @Test
   public void testEnv() {
     String name = System.getenv().keySet().iterator().next();
@@ -87,7 +87,7 @@ public class PasswordConverterTest {
     new JCommander(password, argv);
     assertEquals(System.getenv(name), password.password);
   }
-  
+
   @Test
   public void testFile() throws FileNotFoundException {
     argv[1] = "file:pom.xml";
@@ -97,8 +97,8 @@ public class PasswordConverterTest {
     new JCommander(password, argv);
     assertEquals(expected, password.password);
   }
-  
-  @Test(expected=ParameterException.class)
+
+  @Test(expected = ParameterException.class)
   public void testNoFile() throws FileNotFoundException {
     argv[1] = "file:doesnotexist";
     new JCommander(password, argv);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/ShellConfigTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/ShellConfigTest.java b/shell/src/test/java/org/apache/accumulo/shell/ShellConfigTest.java
index 0e72c8c..ca06c22 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/ShellConfigTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/ShellConfigTest.java
@@ -27,7 +27,6 @@ import java.io.PrintWriter;
 import jline.console.ConsoleReader;
 
 import org.apache.accumulo.core.client.security.tokens.PasswordToken;
-import org.apache.accumulo.shell.Shell;
 import org.apache.accumulo.shell.ShellTest.TestOutputStream;
 import org.apache.log4j.Level;
 import org.junit.After;
@@ -40,11 +39,11 @@ public class ShellConfigTest {
   TestOutputStream output;
   Shell shell;
   PrintStream out;
-  
+
   @Before
   public void setUp() throws Exception {
     Shell.log.setLevel(Level.ERROR);
-    
+
     out = System.out;
     output = new TestOutputStream();
     System.setOut(new PrintStream(output));
@@ -52,37 +51,37 @@ public class ShellConfigTest {
     shell = new Shell(new ConsoleReader(new FileInputStream(FileDescriptor.in), output), new PrintWriter(output));
     shell.setLogErrorsToConsole();
   }
-  
+
   @After
   public void teardown() throws Exception {
     shell.shutdown();
     output.clear();
     System.setOut(out);
   }
-  
+
   @Test
   public void testHelp() {
     assertTrue(shell.config("--help"));
     assertTrue("Did not print usage", output.get().startsWith("Usage"));
   }
-  
+
   @Test
   public void testBadArg() {
     assertTrue(shell.config("--bogus"));
     assertTrue("Did not print usage", output.get().startsWith("Usage"));
   }
-  
+
   @Test
   public void testTokenWithoutOptions() {
     assertTrue(shell.config("--fake", "-tc", PasswordToken.class.getCanonicalName()));
     assertFalse(output.get().contains(ParameterException.class.getCanonicalName()));
   }
-  
+
   @Test
   public void testTokenAndOption() {
     assertFalse(shell.config("--fake", "-tc", PasswordToken.class.getCanonicalName(), "-u", "foo", "-l", "password=foo"));
   }
-  
+
   @Test
   public void testTokenAndOptionAndPassword() {
     assertTrue(shell.config("--fake", "-tc", PasswordToken.class.getCanonicalName(), "-l", "password=foo", "-p", "bar"));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/ShellTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/ShellTest.java b/shell/src/test/java/org/apache/accumulo/shell/ShellTest.java
index 2ce8ed6..82e61fd 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/ShellTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/ShellTest.java
@@ -34,7 +34,6 @@ import java.util.TimeZone;
 import jline.console.ConsoleReader;
 
 import org.apache.accumulo.core.util.format.DateStringFormatter;
-import org.apache.accumulo.shell.Shell;
 import org.apache.log4j.Level;
 import org.junit.After;
 import org.junit.Before;
@@ -286,12 +285,12 @@ public class ShellTest {
 
     input.set("\n\n");
     exec("setiter -scan -class org.apache.accumulo.core.iterators.ColumnFamilyCounter -p 30 -name foo", true);
-    
+
     input.set("bar\nname value\n");
     exec("setiter -scan -class org.apache.accumulo.core.iterators.ColumnFamilyCounter -p 31", true);
-    
-    //TODO can't verify this as config -t fails, functionality verified in ShellServerIT
-    
+
+    // TODO can't verify this as config -t fails, functionality verified in ShellServerIT
+
     exec("deletetable t -f", true, "Table: [t] has been deleted");
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/ShellUtilTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/ShellUtilTest.java b/shell/src/test/java/org/apache/accumulo/shell/ShellUtilTest.java
index c3af70a..9b3be51 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/ShellUtilTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/ShellUtilTest.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.shell;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
 
 import java.io.File;
 import java.io.FileNotFoundException;
@@ -25,7 +25,6 @@ import java.io.IOException;
 import java.util.List;
 
 import org.apache.accumulo.core.util.Base64;
-import org.apache.accumulo.shell.ShellUtil;
 import org.apache.commons.io.FileUtils;
 import org.apache.hadoop.io.Text;
 import org.junit.Rule;
@@ -55,9 +54,7 @@ public class ShellUtilTest {
     File testFile = new File(folder.getRoot(), "testFileWithDecode.txt");
     FileUtils.writeStringToFile(testFile, FILEDATA);
     List<Text> output = ShellUtil.scanFile(testFile.getAbsolutePath(), true);
-    assertEquals(
-        ImmutableList.of(new Text(Base64.decodeBase64("line1".getBytes(UTF_8))), new Text(Base64.decodeBase64("line2".getBytes(UTF_8)))),
-        output);
+    assertEquals(ImmutableList.of(new Text(Base64.decodeBase64("line1".getBytes(UTF_8))), new Text(Base64.decodeBase64("line2".getBytes(UTF_8)))), output);
   }
 
   @Test(expected = FileNotFoundException.class)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/commands/DeleteTableCommandTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/commands/DeleteTableCommandTest.java b/shell/src/test/java/org/apache/accumulo/shell/commands/DeleteTableCommandTest.java
index 3cbb227..4877552 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/commands/DeleteTableCommandTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/commands/DeleteTableCommandTest.java
@@ -35,7 +35,7 @@ public class DeleteTableCommandTest {
     Set<String> tables = new HashSet<String>(Arrays.asList(MetadataTable.NAME, RootTable.NAME, "a1", "a2"));
     DeleteTableCommand cmd = new DeleteTableCommand();
     cmd.pruneTables("a.*", tables);
-    
+
     Assert.assertEquals(new HashSet<String>(Arrays.asList("a1", "a2")), tables);
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/commands/DropUserCommandTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/commands/DropUserCommandTest.java b/shell/src/test/java/org/apache/accumulo/shell/commands/DropUserCommandTest.java
index 2c82bf1..a08edc5 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/commands/DropUserCommandTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/commands/DropUserCommandTest.java
@@ -21,14 +21,13 @@ import jline.console.ConsoleReader;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.admin.SecurityOperations;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.commands.DropUserCommand;
 import org.apache.commons.cli.CommandLine;
 import org.easymock.EasyMock;
 import org.junit.Before;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class DropUserCommandTest {
 
@@ -36,7 +35,7 @@ public class DropUserCommandTest {
 
   @Before
   public void setup() {
-    cmd  = new DropUserCommand();
+    cmd = new DropUserCommand();
 
     // Initialize that internal state
     cmd.getOptions();
@@ -52,7 +51,7 @@ public class DropUserCommandTest {
 
     EasyMock.expect(shellState.getConnector()).andReturn(conn);
 
-    // The user we want to remove 
+    // The user we want to remove
     EasyMock.expect(cli.getArgs()).andReturn(new String[] {"user"});
 
     // We're the root user
@@ -74,7 +73,7 @@ public class DropUserCommandTest {
     EasyMock.expectLastCall();
 
     EasyMock.replay(conn, cli, shellState, reader, secOps);
-    
+
     cmd.execute("dropuser foo -f", cli, shellState);
 
     EasyMock.verify(conn, cli, shellState, reader, secOps);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/commands/FormatterCommandTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/commands/FormatterCommandTest.java b/shell/src/test/java/org/apache/accumulo/shell/commands/FormatterCommandTest.java
index ddcd243..8a174d2 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/commands/FormatterCommandTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/commands/FormatterCommandTest.java
@@ -41,87 +41,87 @@ import org.junit.Test;
 public class FormatterCommandTest {
   ByteArrayOutputStream out = null;
   InputStream in = null;
-  
+
   @Test
   public void test() throws IOException, AccumuloException, AccumuloSecurityException, TableExistsException, ClassNotFoundException {
     // Keep the Shell AUDIT log off the test output
     Logger.getLogger(Shell.class).setLevel(Level.WARN);
-    
+
     final String[] args = new String[] {"--fake", "-u", "root", "-p", ""};
-    
+
     final String[] commands = createCommands();
-    
+
     in = MockShell.makeCommands(commands);
     out = new ByteArrayOutputStream();
-    
+
     final MockShell shell = new MockShell(in, out);
     shell.config(args);
-    
+
     // Can't call createtable in the shell with MockAccumulo
     shell.getConnector().tableOperations().create("test");
-    
+
     try {
       shell.start();
     } catch (Exception e) {
       Assert.fail("Exception while running commands: " + e.getMessage());
     }
-    
+
     shell.getReader().flush();
-    
+
     final String[] output = new String(out.toByteArray()).split("\n\r");
-    
+
     boolean formatterOn = false;
-    
+
     final String[] expectedDefault = new String[] {"row cf:cq []    1234abcd", "row cf1:cq1 []    9876fedc", "row2 cf:cq []    13579bdf",
         "row2 cf1:cq []    2468ace"};
-    
+
     final String[] expectedFormatted = new String[] {"row cf:cq []    0x31 0x32 0x33 0x34 0x61 0x62 0x63 0x64",
         "row cf1:cq1 []    0x39 0x38 0x37 0x36 0x66 0x65 0x64 0x63", "row2 cf:cq []    0x31 0x33 0x35 0x37 0x39 0x62 0x64 0x66",
         "row2 cf1:cq []    0x32 0x34 0x36 0x38 0x61 0x63 0x65"};
-    
+
     int outputIndex = 0;
     while (outputIndex < output.length) {
       final String line = output[outputIndex];
-      
+
       if (line.startsWith("root@mock-instance")) {
         if (line.contains("formatter")) {
           formatterOn = true;
         }
-        
+
         outputIndex++;
       } else if (line.startsWith("row")) {
         int expectedIndex = 0;
         String[] comparisonData;
-        
+
         // Pick the type of data we expect (formatted or default)
         if (formatterOn) {
           comparisonData = expectedFormatted;
         } else {
           comparisonData = expectedDefault;
         }
-        
+
         // Ensure each output is what we expected
         while (expectedIndex + outputIndex < output.length && expectedIndex < expectedFormatted.length) {
           Assert.assertEquals(comparisonData[expectedIndex].trim(), output[expectedIndex + outputIndex].trim());
           expectedIndex++;
         }
-        
+
         outputIndex += expectedIndex;
       }
     }
   }
-  
+
   private String[] createCommands() {
     return new String[] {"table test", "insert row cf cq 1234abcd", "insert row cf1 cq1 9876fedc", "insert row2 cf cq 13579bdf", "insert row2 cf1 cq 2468ace",
         "scan", "formatter -t test -f org.apache.accumulo.core.util.shell.command.FormatterCommandTest$HexFormatter", "scan"};
   }
-  
+
   /**
    * <p>
    * Simple <code>Formatter</code> that will convert each character in the Value from decimal to hexadecimal. Will automatically skip over characters in the
    * value which do not fall within the [0-9,a-f] range.
    * </p>
-   * 
+   *
    * <p>
    * Example: <code>'0'</code> will be displayed as <code>'0x30'</code>
    * </p>
@@ -129,56 +129,56 @@ public class FormatterCommandTest {
   public static class HexFormatter implements Formatter {
     private Iterator<Entry<Key,Value>> iter = null;
     private boolean printTs = false;
-    
+
     private final static String tab = "\t";
     private final static String newline = "\n";
-    
+
     public HexFormatter() {}
-    
+
     @Override
     public boolean hasNext() {
       return this.iter.hasNext();
     }
-    
+
     @Override
     public String next() {
       final Entry<Key,Value> entry = iter.next();
-      
+
       String key;
-      
+
       // Observe the timestamps
       if (printTs) {
         key = entry.getKey().toString();
       } else {
         key = entry.getKey().toStringNoTime();
       }
-      
+
       final Value v = entry.getValue();
-      
+
       // Approximate how much space we'll need
       final StringBuilder sb = new StringBuilder(key.length() + v.getSize() * 5);
-      
+
       sb.append(key).append(tab);
-      
+
       for (byte b : v.get()) {
         if ((b >= 48 && b <= 57) || (b >= 97 || b <= 102)) {
           sb.append(String.format("0x%x ", Integer.valueOf(b)));
         }
       }
-      
+
       sb.append(newline);
-      
+
       return sb.toString();
     }
-    
+
     @Override
     public void remove() {}
-    
+
     @Override
     public void initialize(final Iterable<Entry<Key,Value>> scanner, final boolean printTimestamps) {
       this.iter = scanner.iterator();
       this.printTs = printTimestamps;
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/commands/HistoryCommandTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/commands/HistoryCommandTest.java b/shell/src/test/java/org/apache/accumulo/shell/commands/HistoryCommandTest.java
index 8a5b598..638af3f 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/commands/HistoryCommandTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/commands/HistoryCommandTest.java
@@ -30,7 +30,6 @@ import jline.console.history.History;
 import jline.console.history.MemoryHistory;
 
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.commands.HistoryCommand;
 import org.apache.commons.cli.CommandLine;
 import org.junit.Assume;
 import org.junit.Before;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/test/java/org/apache/accumulo/shell/format/DeleterFormatterTest.java
----------------------------------------------------------------------
diff --git a/shell/src/test/java/org/apache/accumulo/shell/format/DeleterFormatterTest.java b/shell/src/test/java/org/apache/accumulo/shell/format/DeleterFormatterTest.java
index 0229cee..d99c0f2 100644
--- a/shell/src/test/java/org/apache/accumulo/shell/format/DeleterFormatterTest.java
+++ b/shell/src/test/java/org/apache/accumulo/shell/format/DeleterFormatterTest.java
@@ -44,7 +44,6 @@ import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.shell.Shell;
-import org.apache.accumulo.shell.format.DeleterFormatter;
 import org.junit.Before;
 import org.junit.Test;
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/Main.java b/start/src/main/java/org/apache/accumulo/start/Main.java
index 06ea686..55ee48c 100644
--- a/start/src/main/java/org/apache/accumulo/start/Main.java
+++ b/start/src/main/java/org/apache/accumulo/start/Main.java
@@ -27,7 +27,7 @@ import org.apache.accumulo.start.classloader.AccumuloClassLoader;
 import org.apache.log4j.Logger;
 
 public class Main {
-  
+
   private static final Logger log = Logger.getLogger(Main.class);
 
   public static void main(String[] args) {
@@ -136,7 +136,7 @@ public class Main {
       try {
         main = runTMP.getMethod("main", args.getClass());
       } catch (Throwable t) {
-        log.error("Could not run main method on '"+runTMP.getName()+"'.", t);
+        log.error("Could not run main method on '" + runTMP.getName() + "'.", t);
       }
       if (main == null || !Modifier.isPublic(main.getModifiers()) || !Modifier.isStatic(main.getModifiers())) {
         System.out.println(args[0] + " must implement a public static void main(String args[]) method");
@@ -190,7 +190,7 @@ public class Main {
    *          The {@link Throwable} containing a stack trace to print.
    */
   private static void die(Throwable t) {
-    log.error("Thread '"+Thread.currentThread().getName()+"' died.", t);
+    log.error("Thread '" + Thread.currentThread().getName() + "' died.", t);
     System.exit(1);
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/TestMain.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/TestMain.java b/start/src/main/java/org/apache/accumulo/start/TestMain.java
index 9320bfa..e52faba 100644
--- a/start/src/main/java/org/apache/accumulo/start/TestMain.java
+++ b/start/src/main/java/org/apache/accumulo/start/TestMain.java
@@ -16,11 +16,11 @@
  */
 package org.apache.accumulo.start;
 
-/** This program tests that the proper exit code is propagated to the shell.
+/**
+ * This program tests that the proper exit code is propagated to the shell.
  *
- * $ ./bin/accumulo org.apache.accumulo.start.TestMain
- * $ ./bin/accumulo org.apache.accumulo.start.TestMain badExit
- * $ ./bin/accumulo org.apache.accumulo.start.TestMain throw
+ * $ ./bin/accumulo org.apache.accumulo.start.TestMain $ ./bin/accumulo org.apache.accumulo.start.TestMain badExit $ ./bin/accumulo
+ * org.apache.accumulo.start.TestMain throw
  *
  * The last case tests the proper logging of an exception.
  */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
index 2db8a5d..25cf646 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
@@ -42,20 +42,20 @@ import org.w3c.dom.Node;
 import org.w3c.dom.NodeList;
 
 /**
- * 
+ *
  */
 public class AccumuloClassLoader {
-  
+
   public static final String CLASSPATH_PROPERTY_NAME = "general.classpaths";
-  
+
   /* @formatter:off */
-  public static final String ACCUMULO_CLASSPATH_VALUE = 
-      "$ACCUMULO_CONF_DIR,\n" + 
-          "$ACCUMULO_HOME/lib/[^.].*.jar,\n" + 
-          "$ZOOKEEPER_HOME/zookeeper[^.].*.jar,\n" + 
+  public static final String ACCUMULO_CLASSPATH_VALUE =
+      "$ACCUMULO_CONF_DIR,\n" +
+          "$ACCUMULO_HOME/lib/[^.].*.jar,\n" +
+          "$ZOOKEEPER_HOME/zookeeper[^.].*.jar,\n" +
           "$HADOOP_CONF_DIR,\n" +
-          "$HADOOP_PREFIX/[^.].*.jar,\n" + 
-          "$HADOOP_PREFIX/lib/[^.].*.jar,\n" + 
+          "$HADOOP_PREFIX/[^.].*.jar,\n" +
+          "$HADOOP_PREFIX/lib/[^.].*.jar,\n" +
           "$HADOOP_PREFIX/share/hadoop/common/.*.jar,\n" +
           "$HADOOP_PREFIX/share/hadoop/common/lib/.*.jar,\n" +
           "$HADOOP_PREFIX/share/hadoop/hdfs/.*.jar,\n" +
@@ -67,16 +67,16 @@ public class AccumuloClassLoader {
           "/usr/lib/hadoop-yarn/[^.].*.jar,\n"
           ;
   /* @formatter:on */
-  
+
   public static final String MAVEN_PROJECT_BASEDIR_PROPERTY_NAME = "general.maven.project.basedir";
   public static final String DEFAULT_MAVEN_PROJECT_BASEDIR_VALUE = "";
-  
+
   private static String SITE_CONF;
-  
+
   private static URLClassLoader classloader;
-  
+
   private static final Logger log = Logger.getLogger(AccumuloClassLoader.class);
-  
+
   static {
     String configFile = System.getProperty("org.apache.accumulo.config.file", "accumulo-site.xml");
     if (System.getenv("ACCUMULO_CONF_DIR") != null) {
@@ -89,11 +89,11 @@ public class AccumuloClassLoader {
       SITE_CONF = null;
     }
   }
-  
+
   /**
    * Parses and XML Document for a property node for a <name> with the value propertyName if it finds one the function return that property's value for its
    * <value> node. If not found the function will return null
-   * 
+   *
    * @param d
    *          XMLDocument to search through
    */
@@ -111,20 +111,20 @@ public class AccumuloClassLoader {
     }
     return null;
   }
-  
+
   /**
    * Looks for the site configuration file for Accumulo and if it has a property for propertyName return it otherwise returns defaultValue Should throw an
    * exception if the default configuration can not be read;
-   * 
+   *
    * @param propertyName
    *          Name of the property to pull
    * @param defaultValue
    *          Value to default to if not found.
    * @return site or default class path String
    */
-  
+
   public static String getAccumuloString(String propertyName, String defaultValue) {
-    
+
     try {
       DocumentBuilderFactory dbf = DocumentBuilderFactory.newInstance();
       DocumentBuilder db = dbf.newDocumentBuilder();
@@ -142,7 +142,7 @@ public class AccumuloClassLoader {
       throw new IllegalStateException("ClassPath Strings Lookup failed", e);
     }
   }
-  
+
   /**
    * Replace environment variables in the classpath string with their actual value
    */
@@ -161,7 +161,7 @@ public class AccumuloClassLoader {
     }
     return classpath;
   }
-  
+
   /**
    * Populate the list of URLs with the items in the classpath string
    */
@@ -169,9 +169,9 @@ public class AccumuloClassLoader {
     classpath = classpath.trim();
     if (classpath.length() == 0)
       return;
-    
+
     classpath = replaceEnvVars(classpath, System.getenv());
-    
+
     // Try to make a URI out of the classpath
     URI uri = null;
     try {
@@ -179,7 +179,7 @@ public class AccumuloClassLoader {
     } catch (URISyntaxException e) {
       // Not a valid URI
     }
-    
+
     if (null == uri || !uri.isAbsolute() || (null != uri.getScheme() && uri.getScheme().equals("file://"))) {
       // Then treat this URI as a File.
       // This checks to see if the url string is a dir if it expand and get all jars in that directory
@@ -207,9 +207,9 @@ public class AccumuloClassLoader {
     } else {
       urls.add(uri.toURL());
     }
-    
+
   }
-  
+
   private static ArrayList<URL> findAccumuloURLs() throws IOException {
     String cp = getAccumuloString(AccumuloClassLoader.CLASSPATH_PROPERTY_NAME, AccumuloClassLoader.ACCUMULO_CLASSPATH_VALUE);
     if (cp == null)
@@ -225,7 +225,7 @@ public class AccumuloClassLoader {
     }
     return urls;
   }
-  
+
   private static Set<String> getMavenClasspaths() {
     String baseDirname = AccumuloClassLoader.getAccumuloString(MAVEN_PROJECT_BASEDIR_PROPERTY_NAME, DEFAULT_MAVEN_PROJECT_BASEDIR_VALUE);
     if (baseDirname == null || baseDirname.trim().isEmpty())
@@ -234,7 +234,7 @@ public class AccumuloClassLoader {
     findMavenTargetClasses(paths, new File(baseDirname.trim()), 0);
     return paths;
   }
-  
+
   private static void findMavenTargetClasses(Set<String> paths, File file, int depth) {
     if (depth > 3)
       return;
@@ -246,18 +246,18 @@ public class AccumuloClassLoader {
       paths.add(file.getParentFile().getAbsolutePath() + File.separator + "target" + File.separator + "classes");
     }
   }
-  
+
   public static synchronized ClassLoader getClassLoader() throws IOException {
     if (classloader == null) {
       ArrayList<URL> urls = findAccumuloURLs();
-      
+
       ClassLoader parentClassLoader = AccumuloClassLoader.class.getClassLoader();
-      
+
       log.debug("Create 2nd tier ClassLoader using URLs: " + urls.toString());
       URLClassLoader aClassLoader = new URLClassLoader(urls.toArray(new URL[urls.size()]), parentClassLoader) {
         @Override
         protected synchronized Class<?> loadClass(String name, boolean resolve) throws ClassNotFoundException {
-          
+
           if (name.startsWith("org.apache.accumulo.start.classloader.vfs")) {
             Class<?> c = findLoadedClass(name);
             if (c == null) {
@@ -272,7 +272,7 @@ public class AccumuloClassLoader {
       };
       classloader = aClassLoader;
     }
-    
+
     return classloader;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoader.java
index c5e8d8b..88dfd1e 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloReloadingVFSClassLoader.java
@@ -29,18 +29,17 @@ import org.apache.commons.vfs2.impl.VFSClassLoader;
 import org.apache.log4j.Logger;
 
 /**
- * Classloader that delegates operations to a VFSClassLoader object. This class also listens
- * for changes in any of the files/directories that are in the classpath and will recreate
- * the delegate object if there is any change in the classpath.
+ * Classloader that delegates operations to a VFSClassLoader object. This class also listens for changes in any of the files/directories that are in the
+ * classpath and will recreate the delegate object if there is any change in the classpath.
  *
  */
 public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingClassLoader {
-  
+
   private static final Logger log = Logger.getLogger(AccumuloReloadingVFSClassLoader.class);
 
   // set to 5 mins. The rational behind this large time is to avoid a gazillion tservers all asking the name node for info too frequently.
   private static final int DEFAULT_TIMEOUT = 300000;
-  
+
   private String uris;
   private FileObject[] files;
   private FileSystemManager vfs = null;
@@ -48,6 +47,7 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
   private DefaultFileMonitor monitor = null;
   private VFSClassLoader cl = null;
   private boolean preDelegate;
+
   public String stringify(FileObject[] files) {
     StringBuilder sb = new StringBuilder();
     sb.append('[');
@@ -79,13 +79,13 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
         throw new RuntimeException(fse);
       }
     }
-    
+
     return cl;
   }
-  
+
   private synchronized void setClassloader(VFSClassLoader cl) {
     this.cl = cl;
-    
+
   }
 
   public AccumuloReloadingVFSClassLoader(String uris, FileSystemManager vfs, ReloadingClassLoader parent, long monitorDelay, boolean preDelegate)
@@ -95,7 +95,7 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
     this.vfs = vfs;
     this.parent = parent;
     this.preDelegate = preDelegate;
-    
+
     ArrayList<FileObject> pathsToMonitor = new ArrayList<FileObject>();
     files = AccumuloVFSClassLoader.resolve(vfs, uris, pathsToMonitor);
 
@@ -103,7 +103,7 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
       cl = new VFSClassLoader(files, vfs, parent.getClassLoader());
     else
       cl = new PostDelegatingVFSClassLoader(files, vfs, parent.getClassLoader());
-    
+
     monitor = new DefaultFileMonitor(this);
     monitor.setDelay(monitorDelay);
     monitor.setRecursive(false);
@@ -113,16 +113,15 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
     }
     monitor.start();
   }
-  
-  public AccumuloReloadingVFSClassLoader(String uris, FileSystemManager vfs, final ReloadingClassLoader parent, boolean preDelegate)
-      throws FileSystemException {
+
+  public AccumuloReloadingVFSClassLoader(String uris, FileSystemManager vfs, final ReloadingClassLoader parent, boolean preDelegate) throws FileSystemException {
     this(uris, vfs, parent, DEFAULT_TIMEOUT, preDelegate);
   }
 
   synchronized public FileObject[] getFiles() {
     return Arrays.copyOf(this.files, this.files.length);
   }
-  
+
   /**
    * Should be ok if this is not called because the thread started by DefaultFileMonitor is a daemon thread
    */
@@ -148,12 +147,10 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
     setClassloader(null);
   }
 
-  
-  
   @Override
   public String toString() {
     StringBuilder buf = new StringBuilder();
-    
+
     for (FileObject f : files) {
       try {
         buf.append("\t").append(f.getURL().toString()).append("\n");
@@ -161,11 +158,8 @@ public class AccumuloReloadingVFSClassLoader implements FileListener, ReloadingC
         log.error("Error getting URL for file", e);
       }
     }
-    
+
     return buf.toString();
   }
 
-
-
-  
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloVFSClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloVFSClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloVFSClassLoader.java
index 8195dd2..1cee6d7 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloVFSClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/AccumuloVFSClassLoader.java
@@ -46,7 +46,7 @@ import org.apache.log4j.Logger;
 
 /**
  * This class builds a hierarchy of Classloaders in the form of:
- * 
+ *
  * <pre>
  * SystemClassLoader that loads JVM classes
  *       ^
@@ -58,10 +58,10 @@ import org.apache.log4j.Logger;
  *       ^
  *       |
  * AccumuloReloadingVFSClassLoader That loads jars from locations in general.dynamic.classpaths.  Used to load jar dynamically.
- * 
+ *
  * </pre>
- * 
- * 
+ *
+ *
  */
 public class AccumuloVFSClassLoader {
 
@@ -90,7 +90,7 @@ public class AccumuloVFSClassLoader {
   public static final String VFS_CONTEXT_CLASSPATH_PROPERTY = "general.vfs.context.classpath.";
 
   public static final String VFS_CACHE_DIR = "general.vfs.cache.dir";
-  
+
   private static ClassLoader parent = null;
   private static volatile ReloadingClassLoader loader = null;
   private static final Object lock = new Object();
@@ -219,11 +219,11 @@ public class AccumuloVFSClassLoader {
           localLoader = createDynamicClassloader(new VFSClassLoader(vfsCP, vfs, parent));
           loader = localLoader;
 
-          //An HDFS FileSystem and Configuration object were created for each unique HDFS namespace in the call to resolve above.
-          //The HDFS Client did us a favor and cached these objects so that the next time someone calls FileSystem.get(uri), they
-          //get the cached object. However, these objects were created not with the system VFS classloader, but the classloader above
-          //it. We need to override the classloader on the Configuration objects. Ran into an issue were log recovery was being attempted
-          //and SequenceFile$Reader was trying to instantiate the key class via WritableName.getClass(String, Configuration)
+          // An HDFS FileSystem and Configuration object were created for each unique HDFS namespace in the call to resolve above.
+          // The HDFS Client did us a favor and cached these objects so that the next time someone calls FileSystem.get(uri), they
+          // get the cached object. However, these objects were created not with the system VFS classloader, but the classloader above
+          // it. We need to override the classloader on the Configuration objects. Ran into an issue were log recovery was being attempted
+          // and SequenceFile$Reader was trying to instantiate the key class via WritableName.getClass(String, Configuration)
           for (FileObject fo : vfsCP) {
             if (fo instanceof HdfsFileObject) {
               String uri = fo.getName().getRootURI();
@@ -276,7 +276,7 @@ public class AccumuloVFSClassLoader {
     vfs.addMimeTypeMap("application/zip", "zip");
     vfs.setFileContentInfoFactory(new FileContentInfoFilenameFactory());
     vfs.setFilesCache(new SoftRefFilesCache());
-    File cacheDir = computeTopCacheDir(); 
+    File cacheDir = computeTopCacheDir();
     vfs.setReplicator(new UniqueFileReplicator(cacheDir));
     vfs.setCacheStrategy(CacheStrategy.ON_RESOLVE);
     vfs.init();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ContextManager.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ContextManager.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ContextManager.java
index c7bc120..1cf8637 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ContextManager.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ContextManager.java
@@ -27,28 +27,28 @@ import org.apache.commons.vfs2.FileSystemException;
 import org.apache.commons.vfs2.FileSystemManager;
 
 public class ContextManager {
-  
+
   // there is a lock per context so that one context can initialize w/o blocking another context
   private class Context {
     AccumuloReloadingVFSClassLoader loader;
     ContextConfig cconfig;
     boolean closed = false;
-    
+
     Context(ContextConfig cconfig) {
       this.cconfig = cconfig;
     }
-    
+
     synchronized ClassLoader getClassLoader() throws FileSystemException {
       if (closed)
         return null;
-      
+
       if (loader == null) {
         loader = new AccumuloReloadingVFSClassLoader(cconfig.uris, vfs, parent, cconfig.preDelegation);
       }
-      
+
       return loader.getClassLoader();
     }
-    
+
     synchronized void close() {
       closed = true;
       if (loader != null) {
@@ -57,79 +57,79 @@ public class ContextManager {
       loader = null;
     }
   }
-  
+
   private Map<String,Context> contexts = new HashMap<String,Context>();
-  
+
   private volatile ContextsConfig config;
   private FileSystemManager vfs;
   private ReloadingClassLoader parent;
-  
+
   ContextManager(FileSystemManager vfs, ReloadingClassLoader parent) {
     this.vfs = vfs;
     this.parent = parent;
   }
-  
+
   public static class ContextConfig {
     String uris;
     boolean preDelegation;
-    
+
     public ContextConfig(String uris, boolean preDelegation) {
       this.uris = uris;
       this.preDelegation = preDelegation;
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof ContextConfig) {
         ContextConfig oc = (ContextConfig) o;
-        
+
         return uris.equals(oc.uris) && preDelegation == oc.preDelegation;
       }
-      
+
       return false;
     }
-    
+
     @Override
     public int hashCode() {
       return uris.hashCode() + (preDelegation ? Boolean.TRUE : Boolean.FALSE).hashCode();
     }
   }
-  
+
   public interface ContextsConfig {
     ContextConfig getContextConfig(String context);
   }
-  
+
   public static class DefaultContextsConfig implements ContextsConfig {
-    
+
     private Iterable<Entry<String,String>> config;
-    
+
     public DefaultContextsConfig(Iterable<Entry<String,String>> config) {
       this.config = config;
     }
-    
+
     @Override
     public ContextConfig getContextConfig(String context) {
-      
+
       String key = AccumuloVFSClassLoader.VFS_CONTEXT_CLASSPATH_PROPERTY + context;
-      
+
       String uris = null;
       boolean preDelegate = true;
-      
+
       Iterator<Entry<String,String>> iter = config.iterator();
       while (iter.hasNext()) {
         Entry<String,String> entry = iter.next();
         if (entry.getKey().equals(key)) {
           uris = entry.getValue();
         }
-        
+
         if (entry.getKey().equals(key + ".delegation") && entry.getValue().trim().equalsIgnoreCase("post")) {
           preDelegate = false;
         }
       }
-      
+
       if (uris != null)
         return new ContextConfig(uris, preDelegate);
-      
+
       return null;
     }
   }
@@ -142,22 +142,22 @@ public class ContextManager {
       throw new IllegalStateException("Context manager config already set");
     this.config = config;
   }
-  
+
   public ClassLoader getClassLoader(String contextName) throws FileSystemException {
-    
+
     ContextConfig cconfig = config.getContextConfig(contextName);
-    
+
     if (cconfig == null)
       throw new IllegalArgumentException("Unknown context " + contextName);
-    
+
     Context context = null;
     Context contextToClose = null;
-    
+
     synchronized (this) {
       // only manipulate internal data structs in this sync block... avoid creating or closing classloader, reading config, etc... basically avoid operations
       // that may block
       context = contexts.get(contextName);
-      
+
       if (context == null) {
         context = new Context(cconfig);
         contexts.put(contextName, context);
@@ -167,20 +167,20 @@ public class ContextManager {
         contexts.put(contextName, context);
       }
     }
-    
+
     if (contextToClose != null)
       contextToClose.close();
-    
+
     ClassLoader loader = context.getClassLoader();
     if (loader == null) {
       // oops, context was closed by another thread, try again
       return getClassLoader(contextName);
     }
-    
+
     return loader;
-    
+
   }
-  
+
   public <U> Class<? extends U> loadClass(String context, String classname, Class<U> extension) throws ClassNotFoundException {
     try {
       return getClassLoader(context).loadClass(classname).asSubclass(extension);
@@ -188,17 +188,17 @@ public class ContextManager {
       throw new ClassNotFoundException("IO Error loading class " + classname, e);
     }
   }
-  
+
   public void removeUnusedContexts(Set<String> inUse) {
-    
+
     Map<String,Context> unused;
-    
+
     synchronized (this) {
       unused = new HashMap<String,Context>(contexts);
       unused.keySet().removeAll(inUse);
       contexts.keySet().removeAll(unused.keySet());
     }
-    
+
     for (Context context : unused.values()) {
       // close outside of lock
       context.close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/MiniDFSUtil.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/MiniDFSUtil.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/MiniDFSUtil.java
index c822f64..a8e4d60 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/MiniDFSUtil.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/MiniDFSUtil.java
@@ -20,7 +20,7 @@ import java.io.BufferedReader;
 import java.io.InputStreamReader;
 
 public class MiniDFSUtil {
-  
+
   public static String computeDatanodeDirectoryPermission() {
     // MiniDFSCluster will check the permissions on the data directories, but does not
     // do a good job of setting them properly. We need to get the users umask and set
@@ -32,12 +32,12 @@ public class MiniDFSUtil {
       try {
         String line = bri.readLine();
         p.waitFor();
-        
+
         Short umask = Short.parseShort(line.trim(), 8);
         // Need to set permission to 777 xor umask
         // leading zero makes java interpret as base 8
         int newPermission = 0777 ^ umask;
-        
+
         return String.format("%03o", newPermission);
       } finally {
         bri.close();
@@ -46,5 +46,5 @@ public class MiniDFSUtil {
       throw new RuntimeException("Error getting umask from O/S", e);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/PostDelegatingVFSClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/PostDelegatingVFSClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/PostDelegatingVFSClassLoader.java
index 745401b..307ee73 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/PostDelegatingVFSClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/PostDelegatingVFSClassLoader.java
@@ -22,14 +22,14 @@ import org.apache.commons.vfs2.FileSystemManager;
 import org.apache.commons.vfs2.impl.VFSClassLoader;
 
 /**
- * 
+ *
  */
 public class PostDelegatingVFSClassLoader extends VFSClassLoader {
-  
+
   public PostDelegatingVFSClassLoader(FileObject[] files, FileSystemManager manager, ClassLoader parent) throws FileSystemException {
     super(files, manager, parent);
   }
-  
+
   @Override
   protected synchronized Class<?> loadClass(String name, boolean resolve) throws ClassNotFoundException {
     Class<?> c = findLoadedClass(name);
@@ -39,7 +39,7 @@ public class PostDelegatingVFSClassLoader extends VFSClassLoader {
       // try finding this class here instead of parent
       return findClass(name);
     } catch (ClassNotFoundException e) {
-      
+
     }
     return super.loadClass(name, resolve);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ReloadingClassLoader.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ReloadingClassLoader.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ReloadingClassLoader.java
index 9febc45..9de0422 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ReloadingClassLoader.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/ReloadingClassLoader.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.start.classloader.vfs;
 
 /**
- * 
+ *
  */
 public interface ReloadingClassLoader {
   ClassLoader getClassLoader();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/UniqueFileReplicator.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/UniqueFileReplicator.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/UniqueFileReplicator.java
index e11cf47..ff6e87c 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/UniqueFileReplicator.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/UniqueFileReplicator.java
@@ -33,7 +33,7 @@ import org.apache.commons.vfs2.provider.VfsComponentContext;
 import org.apache.log4j.Logger;
 
 /**
- * 
+ *
  */
 public class UniqueFileReplicator implements VfsComponent, FileReplicator {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileAttributes.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileAttributes.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileAttributes.java
index c088956..bfdd561 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileAttributes.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileAttributes.java
@@ -18,43 +18,42 @@ package org.apache.accumulo.start.classloader.vfs.providers;
 
 /**
  * HDFS file content attributes.
- * 
+ *
  * @since 2.1
  */
-public enum HdfsFileAttributes
-{
-    /**
-     * Last access time.
-     */
-    LAST_ACCESS_TIME,
+public enum HdfsFileAttributes {
+  /**
+   * Last access time.
+   */
+  LAST_ACCESS_TIME,
 
-    /**
-     * Block size.
-     */
-    BLOCK_SIZE,
+  /**
+   * Block size.
+   */
+  BLOCK_SIZE,
 
-    /**
-     * Group.
-     */
-    GROUP,
+  /**
+   * Group.
+   */
+  GROUP,
 
-    /**
-     * Owner.
-     */
-    OWNER,
+  /**
+   * Owner.
+   */
+  OWNER,
 
-    /**
-     * Permissions.
-     */
-    PERMISSIONS,
+  /**
+   * Permissions.
+   */
+  PERMISSIONS,
 
-    /**
-     * Length.
-     */
-    LENGTH,
+  /**
+   * Length.
+   */
+  LENGTH,
 
-    /**
-     * Modification time.
-     */
-    MODIFICATION_TIME;
+  /**
+   * Modification time.
+   */
+  MODIFICATION_TIME;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileContentInfoFactory.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileContentInfoFactory.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileContentInfoFactory.java
index c3f3a76..2621255 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileContentInfoFactory.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileContentInfoFactory.java
@@ -24,27 +24,25 @@ import org.apache.commons.vfs2.impl.DefaultFileContentInfo;
 
 /**
  * Creates FileContentInfo instances for HDFS.
- * 
+ *
  * @since 2.1
  */
-public class HdfsFileContentInfoFactory implements FileContentInfoFactory
-{
-    private static final String CONTENT = "text/plain";
-    private static final String ENCODING = "UTF-8";
+public class HdfsFileContentInfoFactory implements FileContentInfoFactory {
+  private static final String CONTENT = "text/plain";
+  private static final String ENCODING = "UTF-8";
 
-    /**
-     * Creates a FileContentInfo for a the given FileContent.
-     * 
-     * @param fileContent
-     *            Use this FileContent to create a matching FileContentInfo
-     * @return a FileContentInfo for the given FileContent with content set to "text/plain" and encoding set to "UTF-8"
-     * @throws FileSystemException
-     *             when a problem occurs creating the FileContentInfo.
-     */
-    @Override
-    public FileContentInfo create(final FileContent fileContent) throws FileSystemException
-    {
-        return new DefaultFileContentInfo(CONTENT, ENCODING);
-    }
+  /**
+   * Creates a FileContentInfo for a the given FileContent.
+   *
+   * @param fileContent
+   *          Use this FileContent to create a matching FileContentInfo
+   * @return a FileContentInfo for the given FileContent with content set to "text/plain" and encoding set to "UTF-8"
+   * @throws FileSystemException
+   *           when a problem occurs creating the FileContentInfo.
+   */
+  @Override
+  public FileContentInfo create(final FileContent fileContent) throws FileSystemException {
+    return new DefaultFileContentInfo(CONTENT, ENCODING);
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileObject.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileObject.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileObject.java
index 2cb0d5f..6849073 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileObject.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileObject.java
@@ -35,322 +35,269 @@ import org.apache.hadoop.fs.Path;
 
 /**
  * A VFS representation of an HDFS file.
- * 
+ *
  * @since 2.1
  */
-public class HdfsFileObject extends AbstractFileObject
-{
-    private final HdfsFileSystem fs;
-    private final FileSystem hdfs;
-    private final Path path;
-    private FileStatus stat;
+public class HdfsFileObject extends AbstractFileObject {
+  private final HdfsFileSystem fs;
+  private final FileSystem hdfs;
+  private final Path path;
+  private FileStatus stat;
 
-    /**
-     * Constructs a new HDFS FileObject
-     * 
-     * @param name
-     *            FileName
-     * @param fs
-     *            HdfsFileSystem instance
-     * @param hdfs
-     *            Hadoop FileSystem instance
-     * @param p
-     *            Path to the file in HDFS
-     */
-    protected HdfsFileObject(final AbstractFileName name, final HdfsFileSystem fs, final FileSystem hdfs, final Path p)
-    {
-        super(name, fs);
-        this.fs = fs;
-        this.hdfs = hdfs;
-        this.path = p;
-    }
+  /**
+   * Constructs a new HDFS FileObject
+   *
+   * @param name
+   *          FileName
+   * @param fs
+   *          HdfsFileSystem instance
+   * @param hdfs
+   *          Hadoop FileSystem instance
+   * @param p
+   *          Path to the file in HDFS
+   */
+  protected HdfsFileObject(final AbstractFileName name, final HdfsFileSystem fs, final FileSystem hdfs, final Path p) {
+    super(name, fs);
+    this.fs = fs;
+    this.hdfs = hdfs;
+    this.path = p;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#canRenameTo(org.apache.commons.vfs2.FileObject)
-     */
-    @Override
-    public boolean canRenameTo(final FileObject newfile)
-    {
-        throw new UnsupportedOperationException();
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#canRenameTo(org.apache.commons.vfs2.FileObject)
+   */
+  @Override
+  public boolean canRenameTo(final FileObject newfile) {
+    throw new UnsupportedOperationException();
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doAttach()
-     */
-    @Override
-    protected void doAttach() throws Exception
-    {
-        try
-        {
-            this.stat = this.hdfs.getFileStatus(this.path);
-        }
-        catch (final FileNotFoundException e)
-        {
-            this.stat = null;
-            return;
-        }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doAttach()
+   */
+  @Override
+  protected void doAttach() throws Exception {
+    try {
+      this.stat = this.hdfs.getFileStatus(this.path);
+    } catch (final FileNotFoundException e) {
+      this.stat = null;
+      return;
     }
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetAttributes()
-     */
-    @Override
-    protected Map<String, Object> doGetAttributes() throws Exception
-    {
-        if (null == this.stat)
-        {
-            return super.doGetAttributes();
-        }
-        else
-        {
-            final Map<String, Object> attrs = new HashMap<String, Object>();
-            attrs.put(HdfsFileAttributes.LAST_ACCESS_TIME.toString(), this.stat.getAccessTime());
-            attrs.put(HdfsFileAttributes.BLOCK_SIZE.toString(), this.stat.getBlockSize());
-            attrs.put(HdfsFileAttributes.GROUP.toString(), this.stat.getGroup());
-            attrs.put(HdfsFileAttributes.OWNER.toString(), this.stat.getOwner());
-            attrs.put(HdfsFileAttributes.PERMISSIONS.toString(), this.stat.getPermission().toString());
-            attrs.put(HdfsFileAttributes.LENGTH.toString(), this.stat.getLen());
-            attrs.put(HdfsFileAttributes.MODIFICATION_TIME.toString(), this.stat.getModificationTime());
-            return attrs;
-        }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetAttributes()
+   */
+  @Override
+  protected Map<String,Object> doGetAttributes() throws Exception {
+    if (null == this.stat) {
+      return super.doGetAttributes();
+    } else {
+      final Map<String,Object> attrs = new HashMap<String,Object>();
+      attrs.put(HdfsFileAttributes.LAST_ACCESS_TIME.toString(), this.stat.getAccessTime());
+      attrs.put(HdfsFileAttributes.BLOCK_SIZE.toString(), this.stat.getBlockSize());
+      attrs.put(HdfsFileAttributes.GROUP.toString(), this.stat.getGroup());
+      attrs.put(HdfsFileAttributes.OWNER.toString(), this.stat.getOwner());
+      attrs.put(HdfsFileAttributes.PERMISSIONS.toString(), this.stat.getPermission().toString());
+      attrs.put(HdfsFileAttributes.LENGTH.toString(), this.stat.getLen());
+      attrs.put(HdfsFileAttributes.MODIFICATION_TIME.toString(), this.stat.getModificationTime());
+      return attrs;
     }
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetContentSize()
-     */
-    @Override
-    protected long doGetContentSize() throws Exception
-    {
-        return stat.getLen();
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetContentSize()
+   */
+  @Override
+  protected long doGetContentSize() throws Exception {
+    return stat.getLen();
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetInputStream()
-     */
-    @Override
-    protected InputStream doGetInputStream() throws Exception
-    {
-        return this.hdfs.open(this.path);
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetInputStream()
+   */
+  @Override
+  protected InputStream doGetInputStream() throws Exception {
+    return this.hdfs.open(this.path);
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetLastModifiedTime()
-     */
-    @Override
-    protected long doGetLastModifiedTime() throws Exception
-    {
-        if (null != this.stat)
-        {
-            return this.stat.getModificationTime();
-        }
-        else
-        {
-            return -1;
-        }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetLastModifiedTime()
+   */
+  @Override
+  protected long doGetLastModifiedTime() throws Exception {
+    if (null != this.stat) {
+      return this.stat.getModificationTime();
+    } else {
+      return -1;
     }
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetRandomAccessContent
-     *      (org.apache.commons.vfs2.util.RandomAccessMode)
-     */
-    @Override
-    protected RandomAccessContent doGetRandomAccessContent(final RandomAccessMode mode) throws Exception
-    {
-        if (mode.equals(RandomAccessMode.READWRITE))
-        {
-            throw new UnsupportedOperationException();
-        }
-        return new HdfsRandomAccessContent(this.path, this.hdfs);
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetRandomAccessContent (org.apache.commons.vfs2.util.RandomAccessMode)
+   */
+  @Override
+  protected RandomAccessContent doGetRandomAccessContent(final RandomAccessMode mode) throws Exception {
+    if (mode.equals(RandomAccessMode.READWRITE)) {
+      throw new UnsupportedOperationException();
     }
+    return new HdfsRandomAccessContent(this.path, this.hdfs);
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetType()
-     */
-    @Override  
-    //TODO Remove deprecation warning suppression when Hadoop1 support is dropped
-    @SuppressWarnings("deprecation")
-    protected FileType doGetType() throws Exception
-    {
-        try
-        {
-            doAttach();
-            if (null == stat)
-            {
-                return FileType.IMAGINARY;
-            }
-            if (stat.isDir())
-            {
-                return FileType.FOLDER;
-            }
-            else
-            {
-                return FileType.FILE;
-            }
-        }
-        catch (final FileNotFoundException fnfe)
-        {
-            return FileType.IMAGINARY;
-        }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doGetType()
+   */
+  @Override
+  // TODO Remove deprecation warning suppression when Hadoop1 support is dropped
+  @SuppressWarnings("deprecation")
+  protected FileType doGetType() throws Exception {
+    try {
+      doAttach();
+      if (null == stat) {
+        return FileType.IMAGINARY;
+      }
+      if (stat.isDir()) {
+        return FileType.FOLDER;
+      } else {
+        return FileType.FILE;
+      }
+    } catch (final FileNotFoundException fnfe) {
+      return FileType.IMAGINARY;
     }
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsHidden()
-     */
-    @Override
-    protected boolean doIsHidden() throws Exception
-    {
-        return false;
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsHidden()
+   */
+  @Override
+  protected boolean doIsHidden() throws Exception {
+    return false;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsReadable()
-     */
-    @Override
-    protected boolean doIsReadable() throws Exception
-    {
-        return true;
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsReadable()
+   */
+  @Override
+  protected boolean doIsReadable() throws Exception {
+    return true;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsSameFile(org.apache.commons.vfs2.FileObject)
-     */
-    @Override
-    protected boolean doIsSameFile(final FileObject destFile) throws FileSystemException
-    {
-        throw new UnsupportedOperationException();
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsSameFile(org.apache.commons.vfs2.FileObject)
+   */
+  @Override
+  protected boolean doIsSameFile(final FileObject destFile) throws FileSystemException {
+    throw new UnsupportedOperationException();
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsWriteable()
-     */
-    @Override
-    protected boolean doIsWriteable() throws Exception
-    {
-        return false;
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doIsWriteable()
+   */
+  @Override
+  protected boolean doIsWriteable() throws Exception {
+    return false;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doListChildren()
-     */
-    @Override
-    protected String[] doListChildren() throws Exception
-    {
-        if (this.doGetType() != FileType.FOLDER)
-        {
-            throw new FileNotFolderException(this);
-        }
-
-        final FileStatus[] files = this.hdfs.listStatus(this.path);
-        final String[] children = new String[files.length];
-        int i = 0;
-        for (final FileStatus status : files)
-        {
-            children[i++] = status.getPath().getName();
-        }
-        return children;
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doListChildren()
+   */
+  @Override
+  protected String[] doListChildren() throws Exception {
+    if (this.doGetType() != FileType.FOLDER) {
+      throw new FileNotFolderException(this);
     }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doListChildrenResolved()
-     */
-    @Override
-    protected FileObject[] doListChildrenResolved() throws Exception
-    {
-        if (this.doGetType() != FileType.FOLDER)
-        {
-            return null;
-        }
-        final String[] children = doListChildren();
-        final FileObject[] fo = new FileObject[children.length];
-        for (int i = 0; i < children.length; i++)
-        {
-            final Path p = new Path(this.path, children[i]);
-            fo[i] = this.fs.resolveFile(p.toUri().toString());
-        }
-        return fo;
+    final FileStatus[] files = this.hdfs.listStatus(this.path);
+    final String[] children = new String[files.length];
+    int i = 0;
+    for (final FileStatus status : files) {
+      children[i++] = status.getPath().getName();
     }
+    return children;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doRemoveAttribute(java.lang.String)
-     */
-    @Override
-    protected void doRemoveAttribute(final String attrName) throws Exception
-    {
-        throw new UnsupportedOperationException();
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doListChildrenResolved()
+   */
+  @Override
+  protected FileObject[] doListChildrenResolved() throws Exception {
+    if (this.doGetType() != FileType.FOLDER) {
+      return null;
     }
-
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doSetAttribute(java.lang.String, java.lang.Object)
-     */
-    @Override
-    protected void doSetAttribute(final String attrName, final Object value) throws Exception
-    {
-        throw new UnsupportedOperationException();
+    final String[] children = doListChildren();
+    final FileObject[] fo = new FileObject[children.length];
+    for (int i = 0; i < children.length; i++) {
+      final Path p = new Path(this.path, children[i]);
+      fo[i] = this.fs.resolveFile(p.toUri().toString());
     }
+    return fo;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#doSetLastModifiedTime(long)
-     */
-    @Override
-    protected boolean doSetLastModifiedTime(final long modtime) throws Exception
-    {
-        throw new UnsupportedOperationException();
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doRemoveAttribute(java.lang.String)
+   */
+  @Override
+  protected void doRemoveAttribute(final String attrName) throws Exception {
+    throw new UnsupportedOperationException();
+  }
 
-    /**
-     * @see java.lang.Object#equals(java.lang.Object)
-     */
-    @Override
-    public boolean equals(final Object o)
-    {
-        if (null == o)
-        {
-            return false;
-        }
-        if (o == this)
-        {
-            return true;
-        }
-        if (o instanceof HdfsFileObject)
-        {
-            final HdfsFileObject other = (HdfsFileObject) o;
-            if (other.path.equals(this.path))
-            {
-                return true;
-            }
-        }
-        return false;
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doSetAttribute(java.lang.String, java.lang.Object)
+   */
+  @Override
+  protected void doSetAttribute(final String attrName, final Object value) throws Exception {
+    throw new UnsupportedOperationException();
+  }
+
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#doSetLastModifiedTime(long)
+   */
+  @Override
+  protected boolean doSetLastModifiedTime(final long modtime) throws Exception {
+    throw new UnsupportedOperationException();
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileObject#exists()
-     * @return boolean true if file exists, false if not
-     */
-    @Override
-    public boolean exists() throws FileSystemException
-    {
-        try
-        {
-            doAttach();
-            return this.stat != null;
-        }
-        catch (final FileNotFoundException fne)
-        {
-            return false;
-        }
-        catch (final Exception e)
-        {
-            throw new FileSystemException("Unable to check existance ", e);
-        }
+  /**
+   * @see java.lang.Object#equals(java.lang.Object)
+   */
+  @Override
+  public boolean equals(final Object o) {
+    if (null == o) {
+      return false;
     }
+    if (o == this) {
+      return true;
+    }
+    if (o instanceof HdfsFileObject) {
+      final HdfsFileObject other = (HdfsFileObject) o;
+      if (other.path.equals(this.path)) {
+        return true;
+      }
+    }
+    return false;
+  }
 
-    /**
-     * @see java.lang.Object#hashCode()
-     */
-    @Override
-    public int hashCode()
-    {
-        return this.path.getName().toString().hashCode();
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileObject#exists()
+   * @return boolean true if file exists, false if not
+   */
+  @Override
+  public boolean exists() throws FileSystemException {
+    try {
+      doAttach();
+      return this.stat != null;
+    } catch (final FileNotFoundException fne) {
+      return false;
+    } catch (final Exception e) {
+      throw new FileSystemException("Unable to check existance ", e);
     }
+  }
+
+  /**
+   * @see java.lang.Object#hashCode()
+   */
+  @Override
+  public int hashCode() {
+    return this.path.getName().toString().hashCode();
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileProvider.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileProvider.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileProvider.java
index 5bea2b9..9ddfab5 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileProvider.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileProvider.java
@@ -31,58 +31,45 @@ import org.apache.commons.vfs2.provider.http.HttpFileNameParser;
 
 /**
  * FileProvider for HDFS files.
- * 
+ *
  * @since 2.1
  */
-public class HdfsFileProvider extends AbstractOriginatingFileProvider
-{
-    protected static final Collection<Capability> CAPABILITIES = Collections.unmodifiableCollection(Arrays
-            .asList(new Capability[]
-            {
-                    Capability.GET_TYPE,
-                    Capability.READ_CONTENT,
-                    Capability.URI,
-                    Capability.GET_LAST_MODIFIED,
-                    Capability.ATTRIBUTES,
-                    Capability.RANDOM_ACCESS_READ,
-                    Capability.DIRECTORY_READ_CONTENT,
-                    Capability.LIST_CHILDREN }));
+public class HdfsFileProvider extends AbstractOriginatingFileProvider {
+  protected static final Collection<Capability> CAPABILITIES = Collections.unmodifiableCollection(Arrays.asList(new Capability[] {Capability.GET_TYPE,
+      Capability.READ_CONTENT, Capability.URI, Capability.GET_LAST_MODIFIED, Capability.ATTRIBUTES, Capability.RANDOM_ACCESS_READ,
+      Capability.DIRECTORY_READ_CONTENT, Capability.LIST_CHILDREN}));
 
-    /**
-     * Constructs a new HdfsFileProvider
-     */
-    public HdfsFileProvider()
-    {
-        super();
-        this.setFileNameParser(HttpFileNameParser.getInstance());
-    }
+  /**
+   * Constructs a new HdfsFileProvider
+   */
+  public HdfsFileProvider() {
+    super();
+    this.setFileNameParser(HttpFileNameParser.getInstance());
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider#doCreateFileSystem(org.apache.commons.vfs2.FileName, org.apache.commons.vfs2.FileSystemOptions)
-     */
-    @Override
-    protected FileSystem doCreateFileSystem(final FileName rootName, final FileSystemOptions fileSystemOptions)
-            throws FileSystemException
-    {
-        return new HdfsFileSystem(rootName, fileSystemOptions);
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider#doCreateFileSystem(org.apache.commons.vfs2.FileName,
+   *      org.apache.commons.vfs2.FileSystemOptions)
+   */
+  @Override
+  protected FileSystem doCreateFileSystem(final FileName rootName, final FileSystemOptions fileSystemOptions) throws FileSystemException {
+    return new HdfsFileSystem(rootName, fileSystemOptions);
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.FileProvider#getCapabilities()
-     */
-    @Override
-    public Collection<Capability> getCapabilities()
-    {
-        return CAPABILITIES;
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.FileProvider#getCapabilities()
+   */
+  @Override
+  public Collection<Capability> getCapabilities() {
+    return CAPABILITIES;
+  }
 
-    /**
-     * @see org.apache.commons.vfs2.provider.AbstractFileProvider#getConfigBuilder()
-     */
-    @Override
-    public FileSystemConfigBuilder getConfigBuilder()
-    {
-        return HdfsFileSystemConfigBuilder.getInstance();
-    }
+  /**
+   * @see org.apache.commons.vfs2.provider.AbstractFileProvider#getConfigBuilder()
+   */
+  @Override
+  public FileSystemConfigBuilder getConfigBuilder() {
+    return HdfsFileSystemConfigBuilder.getInstance();
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystem.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystem.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystem.java
index 2973750..7ad68f9 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystem.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystem.java
@@ -37,123 +37,100 @@ import org.apache.hadoop.fs.Path;
 
 /**
  * A VFS FileSystem that interacts with HDFS.
- * 
+ *
  * @since 2.1
  */
-public class HdfsFileSystem extends AbstractFileSystem
-{
+public class HdfsFileSystem extends AbstractFileSystem {
   private static final Log log = LogFactory.getLog(HdfsFileSystem.class);
-  
+
   private FileSystem fs;
-  
-  protected HdfsFileSystem(final FileName rootName, final FileSystemOptions fileSystemOptions)
-  {
+
+  protected HdfsFileSystem(final FileName rootName, final FileSystemOptions fileSystemOptions) {
     super(rootName, null, fileSystemOptions);
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.provider.AbstractFileSystem#addCapabilities(java.util.Collection)
    */
   @Override
-  protected void addCapabilities(final Collection<Capability> capabilities)
-  {
+  protected void addCapabilities(final Collection<Capability> capabilities) {
     capabilities.addAll(HdfsFileProvider.CAPABILITIES);
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.provider.AbstractFileSystem#close()
    */
   @Override
-  synchronized public void close()
-  {
-    try
-    {
-      if (null != fs)
-      {
+  synchronized public void close() {
+    try {
+      if (null != fs) {
         fs.close();
       }
-    }
-    catch (final IOException e)
-    {
+    } catch (final IOException e) {
       throw new RuntimeException("Error closing HDFS client", e);
     }
     super.close();
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.provider.AbstractFileSystem#createFile(org.apache.commons.vfs2.provider.AbstractFileName)
    */
   @Override
-  protected FileObject createFile(final AbstractFileName name) throws Exception
-  {
+  protected FileObject createFile(final AbstractFileName name) throws Exception {
     throw new FileSystemException("Operation not supported");
   }
-  
+
   /**
    * @see org.apache.commons.vfs2.provider.AbstractFileSystem#resolveFile(org.apache.commons.vfs2.FileName)
    */
   @Override
-  public FileObject resolveFile(final FileName name) throws FileSystemException
-  {
-    
-    synchronized (this)
-    {
-      if (null == this.fs)
-      {
+  public FileObject resolveFile(final FileName name) throws FileSystemException {
+
+    synchronized (this) {
+      if (null == this.fs) {
         final String hdfsUri = name.getRootURI();
         final Configuration conf = new Configuration(true);
         conf.set(org.apache.hadoop.fs.FileSystem.FS_DEFAULT_NAME_KEY, hdfsUri);
         this.fs = null;
-        try
-        {
+        try {
           fs = org.apache.hadoop.fs.FileSystem.get(conf);
-        }
-        catch (final IOException e)
-        {
+        } catch (final IOException e) {
           log.error("Error connecting to filesystem " + hdfsUri, e);
           throw new FileSystemException("Error connecting to filesystem " + hdfsUri, e);
         }
       }
     }
-    
+
     boolean useCache = (null != getContext().getFileSystemManager().getFilesCache());
     FileObject file;
-    if (useCache)
-    {
+    if (useCache) {
       file = this.getFileFromCache(name);
-    }
-    else
-    {
+    } else {
       file = null;
     }
-    if (null == file)
-    {
+    if (null == file) {
       String path = null;
-      try
-      {
+      try {
         path = URLDecoder.decode(name.getPath(), "UTF-8");
-      }
-      catch (final UnsupportedEncodingException e)
-      {
+      } catch (final UnsupportedEncodingException e) {
         path = name.getPath();
       }
       final Path filePath = new Path(path);
       file = new HdfsFileObject((AbstractFileName) name, this, fs, filePath);
-      if (useCache)
-      {
+      if (useCache) {
         this.putFileToCache(file);
       }
-      
+
     }
-    
+
     /**
      * resync the file information if requested
      */
     if (getFileSystemManager().getCacheStrategy().equals(CacheStrategy.ON_RESOLVE)) {
       file.refresh();
     }
-    
+
     return file;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystemConfigBuilder.java
----------------------------------------------------------------------
diff --git a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystemConfigBuilder.java b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystemConfigBuilder.java
index 6aa6d96..0defa56 100644
--- a/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystemConfigBuilder.java
+++ b/start/src/main/java/org/apache/accumulo/start/classloader/vfs/providers/HdfsFileSystemConfigBuilder.java
@@ -21,28 +21,25 @@ import org.apache.commons.vfs2.FileSystemConfigBuilder;
 
 /**
  * Configuration settings for the HdfsFileSystem.
- * 
+ *
  * @since 2.1
  */
-public class HdfsFileSystemConfigBuilder extends FileSystemConfigBuilder
-{
-    private static final HdfsFileSystemConfigBuilder BUILDER = new HdfsFileSystemConfigBuilder();
+public class HdfsFileSystemConfigBuilder extends FileSystemConfigBuilder {
+  private static final HdfsFileSystemConfigBuilder BUILDER = new HdfsFileSystemConfigBuilder();
 
-    /**
-     * @return HdfsFileSystemConfigBuilder instance
-     */
-    public static HdfsFileSystemConfigBuilder getInstance()
-    {
-        return BUILDER;
-    }
+  /**
+   * @return HdfsFileSystemConfigBuilder instance
+   */
+  public static HdfsFileSystemConfigBuilder getInstance() {
+    return BUILDER;
+  }
 
-    /**
-     * @return HDFSFileSystem
-     */
-    @Override
-    protected Class<? extends FileSystem> getConfigClass()
-    {
-        return HdfsFileSystem.class;
-    }
+  /**
+   * @return HDFSFileSystem
+   */
+  @Override
+  protected Class<? extends FileSystem> getConfigClass() {
+    return HdfsFileSystem.class;
+  }
 
 }


[55/66] [abbrv] accumulo git commit: ACCUMULO-3451 Add minimal checkstyle enforcement

Posted by ct...@apache.org.
ACCUMULO-3451 Add minimal checkstyle enforcement


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/d2c116ff
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/d2c116ff
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/d2c116ff

Branch: refs/heads/1.6
Commit: d2c116ffae59672ff995033e81be73260c4d5c25
Parents: c2155f4
Author: Christopher Tubbs <ct...@apache.org>
Authored: Tue Dec 23 18:51:04 2014 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:23:36 2015 -0500

----------------------------------------------------------------------
 pom.xml | 108 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 108 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/d2c116ff/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 8316a33..711a21d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -367,6 +367,11 @@
     <pluginManagement>
       <plugins>
         <plugin>
+          <groupId>org.apache.maven.plugins</groupId>
+          <artifactId>maven-checkstyle-plugin</artifactId>
+          <version>2.13</version>
+        </plugin>
+        <plugin>
           <groupId>com.google.code.sortpom</groupId>
           <artifactId>maven-sortpom-plugin</artifactId>
           <version>2.1.0</version>
@@ -566,6 +571,19 @@
                 <pluginExecution>
                   <pluginExecutionFilter>
                     <groupId>org.apache.maven.plugins</groupId>
+                    <artifactId>maven-checkstyle-plugin</artifactId>
+                    <versionRange>[2.13,)</versionRange>
+                    <goals>
+                      <goal>check</goal>
+                    </goals>
+                  </pluginExecutionFilter>
+                  <action>
+                    <ignore />
+                  </action>
+                </pluginExecution>
+                <pluginExecution>
+                  <pluginExecutionFilter>
+                    <groupId>org.apache.maven.plugins</groupId>
                     <artifactId>maven-dependency-plugin</artifactId>
                     <versionRange>[2.0,)</versionRange>
                     <goals>
@@ -709,6 +727,96 @@
         </executions>
       </plugin>
       <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-checkstyle-plugin</artifactId>
+        <configuration>
+          <checkstyleRules>
+            <module name="Checker">
+              <property name="charset" value="UTF-8" />
+              <property name="severity" value="warning" />
+              <!-- Checks for whitespace                               -->
+              <!-- See http://checkstyle.sf.net/config_whitespace.html -->
+              <module name="FileTabCharacter">
+                <property name="eachLine" value="true" />
+              </module>
+              <module name="TreeWalker">
+                <module name="OuterTypeFilename" />
+                <module name="LineLength">
+                  <!-- needs extra, because Eclipse formatter ignores the ending left brace -->
+                  <property name="max" value="200"/>
+                  <property name="ignorePattern" value="^package.*|^import.*|a href|href|http://|https://|ftp://"/>
+                </module>
+                <module name="AvoidStarImport" />
+                <module name="NoLineWrap" />
+                <module name="LeftCurly">
+                  <property name="maxLineLength" value="160" />
+                </module>
+                <module name="RightCurly" />
+                <module name="RightCurly">
+                  <property name="option" value="alone" />
+                  <property name="tokens" value="CLASS_DEF, METHOD_DEF, CTOR_DEF, LITERAL_FOR, LITERAL_WHILE, LITERAL_DO, STATIC_INIT, INSTANCE_INIT" />
+                </module>
+                <module name="SeparatorWrap">
+                  <property name="tokens" value="DOT" />
+                  <property name="option" value="nl" />
+                </module>
+                <module name="SeparatorWrap">
+                  <property name="tokens" value="COMMA" />
+                  <property name="option" value="EOL" />
+                </module>
+                <module name="PackageName">
+                  <property name="format" value="^[a-z]+(\.[a-z][a-zA-Z0-9]*)*$" />
+                </module>
+                <module name="MethodTypeParameterName">
+                  <property name="format" value="(^[A-Z][0-9]?)$|([A-Z][a-zA-Z0-9]*[T]$)" />
+                </module>
+                <module name="MethodParamPad" />
+                <module name="OperatorWrap">
+                  <property name="option" value="NL" />
+                  <property name="tokens" value="BAND, BOR, BSR, BXOR, DIV, EQUAL, GE, GT, LAND, LE, LITERAL_INSTANCEOF, LOR, LT, MINUS, MOD, NOT_EQUAL, QUESTION, SL, SR, STAR " />
+                </module>
+                <module name="AnnotationLocation">
+                  <property name="tokens" value="CLASS_DEF, INTERFACE_DEF, ENUM_DEF, METHOD_DEF, CTOR_DEF" />
+                </module>
+                <module name="AnnotationLocation">
+                  <property name="tokens" value="VARIABLE_DEF" />
+                  <property name="allowSamelineMultipleAnnotations" value="true" />
+                </module>
+                <module name="NonEmptyAtclauseDescription" />
+                <module name="JavadocTagContinuationIndentation" />
+                <module name="JavadocMethod">
+                  <property name="allowMissingJavadoc" value="true" />
+                  <property name="allowMissingParamTags" value="true" />
+                  <property name="allowMissingThrowsTags" value="true" />
+                  <property name="allowMissingReturnTag" value="true" />
+                  <property name="allowedAnnotations" value="Override,Test,BeforeClass,AfterClass,Before,After" />
+                  <property name="allowThrowsTagsForSubclasses" value="true" />
+                </module>
+                <module name="SingleLineJavadoc" />
+              </module>
+            </module>
+          </checkstyleRules>
+          <violationSeverity>warning</violationSeverity>
+          <includeTestSourceDirectory>true</includeTestSourceDirectory>
+          <excludes>**/thrift/*.java</excludes>
+        </configuration>
+        <dependencies>
+          <dependency>
+            <groupId>com.puppycrawl.tools</groupId>
+            <artifactId>checkstyle</artifactId>
+            <version>6.1.1</version>
+          </dependency>
+        </dependencies>
+        <executions>
+          <execution>
+            <id>check-style</id>
+            <goals>
+              <goal>check</goal>
+            </goals>
+          </execution>
+        </executions>
+      </plugin>
+      <plugin>
         <groupId>com.github.koraktor</groupId>
         <artifactId>mavanagaiata</artifactId>
         <executions>


[61/66] [abbrv] accumulo git commit: Merge branch '1.6'

Posted by ct...@apache.org.
Merge branch '1.6'


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/83a96230
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/83a96230
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/83a96230

Branch: refs/heads/master
Commit: 83a96230da0e3fb931512db15b61b42cea4d1627
Parents: b693b6d 627e525
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:36:18 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:36:18 2015 -0500

----------------------------------------------------------------------
 pom.xml | 108 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 108 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/83a96230/pom.xml
----------------------------------------------------------------------


[42/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/security/tokens/NullToken.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/NullToken.java b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/NullToken.java
index 474b4a1..530b41f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/NullToken.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/NullToken.java
@@ -28,42 +28,39 @@ import javax.security.auth.DestroyFailedException;
  * @since 1.5.0
  */
 public class NullToken implements AuthenticationToken {
-  
+
   @Override
-  public void readFields(DataInput arg0) throws IOException {
-  }
-  
+  public void readFields(DataInput arg0) throws IOException {}
+
   @Override
-  public void write(DataOutput arg0) throws IOException {
-  }
-  
+  public void write(DataOutput arg0) throws IOException {}
+
   @Override
-  public void destroy() throws DestroyFailedException {
-  }
-  
+  public void destroy() throws DestroyFailedException {}
+
   @Override
   public boolean isDestroyed() {
     return false;
   }
-  
+
   @Override
   public NullToken clone() {
     return new NullToken();
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     return obj instanceof NullToken;
   }
-  
+
   @Override
   public void init(Properties properties) {}
-  
+
   @Override
   public Set<TokenProperty> getProperties() {
     return Collections.emptySet();
   }
-  
+
   @Override
   public int hashCode() {
     return 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/security/tokens/PasswordToken.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/PasswordToken.java b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/PasswordToken.java
index 53a0bcf..eb1abe6 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/PasswordToken.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/PasswordToken.java
@@ -39,28 +39,28 @@ import org.apache.hadoop.io.WritableUtils;
 
 public class PasswordToken implements AuthenticationToken {
   private byte[] password = null;
-  
+
   public byte[] getPassword() {
     return Arrays.copyOf(password, password.length);
   }
-  
+
   /**
    * Constructor for use with {@link Writable}. Call {@link #readFields(DataInput)}.
    */
   public PasswordToken() {
     password = new byte[0];
   }
-  
+
   /**
    * Constructs a token from a copy of the password. Destroying the argument after construction will not destroy the copy in this token, and destroying this
    * token will only destroy the copy held inside this token, not the argument.
-   * 
+   *
    * Password tokens created with this constructor will store the password as UTF-8 bytes.
    */
   public PasswordToken(CharSequence password) {
     setPassword(CharBuffer.wrap(password));
   }
-  
+
   /**
    * Constructs a token from a copy of the password. Destroying the argument after construction will not destroy the copy in this token, and destroying this
    * token will only destroy the copy held inside this token, not the argument.
@@ -68,7 +68,7 @@ public class PasswordToken implements AuthenticationToken {
   public PasswordToken(byte[] password) {
     this.password = Arrays.copyOf(password, password.length);
   }
-  
+
   /**
    * Constructs a token from a copy of the password. Destroying the argument after construction will not destroy the copy in this token, and destroying this
    * token will only destroy the copy held inside this token, not the argument.
@@ -76,33 +76,33 @@ public class PasswordToken implements AuthenticationToken {
   public PasswordToken(ByteBuffer password) {
     this.password = ByteBufferUtil.toBytes(password);
   }
-  
+
   @Override
   public void readFields(DataInput arg0) throws IOException {
     password = WritableUtils.readCompressedByteArray(arg0);
   }
-  
+
   @Override
   public void write(DataOutput arg0) throws IOException {
     WritableUtils.writeCompressedByteArray(arg0, password);
   }
-  
+
   @Override
   public void destroy() throws DestroyFailedException {
     Arrays.fill(password, (byte) 0x00);
     password = null;
   }
-  
+
   @Override
   public boolean isDestroyed() {
     return password == null;
   }
-  
+
   @Override
   public int hashCode() {
     return Arrays.hashCode(password);
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     if (this == obj)
@@ -114,7 +114,7 @@ public class PasswordToken implements AuthenticationToken {
     PasswordToken other = (PasswordToken) obj;
     return Arrays.equals(password, other.password);
   }
-  
+
   @Override
   public PasswordToken clone() {
     try {
@@ -144,7 +144,7 @@ public class PasswordToken implements AuthenticationToken {
       }
     }
   }
-  
+
   @Override
   public void init(Properties properties) {
     if (properties.containsKey("password")) {
@@ -152,7 +152,7 @@ public class PasswordToken implements AuthenticationToken {
     } else
       throw new IllegalArgumentException("Missing 'password' property");
   }
-  
+
   @Override
   public Set<TokenProperty> getProperties() {
     Set<TokenProperty> internal = new LinkedHashSet<TokenProperty>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/AccumuloConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/AccumuloConfiguration.java b/core/src/main/java/org/apache/accumulo/core/conf/AccumuloConfiguration.java
index 813633c..cb075a9 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/AccumuloConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/AccumuloConfiguration.java
@@ -42,7 +42,8 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
     /**
      * Determines whether to accept a property based on its key.
      *
-     * @param key property key
+     * @param key
+     *          property key
      * @return true to accept property (pass filter)
      */
     boolean accept(String key);
@@ -68,7 +69,8 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
     /**
      * Creates a new filter.
      *
-     * @param prefix prefix of property keys to accept
+     * @param prefix
+     *          prefix of property keys to accept
      */
     public PrefixFilter(String prefix) {
       this.prefix = prefix;
@@ -81,28 +83,29 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
   }
 
   private static final Logger log = Logger.getLogger(AccumuloConfiguration.class);
-  
+
   /**
    * Gets a property value from this configuration.
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
    */
   public abstract String get(Property property);
-  
+
   /**
-   * Returns property key/value pairs in this configuration. The pairs include
-   * those defined in this configuration which pass the given filter, and those
+   * Returns property key/value pairs in this configuration. The pairs include those defined in this configuration which pass the given filter, and those
    * supplied from the parent configuration which are not included from here.
    *
-   * @param props properties object to populate
-   * @param filter filter for accepting properties from this configuration
+   * @param props
+   *          properties object to populate
+   * @param filter
+   *          filter for accepting properties from this configuration
    */
   public abstract void getProperties(Map<String,String> props, PropertyFilter filter);
 
   /**
-   * Returns an iterator over property key/value pairs in this configuration.
-   * Some implementations may elect to omit properties.
+   * Returns an iterator over property key/value pairs in this configuration. Some implementations may elect to omit properties.
    *
    * @return iterator over properties
    */
@@ -112,7 +115,7 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
     getProperties(entries, new AllFilter());
     return entries.entrySet().iterator();
   }
-  
+
   private void checkType(Property property, PropertyType type) {
     if (!property.getType().equals(type)) {
       String msg = "Configuration method intended for type " + type + " called with a " + property.getType() + " argument (" + property.getKey() + ")";
@@ -121,54 +124,54 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
       throw err;
     }
   }
-  
+
   /**
    * Gets all properties under the given prefix in this configuration.
    *
-   * @param property prefix property, must be of type PropertyType.PREFIX
+   * @param property
+   *          prefix property, must be of type PropertyType.PREFIX
    * @return a map of property keys to values
-   * @throws IllegalArgumentException if property is not a prefix
+   * @throws IllegalArgumentException
+   *           if property is not a prefix
    */
   public Map<String,String> getAllPropertiesWithPrefix(Property property) {
     checkType(property, PropertyType.PREFIX);
-    
+
     Map<String,String> propMap = new HashMap<String,String>();
     getProperties(propMap, new PrefixFilter(property.getKey()));
     return propMap;
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#MEMORY}, interpreting the
-   * value properly.
+   * Gets a property of type {@link PropertyType#MEMORY}, interpreting the value properly.
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    * @see #getMemoryInBytes(String)
    */
   public long getMemoryInBytes(Property property) {
     checkType(property, PropertyType.MEMORY);
-    
+
     String memString = get(property);
     return getMemoryInBytes(memString);
   }
-  
+
   /**
-   * Interprets a string specifying a memory size. A memory size is specified
-   * as a long integer followed by an optional B (bytes), K (KB), M (MB), or
-   * G (GB).
+   * Interprets a string specifying a memory size. A memory size is specified as a long integer followed by an optional B (bytes), K (KB), M (MB), or G (GB).
    *
-   * @param str string value
+   * @param str
+   *          string value
    * @return interpreted memory size
    */
   static public long getMemoryInBytes(String str) {
     int multiplier = 0;
     char lastChar = str.charAt(str.length() - 1);
-    
+
     if (lastChar == 'b') {
-      log.warn("The 'b' in " + str + 
-          " is being considered as bytes. " + 
-          "Setting memory by bits is not supported");
+      log.warn("The 'b' in " + str + " is being considered as bytes. " + "Setting memory by bits is not supported");
     }
     try {
       switch (Character.toUpperCase(lastChar)) {
@@ -184,34 +187,33 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
           return Long.parseLong(str);
       }
     } catch (Exception ex) {
-      throw new IllegalArgumentException("The value '" + str + 
-          "' is not a valid memory setting. A valid value would a number " +
-          "possibily followed by an optional 'G', 'M', 'K', or 'B'.");
+      throw new IllegalArgumentException("The value '" + str + "' is not a valid memory setting. A valid value would a number "
+          + "possibily followed by an optional 'G', 'M', 'K', or 'B'.");
     }
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#TIMEDURATION}, interpreting the
-   * value properly.
+   * Gets a property of type {@link PropertyType#TIMEDURATION}, interpreting the value properly.
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    * @see #getTimeInMillis(String)
    */
   public long getTimeInMillis(Property property) {
     checkType(property, PropertyType.TIMEDURATION);
-    
+
     return getTimeInMillis(get(property));
   }
-  
+
   /**
-   * Interprets a string specifying a time duration. A time duration is
-   * specified as a long integer followed by an optional d (days), h (hours),
-   * m (minutes), s (seconds), or ms (milliseconds). A value without a unit
-   * is interpreted as seconds.
+   * Interprets a string specifying a time duration. A time duration is specified as a long integer followed by an optional d (days), h (hours), m (minutes), s
+   * (seconds), or ms (milliseconds). A value without a unit is interpreted as seconds.
    *
-   * @param str string value
+   * @param str
+   *          string value
    * @return interpreted time duration in milliseconds
    */
   static public long getTimeInMillis(String str) {
@@ -233,40 +235,42 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
         return Long.parseLong(str) * 1000;
     }
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#BOOLEAN}, interpreting the
-   * value properly (using <code>Boolean.parseBoolean()</code>).
+   * Gets a property of type {@link PropertyType#BOOLEAN}, interpreting the value properly (using <code>Boolean.parseBoolean()</code>).
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    */
   public boolean getBoolean(Property property) {
     checkType(property, PropertyType.BOOLEAN);
     return Boolean.parseBoolean(get(property));
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#FRACTION}, interpreting the
-   * value properly.
+   * Gets a property of type {@link PropertyType#FRACTION}, interpreting the value properly.
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    * @see #getFraction(String)
    */
   public double getFraction(Property property) {
     checkType(property, PropertyType.FRACTION);
-    
+
     return getFraction(get(property));
   }
-  
+
   /**
-   * Interprets a string specifying a fraction. A fraction is specified as a
-   * double. An optional % at the end signifies a percentage.
+   * Interprets a string specifying a fraction. A fraction is specified as a double. An optional % at the end signifies a percentage.
    *
-   * @param str string value
+   * @param str
+   *          string value
    * @return interpreted fraction as a decimal value
    */
   public double getFraction(String str) {
@@ -274,19 +278,20 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
       return Double.parseDouble(str.substring(0, str.length() - 1)) / 100.0;
     return Double.parseDouble(str);
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#PORT}, interpreting the
-   * value properly (as an integer within the range of non-privileged ports).
+   * Gets a property of type {@link PropertyType#PORT}, interpreting the value properly (as an integer within the range of non-privileged ports).
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    * @see #getTimeInMillis(String)
    */
   public int getPort(Property property) {
     checkType(property, PropertyType.PORT);
-    
+
     String portString = get(property);
     int port = Integer.parseInt(portString);
     if (port != 0) {
@@ -297,37 +302,40 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
     }
     return port;
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#COUNT}, interpreting the
-   * value properly (as an integer).
+   * Gets a property of type {@link PropertyType#COUNT}, interpreting the value properly (as an integer).
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    * @see #getTimeInMillis(String)
    */
   public int getCount(Property property) {
     checkType(property, PropertyType.COUNT);
-    
+
     String countString = get(property);
     return Integer.parseInt(countString);
   }
-  
+
   /**
-   * Gets a property of type {@link PropertyType#PATH}, interpreting the
-   * value properly, replacing supported environment variables.
+   * Gets a property of type {@link PropertyType#PATH}, interpreting the value properly, replacing supported environment variables.
    *
-   * @param property property to get
+   * @param property
+   *          property to get
    * @return property value
-   * @throws IllegalArgumentException if the property is of the wrong type
+   * @throws IllegalArgumentException
+   *           if the property is of the wrong type
    * @see Constants#PATH_PROPERTY_ENV_VARS
    */
   public String getPath(Property property) {
     checkType(property, PropertyType.PATH);
 
     String pathString = get(property);
-    if (pathString == null) return null;
+    if (pathString == null)
+      return null;
 
     for (String replaceableEnvVar : Constants.PATH_PROPERTY_ENV_VARS) {
       String envValue = System.getenv(replaceableEnvVar);
@@ -347,21 +355,25 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
   public static synchronized DefaultConfiguration getDefaultConfiguration() {
     return DefaultConfiguration.getInstance();
   }
-  
+
   /**
    * Gets the configuration specific to a table.
    *
-   * @param conn connector (used to find table name)
-   * @param tableId table ID
+   * @param conn
+   *          connector (used to find table name)
+   * @param tableId
+   *          table ID
    * @return configuration containing table properties
-   * @throws TableNotFoundException if the table is not found
-   * @throws AccumuloException if there is a problem communicating to Accumulo
+   * @throws TableNotFoundException
+   *           if the table is not found
+   * @throws AccumuloException
+   *           if there is a problem communicating to Accumulo
    */
   public static AccumuloConfiguration getTableConfiguration(Connector conn, String tableId) throws TableNotFoundException, AccumuloException {
     String tableName = Tables.getTableName(conn.getInstance(), tableId);
     return new ConfigurationCopy(conn.tableOperations().getProperties(tableName));
   }
-  
+
   /**
    * Gets the maximum number of files per tablet from this configuration.
    *
@@ -375,26 +387,29 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
       maxFilesPerTablet = getCount(Property.TSERV_SCAN_MAX_OPENFILES) - 1;
       log.debug("Max files per tablet " + maxFilesPerTablet);
     }
-    
+
     return maxFilesPerTablet;
   }
-  
+
   // overridden in ZooConfiguration
   public void invalidateCache() {}
-  
+
   /**
    * Creates a new instance of a class specified in a configuration property.
    *
-   * @param property property specifying class name
-   * @param base base class of type
-   * @param defaultInstance instance to use if creation fails
+   * @param property
+   *          property specifying class name
+   * @param base
+   *          base class of type
+   * @param defaultInstance
+   *          instance to use if creation fails
    * @return new class instance, or default instance if creation failed
    * @see AccumuloVFSClassLoader
    */
   public <T> T instantiateClassProperty(Property property, Class<T> base, T defaultInstance) {
     String clazzName = get(property);
     T instance = null;
-    
+
     try {
       Class<? extends T> clazz = AccumuloVFSClassLoader.loadClass(clazzName, base);
       instance = clazz.newInstance();
@@ -402,12 +417,12 @@ public abstract class AccumuloConfiguration implements Iterable<Entry<String,Str
     } catch (Exception e) {
       log.warn("Failed to load class ", e);
     }
-    
+
     if (instance == null) {
       log.info("Using " + defaultInstance.getClass().getName());
       instance = defaultInstance;
     }
     return instance;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/ConfigSanityCheck.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/ConfigSanityCheck.java b/core/src/main/java/org/apache/accumulo/core/conf/ConfigSanityCheck.java
index 3754b02..99e3920 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/ConfigSanityCheck.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/ConfigSanityCheck.java
@@ -24,19 +24,19 @@ import org.apache.log4j.Logger;
  * A utility class for validating {@link AccumuloConfiguration} instances.
  */
 public class ConfigSanityCheck {
-  
+
   private static final Logger log = Logger.getLogger(ConfigSanityCheck.class);
   private static final String PREFIX = "BAD CONFIG ";
-  
+
   /**
-   * Validates the given configuration entries. A valid configuration contains only
-   * valid properties (i.e., defined or otherwise valid) that are not prefixes
-   * and whose values are formatted correctly for their property types. A valid
-   * configuration also contains a value for property
+   * Validates the given configuration entries. A valid configuration contains only valid properties (i.e., defined or otherwise valid) that are not prefixes
+   * and whose values are formatted correctly for their property types. A valid configuration also contains a value for property
    * {@link Property#INSTANCE_ZK_TIMEOUT} within a valid range.
    *
-   * @param entries iterable through configuration keys and values
-   * @throws SanityCheckException if a fatal configuration error is found
+   * @param entries
+   *          iterable through configuration keys and values
+   * @throws SanityCheckException
+   *           if a fatal configuration error is found
    */
   public static void validate(Iterable<Entry<String,String>> entries) {
     String instanceZkTimeoutKey = Property.INSTANCE_ZK_TIMEOUT.getKey();
@@ -60,42 +60,41 @@ public class ConfigSanityCheck {
     }
 
     if (instanceZkTimeoutValue != null) {
-      checkTimeDuration(Property.INSTANCE_ZK_TIMEOUT, instanceZkTimeoutValue,
-                        new CheckTimeDurationBetween(1000, 300000));
+      checkTimeDuration(Property.INSTANCE_ZK_TIMEOUT, instanceZkTimeoutValue, new CheckTimeDurationBetween(1000, 300000));
     }
   }
-  
+
   private interface CheckTimeDuration {
     boolean check(long propVal);
-    
+
     String getDescription(Property prop);
   }
-  
+
   private static class CheckTimeDurationBetween implements CheckTimeDuration {
     long min, max;
-    
+
     CheckTimeDurationBetween(long x, long y) {
       min = Math.min(x, y);
       max = Math.max(x, y);
     }
-    
+
     @Override
     public boolean check(long propVal) {
       return propVal >= min && propVal <= max;
     }
-    
+
     @Override
     public String getDescription(Property prop) {
       return "ensure " + min + " <= " + prop + " <= " + max;
     }
   }
-  
+
   private static void checkTimeDuration(Property prop, String value, CheckTimeDuration chk) {
     verifyPropertyTypes(PropertyType.TIMEDURATION, prop);
     if (!chk.check(AccumuloConfiguration.getTimeInMillis(value)))
       fatal(PREFIX + chk.getDescription(prop));
   }
-  
+
   private static void verifyPropertyTypes(PropertyType type, Property... properties) {
     for (Property prop : properties)
       if (prop.getType() != type)
@@ -111,7 +110,9 @@ public class ConfigSanityCheck {
     /**
      * Creates a new exception with the given message.
      */
-    public SanityCheckException(String msg) { super(msg); }
+    public SanityCheckException(String msg) {
+      super(msg);
+    }
   }
 
   private static void fatal(String msg) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationCopy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationCopy.java b/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationCopy.java
index 2b1945d..8326725 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationCopy.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationCopy.java
@@ -22,44 +22,45 @@ import java.util.Map;
 import java.util.Map.Entry;
 
 /**
- * An {@link AccumuloConfiguration} which holds a flat copy of properties
- * defined in another configuration
+ * An {@link AccumuloConfiguration} which holds a flat copy of properties defined in another configuration
  */
 public class ConfigurationCopy extends AccumuloConfiguration {
   final Map<String,String> copy = Collections.synchronizedMap(new HashMap<String,String>());
-  
+
   /**
    * Creates a new configuration.
    *
-   * @param config configuration property key/value pairs to copy
+   * @param config
+   *          configuration property key/value pairs to copy
    */
   public ConfigurationCopy(Map<String,String> config) {
     this(config.entrySet());
   }
-  
+
   /**
    * Creates a new configuration.
    *
-   * @param config configuration property iterable to use for copying
+   * @param config
+   *          configuration property iterable to use for copying
    */
   public ConfigurationCopy(Iterable<Entry<String,String>> config) {
     for (Entry<String,String> entry : config) {
       copy.put(entry.getKey(), entry.getValue());
     }
   }
-  
+
   /**
    * Creates a new empty configuration.
    */
   public ConfigurationCopy() {
     this(new HashMap<String,String>());
   }
-  
+
   @Override
   public String get(Property property) {
     return copy.get(property.getKey());
   }
-  
+
   @Override
   public void getProperties(Map<String,String> props, PropertyFilter filter) {
     for (Entry<String,String> entry : copy.entrySet()) {
@@ -68,25 +69,29 @@ public class ConfigurationCopy extends AccumuloConfiguration {
       }
     }
   }
-  
+
   /**
    * Sets a property in this configuration.
    *
-   * @param prop property to set
-   * @param value property value
+   * @param prop
+   *          property to set
+   * @param value
+   *          property value
    */
   public void set(Property prop, String value) {
     copy.put(prop.getKey(), value);
   }
-  
+
   /**
    * Sets a property in this configuration.
    *
-   * @param key key of property to set
-   * @param value property value
+   * @param key
+   *          key of property to set
+   * @param value
+   *          property value
    */
   public void set(String key, String value) {
     copy.put(key, value);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationDocGen.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationDocGen.java b/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationDocGen.java
index 60d1996..f15e6ec 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationDocGen.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationDocGen.java
@@ -356,10 +356,10 @@ class ConfigurationDocGen {
   }
 
   /*
-   * Generates documentation for conf/accumulo-site.xml file usage. Arguments
-   * are: "--generate-doc", file to write to.
+   * Generates documentation for conf/accumulo-site.xml file usage. Arguments are: "--generate-doc", file to write to.
    *
    * @param args command-line arguments
+   *
    * @throws IllegalArgumentException if args is invalid
    */
   public static void main(String[] args) throws FileNotFoundException, UnsupportedEncodingException {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationObserver.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationObserver.java b/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationObserver.java
index 46f8692..8a62d26 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationObserver.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/ConfigurationObserver.java
@@ -18,8 +18,8 @@ package org.apache.accumulo.core.conf;
 
 public interface ConfigurationObserver {
   void propertyChanged(String key);
-  
+
   void propertiesChanged();
-  
+
   void sessionExpired();
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/CredentialProviderFactoryShim.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/CredentialProviderFactoryShim.java b/core/src/main/java/org/apache/accumulo/core/conf/CredentialProviderFactoryShim.java
index 33b1848..81fe540 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/CredentialProviderFactoryShim.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/CredentialProviderFactoryShim.java
@@ -314,10 +314,10 @@ public class CredentialProviderFactoryShim {
     Preconditions.checkNotNull(credentialProviders);
     return getConfiguration(new Configuration(CachedConfiguration.getInstance()), credentialProviders);
   }
-  
+
   /**
    * Adds the Credential Provider configuration elements to the provided {@link Configuration}.
-   * 
+   *
    * @param conf
    *          Existing Hadoop Configuration
    * @param credentialProviders

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/DefaultConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/DefaultConfiguration.java b/core/src/main/java/org/apache/accumulo/core/conf/DefaultConfiguration.java
index e473772..09dc0c9 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/DefaultConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/DefaultConfiguration.java
@@ -22,8 +22,7 @@ import java.util.Map;
 import java.util.Map.Entry;
 
 /**
- * An {@link AccumuloConfiguration} that contains only default values for
- * properties. This class is a singleton.
+ * An {@link AccumuloConfiguration} that contains only default values for properties. This class is a singleton.
  */
 public class DefaultConfiguration extends AccumuloConfiguration {
   private final static Map<String,String> resolvedProps;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/Experimental.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/Experimental.java b/core/src/main/java/org/apache/accumulo/core/conf/Experimental.java
index 0d1059a..e2b4cec 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/Experimental.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/Experimental.java
@@ -26,5 +26,5 @@ import java.lang.annotation.RetentionPolicy;
 @Inherited
 @Retention(RetentionPolicy.RUNTIME)
 @interface Experimental {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/Interpolated.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/Interpolated.java b/core/src/main/java/org/apache/accumulo/core/conf/Interpolated.java
index 1390db2..df5b6f8 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/Interpolated.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/Interpolated.java
@@ -22,7 +22,7 @@ import java.lang.annotation.RetentionPolicy;
 
 /**
  * An annotation to denote {@link AccumuloConfiguration} {@link Property} keys, whose values should be interpolated with system properties.
- * 
+ *
  * Interpolated items need to be careful, as JVM properties could be updates and we may want that propogated when those changes occur. Currently only
  * VFS_CLASSLOADER_CACHE_DIR, which isn't ZK mutable, is interpolated, so this shouldn't be an issue as java.io.tmpdir also shouldn't be changing.
  */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/Property.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/Property.java b/core/src/main/java/org/apache/accumulo/core/conf/Property.java
index e054a5f..ce5de85 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/Property.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/Property.java
@@ -168,7 +168,8 @@ public enum Property {
   @Experimental
   GENERAL_VOLUME_CHOOSER("general.volume.chooser", "org.apache.accumulo.server.fs.PerTableVolumeChooser", PropertyType.CLASSNAME,
       "The class that will be used to select which volume will be used to create new files."),
-  GENERAL_SECURITY_CREDENTIAL_PROVIDER_PATHS("general.security.credential.provider.paths", "", PropertyType.STRING, "Comma-separated list of paths to CredentialProviders"),
+  GENERAL_SECURITY_CREDENTIAL_PROVIDER_PATHS("general.security.credential.provider.paths", "", PropertyType.STRING,
+      "Comma-separated list of paths to CredentialProviders"),
   GENERAL_LEGACY_METRICS("general.legacy.metrics", "false", PropertyType.BOOLEAN,
       "Use the old metric infrastructure configured by accumulo-metrics.xml, instead of Hadoop Metrics2"),
 
@@ -182,7 +183,8 @@ public enum Property {
   MASTER_BULK_RETRIES("master.bulk.retries", "3", PropertyType.COUNT, "The number of attempts to bulk-load a file before giving up."),
   MASTER_BULK_THREADPOOL_SIZE("master.bulk.threadpool.size", "5", PropertyType.COUNT, "The number of threads to use when coordinating a bulk-import."),
   MASTER_BULK_TIMEOUT("master.bulk.timeout", "5m", PropertyType.TIMEDURATION, "The time to wait for a tablet server to process a bulk import request"),
-  MASTER_BULK_RENAME_THREADS("master.bulk.rename.threadpool.size", "20", PropertyType.COUNT, "The number of threads to use when moving user files to bulk ingest directories under accumulo control"),
+  MASTER_BULK_RENAME_THREADS("master.bulk.rename.threadpool.size", "20", PropertyType.COUNT,
+      "The number of threads to use when moving user files to bulk ingest directories under accumulo control"),
   MASTER_MINTHREADS("master.server.threads.minimum", "20", PropertyType.COUNT, "The minimum number of threads to use to handle incoming requests."),
   MASTER_THREADCHECK("master.server.threadcheck.time", "1s", PropertyType.TIMEDURATION, "The time between adjustments of the server thread pool."),
   MASTER_RECOVERY_DELAY("master.recovery.delay", "10s", PropertyType.TIMEDURATION,
@@ -196,8 +198,10 @@ public enum Property {
   MASTER_REPLICATION_SCAN_INTERVAL("master.replication.status.scan.interval", "30s", PropertyType.TIMEDURATION,
       "Amount of time to sleep before scanning the status section of the replication table for new data"),
   MASTER_REPLICATION_COORDINATOR_PORT("master.replication.coordinator.port", "10001", PropertyType.PORT, "Port for the replication coordinator service"),
-  MASTER_REPLICATION_COORDINATOR_MINTHREADS("master.replication.coordinator.minthreads", "4", PropertyType.COUNT, "Minimum number of threads dedicated to answering coordinator requests"),
-  MASTER_REPLICATION_COORDINATOR_THREADCHECK("master.replication.coordinator.threadcheck.time", "5s", PropertyType.TIMEDURATION, "The time between adjustments of the coordinator thread pool"),
+  MASTER_REPLICATION_COORDINATOR_MINTHREADS("master.replication.coordinator.minthreads", "4", PropertyType.COUNT,
+      "Minimum number of threads dedicated to answering coordinator requests"),
+  MASTER_REPLICATION_COORDINATOR_THREADCHECK("master.replication.coordinator.threadcheck.time", "5s", PropertyType.TIMEDURATION,
+      "The time between adjustments of the coordinator thread pool"),
 
   // properties that are specific to tablet server behavior
   TSERV_PREFIX("tserver.", null, PropertyType.PREFIX, "Properties in this category affect the behavior of the tablet servers"),
@@ -208,12 +212,11 @@ public enum Property {
   TSERV_PORTSEARCH("tserver.port.search", "false", PropertyType.BOOLEAN, "if the ports above are in use, search higher ports until one is available"),
   TSERV_CLIENTPORT("tserver.port.client", "9997", PropertyType.PORT, "The port used for handling client connections on the tablet servers"),
   @Deprecated
-  TSERV_MUTATION_QUEUE_MAX("tserver.mutation.queue.max", "1M", PropertyType.MEMORY,
-      "This setting is deprecated. See tserver.total.mutation.queue.max. "
-          + "The amount of memory to use to store write-ahead-log mutations-per-session before flushing them. Since the buffer is per write session, consider the"
-          + " max number of concurrent writer when configuring. When using Hadoop 2, Accumulo will call hsync() on the WAL . For a small number of "
-          + "concurrent writers, increasing this buffer size decreases the frequncy of hsync calls. For a large number of concurrent writers a small buffers "
-          + "size is ok because of group commit."),
+  TSERV_MUTATION_QUEUE_MAX("tserver.mutation.queue.max", "1M", PropertyType.MEMORY, "This setting is deprecated. See tserver.total.mutation.queue.max. "
+      + "The amount of memory to use to store write-ahead-log mutations-per-session before flushing them. Since the buffer is per write session, consider the"
+      + " max number of concurrent writer when configuring. When using Hadoop 2, Accumulo will call hsync() on the WAL . For a small number of "
+      + "concurrent writers, increasing this buffer size decreases the frequncy of hsync calls. For a large number of concurrent writers a small buffers "
+      + "size is ok because of group commit."),
   TSERV_TOTAL_MUTATION_QUEUE_MAX("tserver.total.mutation.queue.max", "50M", PropertyType.MEMORY,
       "The amount of memory used to store write-ahead-log mutations before flushing them."),
   TSERV_TABLET_SPLIT_FINDMIDPOINT_MAXOPEN("tserver.tablet.split.midpoint.files.max", "30", PropertyType.COUNT,
@@ -294,11 +297,14 @@ public enum Property {
   TSERV_WAL_SYNC_METHOD("tserver.wal.sync.method", "hsync", PropertyType.STRING, "This property is deprecated. Use table.durability instead."),
   TSERV_ASSIGNMENT_DURATION_WARNING("tserver.assignment.duration.warning", "10m", PropertyType.TIMEDURATION, "The amount of time an assignment can run "
       + " before the server will print a warning along with the current stack trace. Meant to help debug stuck assignments"),
-  TSERV_REPLICATION_REPLAYERS("tserver.replication.replayer.", null, PropertyType.PREFIX, "Allows configuration of implementation used to apply replicated data"),
+  TSERV_REPLICATION_REPLAYERS("tserver.replication.replayer.", null, PropertyType.PREFIX,
+      "Allows configuration of implementation used to apply replicated data"),
   TSERV_REPLICATION_DEFAULT_HANDLER("tserver.replication.default.replayer", "org.apache.accumulo.tserver.replication.BatchWriterReplicationReplayer",
       PropertyType.CLASSNAME, "Default AccumuloReplicationReplayer implementation"),
-  TSERV_REPLICATION_BW_REPLAYER_MEMORY("tserver.replication.batchwriter.replayer.memory", "50M", PropertyType.MEMORY, "Memory to provide to batchwriter to replay mutations for replication"),
-  TSERV_ASSIGNMENT_MAXCONCURRENT("tserver.assignment.concurrent.max", "2", PropertyType.COUNT, "The number of threads available to load tablets. Recoveries are still performed serially."),
+  TSERV_REPLICATION_BW_REPLAYER_MEMORY("tserver.replication.batchwriter.replayer.memory", "50M", PropertyType.MEMORY,
+      "Memory to provide to batchwriter to replay mutations for replication"),
+  TSERV_ASSIGNMENT_MAXCONCURRENT("tserver.assignment.concurrent.max", "2", PropertyType.COUNT,
+      "The number of threads available to load tablets. Recoveries are still performed serially."),
 
   // properties that are specific to logger server behavior
   LOGGER_PREFIX("logger.", null, PropertyType.PREFIX, "Properties in this category affect the behavior of the write-ahead logger servers"),
@@ -332,8 +338,10 @@ public enum Property {
   MONITOR_SSL_TRUSTSTORE("monitor.ssl.trustStore", "", PropertyType.PATH, "The truststore for enabling monitor SSL."),
   @Sensitive
   MONITOR_SSL_TRUSTSTOREPASS("monitor.ssl.trustStorePassword", "", PropertyType.STRING, "The truststore password for enabling monitor SSL."),
-  MONITOR_SSL_INCLUDE_CIPHERS("monitor.ssl.include.ciphers", "", PropertyType.STRING, "A comma-separated list of allows SSL Ciphers, see monitor.ssl.exclude.ciphers to disallow ciphers"),
-  MONITOR_SSL_EXCLUDE_CIPHERS("monitor.ssl.exclude.ciphers", "", PropertyType.STRING, "A comma-separated list of disallowed SSL Ciphers, see mmonitor.ssl.include.ciphers to allow ciphers"),
+  MONITOR_SSL_INCLUDE_CIPHERS("monitor.ssl.include.ciphers", "", PropertyType.STRING,
+      "A comma-separated list of allows SSL Ciphers, see monitor.ssl.exclude.ciphers to disallow ciphers"),
+  MONITOR_SSL_EXCLUDE_CIPHERS("monitor.ssl.exclude.ciphers", "", PropertyType.STRING,
+      "A comma-separated list of disallowed SSL Ciphers, see mmonitor.ssl.include.ciphers to allow ciphers"),
   MONITOR_SSL_INCLUDE_PROTOCOLS("monitor.ssl.include.protocols", "TLSv1,TLSv1.1,TLSv1.2", PropertyType.STRING, "A comma-separate list of allowed SSL protocols"),
 
   MONITOR_LOCK_CHECK_INTERVAL("monitor.lock.check.interval", "5s", PropertyType.TIMEDURATION,
@@ -342,7 +350,8 @@ public enum Property {
       + "the date shown on the 'Recent Logs' monitor page"),
 
   TRACE_PREFIX("trace.", null, PropertyType.PREFIX, "Properties in this category affect the behavior of distributed tracing."),
-  TRACE_SPAN_RECEIVERS("trace.span.receivers", "org.apache.accumulo.tracer.ZooTraceClient", PropertyType.CLASSNAMELIST, "A list of span receiver classes to send trace spans"),
+  TRACE_SPAN_RECEIVERS("trace.span.receivers", "org.apache.accumulo.tracer.ZooTraceClient", PropertyType.CLASSNAMELIST,
+      "A list of span receiver classes to send trace spans"),
   TRACE_SPAN_RECEIVER_PREFIX("trace.span.receiver.", null, PropertyType.PREFIX, "Prefix for span receiver configuration properties"),
   TRACE_ZK_PATH("trace.span.receiver.zookeeper.path", Constants.ZTRACERS, PropertyType.STRING, "The zookeeper node where tracers are registered"),
   TRACE_PORT("trace.port.client", "12234", PropertyType.PORT, "The listening port for the trace server"),
@@ -413,11 +422,10 @@ public enum Property {
           + ",org.apache.accumulo.core.file.keyfunctor.ColumnFamilyFunctor, and org.apache.accumulo.core.file.keyfunctor.ColumnQualifierFunctor are"
           + " allowable values. One can extend any of the above mentioned classes to perform specialized parsing of the key. "),
   TABLE_BLOOM_HASHTYPE("table.bloom.hash.type", "murmur", PropertyType.STRING, "The bloom filter hash type"),
-  TABLE_DURABILITY("table.durability", "sync", PropertyType.DURABILITY, "The durability used to write to the write-ahead log." +
-      " Legal values are: none, which skips the write-ahead log; " +
-      "log, which sends the data to the write-ahead log, but does nothing to make it durable; " +
-      "flush, which pushes data to the file system; and " +
-      "sync, which ensures the data is written to disk."),
+  TABLE_DURABILITY("table.durability", "sync", PropertyType.DURABILITY, "The durability used to write to the write-ahead log."
+      + " Legal values are: none, which skips the write-ahead log; "
+      + "log, which sends the data to the write-ahead log, but does nothing to make it durable; " + "flush, which pushes data to the file system; and "
+      + "sync, which ensures the data is written to disk."),
   TABLE_FAILURES_IGNORE("table.failures.ignore", "false", PropertyType.BOOLEAN,
       "If you want queries for your table to hang or fail when data is missing from the system, "
           + "then set this to false. When this set to true missing data will be reported but queries "
@@ -447,9 +455,12 @@ public enum Property {
           + "These iterators can take options if additional properties are set that look like this property, "
           + "but are suffixed with a period, followed by 'opt' followed by another period, and a property name.\n"
           + "For example, table.iterator.minc.vers.opt.maxVersions = 3"),
-  TABLE_ITERATOR_SCAN_PREFIX(TABLE_ITERATOR_PREFIX.getKey() + IteratorScope.scan.name() + ".", null, PropertyType.PREFIX, "Convenience prefix to find options for the scan iterator scope"),
-  TABLE_ITERATOR_MINC_PREFIX(TABLE_ITERATOR_PREFIX.getKey() + IteratorScope.minc.name() + ".", null, PropertyType.PREFIX, "Convenience prefix to find options for the minc iterator scope"),
-  TABLE_ITERATOR_MAJC_PREFIX(TABLE_ITERATOR_PREFIX.getKey() + IteratorScope.majc.name() + ".", null, PropertyType.PREFIX, "Convenience prefix to find options for the majc iterator scope"),
+  TABLE_ITERATOR_SCAN_PREFIX(TABLE_ITERATOR_PREFIX.getKey() + IteratorScope.scan.name() + ".", null, PropertyType.PREFIX,
+      "Convenience prefix to find options for the scan iterator scope"),
+  TABLE_ITERATOR_MINC_PREFIX(TABLE_ITERATOR_PREFIX.getKey() + IteratorScope.minc.name() + ".", null, PropertyType.PREFIX,
+      "Convenience prefix to find options for the minc iterator scope"),
+  TABLE_ITERATOR_MAJC_PREFIX(TABLE_ITERATOR_PREFIX.getKey() + IteratorScope.majc.name() + ".", null, PropertyType.PREFIX,
+      "Convenience prefix to find options for the majc iterator scope"),
   TABLE_LOCALITY_GROUP_PREFIX("table.group.", null, PropertyType.PREFIX,
       "Properties in this category are per-table properties that define locality groups in a table. These properties start "
           + "with the category prefix, followed by a name, followed by a period, and followed by a property for that group.\n"
@@ -465,9 +476,9 @@ public enum Property {
   TABLE_COMPACTION_STRATEGY_PREFIX("table.majc.compaction.strategy.opts.", null, PropertyType.PREFIX,
       "Properties in this category are used to configure the compaction strategy."),
   TABLE_REPLICATION("table.replication", "false", PropertyType.BOOLEAN, "Is replication enabled for the given table"),
-  TABLE_REPLICATION_TARGET("table.replication.target.", null, PropertyType.PREFIX, "Enumerate a mapping of other systems which this table should " +
-      "replicate their data to. The key suffix is the identifying cluster name and the value is an identifier for a location on the target system, " +
-      "e.g. the ID of the table on the target to replicate to"),
+  TABLE_REPLICATION_TARGET("table.replication.target.", null, PropertyType.PREFIX, "Enumerate a mapping of other systems which this table should "
+      + "replicate their data to. The key suffix is the identifying cluster name and the value is an identifier for a location on the target system, "
+      + "e.g. the ID of the table on the target to replicate to"),
   @Experimental
   TABLE_VOLUME_CHOOSER("table.volume.chooser", "org.apache.accumulo.server.fs.RandomVolumeChooser", PropertyType.CLASSNAME,
       "The class that will be used to select which volume will be used to create new files for this table."),
@@ -498,21 +509,28 @@ public enum Property {
   REPLICATION_PEER_USER("replication.peer.user.", null, PropertyType.PREFIX, "The username to provide when authenticating with the given peer"),
   @Sensitive
   REPLICATION_PEER_PASSWORD("replication.peer.password.", null, PropertyType.PREFIX, "The password to provide when authenticating with the given peer"),
-  REPLICATION_NAME("replication.name", "", PropertyType.STRING, "Name of this cluster with respect to replication. Used to identify this instance from other peers"),
+  REPLICATION_NAME("replication.name", "", PropertyType.STRING,
+      "Name of this cluster with respect to replication. Used to identify this instance from other peers"),
   REPLICATION_MAX_WORK_QUEUE("replication.max.work.queue", "1000", PropertyType.COUNT, "Upper bound of the number of files queued for replication"),
-  REPLICATION_WORK_ASSIGNMENT_SLEEP("replication.work.assignment.sleep", "30s", PropertyType.TIMEDURATION, "Amount of time to sleep between replication work assignment"),
+  REPLICATION_WORK_ASSIGNMENT_SLEEP("replication.work.assignment.sleep", "30s", PropertyType.TIMEDURATION,
+      "Amount of time to sleep between replication work assignment"),
   REPLICATION_WORKER_THREADS("replication.worker.threads", "4", PropertyType.COUNT, "Size of the threadpool that each tabletserver devotes to replicating data"),
-  REPLICATION_RECEIPT_SERVICE_PORT("replication.receipt.service.port", "10002", PropertyType.PORT, "Listen port used by thrift service in tserver listening for replication"),
-  REPLICATION_WORK_ATTEMPTS("replication.work.attempts", "10", PropertyType.COUNT, "Number of attempts to try to replicate some data before giving up and letting it naturally be retried later"),
+  REPLICATION_RECEIPT_SERVICE_PORT("replication.receipt.service.port", "10002", PropertyType.PORT,
+      "Listen port used by thrift service in tserver listening for replication"),
+  REPLICATION_WORK_ATTEMPTS("replication.work.attempts", "10", PropertyType.COUNT,
+      "Number of attempts to try to replicate some data before giving up and letting it naturally be retried later"),
   REPLICATION_MIN_THREADS("replication.receiver.min.threads", "1", PropertyType.COUNT, "Minimum number of threads for replication"),
-  REPLICATION_THREADCHECK("replication.receiver.threadcheck.time", "30s", PropertyType.TIMEDURATION, "The time between adjustments of the replication thread pool."),
+  REPLICATION_THREADCHECK("replication.receiver.threadcheck.time", "30s", PropertyType.TIMEDURATION,
+      "The time between adjustments of the replication thread pool."),
   REPLICATION_MAX_UNIT_SIZE("replication.max.unit.size", "64M", PropertyType.MEMORY, "Maximum size of data to send in a replication message"),
   REPLICATION_WORK_ASSIGNER("replication.work.assigner", "org.apache.accumulo.master.replication.UnorderedWorkAssigner", PropertyType.CLASSNAME,
       "Replication WorkAssigner implementation to use"),
   REPLICATION_DRIVER_DELAY("replication.driver.delay", "0s", PropertyType.TIMEDURATION,
       "Amount of time to wait before the replication work loop begins in the master."),
-  REPLICATION_WORK_PROCESSOR_DELAY("replication.work.processor.delay", "0s", PropertyType.TIMEDURATION, "Amount of time to wait before first checking for replication work, not useful outside of tests"),
-  REPLICATION_WORK_PROCESSOR_PERIOD("replication.work.processor.period", "0s", PropertyType.TIMEDURATION, "Amount of time to wait before re-checking for replication work, not useful outside of tests"),
+  REPLICATION_WORK_PROCESSOR_DELAY("replication.work.processor.delay", "0s", PropertyType.TIMEDURATION,
+      "Amount of time to wait before first checking for replication work, not useful outside of tests"),
+  REPLICATION_WORK_PROCESSOR_PERIOD("replication.work.processor.period", "0s", PropertyType.TIMEDURATION,
+      "Amount of time to wait before re-checking for replication work, not useful outside of tests"),
 
   ;
 
@@ -820,7 +838,6 @@ public enum Property {
     return instance;
   }
 
-
   /**
    * Creates a new instance of a class specified in a configuration property. The table classpath context is used if set.
    *

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/PropertyType.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/PropertyType.java b/core/src/main/java/org/apache/accumulo/core/conf/PropertyType.java
index bf39da9..09055d5 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/PropertyType.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/PropertyType.java
@@ -22,8 +22,7 @@ import org.apache.accumulo.core.Constants;
 import org.apache.hadoop.fs.Path;
 
 /**
- * Types of {@link Property} values. Each type has a short name, a description,
- * and a regex which valid values match. All of these fields are optional.
+ * Types of {@link Property} values. Each type has a short name, a description, and a regex which valid values match. All of these fields are optional.
  */
 public enum PropertyType {
   PREFIX(null, null, null),
@@ -99,7 +98,7 @@ public enum PropertyType {
   String getFormatDescription() {
     return format;
   }
-  
+
   /**
    * Checks if the given value is valid for this type.
    *

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/Sensitive.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/Sensitive.java b/core/src/main/java/org/apache/accumulo/core/conf/Sensitive.java
index 42371f0..57dfec8 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/Sensitive.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/Sensitive.java
@@ -26,5 +26,5 @@ import java.lang.annotation.RetentionPolicy;
 @Inherited
 @Retention(RetentionPolicy.RUNTIME)
 @interface Sensitive {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/conf/SiteConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/conf/SiteConfiguration.java b/core/src/main/java/org/apache/accumulo/core/conf/SiteConfiguration.java
index b74335c..f0c9e59 100644
--- a/core/src/main/java/org/apache/accumulo/core/conf/SiteConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/conf/SiteConfiguration.java
@@ -25,18 +25,15 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.log4j.Logger;
 
 /**
- * An {@link AccumuloConfiguration} which loads properties from an XML file,
- * usually accumulo-site.xml. This implementation supports defaulting undefined
+ * An {@link AccumuloConfiguration} which loads properties from an XML file, usually accumulo-site.xml. This implementation supports defaulting undefined
  * property values to a parent configuration's definitions.
  * <p>
- * The system property "org.apache.accumulo.config.file" can be used to specify
- * the location of the XML configuration file on the classpath. If the system
+ * The system property "org.apache.accumulo.config.file" can be used to specify the location of the XML configuration file on the classpath. If the system
  * property is not defined, it defaults to "accumulo-site.xml".
  * <p>
  * This class is a singleton.
  * <p>
- * <b>Note</b>: Client code should not use this class, and it may be deprecated
- * in the future.
+ * <b>Note</b>: Client code should not use this class, and it may be deprecated in the future.
  */
 public class SiteConfiguration extends AccumuloConfiguration {
   private static final Logger log = Logger.getLogger(SiteConfiguration.class);
@@ -52,13 +49,14 @@ public class SiteConfiguration extends AccumuloConfiguration {
   SiteConfiguration(AccumuloConfiguration parent) {
     SiteConfiguration.parent = parent;
   }
-  
+
   /**
-   * Gets an instance of this class. A new instance is only created on the first
-   * call, and so the parent configuration cannot be changed later.
+   * Gets an instance of this class. A new instance is only created on the first call, and so the parent configuration cannot be changed later.
    *
-   * @param parent parent (default) configuration
-   * @throws RuntimeException if the configuration is invalid
+   * @param parent
+   *          parent (default) configuration
+   * @throws RuntimeException
+   *           if the configuration is invalid
    */
   synchronized public static SiteConfiguration getInstance(AccumuloConfiguration parent) {
     if (instance == null) {
@@ -67,7 +65,7 @@ public class SiteConfiguration extends AccumuloConfiguration {
     }
     return instance;
   }
-  
+
   synchronized public static SiteConfiguration getInstance() {
     return getInstance(DefaultConfiguration.getInstance());
   }
@@ -161,8 +159,7 @@ public class SiteConfiguration extends AccumuloConfiguration {
   }
 
   /**
-   * Clears the configuration properties in this configuration (but not the
-   * parent). This method supports testing and should not be called.
+   * Clears the configuration properties in this configuration (but not the parent). This method supports testing and should not be called.
    */
   synchronized public static void clearInstance() {
     instance = null;
@@ -175,11 +172,8 @@ public class SiteConfiguration extends AccumuloConfiguration {
     getXmlConfig().clear();
   }
 
-
   /**
-   * Clears the configuration properties in this configuration (but not the
-   * parent) and nulls it. This method supports testing and should not be
-   * called.
+   * Clears the configuration properties in this configuration (but not the parent) and nulls it. This method supports testing and should not be called.
    */
   public synchronized void clearAndNull() {
     if (xmlConfig != null) {
@@ -191,8 +185,10 @@ public class SiteConfiguration extends AccumuloConfiguration {
   /**
    * Sets a property. This method supports testing and should not be called.
    *
-   * @param property property to set
-   * @param value property value
+   * @param property
+   *          property to set
+   * @param value
+   *          property value
    */
   public void set(Property property, String value) {
     set(property.getKey(), value);
@@ -201,8 +197,10 @@ public class SiteConfiguration extends AccumuloConfiguration {
   /**
    * Sets a property. This method supports testing and should not be called.
    *
-   * @param key key of property to set
-   * @param value property value
+   * @param key
+   *          key of property to set
+   * @param value
+   *          property value
    */
   public void set(String key, String value) {
     getXmlConfig().set(key, value);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/constraints/Constraint.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/constraints/Constraint.java b/core/src/main/java/org/apache/accumulo/core/constraints/Constraint.java
index 3c5f797..e95088a 100644
--- a/core/src/main/java/org/apache/accumulo/core/constraints/Constraint.java
+++ b/core/src/main/java/org/apache/accumulo/core/constraints/Constraint.java
@@ -44,7 +44,7 @@ import org.apache.accumulo.core.security.Authorizations;
  * </p>
  */
 public interface Constraint {
-  
+
   /**
    * The environment within which a constraint exists.
    */
@@ -55,14 +55,14 @@ public interface Constraint {
      * @return key extent
      */
     KeyExtent getExtent();
-    
+
     /**
      * Gets the user within the environment.
      *
      * @return user
      */
     String getUser();
-    
+
     /**
      * Gets the authorizations in the environment.
      *
@@ -79,23 +79,25 @@ public interface Constraint {
      */
     AuthorizationContainer getAuthorizationsContainer();
   }
-  
+
   /**
    * Gets a short, one-sentence description of what a given violation code means.
    *
-   * @param violationCode numeric violation code
+   * @param violationCode
+   *          numeric violation code
    * @return matching violation description
    */
   String getViolationDescription(short violationCode);
-  
+
   /**
-   * Checks a mutation for constraint violations. If the mutation contains no violations, returns null. Otherwise, returns
-   * a list of violation codes.
+   * Checks a mutation for constraint violations. If the mutation contains no violations, returns null. Otherwise, returns a list of violation codes.
    *
    * Violation codes must be non-negative. Negative violation codes are reserved for system use.
    *
-   * @param env constraint environment
-   * @param mutation mutation to check
+   * @param env
+   *          constraint environment
+   * @param mutation
+   *          mutation to check
    * @return list of violation codes, or null if none
    */
   List<Short> check(Environment env, Mutation mutation);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraint.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraint.java b/core/src/main/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraint.java
index 88b7eea..7cc42c1 100644
--- a/core/src/main/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraint.java
+++ b/core/src/main/java/org/apache/accumulo/core/constraints/DefaultKeySizeConstraint.java
@@ -26,30 +26,30 @@ import org.apache.accumulo.core.data.Mutation;
  * A constraints that limits the size of keys to 1mb.
  */
 public class DefaultKeySizeConstraint implements Constraint {
-  
+
   protected static final short MAX__KEY_SIZE_EXCEEDED_VIOLATION = 1;
   protected static final long maxSize = 1048576; // 1MB default size
-  
+
   @Override
   public String getViolationDescription(short violationCode) {
-    
+
     switch (violationCode) {
       case MAX__KEY_SIZE_EXCEEDED_VIOLATION:
         return "Key was larger than 1MB";
     }
-    
+
     return null;
   }
-  
+
   final static List<Short> NO_VIOLATIONS = new ArrayList<Short>();
-  
+
   @Override
   public List<Short> check(Environment env, Mutation mutation) {
 
     // fast size check
     if (mutation.numBytes() < maxSize)
       return NO_VIOLATIONS;
-    
+
     List<Short> violations = new ArrayList<Short>();
 
     for (ColumnUpdate cu : mutation.getUpdates()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/constraints/Violations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/constraints/Violations.java b/core/src/main/java/org/apache/accumulo/core/constraints/Violations.java
index eda1a67..1db13cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/constraints/Violations.java
+++ b/core/src/main/java/org/apache/accumulo/core/constraints/Violations.java
@@ -28,42 +28,42 @@ import org.apache.accumulo.core.data.ConstraintViolationSummary;
  * A class for accumulating constraint violations across a number of mutations.
  */
 public class Violations {
-  
+
   private static class CVSKey {
     private String className;
     private short vcode;
-    
+
     CVSKey(ConstraintViolationSummary cvs) {
       this.className = cvs.constrainClass;
       this.vcode = cvs.violationCode;
     }
-    
+
     @Override
     public int hashCode() {
       return className.hashCode() + vcode;
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof CVSKey)
         return equals((CVSKey) o);
       return false;
     }
-    
+
     public boolean equals(CVSKey ocvsk) {
       return className.equals(ocvsk.className) && vcode == ocvsk.vcode;
     }
   }
-  
+
   private HashMap<CVSKey,ConstraintViolationSummary> cvsmap;
-  
+
   /**
    * Creates a new empty object.
    */
   public Violations() {
     cvsmap = new HashMap<CVSKey,ConstraintViolationSummary>();
   }
-  
+
   /**
    * Checks if this object is empty, i.e., that no violations have been added.
    *
@@ -72,54 +72,56 @@ public class Violations {
   public boolean isEmpty() {
     return cvsmap.isEmpty();
   }
-  
+
   private void add(CVSKey cvsk, ConstraintViolationSummary cvs) {
     ConstraintViolationSummary existingCvs = cvsmap.get(cvsk);
-    
+
     if (existingCvs == null) {
       cvsmap.put(cvsk, cvs);
     } else {
       existingCvs.numberOfViolatingMutations += cvs.numberOfViolatingMutations;
     }
   }
-  
+
   /**
-   * Adds a violation. If a matching violation was already added, then its
-   * count is increased.
+   * Adds a violation. If a matching violation was already added, then its count is increased.
    *
-   * @param cvs summary of violation
+   * @param cvs
+   *          summary of violation
    */
   public void add(ConstraintViolationSummary cvs) {
     CVSKey cvsk = new CVSKey(cvs);
     add(cvsk, cvs);
   }
-  
+
   /**
    * Adds all violations from the given object to this one.
    *
-   * @param violations violations to add
+   * @param violations
+   *          violations to add
    */
   public void add(Violations violations) {
     Set<Entry<CVSKey,ConstraintViolationSummary>> es = violations.cvsmap.entrySet();
-    
+
     for (Entry<CVSKey,ConstraintViolationSummary> entry : es) {
       add(entry.getKey(), entry.getValue());
     }
-    
+
   }
-  
+
   /**
    * Adds a list of violations.
    *
-   * @param cvsList list of violation summaries
+   * @param cvsList
+   *          list of violation summaries
    */
   public void add(List<ConstraintViolationSummary> cvsList) {
     for (ConstraintViolationSummary constraintViolationSummary : cvsList) {
       add(constraintViolationSummary);
     }
-    
+
   }
-  
+
   /**
    * Gets the violations as a list of summaries.
    *
@@ -128,5 +130,5 @@ public class Violations {
   public List<ConstraintViolationSummary> asList() {
     return new ArrayList<ConstraintViolationSummary>(cvsmap.values());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/ArrayByteSequence.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/ArrayByteSequence.java b/core/src/main/java/org/apache/accumulo/core/data/ArrayByteSequence.java
index 954fef3..31b1204 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/ArrayByteSequence.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/ArrayByteSequence.java
@@ -25,67 +25,67 @@ import java.nio.ByteBuffer;
  * An implementation of {@link ByteSequence} that uses a backing byte array.
  */
 public class ArrayByteSequence extends ByteSequence implements Serializable {
-  
+
   private static final long serialVersionUID = 1L;
 
   protected byte data[];
   protected int offset;
   protected int length;
-  
+
   /**
-   * Creates a new sequence. The given byte array is used directly as the
-   * backing array, so later changes made to the array reflect into the new
-   * sequence.
+   * Creates a new sequence. The given byte array is used directly as the backing array, so later changes made to the array reflect into the new sequence.
    *
-   * @param data byte data
+   * @param data
+   *          byte data
    */
   public ArrayByteSequence(byte data[]) {
     this.data = data;
     this.offset = 0;
     this.length = data.length;
   }
-  
+
   /**
-   * Creates a new sequence from a subsequence of the given byte array. The
-   * given byte array is used directly as the backing array, so later changes
-   * made to the (relevant portion of the) array reflect into the new sequence.
+   * Creates a new sequence from a subsequence of the given byte array. The given byte array is used directly as the backing array, so later changes made to the
+   * (relevant portion of the) array reflect into the new sequence.
    *
-   * @param data byte data
-   * @param offset starting offset in byte array (inclusive)
-   * @param length number of bytes to include in sequence
-   * @throws IllegalArgumentException if the offset or length are out of bounds
-   * for the given byte array
+   * @param data
+   *          byte data
+   * @param offset
+   *          starting offset in byte array (inclusive)
+   * @param length
+   *          number of bytes to include in sequence
+   * @throws IllegalArgumentException
+   *           if the offset or length are out of bounds for the given byte array
    */
   public ArrayByteSequence(byte data[], int offset, int length) {
-    
+
     if (offset < 0 || offset > data.length || length < 0 || (offset + length) > data.length) {
       throw new IllegalArgumentException(" Bad offset and/or length data.length = " + data.length + " offset = " + offset + " length = " + length);
     }
-    
+
     this.data = data;
     this.offset = offset;
     this.length = length;
-    
+
   }
-  
+
   /**
-   * Creates a new sequence from the given string. The bytes are determined from
-   * the string using the default platform encoding.
+   * Creates a new sequence from the given string. The bytes are determined from the string using the default platform encoding.
    *
-   * @param s string to represent as bytes
+   * @param s
+   *          string to represent as bytes
    */
   public ArrayByteSequence(String s) {
     this(s.getBytes(UTF_8));
   }
-  
+
   /**
-   * Creates a new sequence based on a byte buffer. If the byte buffer has an
-   * array, that array (and the buffer's offset and limit) are used; otherwise,
-   * a new backing array is created and a relative bulk get is performed to
-   * transfer the buffer's contents (starting at its current position and
-   * not beyond its limit).
+   * Creates a new sequence based on a byte buffer. If the byte buffer has an array, that array (and the buffer's offset and limit) are used; otherwise, a new
+   * backing array is created and a relative bulk get is performed to transfer the buffer's contents (starting at its current position and not beyond its
+   * limit).
    *
-   * @param buffer byte buffer
+   * @param buffer
+   *          byte buffer
    */
   public ArrayByteSequence(ByteBuffer buffer) {
     this.length = buffer.remaining();
@@ -102,58 +102,58 @@ public class ArrayByteSequence extends ByteSequence implements Serializable {
 
   @Override
   public byte byteAt(int i) {
-    
+
     if (i < 0) {
       throw new IllegalArgumentException("i < 0, " + i);
     }
-    
+
     if (i >= length) {
       throw new IllegalArgumentException("i >= length, " + i + " >= " + length);
     }
-    
+
     return data[offset + i];
   }
-  
+
   @Override
   public byte[] getBackingArray() {
     return data;
   }
-  
+
   @Override
   public boolean isBackedByArray() {
     return true;
   }
-  
+
   @Override
   public int length() {
     return length;
   }
-  
+
   @Override
   public int offset() {
     return offset;
   }
-  
+
   @Override
   public ByteSequence subSequence(int start, int end) {
-    
+
     if (start > end || start < 0 || end > length) {
       throw new IllegalArgumentException("Bad start and/end start = " + start + " end=" + end + " offset=" + offset + " length=" + length);
     }
-    
+
     return new ArrayByteSequence(data, offset + start, end - start);
   }
-  
+
   @Override
   public byte[] toArray() {
     if (offset == 0 && length == data.length)
       return data;
-    
+
     byte[] copy = new byte[length];
     System.arraycopy(data, offset, copy, 0, length);
     return copy;
   }
-  
+
   public String toString() {
     return new String(data, offset, length, UTF_8);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/ByteSequence.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/ByteSequence.java b/core/src/main/java/org/apache/accumulo/core/data/ByteSequence.java
index 196cb11..ce8dab0 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/ByteSequence.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/ByteSequence.java
@@ -22,93 +22,95 @@ import org.apache.hadoop.io.WritableComparator;
  * A sequence of bytes.
  */
 public abstract class ByteSequence implements Comparable<ByteSequence> {
-  
+
   /**
    * Gets a byte within this sequence.
    *
-   * @param i index into sequence
+   * @param i
+   *          index into sequence
    * @return byte
-   * @throws IllegalArgumentException if i is out of range
+   * @throws IllegalArgumentException
+   *           if i is out of range
    */
   public abstract byte byteAt(int i);
-  
+
   /**
    * Gets the length of this sequence.
    *
    * @return sequence length
    */
   public abstract int length();
-  
+
   /**
    * Returns a portion of this sequence.
    *
-   * @param start index of subsequence start (inclusive)
-   * @param end index of subsequence end (exclusive)
+   * @param start
+   *          index of subsequence start (inclusive)
+   * @param end
+   *          index of subsequence end (exclusive)
    */
   public abstract ByteSequence subSequence(int start, int end);
-  
+
   /**
-   * Returns a byte array containing the bytes in this sequence. This method
-   * may copy the sequence data or may return a backing byte array directly.
+   * Returns a byte array containing the bytes in this sequence. This method may copy the sequence data or may return a backing byte array directly.
    *
    * @return byte array
    */
   public abstract byte[] toArray();
-  
+
   /**
    * Determines whether this sequence is backed by a byte array.
    *
    * @return true if sequence is backed by a byte array
    */
   public abstract boolean isBackedByArray();
-  
+
   /**
    * Gets the backing byte array for this sequence.
    *
    * @return byte array
    */
   public abstract byte[] getBackingArray();
-  
+
   /**
-   * Gets the offset for this sequence. This value represents the starting
-   * point for the sequence in the backing array, if there is one.
+   * Gets the offset for this sequence. This value represents the starting point for the sequence in the backing array, if there is one.
    *
    * @return offset (inclusive)
    */
   public abstract int offset();
-  
+
   /**
-   * Compares the two given byte sequences, byte by byte, returning a negative,
-   * zero, or positive result if the first sequence is less than, equal to, or
-   * greater than the second. The comparison is performed starting with the
-   * first byte of each sequence, and proceeds until a pair of bytes differs,
-   * or one sequence runs out of byte (is shorter). A shorter sequence is
-   * considered less than a longer one.
+   * Compares the two given byte sequences, byte by byte, returning a negative, zero, or positive result if the first sequence is less than, equal to, or
+   * greater than the second. The comparison is performed starting with the first byte of each sequence, and proceeds until a pair of bytes differs, or one
+   * sequence runs out of byte (is shorter). A shorter sequence is considered less than a longer one.
    *
-   * @param bs1 first byte sequence to compare
-   * @param bs2 second byte sequence to compare
+   * @param bs1
+   *          first byte sequence to compare
+   * @param bs2
+   *          second byte sequence to compare
    * @return comparison result
    */
   public static int compareBytes(ByteSequence bs1, ByteSequence bs2) {
-    
+
     int minLen = Math.min(bs1.length(), bs2.length());
-    
+
     for (int i = 0; i < minLen; i++) {
       int a = (bs1.byteAt(i) & 0xff);
       int b = (bs2.byteAt(i) & 0xff);
-      
+
       if (a != b) {
         return a - b;
       }
     }
-    
+
     return bs1.length() - bs2.length();
   }
-  
+
   /**
    * Compares this byte sequence to another.
    *
-   * @param obs byte sequence to compare
+   * @param obs
+   *          byte sequence to compare
    * @return comparison result
    * @see #compareBytes(ByteSequence, ByteSequence)
    */
@@ -116,28 +118,28 @@ public abstract class ByteSequence implements Comparable<ByteSequence> {
     if (isBackedByArray() && obs.isBackedByArray()) {
       return WritableComparator.compareBytes(getBackingArray(), offset(), length(), obs.getBackingArray(), obs.offset(), obs.length());
     }
-    
+
     return compareBytes(this, obs);
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof ByteSequence) {
       ByteSequence obs = (ByteSequence) o;
-      
+
       if (this == o)
         return true;
-      
+
       if (length() != obs.length())
         return false;
-      
+
       return compareTo(obs) == 0;
     }
-    
+
     return false;
-    
+
   }
-  
+
   @Override
   public int hashCode() {
     int hash = 1;
@@ -152,5 +154,5 @@ public abstract class ByteSequence implements Comparable<ByteSequence> {
     }
     return hash;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/Column.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Column.java b/core/src/main/java/org/apache/accumulo/core/data/Column.java
index f259912..294226a 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Column.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Column.java
@@ -23,6 +23,7 @@ import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.IOException;
 import java.nio.ByteBuffer;
+
 import org.apache.accumulo.core.data.thrift.TColumn;
 import org.apache.hadoop.io.WritableComparable;
 import org.apache.hadoop.io.WritableComparator;
@@ -31,7 +32,7 @@ import org.apache.hadoop.io.WritableComparator;
  * A column, specified by family, qualifier, and visibility.
  */
 public class Column implements WritableComparable<Column> {
-  
+
   static private int compareBytes(byte[] a, byte[] b) {
     if (a == null && b == null)
       return 0;
@@ -41,12 +42,12 @@ public class Column implements WritableComparable<Column> {
       return 1;
     return WritableComparator.compareBytes(a, 0, a.length, b, 0, b.length);
   }
-  
+
   /**
-   * Compares this column to another. Column families are compared first, then
-   * qualifiers, then visibilities.
+   * Compares this column to another. Column families are compared first, then qualifiers, then visibilities.
    *
-   * @param that column to compare
+   * @param that
+   *          column to compare
    * @return comparison result
    */
   public int compareTo(Column that) {
@@ -59,7 +60,7 @@ public class Column implements WritableComparable<Column> {
       return result;
     return compareBytes(this.columnVisibility, that.columnVisibility);
   }
-  
+
   public void readFields(DataInput in) throws IOException {
     if (in.readBoolean()) {
       int len = in.readInt();
@@ -68,7 +69,7 @@ public class Column implements WritableComparable<Column> {
     } else {
       columnFamily = null;
     }
-    
+
     if (in.readBoolean()) {
       int len = in.readInt();
       columnQualifier = new byte[len];
@@ -76,7 +77,7 @@ public class Column implements WritableComparable<Column> {
     } else {
       columnQualifier = null;
     }
-    
+
     if (in.readBoolean()) {
       int len = in.readInt();
       columnVisibility = new byte[len];
@@ -85,7 +86,7 @@ public class Column implements WritableComparable<Column> {
       columnVisibility = null;
     }
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     if (columnFamily == null) {
@@ -95,7 +96,7 @@ public class Column implements WritableComparable<Column> {
       out.writeInt(columnFamily.length);
       out.write(columnFamily);
     }
-    
+
     if (columnQualifier == null) {
       out.writeBoolean(false);
     } else {
@@ -103,7 +104,7 @@ public class Column implements WritableComparable<Column> {
       out.writeInt(columnQualifier.length);
       out.write(columnQualifier);
     }
-    
+
     if (columnVisibility == null) {
       out.writeBoolean(false);
     } else {
@@ -112,39 +113,43 @@ public class Column implements WritableComparable<Column> {
       out.write(columnVisibility);
     }
   }
-  
+
   public byte[] columnFamily;
   public byte[] columnQualifier;
   public byte[] columnVisibility;
-  
+
   /**
    * Creates a new blank column.
    */
   public Column() {}
-  
+
   /**
-  * Creates a new column.
-  *
-  * @param columnFamily family
-  * @param columnQualifier qualifier
-  * @param columnVisibility visibility
-  */
+   * Creates a new column.
+   *
+   * @param columnFamily
+   *          family
+   * @param columnQualifier
+   *          qualifier
+   * @param columnVisibility
+   *          visibility
+   */
   public Column(byte[] columnFamily, byte[] columnQualifier, byte[] columnVisibility) {
     this();
     this.columnFamily = columnFamily;
     this.columnQualifier = columnQualifier;
     this.columnVisibility = columnVisibility;
   }
-  
+
   /**
    * Creates a new column.
    *
-   * @param tcol Thrift column
+   * @param tcol
+   *          Thrift column
    */
   public Column(TColumn tcol) {
     this(toBytes(tcol.columnFamily), toBytes(tcol.columnQualifier), toBytes(tcol.columnVisibility));
   }
-  
+
   @Override
   public boolean equals(Object that) {
     if (that == null)
@@ -153,29 +158,30 @@ public class Column implements WritableComparable<Column> {
       return this.equals((Column) that);
     return false;
   }
-  
+
   /**
    * Checks if this column equals another.
    *
-   * @param that column to compare
+   * @param that
+   *          column to compare
    * @return true if this column equals that, false otherwise
    */
   public boolean equals(Column that) {
     return this.compareTo(that) == 0;
   }
-  
+
   private static int hash(byte[] b) {
     if (b == null)
       return 0;
-    
+
     return WritableComparator.hashBytes(b, b.length);
   }
-  
+
   @Override
   public int hashCode() {
     return hash(columnFamily) + hash(columnQualifier) + hash(columnVisibility);
   }
-  
+
   /**
    * Gets the column family. Not a defensive copy.
    *
@@ -184,7 +190,7 @@ public class Column implements WritableComparable<Column> {
   public byte[] getColumnFamily() {
     return columnFamily;
   }
-  
+
   /**
    * Gets the column qualifier. Not a defensive copy.
    *
@@ -193,7 +199,7 @@ public class Column implements WritableComparable<Column> {
   public byte[] getColumnQualifier() {
     return columnQualifier;
   }
-  
+
   /**
    * Gets the column visibility. Not a defensive copy.
    *
@@ -202,19 +208,19 @@ public class Column implements WritableComparable<Column> {
   public byte[] getColumnVisibility() {
     return columnVisibility;
   }
-  
+
   /**
-   * Gets a string representation of this column. The family, qualifier, and
-   * visibility are interpreted as strings using the platform default encoding;
-   * nulls are interpreted as empty strings.
+   * Gets a string representation of this column. The family, qualifier, and visibility are interpreted as strings using the platform default encoding; nulls
+   * are interpreted as empty strings.
    *
    * @return string form of column
    */
   public String toString() {
-    return new String(columnFamily == null ? new byte[0] : columnFamily, UTF_8) + ":" + new String(columnQualifier == null ? new byte[0] : columnQualifier, UTF_8) + ":"
+    return new String(columnFamily == null ? new byte[0] : columnFamily, UTF_8) + ":"
+        + new String(columnQualifier == null ? new byte[0] : columnQualifier, UTF_8) + ":"
         + new String(columnVisibility == null ? new byte[0] : columnVisibility, UTF_8);
   }
-  
+
   /**
    * Converts this column to Thrift.
    *
@@ -224,5 +230,5 @@ public class Column implements WritableComparable<Column> {
     return new TColumn(columnFamily == null ? null : ByteBuffer.wrap(columnFamily), columnQualifier == null ? null : ByteBuffer.wrap(columnQualifier),
         columnVisibility == null ? null : ByteBuffer.wrap(columnVisibility));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/ColumnUpdate.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/ColumnUpdate.java b/core/src/main/java/org/apache/accumulo/core/data/ColumnUpdate.java
index 2768b35..4f52214 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/ColumnUpdate.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/ColumnUpdate.java
@@ -20,10 +20,10 @@ import java.util.Arrays;
 
 /**
  * A single column and value pair within a {@link Mutation}.
- * 
+ *
  */
 public class ColumnUpdate {
-  
+
   private byte[] columnFamily;
   private byte[] columnQualifier;
   private byte[] columnVisibility;
@@ -31,17 +31,24 @@ public class ColumnUpdate {
   private boolean hasTimestamp;
   private byte[] val;
   private boolean deleted;
-  
+
   /**
    * Creates a new column update.
    *
-   * @param cf column family
-   * @param cq column qualifier
-   * @param cv column visibility
-   * @param hasts true if the update specifies a timestamp
-   * @param ts timestamp
-   * @param deleted delete marker
-   * @param val cell value
+   * @param cf
+   *          column family
+   * @param cq
+   *          column qualifier
+   * @param cv
+   *          column visibility
+   * @param hasts
+   *          true if the update specifies a timestamp
+   * @param ts
+   *          timestamp
+   * @param deleted
+   *          delete marker
+   * @param val
+   *          cell value
    */
   public ColumnUpdate(byte[] cf, byte[] cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
     this.columnFamily = cf;
@@ -52,7 +59,7 @@ public class ColumnUpdate {
     this.deleted = deleted;
     this.val = val;
   }
-  
+
   /**
    * Gets whether this update specifies a timestamp.
    *
@@ -61,7 +68,7 @@ public class ColumnUpdate {
   public boolean hasTimestamp() {
     return hasTimestamp;
   }
-  
+
   /**
    * Gets the column family for this update. Not a defensive copy.
    *
@@ -70,7 +77,7 @@ public class ColumnUpdate {
   public byte[] getColumnFamily() {
     return columnFamily;
   }
-  
+
   /**
    * Gets the column qualifier for this update. Not a defensive copy.
    *
@@ -79,7 +86,7 @@ public class ColumnUpdate {
   public byte[] getColumnQualifier() {
     return columnQualifier;
   }
-  
+
   /**
    * Gets the column visibility for this update.
    *
@@ -88,7 +95,7 @@ public class ColumnUpdate {
   public byte[] getColumnVisibility() {
     return columnVisibility;
   }
-  
+
   /**
    * Gets the timestamp for this update.
    *
@@ -97,7 +104,7 @@ public class ColumnUpdate {
   public long getTimestamp() {
     return this.timestamp;
   }
-  
+
   /**
    * Gets the delete marker for this update.
    *
@@ -106,7 +113,7 @@ public class ColumnUpdate {
   public boolean isDeleted() {
     return this.deleted;
   }
-  
+
   /**
    * Gets the cell value for this update.
    *
@@ -115,13 +122,13 @@ public class ColumnUpdate {
   public byte[] getValue() {
     return this.val;
   }
-  
+
   @Override
   public String toString() {
-    return Arrays.toString(columnFamily) + ":" + Arrays.toString(columnQualifier) + " ["
-        + Arrays.toString(columnVisibility) + "] " + (hasTimestamp ? timestamp : "NO_TIME_STAMP") + " " + Arrays.toString(val) + " " + deleted;
+    return Arrays.toString(columnFamily) + ":" + Arrays.toString(columnQualifier) + " [" + Arrays.toString(columnVisibility) + "] "
+        + (hasTimestamp ? timestamp : "NO_TIME_STAMP") + " " + Arrays.toString(val) + " " + deleted;
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     if (!(obj instanceof ColumnUpdate))
@@ -131,7 +138,7 @@ public class ColumnUpdate {
         && Arrays.equals(getColumnVisibility(), upd.getColumnVisibility()) && isDeleted() == upd.isDeleted() && Arrays.equals(getValue(), upd.getValue())
         && hasTimestamp() == upd.hasTimestamp() && getTimestamp() == upd.getTimestamp();
   }
-  
+
   @Override
   public int hashCode() {
     return Arrays.hashCode(columnFamily) + Arrays.hashCode(columnQualifier) + Arrays.hashCode(columnVisibility)


[31/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/ColumnFQ.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/ColumnFQ.java b/core/src/main/java/org/apache/accumulo/core/util/ColumnFQ.java
index 8826bb1..11d97e5 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/ColumnFQ.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/ColumnFQ.java
@@ -27,48 +27,48 @@ import org.apache.hadoop.io.Text;
 public class ColumnFQ implements Comparable<ColumnFQ> {
   private Text colf;
   private Text colq;
-  
+
   public ColumnFQ(Text colf, Text colq) {
     if (colf == null || colq == null) {
       throw new IllegalArgumentException();
     }
-    
+
     this.colf = colf;
     this.colq = colq;
   }
-  
+
   public ColumnFQ(Key k) {
     this(k.getColumnFamily(), k.getColumnQualifier());
   }
-  
+
   public ColumnFQ(ColumnUpdate cu) {
     this(new Text(cu.getColumnFamily()), new Text(cu.getColumnQualifier()));
   }
-  
+
   public Text getColumnQualifier() {
     return colq;
   }
-  
+
   public Text getColumnFamily() {
     return colf;
   }
-  
+
   public Column toColumn() {
     return new Column(TextUtil.getBytes(colf), TextUtil.getBytes(colq), null);
   }
-  
+
   public void fetch(ScannerBase sb) {
     sb.fetchColumn(colf, colq);
   }
-  
+
   public void put(Mutation m, Value v) {
     m.put(colf, colq, v);
   }
-  
+
   public void putDelete(Mutation m) {
     m.putDelete(colf, colq);
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (!(o instanceof ColumnFQ))
@@ -78,33 +78,33 @@ public class ColumnFQ implements Comparable<ColumnFQ> {
     ColumnFQ ocfq = (ColumnFQ) o;
     return ocfq.colf.equals(colf) && ocfq.colq.equals(colq);
   }
-  
+
   @Override
   public int hashCode() {
     return colf.hashCode() + colq.hashCode();
   }
-  
+
   public boolean hasColumns(Key key) {
     return key.compareColumnFamily(colf) == 0 && key.compareColumnQualifier(colq) == 0;
   }
-  
+
   public boolean equals(Text colf, Text colq) {
     return this.colf.equals(colf) && this.colq.equals(colq);
   }
-  
+
   @Override
   public int compareTo(ColumnFQ o) {
     int cmp = colf.compareTo(o.colf);
-    
+
     if (cmp == 0)
       cmp = colq.compareTo(o.colq);
-    
+
     return cmp;
   }
-  
+
   @Override
   public String toString() {
     return colf + ":" + colq;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/ComparablePair.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/ComparablePair.java b/core/src/main/java/org/apache/accumulo/core/util/ComparablePair.java
index 2fc38b6..6663032 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/ComparablePair.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/ComparablePair.java
@@ -17,22 +17,22 @@
 package org.apache.accumulo.core.util;
 
 /**
- * 
+ *
  */
 public class ComparablePair<A extends Comparable<A>,B extends Comparable<B>> extends Pair<A,B> implements Comparable<ComparablePair<A,B>> {
-  
+
   public ComparablePair(A f, B s) {
     super(f, s);
   }
-  
+
   @Override
   public int compareTo(ComparablePair<A,B> abPair) {
     int cmp = first.compareTo(abPair.first);
     if (cmp == 0) {
       cmp = second.compareTo(abPair.second);
     }
-    
+
     return cmp;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/CreateToken.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/CreateToken.java b/core/src/main/java/org/apache/accumulo/core/util/CreateToken.java
index cfac8fe..79b241c 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/CreateToken.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/CreateToken.java
@@ -22,6 +22,7 @@ import java.io.File;
 import java.io.FileOutputStream;
 import java.io.IOException;
 import java.io.PrintStream;
+
 import jline.console.ConsoleReader;
 
 import org.apache.accumulo.core.cli.ClientOpts.Password;
@@ -32,53 +33,52 @@ import org.apache.accumulo.core.client.security.tokens.AuthenticationToken.Authe
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken.Properties;
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken.TokenProperty;
 import org.apache.accumulo.core.client.security.tokens.PasswordToken;
-import org.apache.accumulo.core.util.Base64;
 
 import com.beust.jcommander.Parameter;
 
 public class CreateToken {
-  
+
   private static ConsoleReader reader = null;
-  
+
   private static ConsoleReader getConsoleReader() throws IOException {
     if (reader == null)
       reader = new ConsoleReader();
     return reader;
   }
-  
+
   static class Opts extends Help {
     @Parameter(names = {"-u", "--user"}, description = "Connection user")
     public String principal = null;
-    
+
     @Parameter(names = "-p", converter = PasswordConverter.class, description = "Connection password")
     public Password password = null;
-    
+
     @Parameter(names = "--password", converter = PasswordConverter.class, description = "Enter the connection password", password = true)
     public Password securePassword = null;
-    
+
     @Parameter(names = {"-tc", "--tokenClass"}, description = "The class of the authentication token")
     public String tokenClassName = PasswordToken.class.getName();
-    
+
     @Parameter(names = {"-f", "--file"}, description = "The filename to save the auth token to. Multiple tokens can be stored in the same file,"
         + " but only the first for each user will be recognized.")
     public String tokenFile = null;
   }
-  
+
   public static void main(String[] args) {
     Opts opts = new Opts();
     opts.parseArgs(CreateToken.class.getName(), args);
-    
+
     Password pass = opts.password;
     if (pass == null && opts.securePassword != null) {
       pass = opts.securePassword;
     }
-    
+
     try {
       String principal = opts.principal;
       if (principal == null) {
         principal = getConsoleReader().readLine("Username (aka principal): ");
       }
-      
+
       AuthenticationToken token = Class.forName(opts.tokenClassName).asSubclass(AuthenticationToken.class).newInstance();
       Properties props = new Properties();
       for (TokenProperty tp : token.getProperties()) {
@@ -96,7 +96,7 @@ public class CreateToken {
         token.init(props);
       }
       String tokenBase64 = Base64.encodeBase64String(AuthenticationTokenSerializer.serialize(token));
-      
+
       String tokenFile = opts.tokenFile;
       if (tokenFile == null) {
         tokenFile = getConsoleReader().readLine("File to save auth token to: ");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/Daemon.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/Daemon.java b/core/src/main/java/org/apache/accumulo/core/util/Daemon.java
index 7ce46eb..a2c9e79 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/Daemon.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/Daemon.java
@@ -17,44 +17,44 @@
 package org.apache.accumulo.core.util;
 
 public class Daemon extends Thread {
-  
+
   public Daemon() {
     setDaemon(true);
   }
-  
+
   public Daemon(Runnable target) {
     super(target);
     setDaemon(true);
   }
-  
+
   public Daemon(String name) {
     super(name);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, Runnable target) {
     super(group, target);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, String name) {
     super(group, name);
     setDaemon(true);
   }
-  
+
   public Daemon(Runnable target, String name) {
     super(target, name);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, Runnable target, String name) {
     super(group, target, name);
     setDaemon(true);
   }
-  
+
   public Daemon(ThreadGroup group, Runnable target, String name, long stackSize) {
     super(group, target, name, stackSize);
     setDaemon(true);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/Duration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/Duration.java b/core/src/main/java/org/apache/accumulo/core/util/Duration.java
index 91ae089..b1b8572 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/Duration.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/Duration.java
@@ -17,15 +17,15 @@
 package org.apache.accumulo.core.util;
 
 public class Duration {
-  
+
   public static String format(long time) {
     return format(time, "&nbsp;");
   }
-  
+
   public static String format(long time, String space) {
     return format(time, space, "&mdash;");
   }
-  
+
   public static String format(long time, String space, String zero) {
     long ms, sec, min, hr, day, yr;
     ms = sec = min = hr = day = yr = -1;
@@ -53,7 +53,7 @@ public class Duration {
       return String.format("%dd" + space + "%dh", day, hr);
     yr = time;
     return String.format("%dy" + space + "%dd", yr, day);
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/Encoding.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/Encoding.java b/core/src/main/java/org/apache/accumulo/core/util/Encoding.java
index 259f783..524f377 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/Encoding.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/Encoding.java
@@ -18,27 +18,26 @@ package org.apache.accumulo.core.util;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
 
-import org.apache.accumulo.core.util.Base64;
 import org.apache.hadoop.io.Text;
 
 public class Encoding {
-  
+
   public static String encodeAsBase64FileName(Text data) {
     String encodedRow = Base64.encodeBase64URLSafeString(TextUtil.getBytes(data));
-    
+
     int index = encodedRow.length() - 1;
     while (index >= 0 && encodedRow.charAt(index) == '=')
       index--;
-    
+
     encodedRow = encodedRow.substring(0, index + 1);
     return encodedRow;
   }
-  
+
   public static byte[] decodeBase64FileName(String node) {
     while (node.length() % 4 != 0)
       node += "=";
     /* decode transparently handles URLSafe encodings */
     return Base64.decodeBase64(node.getBytes(UTF_8));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/FastFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/FastFormat.java b/core/src/main/java/org/apache/accumulo/core/util/FastFormat.java
index e103ac6..8d76f27 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/FastFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/FastFormat.java
@@ -26,28 +26,28 @@ public class FastFormat {
       throw new RuntimeException(" Did not format to expected width " + num + " " + width + " " + radix + " " + new String(prefix, UTF_8));
     return ret;
   }
-  
+
   public static int toZeroPaddedString(byte output[], int outputOffset, long num, int width, int radix, byte[] prefix) {
     if (num < 0)
       throw new IllegalArgumentException();
-    
+
     String s = Long.toString(num, radix);
-    
+
     int index = outputOffset;
-    
+
     for (int i = 0; i < prefix.length; i++) {
       output[index++] = prefix[i];
     }
-    
+
     int end = width - s.length() + index;
-    
+
     while (index < end)
       output[index++] = '0';
-    
+
     for (int i = 0; i < s.length(); i++) {
       output[index++] = (byte) s.charAt(i);
     }
-    
+
     return index - outputOffset;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/LocalityGroupUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/LocalityGroupUtil.java b/core/src/main/java/org/apache/accumulo/core/util/LocalityGroupUtil.java
index 9696025..d590ecd 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/LocalityGroupUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/LocalityGroupUtil.java
@@ -43,11 +43,11 @@ import org.apache.hadoop.io.Text;
 import com.google.common.base.Joiner;
 
 public class LocalityGroupUtil {
-  
+
   // private static final Logger log = Logger.getLogger(ColumnFamilySet.class);
-  
+
   public static final Set<ByteSequence> EMPTY_CF_SET = Collections.emptySet();
-  
+
   public static Set<ByteSequence> families(Collection<Column> columns) {
     Set<ByteSequence> result = new HashSet<ByteSequence>(columns.size());
     for (Column col : columns) {
@@ -55,14 +55,14 @@ public class LocalityGroupUtil {
     }
     return result;
   }
-  
+
   @SuppressWarnings("serial")
   static public class LocalityGroupConfigurationError extends AccumuloException {
     LocalityGroupConfigurationError(String why) {
       super(why);
     }
   }
-  
+
   public static Map<String,Set<ByteSequence>> getLocalityGroups(AccumuloConfiguration acuconf) throws LocalityGroupConfigurationError {
     Map<String,Set<ByteSequence>> result = new HashMap<String,Set<ByteSequence>>();
     String[] groups = acuconf.get(Property.TABLE_LOCALITY_GROUPS).split(",");
@@ -87,7 +87,7 @@ public class LocalityGroupUtil {
               colFamsSet.retainAll(all);
               throw new LocalityGroupConfigurationError("Column families " + colFamsSet + " in group " + group + " is already used by another locality group");
             }
-            
+
             all.addAll(colFamsSet);
             result.put(group, colFamsSet);
           }
@@ -97,35 +97,35 @@ public class LocalityGroupUtil {
     // result.put("", all);
     return result;
   }
-  
+
   public static Set<ByteSequence> decodeColumnFamilies(String colFams) throws LocalityGroupConfigurationError {
     HashSet<ByteSequence> colFamsSet = new HashSet<ByteSequence>();
-    
+
     for (String family : colFams.split(",")) {
       ByteSequence cfbs = decodeColumnFamily(family);
       colFamsSet.add(cfbs);
     }
-    
+
     return colFamsSet;
   }
-  
+
   public static ByteSequence decodeColumnFamily(String colFam) throws LocalityGroupConfigurationError {
     byte output[] = new byte[colFam.length()];
     int pos = 0;
-    
+
     for (int i = 0; i < colFam.length(); i++) {
       char c = colFam.charAt(i);
-      
+
       if (c == '\\') {
         // next char must be 'x' or '\'
         i++;
-        
+
         if (i >= colFam.length()) {
           throw new LocalityGroupConfigurationError("Expected 'x' or '\' after '\'  in " + colFam);
         }
-        
+
         char nc = colFam.charAt(i);
-        
+
         switch (nc) {
           case '\\':
             output[pos++] = '\\';
@@ -142,36 +142,36 @@ public class LocalityGroupUtil {
       } else {
         output[pos++] = (byte) (0xff & c);
       }
-      
+
     }
-    
+
     return new ArrayByteSequence(output, 0, pos);
-    
+
   }
-  
+
   public static String encodeColumnFamilies(Set<Text> colFams) {
     SortedSet<String> ecfs = new TreeSet<String>();
-    
+
     StringBuilder sb = new StringBuilder();
-    
+
     for (Text text : colFams) {
       String ecf = encodeColumnFamily(sb, text.getBytes(), text.getLength());
       ecfs.add(ecf);
     }
-    
+
     return Joiner.on(",").join(ecfs);
   }
-  
+
   public static String encodeColumnFamily(ByteSequence bs) {
     if (bs.offset() != 0) {
       throw new IllegalArgumentException("The offset cannot be non-zero.");
     }
     return encodeColumnFamily(new StringBuilder(), bs.getBackingArray(), bs.length());
   }
-  
+
   private static String encodeColumnFamily(StringBuilder sb, byte[] ba, int len) {
     sb.setLength(0);
-    
+
     for (int i = 0; i < len; i++) {
       int c = 0xff & ba[i];
       if (c == '\\')
@@ -181,45 +181,45 @@ public class LocalityGroupUtil {
       else
         sb.append("\\x").append(String.format("%02X", c));
     }
-    
+
     String ecf = sb.toString();
     return ecf;
   }
-  
+
   private static class PartitionedMutation extends Mutation {
     private byte[] row;
     private List<ColumnUpdate> updates;
-    
+
     PartitionedMutation(byte[] row, List<ColumnUpdate> updates) {
       this.row = row;
       this.updates = updates;
     }
-    
+
     @Override
     public byte[] getRow() {
       return row;
     }
-    
+
     @Override
     public List<ColumnUpdate> getUpdates() {
       return updates;
     }
-    
+
     @Override
     public TMutation toThrift() {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public int hashCode() {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public boolean equals(Object o) {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public boolean equals(Mutation m) {
       throw new UnsupportedOperationException();
@@ -227,28 +227,28 @@ public class LocalityGroupUtil {
   }
 
   public static class Partitioner {
-    
+
     private Map<ByteSequence,Integer> colfamToLgidMap;
     private Map<ByteSequence,MutableLong>[] groups;
-    
+
     public Partitioner(Map<ByteSequence,MutableLong> groups[]) {
       this.groups = groups;
       this.colfamToLgidMap = new HashMap<ByteSequence,Integer>();
-      
+
       for (int i = 0; i < groups.length; i++) {
         for (ByteSequence cf : groups[i].keySet()) {
           colfamToLgidMap.put(cf, i);
         }
       }
     }
-    
+
     public void partition(List<Mutation> mutations, List<Mutation> partitionedMutations[]) {
 
       MutableByteSequence mbs = new MutableByteSequence(new byte[0], 0, 0);
-      
+
       @SuppressWarnings("unchecked")
       List<ColumnUpdate> parts[] = new List[groups.length + 1];
-      
+
       for (Mutation mutation : mutations) {
         if (mutation.getUpdates().size() == 1) {
           int lgid = getLgid(mbs, mutation.getUpdates().get(0));
@@ -257,7 +257,7 @@ public class LocalityGroupUtil {
           for (int i = 0; i < parts.length; i++) {
             parts[i] = null;
           }
-          
+
           int lgcount = 0;
 
           for (ColumnUpdate cu : mutation.getUpdates()) {
@@ -267,10 +267,10 @@ public class LocalityGroupUtil {
               parts[lgid] = new ArrayList<ColumnUpdate>();
               lgcount++;
             }
-            
+
             parts[lgid].add(cu);
           }
-          
+
           if (lgcount == 1) {
             for (int i = 0; i < parts.length; i++)
               if (parts[i] != null) {
@@ -285,7 +285,7 @@ public class LocalityGroupUtil {
         }
       }
     }
-    
+
     private Integer getLgid(MutableByteSequence mbs, ColumnUpdate cu) {
       mbs.setArray(cu.getColumnFamily(), 0, cu.getColumnFamily().length);
       Integer lgid = colfamToLgidMap.get(mbs);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/LoggingRunnable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/LoggingRunnable.java b/core/src/main/java/org/apache/accumulo/core/util/LoggingRunnable.java
index 5d486b9..081a433 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/LoggingRunnable.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/LoggingRunnable.java
@@ -23,7 +23,7 @@ import org.slf4j.Logger;
 public class LoggingRunnable implements Runnable {
   private Runnable runnable;
   private Logger log;
-  
+
   public LoggingRunnable(Logger log, Runnable r) {
     this.runnable = r;
     this.log = log;
@@ -40,7 +40,7 @@ public class LoggingRunnable implements Runnable {
         // maybe the logging system is screwed up OR there is a bug in the exception, like t.getMessage() throws a NPE
         System.err.println("ERROR " + new Date() + " Failed to log message about thread death " + t2.getMessage());
         t2.printStackTrace();
-        
+
         // try to print original exception
         System.err.println("ERROR " + new Date() + " Exception that failed to log : " + t.getMessage());
         t.printStackTrace();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/MapCounter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/MapCounter.java b/core/src/main/java/org/apache/accumulo/core/util/MapCounter.java
index 30f9f1d..f6f3ff7 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/MapCounter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/MapCounter.java
@@ -22,64 +22,64 @@ import java.util.HashMap;
 import java.util.Set;
 
 public class MapCounter<KT> {
-  
+
   static class MutableLong {
     long l = 0l;
   }
-  
+
   private HashMap<KT,MutableLong> map;
-  
+
   public MapCounter() {
     map = new HashMap<KT,MutableLong>();
   }
-  
+
   public long increment(KT key, long l) {
     MutableLong ml = map.get(key);
     if (ml == null) {
       ml = new MutableLong();
       map.put(key, ml);
     }
-    
+
     ml.l += l;
-    
+
     if (ml.l == 0) {
       map.remove(key);
     }
-    
+
     return ml.l;
   }
-  
+
   public long decrement(KT key, long l) {
     return increment(key, -1 * l);
   }
-  
+
   public boolean contains(KT key) {
     return map.containsKey(key);
   }
-  
+
   public long get(KT key) {
     MutableLong ml = map.get(key);
     if (ml == null) {
       return 0;
     }
-    
+
     return ml.l;
   }
-  
+
   public Set<KT> keySet() {
     return map.keySet();
   }
-  
+
   public Collection<Long> values() {
     Collection<MutableLong> vals = map.values();
     ArrayList<Long> ret = new ArrayList<Long>(vals.size());
     for (MutableLong ml : vals) {
       ret.add(ml.l);
     }
-    
+
     return ret;
   }
-  
+
   public int size() {
     return map.size();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/MonitorUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/MonitorUtil.java b/core/src/main/java/org/apache/accumulo/core/util/MonitorUtil.java
index 4a0f9ef..4497981 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/MonitorUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/MonitorUtil.java
@@ -28,6 +28,6 @@ public class MonitorUtil {
   public static String getLocation(Instance instance) throws KeeperException, InterruptedException {
     ZooReader zr = new ZooReader(instance.getZooKeepers(), 5000);
     byte[] loc = zr.getData(ZooUtil.getRoot(instance) + Constants.ZMONITOR_HTTP_ADDR, null);
-    return loc==null ? null : new String(loc, UTF_8);
+    return loc == null ? null : new String(loc, UTF_8);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/MutableByteSequence.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/MutableByteSequence.java b/core/src/main/java/org/apache/accumulo/core/util/MutableByteSequence.java
index 6db7170..f9a0183 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/MutableByteSequence.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/MutableByteSequence.java
@@ -19,28 +19,27 @@ package org.apache.accumulo.core.util;
 import org.apache.accumulo.core.data.ArrayByteSequence;
 import org.apache.accumulo.core.data.ByteSequence;
 
-
 public class MutableByteSequence extends ArrayByteSequence {
   private static final long serialVersionUID = 1L;
 
   public MutableByteSequence(byte[] data, int offset, int length) {
     super(data, offset, length);
   }
-  
+
   public MutableByteSequence(ByteSequence bs) {
     super(new byte[Math.max(64, bs.length())]);
     System.arraycopy(bs.getBackingArray(), bs.offset(), data, 0, bs.length());
     this.length = bs.length();
     this.offset = 0;
   }
-  
+
   public void setArray(byte[] data, int offset, int len) {
     this.data = data;
     this.offset = offset;
     this.length = len;
   }
-  
+
   public void setLength(int len) {
     this.length = len;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/NamingThreadFactory.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/NamingThreadFactory.java b/core/src/main/java/org/apache/accumulo/core/util/NamingThreadFactory.java
index c37d0af..ebe9002 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/NamingThreadFactory.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/NamingThreadFactory.java
@@ -28,13 +28,13 @@ public class NamingThreadFactory implements ThreadFactory {
 
   private AtomicInteger threadNum = new AtomicInteger(1);
   private String name;
-  
+
   public NamingThreadFactory(String name) {
     this.name = name;
   }
-  
+
   public Thread newThread(Runnable r) {
     return new Daemon(new LoggingRunnable(log, new TraceRunnable(r)), name + " " + threadNum.getAndIncrement());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/NumUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/NumUtil.java b/core/src/main/java/org/apache/accumulo/core/util/NumUtil.java
index ebc8c4f..ef94f0c 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/NumUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/NumUtil.java
@@ -21,7 +21,7 @@ import java.text.DecimalFormat;
 public class NumUtil {
 
   private static final String QUANTITY_SUFFIX[] = {"", "K", "M", "B", "T", "e15", "e18", "e21"};
-  private static final String SIZE_SUFFIX[]     = {"", "K", "M", "G", "T", "P", "E", "Z"};
+  private static final String SIZE_SUFFIX[] = {"", "K", "M", "G", "T", "P", "E", "Z"};
 
   private static DecimalFormat df = new DecimalFormat("#,###,##0");
   private static DecimalFormat df_mantissa = new DecimalFormat("#,###,##0.00");
@@ -39,14 +39,16 @@ public class NumUtil {
   }
 
   private static String bigNumber(long big, String[] SUFFIXES, long base) {
-    if (big < base) return df.format(big) + SUFFIXES[0];
+    if (big < base)
+      return df.format(big) + SUFFIXES[0];
     int exp = (int) (Math.log(big) / Math.log(base));
     double val = big / Math.pow(base, exp);
-    return df_mantissa.format(val) +  SUFFIXES[exp];
+    return df_mantissa.format(val) + SUFFIXES[exp];
   }
 
   private static String bigNumber(double big, String[] SUFFIXES, long base) {
-    if (big < base) return df_mantissa.format(big) + SUFFIXES[0];
+    if (big < base)
+      return df_mantissa.format(big) + SUFFIXES[0];
     int exp = (int) (Math.log(big) / Math.log(base));
     double val = big / Math.pow(base, exp);
     return df_mantissa.format(val) + SUFFIXES[exp];

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/OpTimer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/OpTimer.java b/core/src/main/java/org/apache/accumulo/core/util/OpTimer.java
index 205b043..564a824 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/OpTimer.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/OpTimer.java
@@ -27,12 +27,12 @@ public class OpTimer {
   private long t1;
   private long opid;
   private static AtomicLong nextOpid = new AtomicLong();
-  
+
   public OpTimer(Logger log, Level level) {
     this.log = log;
     this.level = level;
   }
-  
+
   public OpTimer start(String msg) {
     opid = nextOpid.getAndIncrement();
     if (log.isEnabledFor(level))
@@ -40,7 +40,7 @@ public class OpTimer {
     t1 = System.currentTimeMillis();
     return this;
   }
-  
+
   public void stop(String msg) {
     if (log.isEnabledFor(level)) {
       long t2 = System.currentTimeMillis();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/PeekingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/PeekingIterator.java b/core/src/main/java/org/apache/accumulo/core/util/PeekingIterator.java
index 04c2b86..6d25a0b 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/PeekingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/PeekingIterator.java
@@ -19,11 +19,11 @@ package org.apache.accumulo.core.util;
 import java.util.Iterator;
 
 public class PeekingIterator<E> implements Iterator<E> {
-  
+
   boolean isInitialized;
   Iterator<E> source;
   E top;
-  
+
   public PeekingIterator(Iterator<E> source) {
     this.source = source;
     if (source.hasNext())
@@ -32,14 +32,14 @@ public class PeekingIterator<E> implements Iterator<E> {
       top = null;
     isInitialized = true;
   }
-  
+
   /**
    * Creates an uninitialized instance. This should be used in conjunction with {@link #initialize(Iterator)}.
    */
   public PeekingIterator() {
     isInitialized = false;
   }
-  
+
   /**
    * Initializes this iterator, to be used with {@link #PeekingIterator()}.
    */
@@ -52,13 +52,13 @@ public class PeekingIterator<E> implements Iterator<E> {
     isInitialized = true;
     return this;
   }
-  
+
   public E peek() {
     if (!isInitialized)
       throw new IllegalStateException("Iterator has not yet been initialized");
     return top;
   }
-  
+
   @Override
   public E next() {
     if (!isInitialized)
@@ -70,12 +70,12 @@ public class PeekingIterator<E> implements Iterator<E> {
       top = null;
     return lastPeeked;
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public boolean hasNext() {
     if (!isInitialized)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/ServerServices.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/ServerServices.java b/core/src/main/java/org/apache/accumulo/core/util/ServerServices.java
index 88c4ebf..a07ef76 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/ServerServices.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/ServerServices.java
@@ -24,35 +24,35 @@ public class ServerServices implements Comparable<ServerServices> {
   public static enum Service {
     TSERV_CLIENT, GC_CLIENT;
   }
-  
+
   public static final String SERVICE_SEPARATOR = ";";
   public static final String SEPARATOR_CHAR = "=";
-  
+
   private EnumMap<Service,String> services;
   private String stringForm = null;
-  
+
   public ServerServices(String services) {
     this.services = new EnumMap<Service,String>(Service.class);
-    
+
     String[] addresses = services.split(SERVICE_SEPARATOR);
     for (String address : addresses) {
       String[] sa = address.split(SEPARATOR_CHAR, 2);
       this.services.put(Service.valueOf(sa[0]), sa[1]);
     }
   }
-  
+
   public ServerServices(String address, Service service) {
     this(service.name() + SEPARATOR_CHAR + address);
   }
-  
+
   public String getAddressString(Service service) {
     return services.get(service);
   }
-  
+
   public HostAndPort getAddress(Service service) {
     return AddressUtil.parseAddress(getAddressString(service), false);
   }
-  
+
   // DON'T CHANGE THIS; WE'RE USING IT FOR SERIALIZATION!!!
   @Override
   public String toString() {
@@ -69,19 +69,19 @@ public class ServerServices implements Comparable<ServerServices> {
     }
     return stringForm;
   }
-  
+
   @Override
   public int hashCode() {
     return toString().hashCode();
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof ServerServices)
       return toString().equals(((ServerServices) o).toString());
     return false;
   }
-  
+
   @Override
   public int compareTo(ServerServices other) {
     return toString().compareTo(other.toString());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/SimpleThreadPool.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/SimpleThreadPool.java b/core/src/main/java/org/apache/accumulo/core/util/SimpleThreadPool.java
index cbac519..a406233 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/SimpleThreadPool.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/SimpleThreadPool.java
@@ -20,15 +20,14 @@ import java.util.concurrent.LinkedBlockingQueue;
 import java.util.concurrent.ThreadPoolExecutor;
 import java.util.concurrent.TimeUnit;
 
-
 /**
  * Create a simple thread pool using common parameters.
  */
 public class SimpleThreadPool extends ThreadPoolExecutor {
-  
+
   public SimpleThreadPool(int max, final String name) {
     super(max, max, 4l, TimeUnit.SECONDS, new LinkedBlockingQueue<Runnable>(), new NamingThreadFactory(name));
     allowCoreThreadTimeOut(true);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/StopWatch.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/StopWatch.java b/core/src/main/java/org/apache/accumulo/core/util/StopWatch.java
index 4f30d4a..ddb612f 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/StopWatch.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/StopWatch.java
@@ -21,62 +21,62 @@ import java.util.EnumMap;
 public class StopWatch<K extends Enum<K>> {
   EnumMap<K,Long> startTime;
   EnumMap<K,Long> totalTime;
-  
+
   public StopWatch(Class<K> k) {
     startTime = new EnumMap<K,Long>(k);
     totalTime = new EnumMap<K,Long>(k);
   }
-  
+
   public synchronized void start(K timer) {
     if (startTime.containsKey(timer)) {
       throw new IllegalStateException(timer + " already started");
     }
     startTime.put(timer, System.currentTimeMillis());
   }
-  
+
   public synchronized void stopIfActive(K timer) {
     if (startTime.containsKey(timer))
       stop(timer);
   }
-  
+
   public synchronized void stop(K timer) {
-    
+
     Long st = startTime.get(timer);
-    
+
     if (st == null) {
       throw new IllegalStateException(timer + " not started");
     }
-    
+
     Long existingTime = totalTime.get(timer);
     if (existingTime == null)
       existingTime = 0L;
-    
+
     totalTime.put(timer, existingTime + (System.currentTimeMillis() - st));
     startTime.remove(timer);
   }
-  
+
   public synchronized void reset(K timer) {
     totalTime.remove(timer);
   }
-  
+
   public synchronized long get(K timer) {
     Long existingTime = totalTime.get(timer);
     if (existingTime == null)
       existingTime = 0L;
     return existingTime;
   }
-  
+
   public synchronized double getSecs(K timer) {
     Long existingTime = totalTime.get(timer);
     if (existingTime == null)
       existingTime = 0L;
     return existingTime / 1000.0;
   }
-  
+
   public synchronized void print() {
     for (K timer : totalTime.keySet()) {
       System.out.printf("%20s : %,6.4f secs%n", timer.toString(), get(timer) / 1000.0);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/TextUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/TextUtil.java b/core/src/main/java/org/apache/accumulo/core/util/TextUtil.java
index 66ad8f5..d2cd6cc 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/TextUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/TextUtil.java
@@ -19,6 +19,7 @@ package org.apache.accumulo.core.util;
 import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.nio.ByteBuffer;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.hadoop.io.Text;
 
@@ -31,14 +32,14 @@ public final class TextUtil {
     }
     return bytes;
   }
-  
+
   public static ByteBuffer getByteBuffer(Text text) {
     if (text == null)
       return null;
     byte[] bytes = text.getBytes();
     return ByteBuffer.wrap(bytes, 0, text.getLength());
   }
-  
+
   public static Text truncate(Text text, int maxLen) {
     if (text.getLength() > maxLen) {
       Text newText = new Text();
@@ -47,10 +48,10 @@ public final class TextUtil {
       newText.append(suffix.getBytes(UTF_8), 0, suffix.length());
       return newText;
     }
-    
+
     return text;
   }
-  
+
   public static Text truncate(Text row) {
     return truncate(row, Constants.MAX_DATA_TO_PRINT);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/UnsynchronizedBuffer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/UnsynchronizedBuffer.java b/core/src/main/java/org/apache/accumulo/core/util/UnsynchronizedBuffer.java
index 36aa473..e07ee10 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/UnsynchronizedBuffer.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/UnsynchronizedBuffer.java
@@ -21,8 +21,7 @@ import java.nio.ByteBuffer;
 import org.apache.hadoop.io.WritableUtils;
 
 /**
- * A utility class for reading and writing bytes to byte buffers without
- * synchronization.
+ * A utility class for reading and writing bytes to byte buffers without synchronization.
  */
 public class UnsynchronizedBuffer {
   // created this little class instead of using ByteArrayOutput stream and DataOutputStream
@@ -31,26 +30,27 @@ public class UnsynchronizedBuffer {
    * A byte buffer writer.
    */
   public static class Writer {
-    
+
     int offset = 0;
     byte data[];
-    
+
     /**
      * Creates a new writer.
      */
     public Writer() {
       data = new byte[64];
     }
-    
+
     /**
      * Creates a new writer.
      *
-     * @param initialCapacity initial byte capacity
+     * @param initialCapacity
+     *          initial byte capacity
      */
     public Writer(int initialCapacity) {
       data = new byte[initialCapacity];
     }
-    
+
     private void reserve(int l) {
       if (offset + l > data.length) {
         int newSize = UnsynchronizedBuffer.nextArraySize(offset + l);
@@ -59,27 +59,32 @@ public class UnsynchronizedBuffer {
         System.arraycopy(data, 0, newData, 0, offset);
         data = newData;
       }
-      
+
     }
-    
+
     /**
      * Adds bytes to this writer's buffer.
      *
-     * @param bytes byte array
-     * @param off offset into array to start copying bytes
-     * @param length number of bytes to add
-     * @throws IndexOutOfBoundsException if off or length are invalid
+     * @param bytes
+     *          byte array
+     * @param off
+     *          offset into array to start copying bytes
+     * @param length
+     *          number of bytes to add
+     * @throws IndexOutOfBoundsException
+     *           if off or length are invalid
      */
     public void add(byte[] bytes, int off, int length) {
       reserve(length);
       System.arraycopy(bytes, off, data, offset, length);
       offset += length;
     }
-    
+
     /**
      * Adds a Boolean value to this writer's buffer.
      *
-     * @param b Boolean value
+     * @param b
+     *          Boolean value
      */
     public void add(boolean b) {
       reserve(1);
@@ -88,7 +93,7 @@ public class UnsynchronizedBuffer {
       else
         data[offset++] = 0;
     }
-    
+
     /**
      * Gets (a copy of) the contents of this writer's buffer.
      *
@@ -99,7 +104,7 @@ public class UnsynchronizedBuffer {
       System.arraycopy(data, 0, ret, 0, offset);
       return ret;
     }
-    
+
     /**
      * Gets a <code>ByteBuffer</code> wrapped around this writer's buffer.
      *
@@ -110,24 +115,23 @@ public class UnsynchronizedBuffer {
     }
 
     /**
-     * Adds an integer value to this writer's buffer. The integer is encoded as
-     * a variable-length list of bytes. See {@link #writeVLong(long)} for a
-     * description of the encoding.
+     * Adds an integer value to this writer's buffer. The integer is encoded as a variable-length list of bytes. See {@link #writeVLong(long)} for a description
+     * of the encoding.
      *
-     * @param i integer value
+     * @param i
+     *          integer value
      */
     public void writeVInt(int i) {
       writeVLong(i);
     }
 
     /**
-     * Adds a long value to this writer's buffer. The long is encoded as
-     * a variable-length list of bytes. For a description of the encoding
-     * scheme, see <code>WritableUtils.writeVLong()</code> in the Hadoop
-     * API.
-     * [<a href="http://hadoop.apache.org/docs/stable/api/org/apache/hadoop/io/WritableUtils.html#writeVLong%28java.io.DataOutput,%20long%29">link</a>]
+     * Adds a long value to this writer's buffer. The long is encoded as a variable-length list of bytes. For a description of the encoding scheme, see
+     * <code>WritableUtils.writeVLong()</code> in the Hadoop API. [<a
+     * href="http://hadoop.apache.org/docs/stable/api/org/apache/hadoop/io/WritableUtils.html#writeVLong%28java.io.DataOutput,%20long%29">link</a>]
      *
-     * @param i long value
+     * @param i
+     *          long value
      */
     public void writeVLong(long i) {
       reserve(9);
@@ -135,23 +139,23 @@ public class UnsynchronizedBuffer {
         data[offset++] = (byte) i;
         return;
       }
-      
+
       int len = -112;
       if (i < 0) {
         i ^= -1L; // take one's complement'
         len = -120;
       }
-      
+
       long tmp = i;
       while (tmp != 0) {
         tmp = tmp >> 8;
         len--;
       }
-      
+
       data[offset++] = (byte) len;
-      
+
       len = (len < -120) ? -(len + 120) : -(len + 112);
-      
+
       for (int idx = len; idx != 0; idx--) {
         int shiftbits = (idx - 1) * 8;
         long mask = 0xFFL << shiftbits;
@@ -159,27 +163,29 @@ public class UnsynchronizedBuffer {
       }
     }
   }
-  
+
   /**
    * A byte buffer reader.
    */
   public static class Reader {
     int offset;
     byte data[];
-    
+
     /**
      * Creates a new reader.
      *
-     * @param b bytes to read
+     * @param b
+     *          bytes to read
      */
     public Reader(byte b[]) {
       this.data = b;
     }
-    
+
     /**
      * Creates a new reader.
      *
-     * @param buffer byte buffer containing bytes to read
+     * @param buffer
+     *          byte buffer containing bytes to read
      */
     public Reader(ByteBuffer buffer) {
       if (buffer.hasArray()) {
@@ -199,7 +205,7 @@ public class UnsynchronizedBuffer {
     public int readInt() {
       return (data[offset++] << 24) + ((data[offset++] & 255) << 16) + ((data[offset++] & 255) << 8) + ((data[offset++] & 255) << 0);
     }
-    
+
     /**
      * Reads a long value from this reader's buffer.
      *
@@ -209,17 +215,18 @@ public class UnsynchronizedBuffer {
       return (((long) data[offset++] << 56) + ((long) (data[offset++] & 255) << 48) + ((long) (data[offset++] & 255) << 40)
           + ((long) (data[offset++] & 255) << 32) + ((long) (data[offset++] & 255) << 24) + ((data[offset++] & 255) << 16) + ((data[offset++] & 255) << 8) + ((data[offset++] & 255) << 0));
     }
-    
+
     /**
      * Reads bytes from this reader's buffer, filling the given byte array.
      *
-     * @param b byte array to fill
+     * @param b
+     *          byte array to fill
      */
     public void readBytes(byte b[]) {
       System.arraycopy(data, offset, b, 0, b.length);
       offset += b.length;
     }
-    
+
     /**
      * Reads a Boolean value from this reader's buffer.
      *
@@ -228,10 +235,9 @@ public class UnsynchronizedBuffer {
     public boolean readBoolean() {
       return (data[offset++] == 1);
     }
-    
+
     /**
-     * Reads an integer value from this reader's buffer, assuming the integer
-     * was encoded as a variable-length list of bytes.
+     * Reads an integer value from this reader's buffer, assuming the integer was encoded as a variable-length list of bytes.
      *
      * @return integer value
      */
@@ -240,8 +246,7 @@ public class UnsynchronizedBuffer {
     }
 
     /**
-     * Reads a long value from this reader's buffer, assuming the long
-     * was encoded as a variable-length list of bytes.
+     * Reads a long value from this reader's buffer, assuming the long was encoded as a variable-length list of bytes.
      *
      * @return long value
      */
@@ -264,21 +269,23 @@ public class UnsynchronizedBuffer {
   /**
    * Determines what next array size should be by rounding up to next power of two.
    *
-   * @param i current array size
+   * @param i
+   *          current array size
    * @return next array size
-   * @throws IllegalArgumentException if i is negative
+   * @throws IllegalArgumentException
+   *           if i is negative
    */
   public static int nextArraySize(int i) {
     if (i < 0)
       throw new IllegalArgumentException();
-    
+
     if (i > (1 << 30))
       return Integer.MAX_VALUE; // this is the next power of 2 minus one... a special case
-  
+
     if (i == 0) {
       return 1;
     }
-    
+
     // round up to next power of two
     int ret = i;
     ret--;
@@ -288,7 +295,7 @@ public class UnsynchronizedBuffer {
     ret |= ret >> 8;
     ret |= ret >> 16;
     ret++;
-    
+
     return ret;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/UtilWaitThread.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/UtilWaitThread.java b/core/src/main/java/org/apache/accumulo/core/util/UtilWaitThread.java
index 36a4279..b559997 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/UtilWaitThread.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/UtilWaitThread.java
@@ -20,7 +20,7 @@ import org.apache.log4j.Logger;
 
 public class UtilWaitThread {
   private static final Logger log = Logger.getLogger(UtilWaitThread.class);
-  
+
   public static void sleep(long millis) {
     try {
       Thread.sleep(millis);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/Validator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/Validator.java b/core/src/main/java/org/apache/accumulo/core/util/Validator.java
index efb70c0..a5ae156 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/Validator.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/Validator.java
@@ -17,14 +17,14 @@
 package org.apache.accumulo.core.util;
 
 /**
- * A class that validates arguments of a particular type. Implementations must
- * implement {@link #isValid(Object)} and should override {@link #invalidMessage(Object)}.
+ * A class that validates arguments of a particular type. Implementations must implement {@link #isValid(Object)} and should override
+ * {@link #invalidMessage(Object)}.
  */
 public abstract class Validator<T> {
 
   /**
    * Validates an argument.
-   * 
+   *
    * @param argument
    *          argument to validate
    * @return the argument, if validation passes
@@ -39,7 +39,7 @@ public abstract class Validator<T> {
 
   /**
    * Checks an argument for validity.
-   * 
+   *
    * @param argument
    *          argument to validate
    * @return true if valid, false if invalid
@@ -48,7 +48,7 @@ public abstract class Validator<T> {
 
   /**
    * Formulates an exception message for invalid values.
-   * 
+   *
    * @param argument
    *          argument that failed validation
    * @return exception message
@@ -60,7 +60,7 @@ public abstract class Validator<T> {
   /**
    * Creates a new validator that is the conjunction of this one and the given one. An argument passed to the returned validator is valid only if it passes both
    * validators.
-   * 
+   *
    * @param other
    *          other validator
    * @return combined validator
@@ -87,7 +87,7 @@ public abstract class Validator<T> {
   /**
    * Creates a new validator that is the disjunction of this one and the given one. An argument passed to the returned validator is valid only if it passes at
    * least one of the validators.
-   * 
+   *
    * @param other
    *          other validator
    * @return combined validator
@@ -113,7 +113,7 @@ public abstract class Validator<T> {
 
   /**
    * Creates a new validator that is the negation of this one. An argument passed to the returned validator is valid only if it fails this one.
-   * 
+   *
    * @return negated validator
    */
   public final Validator<T> not() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/Version.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/Version.java b/core/src/main/java/org/apache/accumulo/core/util/Version.java
index 7347227..ee645ff 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/Version.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/Version.java
@@ -25,17 +25,17 @@ public class Version {
   int minor = 0;
   int release = 0;
   String etcetera = null;
-  
+
   public Version(String everything) {
     parse(everything);
   }
-  
+
   private void parse(String everything) {
     Pattern pattern = Pattern.compile("(([^-]*)-)?(\\d+)(\\.(\\d+)(\\.(\\d+))?)?(-(.*))?");
     Matcher parser = pattern.matcher(everything);
     if (!parser.matches())
       throw new IllegalArgumentException("Unable to parse: " + everything + " as a version");
-    
+
     if (parser.group(1) != null)
       package_ = parser.group(2);
     major = Integer.valueOf(parser.group(3));
@@ -46,29 +46,29 @@ public class Version {
       release = Integer.valueOf(parser.group(7));
     if (parser.group(9) != null)
       etcetera = parser.group(9);
-    
+
   }
-  
+
   public String getPackage() {
     return package_;
   }
-  
+
   public int getMajorVersion() {
     return major;
   }
-  
+
   public int getMinorVersion() {
     return minor;
   }
-  
+
   public int getReleaseVersion() {
     return release;
   }
-  
+
   public String getEtcetera() {
     return etcetera;
   }
-  
+
   @Override
   public String toString() {
     StringBuilder result = new StringBuilder();
@@ -87,5 +87,5 @@ public class Version {
     }
     return result.toString();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
index b31df18..ec20da5 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
@@ -25,17 +25,17 @@ import org.apache.hadoop.io.Text;
 
 public class BinaryFormatter extends DefaultFormatter {
   private static int showLength;
-  
+
   public String next() {
     checkState(true);
     return formatEntry(getScannerIterator().next(), isDoTimestamps());
   }
-  
+
   // this should be replaced with something like Record.toString();
   // it would be great if we were able to combine code with DefaultFormatter.formatEntry, but that currently does not respect the showLength option.
   public static String formatEntry(Entry<Key,Value> entry, boolean showTimestamps) {
     StringBuilder sb = new StringBuilder();
-    
+
     Key key = entry.getKey();
 
     // append row
@@ -49,11 +49,11 @@ public class BinaryFormatter extends DefaultFormatter {
 
     // append visibility expression
     sb.append(new ColumnVisibility(key.getColumnVisibility()));
-    
+
     // append timestamp
     if (showTimestamps)
       sb.append(" ").append(entry.getKey().getTimestamp());
-    
+
     // append value
     Value value = entry.getValue();
     if (value != null && value.getSize() > 0) {
@@ -62,20 +62,20 @@ public class BinaryFormatter extends DefaultFormatter {
     }
     return sb.toString();
   }
-  
+
   public static StringBuilder appendText(StringBuilder sb, Text t) {
     return appendBytes(sb, t.getBytes(), 0, t.getLength());
   }
-  
+
   static StringBuilder appendValue(StringBuilder sb, Value value) {
     return appendBytes(sb, value.get(), 0, value.get().length);
   }
-  
+
   static StringBuilder appendBytes(StringBuilder sb, byte ba[], int offset, int len) {
     int length = Math.min(len, showLength);
     return DefaultFormatter.appendBytes(sb, ba, offset, length);
   }
-  
+
   public static void getlength(int length) {
     showLength = length;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/DateStringFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/DateStringFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/DateStringFormatter.java
index 037ddb0..5bcd4a3 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/DateStringFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/DateStringFormatter.java
@@ -36,26 +36,29 @@ public class DateStringFormatter implements Formatter {
       return new SimpleDateFormat(DATE_FORMAT);
     }
   };
-  
+
   @Override
   public void initialize(Iterable<Entry<Key,Value>> scanner, boolean printTimestamps) {
     this.printTimestamps = printTimestamps;
     defaultFormatter.initialize(scanner, printTimestamps);
   }
+
   @Override
   public boolean hasNext() {
     return defaultFormatter.hasNext();
   }
+
   @Override
   public String next() {
     DateFormat timestampformat = null;
-    
-    if(printTimestamps) {
+
+    if (printTimestamps) {
       timestampformat = formatter.get();
     }
-    
+
     return defaultFormatter.next(timestampformat);
   }
+
   @Override
   public void remove() {
     defaultFormatter.remove();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/DefaultFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/DefaultFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/DefaultFormatter.java
index 931b59f..f104610 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/DefaultFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/DefaultFormatter.java
@@ -36,7 +36,7 @@ public class DefaultFormatter implements Formatter {
     protected DateFormat initialValue() {
       return new DefaultDateFormat();
     }
-    
+
     class DefaultDateFormat extends DateFormat {
       private static final long serialVersionUID = 1L;
 
@@ -50,42 +50,42 @@ public class DefaultFormatter implements Formatter {
       public Date parse(String source, ParsePosition pos) {
         return new Date(Long.parseLong(source));
       }
-      
+
     }
   };
-  
+
   @Override
   public void initialize(Iterable<Entry<Key,Value>> scanner, boolean printTimestamps) {
     checkState(false);
     si = scanner.iterator();
     doTimestamps = printTimestamps;
   }
-  
+
   public boolean hasNext() {
     checkState(true);
     return si.hasNext();
   }
-  
+
   public String next() {
     DateFormat timestampFormat = null;
-    
-    if(doTimestamps) {
+
+    if (doTimestamps) {
       timestampFormat = formatter.get();
     }
-    
+
     return next(timestampFormat);
   }
-  
+
   protected String next(DateFormat timestampFormat) {
     checkState(true);
     return formatEntry(si.next(), timestampFormat);
   }
-  
+
   public void remove() {
     checkState(true);
     si.remove();
   }
-  
+
   protected void checkState(boolean expectInitialized) {
     if (expectInitialized && si == null)
       throw new IllegalStateException("Not initialized");
@@ -96,22 +96,22 @@ public class DefaultFormatter implements Formatter {
   // this should be replaced with something like Record.toString();
   public static String formatEntry(Entry<Key,Value> entry, boolean showTimestamps) {
     DateFormat timestampFormat = null;
-    
-    if(showTimestamps) {
+
+    if (showTimestamps) {
       timestampFormat = formatter.get();
     }
-    
+
     return formatEntry(entry, timestampFormat);
   }
-  
+
   /* so a new date object doesn't get created for every record in the scan result */
   private static ThreadLocal<Date> tmpDate = new ThreadLocal<Date>() {
     @Override
-    protected Date initialValue() { 
+    protected Date initialValue() {
       return new Date();
     }
   };
-  
+
   public static String formatEntry(Entry<Key,Value> entry, DateFormat timestampFormat) {
     StringBuilder sb = new StringBuilder();
     Key key = entry.getKey();
@@ -119,16 +119,16 @@ public class DefaultFormatter implements Formatter {
 
     // append row
     appendText(sb, key.getRow(buffer)).append(" ");
-    
+
     // append column family
     appendText(sb, key.getColumnFamily(buffer)).append(":");
-    
+
     // append column qualifier
     appendText(sb, key.getColumnQualifier(buffer)).append(" ");
-    
+
     // append visibility expression
     sb.append(new ColumnVisibility(key.getColumnVisibility(buffer)));
-    
+
     // append timestamp
     if (timestampFormat != null) {
       tmpDate.get().setTime(entry.getKey().getTimestamp());
@@ -142,18 +142,18 @@ public class DefaultFormatter implements Formatter {
       sb.append("\t");
       appendValue(sb, value);
     }
-    
+
     return sb.toString();
   }
 
   static StringBuilder appendText(StringBuilder sb, Text t) {
     return appendBytes(sb, t.getBytes(), 0, t.getLength());
   }
-  
+
   static StringBuilder appendValue(StringBuilder sb, Value value) {
     return appendBytes(sb, value.get(), 0, value.get().length);
   }
-  
+
   static StringBuilder appendBytes(StringBuilder sb, byte ba[], int offset, int len) {
     for (int i = 0; i < len; i++) {
       int c = 0xff & ba[offset + i];
@@ -166,7 +166,7 @@ public class DefaultFormatter implements Formatter {
     }
     return sb;
   }
-  
+
   public Iterator<Entry<Key,Value>> getScannerIterator() {
     return si;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/FormatterFactory.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/FormatterFactory.java b/core/src/main/java/org/apache/accumulo/core/util/format/FormatterFactory.java
index 27299ee..e8a9e7d 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/FormatterFactory.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/FormatterFactory.java
@@ -24,7 +24,7 @@ import org.apache.log4j.Logger;
 
 public class FormatterFactory {
   private static final Logger log = Logger.getLogger(FormatterFactory.class);
-  
+
   public static Formatter getFormatter(Class<? extends Formatter> formatterClass, Iterable<Entry<Key,Value>> scanner, boolean printTimestamps) {
     Formatter formatter = null;
     try {
@@ -36,7 +36,7 @@ public class FormatterFactory {
     formatter.initialize(scanner, printTimestamps);
     return formatter;
   }
-  
+
   public static Formatter getDefaultFormatter(Iterable<Entry<Key,Value>> scanner, boolean printTimestamps) {
     return getFormatter(DefaultFormatter.class, scanner, printTimestamps);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/HexFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/HexFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/HexFormatter.java
index b636278..65e52d3 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/HexFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/HexFormatter.java
@@ -28,11 +28,11 @@ import org.apache.hadoop.io.Text;
  * A simple formatter that print the row, column family, column qualifier, and value as hex
  */
 public class HexFormatter implements Formatter, ScanInterpreter {
-  
+
   private char chars[] = new char[] {'0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f'};
   private Iterator<Entry<Key,Value>> iter;
   private boolean printTimestamps;
-  
+
   private void toHex(StringBuilder sb, byte[] bin) {
 
     for (int i = 0; i < bin.length; i++) {
@@ -42,22 +42,22 @@ public class HexFormatter implements Formatter, ScanInterpreter {
       sb.append(chars[0x0f & bin[i]]);
     }
   }
-  
+
   private int fromChar(char b) {
     if (b >= '0' && b <= '9') {
       return (b - '0');
     } else if (b >= 'a' && b <= 'f') {
       return (b - 'a' + 10);
     }
-    
+
     throw new IllegalArgumentException("Bad char " + b);
   }
-  
+
   private byte[] toBinary(String hex) {
     hex = hex.replace("-", "");
 
     byte[] bin = new byte[(hex.length() / 2) + (hex.length() % 2)];
-    
+
     int j = 0;
     for (int i = 0; i < bin.length; i++) {
       bin[i] = (byte) (fromChar(hex.charAt(j++)) << 4);
@@ -65,22 +65,21 @@ public class HexFormatter implements Formatter, ScanInterpreter {
         break;
       bin[i] |= (byte) fromChar(hex.charAt(j++));
     }
-    
+
     return bin;
   }
 
-
   @Override
   public boolean hasNext() {
     return iter.hasNext();
   }
-  
+
   @Override
   public String next() {
     Entry<Key,Value> entry = iter.next();
-    
+
     StringBuilder sb = new StringBuilder();
-    
+
     toHex(sb, entry.getKey().getRowData().toArray());
     sb.append("  ");
     toHex(sb, entry.getKey().getColumnFamilyData().toArray());
@@ -94,26 +93,26 @@ public class HexFormatter implements Formatter, ScanInterpreter {
       sb.append("  ");
     }
     toHex(sb, entry.getValue().get());
-    
+
     return sb.toString();
   }
-  
+
   @Override
   public void remove() {
     iter.remove();
   }
-  
+
   @Override
   public void initialize(Iterable<Entry<Key,Value>> scanner, boolean printTimestamps) {
     this.iter = scanner.iterator();
     this.printTimestamps = printTimestamps;
   }
-  
+
   @Override
   public Text interpretRow(Text row) {
     return new Text(toBinary(row.toString()));
   }
-  
+
   @Override
   public Text interpretBeginRow(Text row) {
     return interpretRow(row);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/ShardedTableDistributionFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/ShardedTableDistributionFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/ShardedTableDistributionFormatter.java
index f81209f..877f164 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/ShardedTableDistributionFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/ShardedTableDistributionFormatter.java
@@ -27,15 +27,15 @@ import org.apache.accumulo.core.data.Value;
 /**
  * Formats the rows in a METADATA table scan to show distribution of shards over servers per day. This can be used to determine the effectiveness of the
  * ShardedTableLoadBalancer
- * 
+ *
  * Use this formatter with the following scan command in the shell:
- * 
+ *
  * scan -b tableId -c ~tab:loc
  */
 public class ShardedTableDistributionFormatter extends AggregatingFormatter {
-  
+
   private Map<String,HashSet<String>> countsByDay = new HashMap<String,HashSet<String>>();
-  
+
   @Override
   protected void aggregateStats(Entry<Key,Value> entry) {
     if (entry.getKey().getColumnFamily().toString().equals("~tab") && entry.getKey().getColumnQualifier().toString().equals("loc")) {
@@ -55,7 +55,7 @@ public class ShardedTableDistributionFormatter extends AggregatingFormatter {
       countsByDay.get(day).add(server);
     }
   }
-  
+
   @Override
   protected String getStats() {
     StringBuilder buf = new StringBuilder();
@@ -65,5 +65,5 @@ public class ShardedTableDistributionFormatter extends AggregatingFormatter {
       buf.append(day + "\t\t" + countsByDay.get(day).size() + "\n");
     return buf.toString();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/format/StatisticsDisplayFormatter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/format/StatisticsDisplayFormatter.java b/core/src/main/java/org/apache/accumulo/core/util/format/StatisticsDisplayFormatter.java
index 98d4d28..b3ee3ee 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/format/StatisticsDisplayFormatter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/format/StatisticsDisplayFormatter.java
@@ -32,27 +32,27 @@ public class StatisticsDisplayFormatter extends AggregatingFormatter {
   private Map<String,Long> columnFamilies = new HashMap<String,Long>();
   private Map<String,Long> columnQualifiers = new HashMap<String,Long>();
   private long total = 0;
-  
+
   @Override
   protected void aggregateStats(Entry<Key,Value> entry) {
     String key;
     Long count;
-    
+
     key = entry.getKey().getColumnVisibility().toString();
     count = classifications.get(key);
     classifications.put(key, count != null ? count + 1 : 0L);
-    
+
     key = entry.getKey().getColumnFamily().toString();
     count = columnFamilies.get(key);
     columnFamilies.put(key, count != null ? count + 1 : 0L);
-    
+
     key = entry.getKey().getColumnQualifier().toString();
     count = columnQualifiers.get(key);
     columnQualifiers.put(key, count != null ? count + 1 : 0L);
-    
+
     ++total;
   }
-  
+
   @Override
   protected String getStats() {
     StringBuilder buf = new StringBuilder();
@@ -68,13 +68,13 @@ public class StatisticsDisplayFormatter extends AggregatingFormatter {
     buf.append("------------------\n");
     for (String key : columnQualifiers.keySet())
       buf.append("\t").append(key).append(": ").append(columnQualifiers.get(key)).append("\n");
-    
+
     buf.append(total).append(" entries matched.");
     total = 0;
     classifications = new HashMap<String,Long>();
     columnFamilies = new HashMap<String,Long>();
     columnQualifiers = new HashMap<String,Long>();
-    
+
     return buf.toString();
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/interpret/DefaultScanInterpreter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/interpret/DefaultScanInterpreter.java b/core/src/main/java/org/apache/accumulo/core/util/interpret/DefaultScanInterpreter.java
index b1d7d36..761756f 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/interpret/DefaultScanInterpreter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/interpret/DefaultScanInterpreter.java
@@ -19,33 +19,33 @@ package org.apache.accumulo.core.util.interpret;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class DefaultScanInterpreter implements ScanInterpreter {
-  
+
   @Override
   public Text interpretRow(Text row) {
     return row;
   }
-  
+
   @Override
   public Text interpretBeginRow(Text row) {
     return row;
   }
-  
+
   @Override
   public Text interpretEndRow(Text row) {
     return row;
   }
-  
+
   @Override
   public Text interpretColumnFamily(Text cf) {
     return cf;
   }
-  
+
   @Override
   public Text interpretColumnQualifier(Text cq) {
     return cq;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/interpret/HexScanInterpreter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/interpret/HexScanInterpreter.java b/core/src/main/java/org/apache/accumulo/core/util/interpret/HexScanInterpreter.java
index 78c0ce2..964e8c6 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/interpret/HexScanInterpreter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/interpret/HexScanInterpreter.java
@@ -23,5 +23,5 @@ import org.apache.accumulo.core.util.format.HexFormatter;
  * dashes (because {@link HexFormatter} outputs dashes) which are ignored.
  */
 public class HexScanInterpreter extends HexFormatter implements ScanInterpreter {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/util/interpret/ScanInterpreter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/interpret/ScanInterpreter.java b/core/src/main/java/org/apache/accumulo/core/util/interpret/ScanInterpreter.java
index 315767e..7dfcf22 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/interpret/ScanInterpreter.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/interpret/ScanInterpreter.java
@@ -22,14 +22,14 @@ import org.apache.hadoop.io.Text;
  * A simple interface for creating shell plugins that translate the range and column arguments for the shell's scan command.
  */
 public interface ScanInterpreter {
-  
+
   Text interpretRow(Text row);
 
   Text interpretBeginRow(Text row);
-  
+
   Text interpretEndRow(Text row);
-  
+
   Text interpretColumnFamily(Text cf);
-  
+
   Text interpretColumnQualifier(Text cq);
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/volume/Volume.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/volume/Volume.java b/core/src/main/java/org/apache/accumulo/core/volume/Volume.java
index 58b0ada..5a3ea53 100644
--- a/core/src/main/java/org/apache/accumulo/core/volume/Volume.java
+++ b/core/src/main/java/org/apache/accumulo/core/volume/Volume.java
@@ -20,8 +20,7 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
 /**
- * Encapsulates a {@link FileSystem} and a base {@link Path} within that filesystem. This
- * also avoid the necessity to pass around a Configuration.
+ * Encapsulates a {@link FileSystem} and a base {@link Path} within that filesystem. This also avoid the necessity to pass around a Configuration.
  */
 public interface Volume {
 
@@ -37,21 +36,24 @@ public interface Volume {
 
   /**
    * Convert the given Path into a Path that is relative to the base path for this Volume
-   * @param p The suffix to use
+   *
+   * @param p
+   *          The suffix to use
    * @return A Path for this Volume with the provided suffix
    */
   public Path prefixChild(Path p);
 
   /**
    * Convert the given child path into a Path that is relative to the base path for this Volume
-   * @param p The suffix to use
+   *
+   * @param p
+   *          The suffix to use
    * @return A Path for this Volume with the provided suffix
    */
   public Path prefixChild(String p);
 
   /**
-   * Determine if the Path is valid on this Volume. A Path is valid if it is contained
-   * in the Volume's FileSystem and is rooted beneath the basePath
+   * Determine if the Path is valid on this Volume. A Path is valid if it is contained in the Volume's FileSystem and is rooted beneath the basePath
    */
   public boolean isValidPath(Path p);
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/volume/VolumeConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/volume/VolumeConfiguration.java b/core/src/main/java/org/apache/accumulo/core/volume/VolumeConfiguration.java
index 8d56d9e..99032ad 100644
--- a/core/src/main/java/org/apache/accumulo/core/volume/VolumeConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/volume/VolumeConfiguration.java
@@ -84,12 +84,12 @@ public class VolumeConfiguration {
 
   /**
    * Compute the URIs to be used by Accumulo
-   * 
+   *
    */
   public static String[] getVolumeUris(AccumuloConfiguration conf) {
     return getVolumeUris(conf, CachedConfiguration.getInstance());
   }
-  
+
   public static String[] getVolumeUris(AccumuloConfiguration conf, Configuration hadoopConfig) {
     String ns = conf.get(Property.INSTANCE_VOLUMES);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/volume/VolumeImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/volume/VolumeImpl.java b/core/src/main/java/org/apache/accumulo/core/volume/VolumeImpl.java
index 9a324a0..9b9a80e 100644
--- a/core/src/main/java/org/apache/accumulo/core/volume/VolumeImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/volume/VolumeImpl.java
@@ -27,8 +27,7 @@ import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
 /**
- * Basic Volume implementation that contains a FileSystem and a base path
- * that should be used within that filesystem.
+ * Basic Volume implementation that contains a FileSystem and a base path that should be used within that filesystem.
  */
 public class VolumeImpl implements Volume {
   private static final Logger log = Logger.getLogger(VolumeImpl.class);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/cli/TestHelp.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/cli/TestHelp.java b/core/src/test/java/org/apache/accumulo/core/cli/TestHelp.java
index 7b29986..5cb5c32 100644
--- a/core/src/test/java/org/apache/accumulo/core/cli/TestHelp.java
+++ b/core/src/test/java/org/apache/accumulo/core/cli/TestHelp.java
@@ -1,4 +1,5 @@
 package org.apache.accumulo.core.cli;
+
 /*
  * Licensed to the Apache Software Foundation (ASF) under one or more
  * contributor license agreements.  See the NOTICE file distributed with
@@ -17,12 +18,13 @@ package org.apache.accumulo.core.cli;
  */
 
 import static org.junit.Assert.assertEquals;
+
 import org.junit.Test;
 
 public class TestHelp {
   protected class HelpStub extends Help {
     @Override
-    public void parseArgs(String programName, String[] args, Object ... others) {
+    public void parseArgs(String programName, String[] args, Object... others) {
       super.parseArgs(programName, args, others);
     }
 
@@ -31,7 +33,7 @@ public class TestHelp {
       throw new RuntimeException(Integer.toString(status));
     }
   }
-  
+
   @Test
   public void testInvalidArgs() {
     String[] args = {"foo"};

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/BatchWriterConfigTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/BatchWriterConfigTest.java b/core/src/test/java/org/apache/accumulo/core/client/BatchWriterConfigTest.java
index a386c04..9a32c26 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/BatchWriterConfigTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/BatchWriterConfigTest.java
@@ -30,10 +30,10 @@ import java.util.concurrent.TimeUnit;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class BatchWriterConfigTest {
-  
+
   @Test
   public void testReasonableDefaults() {
     long expectedMaxMemory = 50 * 1024 * 1024l;
@@ -41,7 +41,7 @@ public class BatchWriterConfigTest {
     long expectedTimeout = Long.MAX_VALUE;
     int expectedMaxWriteThreads = 3;
     Durability expectedDurability = Durability.DEFAULT;
-    
+
     BatchWriterConfig defaults = new BatchWriterConfig();
     assertEquals(expectedMaxMemory, defaults.getMaxMemory());
     assertEquals(expectedMaxLatency, defaults.getMaxLatency(TimeUnit.MILLISECONDS));
@@ -49,7 +49,7 @@ public class BatchWriterConfigTest {
     assertEquals(expectedMaxWriteThreads, defaults.getMaxWriteThreads());
     assertEquals(expectedDurability, defaults.getDurability());
   }
-  
+
   @Test
   public void testOverridingDefaults() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
@@ -58,77 +58,77 @@ public class BatchWriterConfigTest {
     bwConfig.setTimeout(33, TimeUnit.DAYS);
     bwConfig.setMaxWriteThreads(42);
     bwConfig.setDurability(Durability.NONE);
-    
+
     assertEquals(1123581321l, bwConfig.getMaxMemory());
     assertEquals(22 * 60 * 60 * 1000l, bwConfig.getMaxLatency(TimeUnit.MILLISECONDS));
     assertEquals(33 * 24 * 60 * 60 * 1000l, bwConfig.getTimeout(TimeUnit.MILLISECONDS));
     assertEquals(42, bwConfig.getMaxWriteThreads());
     assertEquals(Durability.NONE, bwConfig.getDurability());
   }
-  
+
   @Test
   public void testZeroValues() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxLatency(0, TimeUnit.MILLISECONDS);
     bwConfig.setTimeout(0, TimeUnit.MILLISECONDS);
     bwConfig.setMaxMemory(0);
-    
+
     assertEquals(Long.MAX_VALUE, bwConfig.getMaxLatency(TimeUnit.MILLISECONDS));
     assertEquals(Long.MAX_VALUE, bwConfig.getTimeout(TimeUnit.MILLISECONDS));
     assertEquals(0, bwConfig.getMaxMemory());
   }
-  
+
   @Test(expected = IllegalArgumentException.class)
   public void testNegativeMaxMemory() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxMemory(-1);
   }
-  
+
   @Test(expected = IllegalArgumentException.class)
   public void testNegativeMaxLatency() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxLatency(-1, TimeUnit.DAYS);
   }
-  
+
   @Test
   public void testTinyTimeConversions() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxLatency(999, TimeUnit.MICROSECONDS);
     bwConfig.setTimeout(999, TimeUnit.MICROSECONDS);
-    
+
     assertEquals(1000, bwConfig.getMaxLatency(TimeUnit.MICROSECONDS));
     assertEquals(1000, bwConfig.getTimeout(TimeUnit.MICROSECONDS));
     assertEquals(1, bwConfig.getMaxLatency(TimeUnit.MILLISECONDS));
     assertEquals(1, bwConfig.getTimeout(TimeUnit.MILLISECONDS));
-    
+
     bwConfig.setMaxLatency(10, TimeUnit.NANOSECONDS);
     bwConfig.setTimeout(10, TimeUnit.NANOSECONDS);
-    
+
     assertEquals(1000000, bwConfig.getMaxLatency(TimeUnit.NANOSECONDS));
     assertEquals(1000000, bwConfig.getTimeout(TimeUnit.NANOSECONDS));
     assertEquals(1, bwConfig.getMaxLatency(TimeUnit.MILLISECONDS));
     assertEquals(1, bwConfig.getTimeout(TimeUnit.MILLISECONDS));
-    
+
   }
-  
+
   @Test(expected = IllegalArgumentException.class)
   public void testNegativeTimeout() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setTimeout(-1, TimeUnit.DAYS);
   }
-  
+
   @Test(expected = IllegalArgumentException.class)
   public void testZeroMaxWriteThreads() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxWriteThreads(0);
   }
-  
+
   @Test(expected = IllegalArgumentException.class)
   public void testNegativeMaxWriteThreads() {
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxWriteThreads(-1);
   }
-  
+
   @Test
   public void testSerialize() throws IOException {
     // make sure we aren't testing defaults
@@ -138,7 +138,7 @@ public class BatchWriterConfigTest {
     assertNotEquals(42, bwDefaults.getMaxWriteThreads());
     assertNotEquals(1123581321l, bwDefaults.getMaxMemory());
     assertNotEquals(Durability.FLUSH, bwDefaults.getDurability());
-    
+
     // test setting all fields
     BatchWriterConfig bwConfig = new BatchWriterConfig();
     bwConfig.setMaxLatency(7654321l, TimeUnit.MILLISECONDS);
@@ -148,14 +148,14 @@ public class BatchWriterConfigTest {
     bwConfig.setDurability(Durability.FLUSH);
     byte[] bytes = createBytes(bwConfig);
     checkBytes(bwConfig, bytes);
-    
+
     // test human-readable serialization
     bwConfig = new BatchWriterConfig();
     bwConfig.setMaxWriteThreads(42);
     bytes = createBytes(bwConfig);
     assertEquals("     i#maxWriteThreads=42", new String(bytes, UTF_8));
     checkBytes(bwConfig, bytes);
-    
+
     // test human-readable with 2 fields
     bwConfig = new BatchWriterConfig();
     bwConfig.setMaxWriteThreads(24);
@@ -163,7 +163,7 @@ public class BatchWriterConfigTest {
     bytes = createBytes(bwConfig);
     assertEquals("     v#maxWriteThreads=24,timeout=3000", new String(bytes, UTF_8));
     checkBytes(bwConfig, bytes);
-    
+
     // test human-readable durability
     bwConfig = new BatchWriterConfig();
     bwConfig.setDurability(Durability.LOG);
@@ -195,27 +195,27 @@ public class BatchWriterConfigTest {
 
     cfg1.setTimeout(10, TimeUnit.SECONDS);
     cfg2.setTimeout(10000, TimeUnit.MILLISECONDS);
-    
+
     assertEquals(cfg1, cfg2);
 
     assertEquals(cfg1.hashCode(), cfg2.hashCode());
   }
-  
+
   private byte[] createBytes(BatchWriterConfig bwConfig) throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     bwConfig.write(new DataOutputStream(baos));
     return baos.toByteArray();
   }
-  
+
   private void checkBytes(BatchWriterConfig bwConfig, byte[] bytes) throws IOException {
     ByteArrayInputStream bais = new ByteArrayInputStream(bytes);
     BatchWriterConfig createdConfig = new BatchWriterConfig();
     createdConfig.readFields(new DataInputStream(bais));
-    
+
     assertEquals(bwConfig.getMaxMemory(), createdConfig.getMaxMemory());
     assertEquals(bwConfig.getMaxLatency(TimeUnit.MILLISECONDS), createdConfig.getMaxLatency(TimeUnit.MILLISECONDS));
     assertEquals(bwConfig.getTimeout(TimeUnit.MILLISECONDS), createdConfig.getTimeout(TimeUnit.MILLISECONDS));
     assertEquals(bwConfig.getMaxWriteThreads(), createdConfig.getMaxWriteThreads());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/client/ClientSideIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/client/ClientSideIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/client/ClientSideIteratorTest.java
index 74fab9c..60f668b 100644
--- a/core/src/test/java/org/apache/accumulo/core/client/ClientSideIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/client/ClientSideIteratorTest.java
@@ -53,7 +53,7 @@ public class ClientSideIteratorTest {
     resultSet3.add(new Key("part1", "", "doc2"));
     resultSet3.add(new Key("part2", "", "DOC2"));
   }
-  
+
   public void checkResults(final Iterable<Entry<Key,Value>> scanner, final List<Key> results, final PartialKey pk) {
     int i = 0;
     for (Entry<Key,Value> entry : scanner) {
@@ -61,7 +61,7 @@ public class ClientSideIteratorTest {
     }
     assertEquals(i, results.size());
   }
-  
+
   @Test
   public void testIntersect() throws Exception {
     Instance instance = new MockInstance("local");
@@ -83,15 +83,15 @@ public class ClientSideIteratorTest {
     m.put("foo", "DOC3", "value");
     bw.addMutation(m);
     bw.flush();
-    
+
     final ClientSideIteratorScanner csis = new ClientSideIteratorScanner(conn.createScanner("intersect", new Authorizations()));
     final IteratorSetting si = new IteratorSetting(10, "intersect", IntersectingIterator.class);
     IntersectingIterator.setColumnFamilies(si, new Text[] {new Text("bar"), new Text("foo")});
     csis.addScanIterator(si);
-    
+
     checkResults(csis, resultSet3, PartialKey.ROW_COLFAM_COLQUAL);
   }
-  
+
   @Test
   public void testVersioning() throws Exception {
     final Instance instance = new MockInstance("local");
@@ -111,17 +111,17 @@ public class ClientSideIteratorTest {
     m.put("colf", "colq", 4l, "value");
     bw.addMutation(m);
     bw.flush();
-    
+
     final Scanner scanner = conn.createScanner("table", new Authorizations());
-    
+
     final ClientSideIteratorScanner csis = new ClientSideIteratorScanner(scanner);
     final IteratorSetting si = new IteratorSetting(10, "localvers", VersioningIterator.class);
     si.addOption("maxVersions", "2");
     csis.addScanIterator(si);
-    
+
     checkResults(csis, resultSet1, PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME);
     checkResults(scanner, resultSet2, PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME);
-    
+
     csis.fetchColumnFamily(new Text("colf"));
     checkResults(csis, resultSet1, PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME);
     csis.clearColumns();


[36/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/FirstEntryInRowIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/FirstEntryInRowIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/FirstEntryInRowIterator.java
index dd693f1..fcca805 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/FirstEntryInRowIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/FirstEntryInRowIterator.java
@@ -30,48 +30,48 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.hadoop.io.Text;
 
 public class FirstEntryInRowIterator extends SkippingIterator implements OptionDescriber {
-  
+
   // options
   static final String NUM_SCANS_STRING_NAME = "scansBeforeSeek";
-  
+
   // iterator predecessor seek options to pass through
   private Range latestRange;
   private Collection<ByteSequence> latestColumnFamilies;
   private boolean latestInclusive;
-  
+
   // private fields
   private Text lastRowFound;
   private int numscans;
-  
+
   /**
    * convenience method to set the option to optimize the frequency of scans vs. seeks
    */
   public static void setNumScansBeforeSeek(IteratorSetting cfg, int num) {
     cfg.addOption(NUM_SCANS_STRING_NAME, Integer.toString(num));
   }
-  
+
   // this must be public for OptionsDescriber
   public FirstEntryInRowIterator() {
     super();
   }
-  
+
   public FirstEntryInRowIterator(FirstEntryInRowIterator other, IteratorEnvironment env) {
     super();
     setSource(other.getSource().deepCopy(env));
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new FirstEntryInRowIterator(this, env);
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     String o = options.get(NUM_SCANS_STRING_NAME);
     numscans = o == null ? 10 : Integer.parseInt(o);
   }
-  
+
   // this is only ever called immediately after getting "next" entry
   @Override
   protected void consume() throws IOException {
@@ -79,7 +79,7 @@ public class FirstEntryInRowIterator extends SkippingIterator implements OptionD
       return;
     int count = 0;
     while (getSource().hasTop() && lastRowFound.equals(getSource().getTopKey().getRow())) {
-      
+
       // try to efficiently jump to the next matching key
       if (count < numscans) {
         ++count;
@@ -87,7 +87,7 @@ public class FirstEntryInRowIterator extends SkippingIterator implements OptionD
       } else {
         // too many scans, just seek
         count = 0;
-        
+
         // determine where to seek to, but don't go beyond the user-specified range
         Key nextKey = getSource().getTopKey().followingKey(PartialKey.ROW);
         if (!latestRange.afterEndKey(nextKey))
@@ -100,14 +100,14 @@ public class FirstEntryInRowIterator extends SkippingIterator implements OptionD
     }
     lastRowFound = getSource().hasTop() ? getSource().getTopKey().getRow(lastRowFound) : null;
   }
-  
+
   private boolean finished = true;
-  
+
   @Override
   public boolean hasTop() {
     return !finished && getSource().hasTop();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     // save parameters for future internal seeks
@@ -115,19 +115,19 @@ public class FirstEntryInRowIterator extends SkippingIterator implements OptionD
     latestColumnFamilies = columnFamilies;
     latestInclusive = inclusive;
     lastRowFound = null;
-    
+
     Key startKey = range.getStartKey();
     Range seekRange = new Range(startKey == null ? null : new Key(startKey.getRow()), true, range.getEndKey(), range.isEndKeyInclusive());
     super.seek(seekRange, columnFamilies, inclusive);
     finished = false;
-    
+
     if (getSource().hasTop()) {
       lastRowFound = getSource().getTopKey().getRow();
       if (range.beforeStartKey(getSource().getTopKey()))
         consume();
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     String name = "firstEntry";
@@ -136,7 +136,7 @@ public class FirstEntryInRowIterator extends SkippingIterator implements OptionD
     namedOptions.put(NUM_SCANS_STRING_NAME, "Number of scans to try before seeking [10]");
     return new IteratorOptions(name, desc, namedOptions, null);
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     try {
@@ -148,5 +148,5 @@ public class FirstEntryInRowIterator extends SkippingIterator implements OptionD
     }
     return true;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/GrepIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/GrepIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/GrepIterator.java
index f7a68e9..5c44c31 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/GrepIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/GrepIterator.java
@@ -18,10 +18,10 @@ package org.apache.accumulo.core.iterators;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.GrepIterator}
  */
 @Deprecated
 public class GrepIterator extends org.apache.accumulo.core.iterators.user.GrepIterator {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/IntersectingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/IntersectingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/IntersectingIterator.java
index 3f6ee9f..5765982 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/IntersectingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/IntersectingIterator.java
@@ -18,10 +18,10 @@ package org.apache.accumulo.core.iterators;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.IntersectingIterator}
  */
 @Deprecated
 public class IntersectingIterator extends org.apache.accumulo.core.iterators.user.IntersectingIterator {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/IterationInterruptedException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/IterationInterruptedException.java b/core/src/main/java/org/apache/accumulo/core/iterators/IterationInterruptedException.java
index 63f6d01..4bfc4aa 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/IterationInterruptedException.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/IterationInterruptedException.java
@@ -17,13 +17,13 @@
 package org.apache.accumulo.core.iterators;
 
 public class IterationInterruptedException extends RuntimeException {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   public IterationInterruptedException() {
     super();
   }
-  
+
   public IterationInterruptedException(String msg) {
     super(msg);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/IteratorEnvironment.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/IteratorEnvironment.java b/core/src/main/java/org/apache/accumulo/core/iterators/IteratorEnvironment.java
index 9e20cb1..f708db7 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/IteratorEnvironment.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/IteratorEnvironment.java
@@ -24,15 +24,15 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 
 public interface IteratorEnvironment {
-  
+
   SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException;
-  
+
   AccumuloConfiguration getConfig();
-  
+
   IteratorScope getIteratorScope();
-  
+
   boolean isFullMajorCompaction();
-  
+
   void registerSideChannel(SortedKeyValueIterator<Key,Value> iter);
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/IteratorUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/IteratorUtil.java b/core/src/main/java/org/apache/accumulo/core/iterators/IteratorUtil.java
index 2b2c699..98392f6 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/IteratorUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/IteratorUtil.java
@@ -50,16 +50,16 @@ import org.apache.thrift.TSerializer;
 import org.apache.thrift.protocol.TBinaryProtocol;
 
 import com.google.common.base.Preconditions;
+
 public class IteratorUtil {
-  
+
   private static final Logger log = Logger.getLogger(IteratorUtil.class);
-  
+
   public static enum IteratorScope {
     majc, minc, scan;
 
     /**
-     * Fetch the correct configuration key prefix for the given scope. Throws an
-     * IllegalArgumentException if no property exists for the given scope.
+     * Fetch the correct configuration key prefix for the given scope. Throws an IllegalArgumentException if no property exists for the given scope.
      */
     public static Property getProperty(IteratorScope scope) {
       Preconditions.checkNotNull(scope);
@@ -75,7 +75,7 @@ public class IteratorUtil {
       }
     }
   }
-  
+
   public static class IterInfoComparator implements Comparator<IterInfo>, Serializable {
     private static final long serialVersionUID = 1L;
 
@@ -83,44 +83,45 @@ public class IteratorUtil {
     public int compare(IterInfo o1, IterInfo o2) {
       return (o1.priority < o2.priority ? -1 : (o1.priority == o2.priority ? 0 : 1));
     }
-    
+
   }
-  
+
   /**
    * Generate the initial (default) properties for a table
+   *
    * @param limitVersion
-   *   include a VersioningIterator at priority 20 that retains a single version of a given K/V pair.
+   *          include a VersioningIterator at priority 20 that retains a single version of a given K/V pair.
    * @return A map of Table properties
    */
   public static Map<String,String> generateInitialTableProperties(boolean limitVersion) {
     TreeMap<String,String> props = new TreeMap<String,String>();
-    
+
     if (limitVersion) {
-        for (IteratorScope iterScope : IteratorScope.values()) {
-          props.put(Property.TABLE_ITERATOR_PREFIX + iterScope.name() + ".vers", "20," + VersioningIterator.class.getName());
-          props.put(Property.TABLE_ITERATOR_PREFIX + iterScope.name() + ".vers.opt.maxVersions", "1");
-        }
+      for (IteratorScope iterScope : IteratorScope.values()) {
+        props.put(Property.TABLE_ITERATOR_PREFIX + iterScope.name() + ".vers", "20," + VersioningIterator.class.getName());
+        props.put(Property.TABLE_ITERATOR_PREFIX + iterScope.name() + ".vers.opt.maxVersions", "1");
+      }
     }
 
     props.put(Property.TABLE_CONSTRAINT_PREFIX.toString() + "1", DefaultKeySizeConstraint.class.getName());
 
     return props;
   }
-  
+
   public static int getMaxPriority(IteratorScope scope, AccumuloConfiguration conf) {
     List<IterInfo> iters = new ArrayList<IterInfo>();
     parseIterConf(scope, iters, new HashMap<String,Map<String,String>>(), conf);
-    
+
     int max = 0;
-    
+
     for (IterInfo iterInfo : iters) {
       if (iterInfo.priority > max)
         max = iterInfo.priority;
     }
-    
+
     return max;
   }
-  
+
   protected static void parseIterConf(IteratorScope scope, List<IterInfo> iters, Map<String,Map<String,String>> allOptions, AccumuloConfiguration conf) {
     final Property scopeProperty = IteratorScope.getProperty(scope);
     final String scopePropertyKey = scopeProperty.getKey();
@@ -128,7 +129,7 @@ public class IteratorUtil {
     for (Entry<String,String> entry : conf.getAllPropertiesWithPrefix(scopeProperty).entrySet()) {
       String suffix = entry.getKey().substring(scopePropertyKey.length());
       String suffixSplit[] = suffix.split("\\.", 3);
-      
+
       if (suffixSplit.length == 1) {
         String sa[] = entry.getValue().split(",");
         int prio = Integer.parseInt(sa[0]);
@@ -137,15 +138,15 @@ public class IteratorUtil {
       } else if (suffixSplit.length == 3 && suffixSplit[1].equals("opt")) {
         String iterName = suffixSplit[0];
         String optName = suffixSplit[2];
-        
+
         Map<String,String> options = allOptions.get(iterName);
         if (options == null) {
           options = new HashMap<String,String>();
           allOptions.put(iterName, options);
         }
-        
+
         options.put(optName, entry.getValue());
-        
+
       } else {
         log.warn("Unrecognizable option: " + entry.getKey());
       }
@@ -153,13 +154,13 @@ public class IteratorUtil {
 
     Collections.sort(iters, new IterInfoComparator());
   }
-  
+
   public static String findIterator(IteratorScope scope, String className, AccumuloConfiguration conf, Map<String,String> opts) {
     ArrayList<IterInfo> iters = new ArrayList<IterInfo>();
     Map<String,Map<String,String>> allOptions = new HashMap<String,Map<String,String>>();
-    
+
     parseIterConf(scope, iters, allOptions, conf);
-    
+
     for (IterInfo iterInfo : iters)
       if (iterInfo.className.equals(className)) {
         Map<String,String> tmpOpts = allOptions.get(iterInfo.iterName);
@@ -168,46 +169,46 @@ public class IteratorUtil {
         }
         return iterInfo.iterName;
       }
-    
+
     return null;
   }
-  
+
   public static <K extends WritableComparable<?>,V extends Writable> SortedKeyValueIterator<K,V> loadIterators(IteratorScope scope,
       SortedKeyValueIterator<K,V> source, KeyExtent extent, AccumuloConfiguration conf, IteratorEnvironment env) throws IOException {
     List<IterInfo> emptyList = Collections.emptyList();
     Map<String,Map<String,String>> emptyMap = Collections.emptyMap();
     return loadIterators(scope, source, extent, conf, emptyList, emptyMap, env);
   }
-  
+
   public static <K extends WritableComparable<?>,V extends Writable> SortedKeyValueIterator<K,V> loadIterators(IteratorScope scope,
       SortedKeyValueIterator<K,V> source, KeyExtent extent, AccumuloConfiguration conf, List<IteratorSetting> iterators, IteratorEnvironment env)
       throws IOException {
-    
+
     List<IterInfo> ssiList = new ArrayList<IterInfo>();
     Map<String,Map<String,String>> ssio = new HashMap<String,Map<String,String>>();
-    
+
     for (IteratorSetting is : iterators) {
       ssiList.add(new IterInfo(is.getPriority(), is.getIteratorClass(), is.getName()));
       ssio.put(is.getName(), is.getOptions());
     }
-    
+
     return loadIterators(scope, source, extent, conf, ssiList, ssio, env, true);
   }
-  
+
   public static <K extends WritableComparable<?>,V extends Writable> SortedKeyValueIterator<K,V> loadIterators(IteratorScope scope,
       SortedKeyValueIterator<K,V> source, KeyExtent extent, AccumuloConfiguration conf, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio,
       IteratorEnvironment env) throws IOException {
     return loadIterators(scope, source, extent, conf, ssiList, ssio, env, true);
   }
-  
+
   public static <K extends WritableComparable<?>,V extends Writable> SortedKeyValueIterator<K,V> loadIterators(IteratorScope scope,
       SortedKeyValueIterator<K,V> source, KeyExtent extent, AccumuloConfiguration conf, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio,
       IteratorEnvironment env, boolean useAccumuloClassLoader) throws IOException {
     List<IterInfo> iters = new ArrayList<IterInfo>(ssiList);
     Map<String,Map<String,String>> allOptions = new HashMap<String,Map<String,String>>();
-    
+
     parseIterConf(scope, iters, allOptions, conf);
-    
+
     for (Entry<String,Map<String,String>> entry : ssio.entrySet()) {
       if (entry.getValue() == null)
         continue;
@@ -218,37 +219,37 @@ public class IteratorUtil {
         options.putAll(entry.getValue());
       }
     }
-    
+
     return loadIterators(source, iters, allOptions, env, useAccumuloClassLoader, conf.get(Property.TABLE_CLASSPATH));
   }
-  
+
   @SuppressWarnings("unchecked")
   public static <K extends WritableComparable<?>,V extends Writable> SortedKeyValueIterator<K,V> loadIterators(SortedKeyValueIterator<K,V> source,
       Collection<IterInfo> iters, Map<String,Map<String,String>> iterOpts, IteratorEnvironment env, boolean useAccumuloClassLoader, String context)
       throws IOException {
     // wrap the source in a SynchronizedIterator in case any of the additional configured iterators want to use threading
     SortedKeyValueIterator<K,V> prev = new SynchronizedIterator<K,V>(source);
-    
+
     try {
       for (IterInfo iterInfo : iters) {
-       
+
         Class<? extends SortedKeyValueIterator<K,V>> clazz;
-        if (useAccumuloClassLoader){
+        if (useAccumuloClassLoader) {
           if (context != null && !context.equals(""))
             clazz = (Class<? extends SortedKeyValueIterator<K,V>>) AccumuloVFSClassLoader.getContextManager().loadClass(context, iterInfo.className,
                 SortedKeyValueIterator.class);
           else
             clazz = (Class<? extends SortedKeyValueIterator<K,V>>) AccumuloVFSClassLoader.loadClass(iterInfo.className, SortedKeyValueIterator.class);
-        }else{
+        } else {
           clazz = (Class<? extends SortedKeyValueIterator<K,V>>) Class.forName(iterInfo.className).asSubclass(SortedKeyValueIterator.class);
         }
         SortedKeyValueIterator<K,V> skvi = clazz.newInstance();
-        
+
         Map<String,String> options = iterOpts.get(iterInfo.iterName);
-        
+
         if (options == null)
           options = Collections.emptyMap();
-        
+
         skvi.init(prev, options, env);
         prev = skvi;
       }
@@ -264,73 +265,72 @@ public class IteratorUtil {
     }
     return prev;
   }
-  
+
   public static Range maximizeStartKeyTimeStamp(Range range) {
     Range seekRange = range;
-    
+
     if (range.getStartKey() != null && range.getStartKey().getTimestamp() != Long.MAX_VALUE) {
       Key seekKey = new Key(seekRange.getStartKey());
       seekKey.setTimestamp(Long.MAX_VALUE);
       seekRange = new Range(seekKey, true, range.getEndKey(), range.isEndKeyInclusive());
     }
-    
+
     return seekRange;
   }
-  
+
   public static Range minimizeEndKeyTimeStamp(Range range) {
     Range seekRange = range;
-    
+
     if (range.getEndKey() != null && range.getEndKey().getTimestamp() != Long.MIN_VALUE) {
       Key seekKey = new Key(seekRange.getEndKey());
       seekKey.setTimestamp(Long.MIN_VALUE);
       seekRange = new Range(range.getStartKey(), range.isStartKeyInclusive(), seekKey, true);
     }
-    
+
     return seekRange;
   }
-  
+
   public static TIteratorSetting toTIteratorSetting(IteratorSetting is) {
     return new TIteratorSetting(is.getPriority(), is.getName(), is.getIteratorClass(), is.getOptions());
   }
-  
+
   public static IteratorSetting toIteratorSetting(TIteratorSetting tis) {
     return new IteratorSetting(tis.getPriority(), tis.getName(), tis.getIteratorClass(), tis.getProperties());
   }
 
   public static IteratorConfig toIteratorConfig(List<IteratorSetting> iterators) {
     ArrayList<TIteratorSetting> tisList = new ArrayList<TIteratorSetting>();
-    
+
     for (IteratorSetting iteratorSetting : iterators) {
       tisList.add(toTIteratorSetting(iteratorSetting));
     }
-    
+
     return new IteratorConfig(tisList);
   }
-  
+
   public static List<IteratorSetting> toIteratorSettings(IteratorConfig ic) {
     List<IteratorSetting> ret = new ArrayList<IteratorSetting>();
     for (TIteratorSetting tIteratorSetting : ic.getIterators()) {
       ret.add(toIteratorSetting(tIteratorSetting));
     }
-    
+
     return ret;
   }
 
   public static byte[] encodeIteratorSettings(IteratorConfig iterators) {
     TSerializer tser = new TSerializer(new TBinaryProtocol.Factory());
-    
+
     try {
       return tser.serialize(iterators);
     } catch (TException e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   public static byte[] encodeIteratorSettings(List<IteratorSetting> iterators) {
     return encodeIteratorSettings(toIteratorConfig(iterators));
   }
 
-
   public static List<IteratorSetting> decodeIteratorSettings(byte[] enc) {
     TDeserializer tdser = new TDeserializer(new TBinaryProtocol.Factory());
     IteratorConfig ic = new IteratorConfig();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/LargeRowFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/LargeRowFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/LargeRowFilter.java
index 4194202..75155f9 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/LargeRowFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/LargeRowFilter.java
@@ -18,10 +18,10 @@ package org.apache.accumulo.core.iterators;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.LargeRowFilter}
  */
 @Deprecated
 public class LargeRowFilter extends org.apache.accumulo.core.iterators.user.LargeRowFilter {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
index 3c6c6d2..ad44b15 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
@@ -32,11 +32,11 @@ import org.apache.hadoop.io.WritableUtils;
 
 /**
  * A TypedValueCombiner that translates each Value to a Long before reducing, then encodes the reduced Long back to a Value.
- * 
+ *
  * Subclasses must implement a typedReduce method: public Long typedReduce(Key key, Iterator<Long> iter);
- * 
+ *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Longs) for all non-deleted versions of that Key.
- * 
+ *
  * A required option for this Combiner is "type" which indicates which type of Encoder to use to encode and decode Longs into Values. Supported types are
  * VARNUM, LONG, and STRING which indicate the VarNumEncoder, LongEncoder, and StringEncoder respectively.
  */
@@ -44,10 +44,10 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
   public static final Encoder<Long> FIXED_LEN_ENCODER = new FixedLenEncoder();
   public static final Encoder<Long> VAR_LEN_ENCODER = new VarLenEncoder();
   public static final Encoder<Long> STRING_ENCODER = new StringEncoder();
-  
+
   protected static final String TYPE = "type";
   protected static final String CLASS_PREFIX = "class:";
-  
+
   public static enum Type {
     /**
      * indicates a variable-length encoding of a Long using {@link LongCombiner.VarLenEncoder}
@@ -62,13 +62,13 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
      */
     STRING
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     setEncoder(options);
   }
-  
+
   private void setEncoder(Map<String,String> options) {
     String type = options.get(TYPE);
     if (type == null)
@@ -92,7 +92,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
       }
     }
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -101,7 +101,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
     io.addNamedOption(TYPE, "<VARLEN|FIXEDLEN|STRING|fullClassName>");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
@@ -113,7 +113,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
     }
     return true;
   }
-  
+
   /**
    * An Encoder that uses a variable-length encoding for Longs. It uses WritableUtils.writeVLong and WritableUtils.readVLong for encoding and decoding.
    */
@@ -122,16 +122,16 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
     public byte[] encode(Long v) {
       ByteArrayOutputStream baos = new ByteArrayOutputStream();
       DataOutputStream dos = new DataOutputStream(baos);
-      
+
       try {
         WritableUtils.writeVLong(dos, v);
       } catch (IOException e) {
         throw new NumberFormatException(e.getMessage());
       }
-      
+
       return baos.toByteArray();
     }
-    
+
     @Override
     public Long decode(byte[] b) {
       DataInputStream dis = new DataInputStream(new ByteArrayInputStream(b));
@@ -142,7 +142,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
       }
     }
   }
-  
+
   /**
    * An Encoder that uses an 8-byte encoding for Longs.
    */
@@ -160,12 +160,12 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
       b[7] = (byte) (l >>> 0);
       return b;
     }
-    
+
     @Override
     public Long decode(byte[] b) {
       return decode(b, 0);
     }
-    
+
     public static long decode(byte[] b, int offset) {
       if (b.length < offset + 8)
         throw new ValueFormatException("trying to convert to long, but byte array isn't long enough, wanted " + (offset + 8) + " found " + b.length);
@@ -173,7 +173,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
           + ((long) (b[offset + 4] & 255) << 24) + ((b[offset + 5] & 255) << 16) + ((b[offset + 6] & 255) << 8) + ((b[offset + 7] & 255) << 0));
     }
   }
-  
+
   /**
    * An Encoder that uses a String representation of Longs. It uses Long.toString and Long.parseLong for encoding and decoding.
    */
@@ -182,7 +182,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
     public byte[] encode(Long v) {
       return Long.toString(v).getBytes(UTF_8);
     }
-    
+
     @Override
     public Long decode(byte[] b) {
       try {
@@ -192,7 +192,7 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
       }
     }
   }
-  
+
   public static long safeAdd(long a, long b) {
     long aSign = Long.signum(a);
     long bSign = Long.signum(b);
@@ -207,10 +207,10 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
     }
     return a + b;
   }
-  
+
   /**
    * A convenience method for setting the long encoding type.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param type
@@ -219,10 +219,10 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
   public static void setEncodingType(IteratorSetting is, LongCombiner.Type type) {
     is.addOption(TYPE, type.toString());
   }
-  
+
   /**
    * A convenience method for setting the long encoding type.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClass
@@ -231,10 +231,10 @@ public abstract class LongCombiner extends TypedValueCombiner<Long> {
   public static void setEncodingType(IteratorSetting is, Class<? extends Encoder<Long>> encoderClass) {
     is.addOption(TYPE, CLASS_PREFIX + encoderClass.getName());
   }
-  
+
   /**
    * A convenience method for setting the long encoding type.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param encoderClassName

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
index 6069a39..97fd29a 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
@@ -25,10 +25,10 @@ import java.util.Map;
  * The OptionDescriber interface allows you to set up iterator properties interactively in the accumulo shell. If your iterator and/or filter must implement
  * this interface for the interactive part. The alternative would be to manually set configuration options with the config -t tableName property=value. If you
  * go the manual route, be careful to use the correct structure for the property and to set all the properties required for the iterator.
- * 
+ *
  * OptionDescribers will need to implement two methods: {@code describeOptions()} which returns an instance of {@link IteratorOptions} and
  * {@code validateOptions(Map<String,String> options)} which is intended to throw an exception or return false if the options are not acceptable.
- * 
+ *
  */
 public interface OptionDescriber {
   public class IteratorOptions {
@@ -36,10 +36,10 @@ public interface OptionDescriber {
     public ArrayList<String> unnamedOptionDescriptions;
     public String name;
     public String description;
-    
+
     /**
      * IteratorOptions holds the name, description, and option information for an iterator.
-     * 
+     *
      * @param name
      *          is the distinguishing name for the iterator or filter
      * @param description
@@ -63,63 +63,63 @@ public interface OptionDescriber {
         this.unnamedOptionDescriptions = new ArrayList<String>(unnamedOptionDescriptions);
       this.description = description;
     }
-    
+
     public Map<String,String> getNamedOptions() {
       return namedOptions;
     }
-    
+
     public List<String> getUnnamedOptionDescriptions() {
       return unnamedOptionDescriptions;
     }
-    
+
     public String getName() {
       return name;
     }
-    
+
     public String getDescription() {
       return description;
     }
-    
+
     public void setNamedOptions(Map<String,String> namedOptions) {
       this.namedOptions = new LinkedHashMap<String,String>(namedOptions);
     }
-    
+
     public void setUnnamedOptionDescriptions(List<String> unnamedOptionDescriptions) {
       this.unnamedOptionDescriptions = new ArrayList<String>(unnamedOptionDescriptions);
     }
-    
+
     public void setName(String name) {
       this.name = name;
     }
-    
+
     public void setDescription(String description) {
       this.description = description;
     }
-    
+
     public void addNamedOption(String name, String description) {
       if (namedOptions == null)
         namedOptions = new LinkedHashMap<String,String>();
       namedOptions.put(name, description);
     }
-    
+
     public void addUnnamedOption(String description) {
       if (unnamedOptionDescriptions == null)
         unnamedOptionDescriptions = new ArrayList<String>();
       unnamedOptionDescriptions.add(description);
     }
   }
-  
+
   /**
    * Gets an iterator options object that contains information needed to configure this iterator. This object will be used by the accumulo shell to prompt the
    * user to input the appropriate information.
-   * 
+   *
    * @return an iterator options object
    */
   IteratorOptions describeOptions();
-  
+
   /**
    * Check to see if an options map contains all options required by an iterator and that the option values are in the expected formats.
-   * 
+   *
    * @param options
    *          a map of option names to option values
    * @return true if options are valid, false otherwise (IllegalArgumentException preferred)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/OrIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/OrIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/OrIterator.java
index 361c494..98c8cac 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/OrIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/OrIterator.java
@@ -36,31 +36,31 @@ import org.apache.hadoop.io.Text;
  */
 
 public class OrIterator implements SortedKeyValueIterator<Key,Value> {
-  
+
   private TermSource currentTerm;
   private ArrayList<TermSource> sources;
   private PriorityQueue<TermSource> sorted = new PriorityQueue<TermSource>(5);
   private static final Text nullText = new Text();
   private static final Key nullKey = new Key();
-  
+
   protected static class TermSource implements Comparable<TermSource> {
     public SortedKeyValueIterator<Key,Value> iter;
     public Text term;
     public Collection<ByteSequence> seekColfams;
-    
+
     public TermSource(TermSource other) {
       this.iter = other.iter;
       this.term = other.term;
       this.seekColfams = other.seekColfams;
     }
-    
+
     public TermSource(SortedKeyValueIterator<Key,Value> iter, Text term) {
       this.iter = iter;
       this.term = term;
       // The desired column families for this source is the term itself
-      this.seekColfams = Collections.<ByteSequence>singletonList(new ArrayByteSequence(term.getBytes(), 0, term.getLength()));
+      this.seekColfams = Collections.<ByteSequence> singletonList(new ArrayByteSequence(term.getBytes(), 0, term.getLength()));
     }
-    
+
     public int compareTo(TermSource o) {
       // NOTE: If your implementation can have more than one row in a tablet,
       // you must compare row key here first, then column qualifier.
@@ -69,40 +69,40 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
       return this.iter.getTopKey().compareColumnQualifier(o.iter.getTopKey().getColumnQualifier());
     }
   }
-  
+
   public OrIterator() {
     this.sources = new ArrayList<TermSource>();
   }
-  
+
   private OrIterator(OrIterator other, IteratorEnvironment env) {
     this.sources = new ArrayList<TermSource>();
-    
+
     for (TermSource TS : other.sources)
       this.sources.add(new TermSource(TS.iter.deepCopy(env), TS.term));
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new OrIterator(this, env);
   }
-  
+
   public void addTerm(SortedKeyValueIterator<Key,Value> source, Text term, IteratorEnvironment env) {
     this.sources.add(new TermSource(source.deepCopy(env), term));
   }
-  
+
   @Override
   final public void next() throws IOException {
-    
+
     if (currentTerm == null)
       return;
-    
+
     // Advance currentTerm
     currentTerm.iter.next();
-    
+
     // See if currentTerm is still valid, remove if not
     if (!(currentTerm.iter.hasTop()) || ((currentTerm.term != null) && (currentTerm.term.compareTo(currentTerm.iter.getTopKey().getColumnFamily()) != 0)))
       currentTerm = null;
-    
+
     // optimization.
     // if size == 0, currentTerm is the only item left,
     // OR there are no items left.
@@ -115,25 +115,25 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
       currentTerm = sorted.poll();
     }
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-    
+
     // If sources.size is 0, there is nothing to process, so just return.
     if (sources.size() == 0) {
       currentTerm = null;
       return;
     }
-    
+
     // Optimization for when there is only one term.
     // Yes, this is lots of duplicate code, but the speed works...
     // and we don't have a priority queue of size 0 or 1.
     if (sources.size() == 1) {
-      
+
       if (currentTerm == null)
         currentTerm = sources.get(0);
       Range newRange = null;
-      
+
       if (range != null) {
         if ((range.getStartKey() == null) || (range.getStartKey().getRow() == null))
           newRange = range;
@@ -147,7 +147,7 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
         }
       }
       currentTerm.iter.seek(newRange, currentTerm.seekColfams, true);
-      
+
       // If there is no top key
       // OR we are:
       // 1) NOT an iterator
@@ -155,14 +155,14 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
       // then ignore it as a valid source
       if (!(currentTerm.iter.hasTop()) || ((currentTerm.term != null) && (currentTerm.term.compareTo(currentTerm.iter.getTopKey().getColumnFamily()) != 0)))
         currentTerm = null;
-      
+
       // Otherwise, source is valid.
       return;
     }
-    
+
     // Clear the PriorityQueue so that we can re-populate it.
     sorted.clear();
-    
+
     // This check is put in here to guard against the "initial seek"
     // crashing us because the topkey term does not match.
     // Note: It is safe to do the "sources.size() == 1" above
@@ -170,14 +170,14 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
     if (currentTerm == null) {
       for (TermSource TS : sources) {
         TS.iter.seek(range, TS.seekColfams, true);
-        
+
         if ((TS.iter.hasTop()) && ((TS.term != null) && (TS.term.compareTo(TS.iter.getTopKey().getColumnFamily()) == 0)))
           sorted.add(TS);
       }
       currentTerm = sorted.poll();
       return;
     }
-    
+
     TermSource TS = null;
     Iterator<TermSource> iter = sources.iterator();
     // For each term, seek forward.
@@ -185,7 +185,7 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
     while (iter.hasNext()) {
       TS = iter.next();
       Range newRange = null;
-      
+
       if (range != null) {
         if ((range.getStartKey() == null) || (range.getStartKey().getRow() == null))
           newRange = range;
@@ -198,10 +198,10 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
           newRange = new Range((newKey == null) ? nullKey : newKey, true, range.getEndKey(), false);
         }
       }
-      
+
       // Seek only to the term for this source as a column family
       TS.iter.seek(newRange, TS.seekColfams, true);
-      
+
       // If there is no top key
       // OR we are:
       // 1) NOT an iterator
@@ -209,30 +209,30 @@ public class OrIterator implements SortedKeyValueIterator<Key,Value> {
       // then ignore it as a valid source
       if (!(TS.iter.hasTop()) || ((TS.term != null) && (TS.term.compareTo(TS.iter.getTopKey().getColumnFamily()) != 0)))
         iter.remove();
-      
+
       // Otherwise, source is valid. Add it to the sources.
       sorted.add(TS);
     }
-    
+
     // And set currentTerm = the next valid key/term.
     currentTerm = sorted.poll();
   }
-  
+
   @Override
   final public Key getTopKey() {
     return currentTerm.iter.getTopKey();
   }
-  
+
   @Override
   final public Value getTopValue() {
     return currentTerm.iter.getTopValue();
   }
-  
+
   @Override
   final public boolean hasTop() {
     return currentTerm != null;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/RowDeletingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/RowDeletingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/RowDeletingIterator.java
index a79a6ee..ee6989f 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/RowDeletingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/RowDeletingIterator.java
@@ -18,10 +18,10 @@ package org.apache.accumulo.core.iterators;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.RowDeletingIterator}
  */
 @Deprecated
 public class RowDeletingIterator extends org.apache.accumulo.core.iterators.user.RowDeletingIterator {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyIterator.java
index ba45e27..896eb67 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyIterator.java
@@ -23,28 +23,28 @@ import org.apache.accumulo.core.data.Value;
 
 public class SortedKeyIterator extends WrappingIterator implements OptionDescriber {
   private static final Value NOVALUE = new Value(new byte[0]);
-  
+
   public SortedKeyIterator() {}
-  
+
   public SortedKeyIterator(SortedKeyIterator other, IteratorEnvironment env) {
     setSource(other.getSource().deepCopy(env));
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new SortedKeyIterator(this, env);
   }
-  
+
   @Override
   public Value getTopValue() {
     return NOVALUE;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     return new IteratorOptions("keyset", SortedKeyIterator.class.getSimpleName() + " filters out values, but leaves keys intact", null, null);
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     return true;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyValueIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyValueIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyValueIterator.java
index 5b6c4ec..7c65565 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyValueIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/SortedKeyValueIterator.java
@@ -33,7 +33,7 @@ import org.apache.hadoop.io.WritableComparable;
 public interface SortedKeyValueIterator<K extends WritableComparable<?>,V extends Writable> {
   /**
    * Initializes the iterator. Data should not be read from the source in this method.
-   * 
+   *
    * @param source
    *          <tt>SortedKeyValueIterator</tt> source to read data from.
    * @param options
@@ -48,21 +48,21 @@ public interface SortedKeyValueIterator<K extends WritableComparable<?>,V extend
    *              if not supported.
    */
   void init(SortedKeyValueIterator<K,V> source, Map<String,String> options, IteratorEnvironment env) throws IOException;
-  
+
   /**
    * Returns true if the iterator has more elements.
-   * 
+   *
    * @return <tt>true</tt> if the iterator has more elements.
    * @exception IllegalStateException
    *              if called before seek.
    */
   boolean hasTop();
-  
+
   /**
    * Advances to the next K,V pair. Note that in minor compaction scope and in non-full major compaction scopes the iterator may see deletion entries. These
    * entries should be preserved by all iterators except ones that are strictly scan-time iterators that will never be configured for the minc or majc scopes.
    * Deletion entries are only removed during full major compactions.
-   * 
+   *
    * @throws IOException
    *           if an I/O error occurs.
    * @exception IllegalStateException
@@ -71,24 +71,24 @@ public interface SortedKeyValueIterator<K extends WritableComparable<?>,V extend
    *              if next element doesn't exist.
    */
   void next() throws IOException;
-  
+
   /**
    * Seeks to the first key in the Range, restricting the resulting K,V pairs to those with the specified columns. An iterator does not have to stop at the end
    * of the range. The whole range is provided so that iterators can make optimizations.
-   * 
+   *
    * Seek may be called multiple times with different parameters after {@link #init} is called.
-   * 
+   *
    * Iterators that examine groups of adjacent key/value pairs (e.g. rows) to determine their top key and value should be sure that they properly handle a seek
    * to a key in the middle of such a group (e.g. the middle of a row). Even if the client always seeks to a range containing an entire group (a,c), the tablet
    * server could send back a batch of entries corresponding to (a,b], then reseek the iterator to range (b,c) when the scan is continued.
    *
-   * {@code columnFamilies} is used, at the lowest level, to determine which data blocks inside of an RFile need to be opened for this iterator. This set of data
-   * blocks is also the set of locality groups defined for the given table. If no columnFamilies are provided, the data blocks for all locality groups inside of
-   * the correct RFile will be opened and seeked in an attempt to find the correct start key, regardless of the startKey in the {@code range}.
+   * {@code columnFamilies} is used, at the lowest level, to determine which data blocks inside of an RFile need to be opened for this iterator. This set of
+   * data blocks is also the set of locality groups defined for the given table. If no columnFamilies are provided, the data blocks for all locality groups
+   * inside of the correct RFile will be opened and seeked in an attempt to find the correct start key, regardless of the startKey in the {@code range}.
    *
    * In an Accumulo instance in which multiple locality groups exist for a table, it is important to ensure that {@code columnFamilies} is properly set to the
    * minimum required column families to ensure that data from separate locality groups is not inadvertently read.
-   * 
+   *
    * @param range
    *          <tt>Range</tt> of keys to iterate over.
    * @param columnFamilies
@@ -101,12 +101,12 @@ public interface SortedKeyValueIterator<K extends WritableComparable<?>,V extend
    *              if there are problems with the parameters.
    */
   void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException;
-  
+
   /**
    * Returns top key. Can be called 0 or more times without affecting behavior of next() or hasTop(). Note that in minor compaction scope and in non-full major
    * compaction scopes the iterator may see deletion entries. These entries should be preserved by all iterators except ones that are strictly scan-time
    * iterators that will never be configured for the minc or majc scopes. Deletion entries are only removed during full major compactions.
-   * 
+   *
    * @return <tt>K</tt>
    * @exception IllegalStateException
    *              if called before seek.
@@ -114,10 +114,10 @@ public interface SortedKeyValueIterator<K extends WritableComparable<?>,V extend
    *              if top element doesn't exist.
    */
   K getTopKey();
-  
+
   /**
    * Returns top value. Can be called 0 or more times without affecting behavior of next() or hasTop().
-   * 
+   *
    * @return <tt>V</tt>
    * @exception IllegalStateException
    *              if called before seek.
@@ -125,12 +125,12 @@ public interface SortedKeyValueIterator<K extends WritableComparable<?>,V extend
    *              if top element doesn't exist.
    */
   V getTopValue();
-  
+
   /**
    * Creates a deep copy of this iterator as though seek had not yet been called. init should be called on an iterator before deepCopy is called. init should
    * not need to be called on the copy that is returned by deepCopy; that is, when necessary init should be called in the deepCopy method on the iterator it
    * returns. The behavior is unspecified if init is called after deepCopy either on the original or the copy.
-   * 
+   *
    * @param env
    *          <tt>IteratorEnvironment</tt> environment in which iterator is being run.
    * @return <tt>SortedKeyValueIterator</tt> a copy of this iterator (with the same source and settings).

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/SortedMapIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/SortedMapIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/SortedMapIterator.java
index 2371e4d..cf385a9 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/SortedMapIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/SortedMapIterator.java
@@ -32,67 +32,66 @@ import org.apache.accumulo.core.iterators.system.InterruptibleIterator;
 
 /**
  * A simple iterator over a Java SortedMap
- * 
- * Note that this class is intended as an in-memory replacement for RFile$Reader, so its behavior reflects
- * the same assumptions; namely, that this iterator is not responsible for respecting the columnFamilies
- * passed into seek().  If you want a Map-backed Iterator that returns only sought CFs, construct a new 
+ *
+ * Note that this class is intended as an in-memory replacement for RFile$Reader, so its behavior reflects the same assumptions; namely, that this iterator is
+ * not responsible for respecting the columnFamilies passed into seek(). If you want a Map-backed Iterator that returns only sought CFs, construct a new
  * ColumnFamilySkippingIterator(new SortedMapIterator(map)).
- * 
+ *
  * @see org.apache.accumulo.core.iterators.system.ColumnFamilySkippingIterator
- * 
+ *
  */
 
 public class SortedMapIterator implements InterruptibleIterator {
   private Iterator<Entry<Key,Value>> iter;
   private Entry<Key,Value> entry;
-  
+
   private SortedMap<Key,Value> map;
   private Range range;
-  
+
   private AtomicBoolean interruptFlag;
   private int interruptCheckCount = 0;
-  
+
   public SortedMapIterator deepCopy(IteratorEnvironment env) {
     return new SortedMapIterator(map, interruptFlag);
   }
-  
+
   private SortedMapIterator(SortedMap<Key,Value> map, AtomicBoolean interruptFlag) {
     this.map = map;
     iter = null;
     this.range = new Range();
     entry = null;
-    
+
     this.interruptFlag = interruptFlag;
   }
-  
+
   public SortedMapIterator(SortedMap<Key,Value> map) {
     this(map, null);
   }
-  
+
   @Override
   public Key getTopKey() {
     return entry.getKey();
   }
-  
+
   @Override
   public Value getTopValue() {
     return entry.getValue();
   }
-  
+
   @Override
   public boolean hasTop() {
     return entry != null;
   }
-  
+
   @Override
   public void next() throws IOException {
-    
+
     if (entry == null)
       throw new IllegalStateException();
-    
+
     if (interruptFlag != null && interruptCheckCount++ % 100 == 0 && interruptFlag.get())
       throw new IterationInterruptedException();
-    
+
     if (iter.hasNext()) {
       entry = iter.next();
       if (range.afterEndKey((Key) entry.getKey())) {
@@ -100,22 +99,22 @@ public class SortedMapIterator implements InterruptibleIterator {
       }
     } else
       entry = null;
-    
+
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-    
+
     if (interruptFlag != null && interruptFlag.get())
       throw new IterationInterruptedException();
-    
+
     this.range = range;
-    
+
     Key key = range.getStartKey();
     if (key == null) {
       key = new Key();
     }
-    
+
     iter = map.tailMap(key).entrySet().iterator();
     if (iter.hasNext()) {
       entry = iter.next();
@@ -124,16 +123,16 @@ public class SortedMapIterator implements InterruptibleIterator {
       }
     } else
       entry = null;
-    
+
     while (hasTop() && range.beforeStartKey(getTopKey())) {
       next();
     }
   }
-  
+
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     this.interruptFlag = flag;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
index 6577f51..63fff81 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
@@ -28,19 +28,19 @@ import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
 
 /**
  * A Combiner that decodes each Value to type V before reducing, then encodes the result of typedReduce back to Value.
- * 
+ *
  * Subclasses must implement a typedReduce method: public V typedReduce(Key key, Iterator<V> iter);
- * 
+ *
  * This typedReduce method will be passed the most recent Key and an iterator over the Values (translated to Vs) for all non-deleted versions of that Key.
- * 
+ *
  * Subclasses may implement a switch on the "type" variable to choose an Encoder in their init method.
  */
 public abstract class TypedValueCombiner<V> extends Combiner {
   private Encoder<V> encoder = null;
   private boolean lossy = false;
-  
+
   protected static final String LOSSY = "lossy";
-  
+
   /**
    * A Java Iterator that translates an Iterator<Value> to an Iterator<V> using the decode method of an Encoder.
    */
@@ -48,16 +48,16 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     private Iterator<Value> source;
     private Encoder<V> encoder;
     private boolean lossy;
-    
+
     /**
      * Constructs an Iterator<V> from an Iterator<Value>
-     * 
+     *
      * @param iter
      *          The source iterator
-     * 
+     *
      * @param encoder
      *          The Encoder whose decode method is used to translate from Value to V
-     * 
+     *
      * @param lossy
      *          Determines whether to error on failure to decode or ignore and move on
      */
@@ -66,15 +66,15 @@ public abstract class TypedValueCombiner<V> extends Combiner {
       this.encoder = encoder;
       this.lossy = lossy;
     }
-    
+
     V next = null;
     boolean hasNext = false;
-    
+
     @Override
     public boolean hasNext() {
       if (hasNext)
         return true;
-      
+
       while (true) {
         if (!source.hasNext())
           return false;
@@ -87,7 +87,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
         }
       }
     }
-    
+
     @Override
     public V next() {
       if (!hasNext && !hasNext())
@@ -97,32 +97,32 @@ public abstract class TypedValueCombiner<V> extends Combiner {
       hasNext = false;
       return toRet;
     }
-    
+
     @Override
     public void remove() {
       source.remove();
     }
   }
-  
+
   /**
    * An interface for translating from byte[] to V and back.
    */
   public interface Encoder<V> {
     byte[] encode(V v);
-    
+
     V decode(byte[] b) throws ValueFormatException;
   }
-  
+
   /**
    * Sets the Encoder<V> used to translate Values to V and back.
    */
   protected void setEncoder(Encoder<V> encoder) {
     this.encoder = encoder;
   }
-  
+
   /**
    * Instantiates and sets the Encoder<V> used to translate Values to V and back.
-   * 
+   *
    * @throws IllegalArgumentException
    *           if ClassNotFoundException, InstantiationException, or IllegalAccessException occurs
    */
@@ -139,10 +139,10 @@ public abstract class TypedValueCombiner<V> extends Combiner {
       throw new IllegalArgumentException(e);
     }
   }
-  
+
   /**
    * Tests whether v remains the same when encoded and decoded with the current encoder.
-   * 
+   *
    * @throws IllegalStateException
    *           if an encoder has not been set.
    * @throws IllegalArgumentException
@@ -153,10 +153,10 @@ public abstract class TypedValueCombiner<V> extends Combiner {
       throw new IllegalStateException("encoder has not been initialized");
     testEncoder(encoder, v);
   }
-  
+
   /**
    * Tests whether v remains the same when encoded and decoded with the given encoder.
-   * 
+   *
    * @throws IllegalArgumentException
    *           if the test fails.
    */
@@ -168,7 +168,7 @@ public abstract class TypedValueCombiner<V> extends Combiner {
       throw new IllegalArgumentException(encoder.getClass().getName() + " doesn't encode " + v.getClass().getName());
     }
   }
-  
+
   @SuppressWarnings("unchecked")
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
@@ -176,18 +176,18 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     newInstance.setEncoder(encoder);
     return newInstance;
   }
-  
+
   @Override
   public Value reduce(Key key, Iterator<Value> iter) {
     return new Value(encoder.encode(typedReduce(key, new VIterator<V>(iter, encoder, lossy))));
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     setLossyness(options);
   }
-  
+
   private void setLossyness(Map<String,String> options) {
     String loss = options.get(LOSSY);
     if (loss == null)
@@ -195,14 +195,14 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     else
       lossy = Boolean.parseBoolean(loss);
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
     io.addNamedOption(LOSSY, "if true, failed decodes are ignored. Otherwise combiner will error on failed decodes (default false): <TRUE|FALSE>");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
@@ -214,11 +214,11 @@ public abstract class TypedValueCombiner<V> extends Combiner {
     }
     return true;
   }
-  
+
   /**
    * A convenience method to set the "lossy" option on a TypedValueCombiner. If true, the combiner will ignore any values which fail to decode. Otherwise, the
    * combiner will throw an error which will interrupt the action (and prevent potential data loss). False is the default behavior.
-   * 
+   *
    * @param is
    *          iterator settings object to configure
    * @param lossy
@@ -227,16 +227,16 @@ public abstract class TypedValueCombiner<V> extends Combiner {
   public static void setLossyness(IteratorSetting is, boolean lossy) {
     is.addOption(LOSSY, Boolean.toString(lossy));
   }
-  
+
   /**
    * Reduces a list of V into a single V.
-   * 
+   *
    * @param key
    *          The most recent version of the Key being reduced.
-   * 
+   *
    * @param iter
    *          An iterator over the V for different versions of the key.
-   * 
+   *
    * @return The combined V.
    */
   public abstract V typedReduce(Key key, Iterator<V> iter);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/ValueFormatException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/ValueFormatException.java b/core/src/main/java/org/apache/accumulo/core/iterators/ValueFormatException.java
index 7ede7fe..6e52794 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/ValueFormatException.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/ValueFormatException.java
@@ -20,7 +20,7 @@ package org.apache.accumulo.core.iterators;
  * Exception used for TypedValueCombiner and it's Encoders decode() function
  */
 public class ValueFormatException extends IllegalArgumentException {
-  
+
   public ValueFormatException(String string) {
     super(string);
   }
@@ -30,5 +30,5 @@ public class ValueFormatException extends IllegalArgumentException {
   }
 
   private static final long serialVersionUID = 4170291568272971821L;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/VersioningIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/VersioningIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/VersioningIterator.java
index cc37067..d849275 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/VersioningIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/VersioningIterator.java
@@ -21,13 +21,13 @@ import org.apache.accumulo.core.data.Value;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.VersioningIterator}
  */
 @Deprecated
 public class VersioningIterator extends org.apache.accumulo.core.iterators.user.VersioningIterator {
   public VersioningIterator() {}
-  
+
   public VersioningIterator(SortedKeyValueIterator<Key,Value> iterator, int maxVersions) {
     super();
     this.setSource(iterator);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/WholeRowIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/WholeRowIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/WholeRowIterator.java
index 6bc0bde..7432a88 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/WholeRowIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/WholeRowIterator.java
@@ -18,10 +18,10 @@ package org.apache.accumulo.core.iterators;
 
 /**
  * This class remains here for backwards compatibility.
- * 
+ *
  * @deprecated since 1.4, replaced by {@link org.apache.accumulo.core.iterators.user.WholeRowIterator}
  */
 @Deprecated
 public class WholeRowIterator extends org.apache.accumulo.core.iterators.user.WholeRowIterator {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/WrappingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/WrappingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/WrappingIterator.java
index 060fa76..7723ef1 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/WrappingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/WrappingIterator.java
@@ -28,32 +28,32 @@ import org.apache.accumulo.core.data.Value;
 /**
  * A convenience class for implementing iterators that select, but do not modify, entries read from a source iterator. Default implementations exist for all
  * methods, but {@link #deepCopy} will throw an <code>UnsupportedOperationException</code>.
- * 
+ *
  * This iterator has some checks in place to enforce the iterator contract. Specifically, it verifies that it has a source iterator and that {@link #seek} has
  * been called before any data is read. If either of these conditions does not hold true, an <code>IllegalStateException</code> will be thrown. In particular,
  * this means that <code>getSource().seek</code> and <code>super.seek</code> no longer perform identical actions. Implementors should take note of this and if
  * <code>seek</code> is overridden, ensure that <code>super.seek</code> is called before data is read.
  */
 public abstract class WrappingIterator implements SortedKeyValueIterator<Key,Value> {
-  
+
   private SortedKeyValueIterator<Key,Value> source = null;
   boolean seenSeek = false;
-  
+
   protected void setSource(SortedKeyValueIterator<Key,Value> source) {
     this.source = source;
   }
-  
+
   protected SortedKeyValueIterator<Key,Value> getSource() {
     if (source == null)
       throw new IllegalStateException("getting null source");
     return source;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public Key getTopKey() {
     if (source == null)
@@ -62,7 +62,7 @@ public abstract class WrappingIterator implements SortedKeyValueIterator<Key,Val
       throw new IllegalStateException("never been seeked");
     return getSource().getTopKey();
   }
-  
+
   @Override
   public Value getTopValue() {
     if (source == null)
@@ -71,7 +71,7 @@ public abstract class WrappingIterator implements SortedKeyValueIterator<Key,Val
       throw new IllegalStateException("never been seeked");
     return getSource().getTopValue();
   }
-  
+
   @Override
   public boolean hasTop() {
     if (source == null)
@@ -80,13 +80,13 @@ public abstract class WrappingIterator implements SortedKeyValueIterator<Key,Val
       throw new IllegalStateException("never been seeked");
     return getSource().hasTop();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     this.setSource(source);
-    
+
   }
-  
+
   @Override
   public void next() throws IOException {
     if (source == null)
@@ -95,11 +95,11 @@ public abstract class WrappingIterator implements SortedKeyValueIterator<Key,Val
       throw new IllegalStateException("never been seeked");
     getSource().next();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     getSource().seek(range, columnFamilies, inclusive);
     seenSeek = true;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/Aggregator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/Aggregator.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/Aggregator.java
index 2264bd9..f9183dc 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/Aggregator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/Aggregator.java
@@ -24,8 +24,8 @@ import org.apache.accumulo.core.data.Value;
 @Deprecated
 public interface Aggregator {
   void reset();
-  
+
   void collect(Value value);
-  
+
   Value aggregate();
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/LongSummation.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/LongSummation.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/LongSummation.java
index bee96ac..de86ec4 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/LongSummation.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/LongSummation.java
@@ -29,11 +29,11 @@ import org.apache.log4j.Logger;
 public class LongSummation implements Aggregator {
   private static final Logger log = Logger.getLogger(LongSummation.class);
   long sum = 0;
-  
+
   public Value aggregate() {
     return new Value(longToBytes(sum));
   }
-  
+
   public void collect(Value value) {
     try {
       sum += bytesToLong(value.get());
@@ -41,22 +41,22 @@ public class LongSummation implements Aggregator {
       log.error(LongSummation.class.getSimpleName() + " trying to convert bytes to long, but byte array isn't length 8");
     }
   }
-  
+
   public void reset() {
     sum = 0;
   }
-  
+
   public static long bytesToLong(byte[] b) throws IOException {
     return bytesToLong(b, 0);
   }
-  
+
   public static long bytesToLong(byte[] b, int offset) throws IOException {
     if (b.length < offset + 8)
       throw new IOException("trying to convert to long, but byte array isn't long enough, wanted " + (offset + 8) + " found " + b.length);
     return (((long) b[offset + 0] << 56) + ((long) (b[offset + 1] & 255) << 48) + ((long) (b[offset + 2] & 255) << 40) + ((long) (b[offset + 3] & 255) << 32)
         + ((long) (b[offset + 4] & 255) << 24) + ((b[offset + 5] & 255) << 16) + ((b[offset + 6] & 255) << 8) + ((b[offset + 7] & 255) << 0));
   }
-  
+
   public static byte[] longToBytes(long l) {
     byte[] b = new byte[8];
     b[0] = (byte) (l >>> 56);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumArraySummation.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumArraySummation.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumArraySummation.java
index 1d1e4cb..ca00337 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumArraySummation.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumArraySummation.java
@@ -32,7 +32,7 @@ import org.apache.hadoop.io.WritableUtils;
 @Deprecated
 public class NumArraySummation implements Aggregator {
   long[] sum = new long[0];
-  
+
   public Value aggregate() {
     try {
       return new Value(NumArraySummation.longArrayToBytes(sum));
@@ -40,7 +40,7 @@ public class NumArraySummation implements Aggregator {
       throw new RuntimeException(e);
     }
   }
-  
+
   public void collect(Value value) {
     long[] la;
     try {
@@ -48,7 +48,7 @@ public class NumArraySummation implements Aggregator {
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
     if (la.length > sum.length) {
       for (int i = 0; i < sum.length; i++) {
         la[i] = NumSummation.safeAdd(la[i], sum[i]);
@@ -60,34 +60,34 @@ public class NumArraySummation implements Aggregator {
       }
     }
   }
-  
+
   public static byte[] longArrayToBytes(long[] la) throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
-    
+
     WritableUtils.writeVInt(dos, la.length);
     for (int i = 0; i < la.length; i++) {
       WritableUtils.writeVLong(dos, la[i]);
     }
-    
+
     return baos.toByteArray();
   }
-  
+
   public static long[] bytesToLongArray(byte[] b) throws IOException {
     DataInputStream dis = new DataInputStream(new ByteArrayInputStream(b));
     int len = WritableUtils.readVInt(dis);
-    
+
     long[] la = new long[len];
-    
+
     for (int i = 0; i < len; i++) {
       la[i] = WritableUtils.readVLong(dis);
     }
-    
+
     return la;
   }
-  
+
   public void reset() {
     sum = new long[0];
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumSummation.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumSummation.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumSummation.java
index 7994195..9c28776 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumSummation.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/NumSummation.java
@@ -32,7 +32,7 @@ import org.apache.hadoop.io.WritableUtils;
 @Deprecated
 public class NumSummation implements Aggregator {
   long sum = 0l;
-  
+
   public Value aggregate() {
     try {
       return new Value(NumSummation.longToBytes(sum));
@@ -40,7 +40,7 @@ public class NumSummation implements Aggregator {
       throw new RuntimeException(e);
     }
   }
-  
+
   public void collect(Value value) {
     long l;
     try {
@@ -48,24 +48,24 @@ public class NumSummation implements Aggregator {
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
     sum = NumSummation.safeAdd(sum, l);
   }
-  
+
   public static byte[] longToBytes(long l) throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
-    
+
     WritableUtils.writeVLong(dos, l);
-    
+
     return baos.toByteArray();
   }
-  
+
   public static long bytesToLong(byte[] b) throws IOException {
     DataInputStream dis = new DataInputStream(new ByteArrayInputStream(b));
     return WritableUtils.readVLong(dis);
   }
-  
+
   public static long safeAdd(long a, long b) {
     long aSign = Long.signum(a);
     long bSign = Long.signum(b);
@@ -80,9 +80,9 @@ public class NumSummation implements Aggregator {
     }
     return a + b;
   }
-  
+
   public void reset() {
     sum = 0l;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMax.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMax.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMax.java
index 36c777a..e086b22 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMax.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMax.java
@@ -24,22 +24,22 @@ import org.apache.accumulo.core.data.Value;
  */
 @Deprecated
 public class StringMax implements Aggregator {
-  
+
   long max = Long.MIN_VALUE;
-  
+
   public Value aggregate() {
     return new Value(Long.toString(max).getBytes());
   }
-  
+
   public void collect(Value value) {
     long l = Long.parseLong(new String(value.get()));
     if (l > max) {
       max = l;
     }
   }
-  
+
   public void reset() {
     max = Long.MIN_VALUE;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMin.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMin.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMin.java
index 1e6b7c7..48855b3 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMin.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringMin.java
@@ -24,22 +24,22 @@ import org.apache.accumulo.core.data.Value;
  */
 @Deprecated
 public class StringMin implements Aggregator {
-  
+
   long min = Long.MAX_VALUE;
-  
+
   public Value aggregate() {
     return new Value(Long.toString(min).getBytes());
   }
-  
+
   public void collect(Value value) {
     long l = Long.parseLong(new String(value.get()));
     if (l < min) {
       min = l;
     }
   }
-  
+
   public void reset() {
     min = Long.MAX_VALUE;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringSummation.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringSummation.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringSummation.java
index 00bfd10..63a8297 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringSummation.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/StringSummation.java
@@ -24,19 +24,19 @@ import org.apache.accumulo.core.data.Value;
  */
 @Deprecated
 public class StringSummation implements Aggregator {
-  
+
   long sum = 0;
-  
+
   public Value aggregate() {
     return new Value(Long.toString(sum).getBytes());
   }
-  
+
   public void collect(Value value) {
     sum += Long.parseLong(new String(value.get()));
   }
-  
+
   public void reset() {
     sum = 0;
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfiguration.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfiguration.java
index a74dd6d..41c0374 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorConfiguration.java
@@ -24,11 +24,11 @@ import org.apache.hadoop.io.Text;
  */
 @Deprecated
 public class AggregatorConfiguration extends PerColumnIteratorConfig {
-  
+
   public AggregatorConfiguration(Text columnFamily, String aggClassName) {
     super(columnFamily, aggClassName);
   }
-  
+
   public AggregatorConfiguration(Text columnFamily, Text columnQualifier, String aggClassName) {
     super(columnFamily, columnQualifier, aggClassName);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorSet.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorSet.java b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorSet.java
index ad33fa2..c874cc8 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorSet.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/aggregation/conf/AggregatorSet.java
@@ -31,11 +31,11 @@ public class AggregatorSet extends ColumnToClassMapping<Aggregator> {
   public AggregatorSet(Map<String,String> opts) throws InstantiationException, IllegalAccessException, ClassNotFoundException, IOException {
     super(opts, Aggregator.class);
   }
-  
+
   public AggregatorSet() {
     super();
   }
-  
+
   public Aggregator getAggregator(Key k) {
     return getObject(k);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnSet.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnSet.java b/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnSet.java
index 01ed523..c1edf5d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnSet.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnSet.java
@@ -31,21 +31,21 @@ import org.apache.hadoop.io.Text;
 public class ColumnSet {
   private Set<ColFamHashKey> objectsCF;
   private Set<ColHashKey> objectsCol;
-  
+
   private ColHashKey lookupCol = new ColHashKey();
   private ColFamHashKey lookupCF = new ColFamHashKey();
-  
+
   public ColumnSet() {
     objectsCF = new HashSet<ColFamHashKey>();
     objectsCol = new HashSet<ColHashKey>();
   }
-  
+
   public ColumnSet(Collection<String> objectStrings) {
     this();
-    
+
     for (String column : objectStrings) {
       Pair<Text,Text> pcic = ColumnSet.decodeColumns(column);
-      
+
       if (pcic.getSecond() == null) {
         add(pcic.getFirst());
       } else {
@@ -53,15 +53,15 @@ public class ColumnSet {
       }
     }
   }
-  
+
   protected void add(Text colf) {
     objectsCF.add(new ColFamHashKey(new Text(colf)));
   }
-  
+
   protected void add(Text colf, Text colq) {
     objectsCol.add(new ColHashKey(colf, colq));
   }
-  
+
   public boolean contains(Key key) {
     // lookup column family and column qualifier
     if (objectsCol.size() > 0) {
@@ -69,36 +69,36 @@ public class ColumnSet {
       if (objectsCol.contains(lookupCol))
         return true;
     }
-    
+
     // lookup just column family
     if (objectsCF.size() > 0) {
       lookupCF.set(key);
       return objectsCF.contains(lookupCF);
     }
-    
+
     return false;
   }
-  
+
   public boolean isEmpty() {
     return objectsCol.size() == 0 && objectsCF.size() == 0;
   }
 
   public static String encodeColumns(Text columnFamily, Text columnQualifier) {
     StringBuilder sb = new StringBuilder();
-    
+
     encode(sb, columnFamily);
     if (columnQualifier != null) {
       sb.append(':');
       encode(sb, columnQualifier);
     }
-    
+
     return sb.toString();
   }
-  
+
   static void encode(StringBuilder sb, Text t) {
     for (int i = 0; i < t.getLength(); i++) {
       int b = (0xff & t.getBytes()[i]);
-      
+
       // very inefficient code
       if ((b >= 'a' && b <= 'z') || (b >= 'A' && b <= 'Z') || (b >= '0' && b <= '9') || b == '_' || b == '-') {
         sb.append((char) b);
@@ -115,16 +115,16 @@ public class ColumnSet {
       if (!validChar)
         return false;
     }
-    
+
     return true;
   }
 
   public static Pair<Text,Text> decodeColumns(String columns) {
     if (!isValidEncoding(columns))
       throw new IllegalArgumentException("Invalid encoding " + columns);
-  
+
     String[] cols = columns.split(":");
-    
+
     if (cols.length == 1) {
       return new Pair<Text,Text>(decode(cols[0]), null);
     } else if (cols.length == 2) {
@@ -136,9 +136,9 @@ public class ColumnSet {
 
   static Text decode(String s) {
     Text t = new Text();
-    
+
     byte[] sb = s.getBytes(UTF_8);
-    
+
     // very inefficient code
     for (int i = 0; i < sb.length; i++) {
       if (sb[i] != '%') {
@@ -150,7 +150,7 @@ public class ColumnSet {
         t.append(new byte[] {(byte) b}, 0, 1);
       }
     }
-    
+
     return t;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnToClassMapping.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnToClassMapping.java b/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnToClassMapping.java
index 59063d9..84a6996 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnToClassMapping.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnToClassMapping.java
@@ -29,39 +29,39 @@ import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
 import org.apache.hadoop.io.Text;
 
 public class ColumnToClassMapping<K> {
-  
+
   private HashMap<ColFamHashKey,K> objectsCF;
   private HashMap<ColHashKey,K> objectsCol;
-  
+
   private ColHashKey lookupCol = new ColHashKey();
   private ColFamHashKey lookupCF = new ColFamHashKey();
-  
+
   public ColumnToClassMapping() {
     objectsCF = new HashMap<ColFamHashKey,K>();
     objectsCol = new HashMap<ColHashKey,K>();
   }
-  
+
   public ColumnToClassMapping(Map<String,String> objectStrings, Class<? extends K> c) throws InstantiationException, IllegalAccessException,
       ClassNotFoundException, IOException {
-	  this(objectStrings, c, null);
+    this(objectStrings, c, null);
   }
 
   public ColumnToClassMapping(Map<String,String> objectStrings, Class<? extends K> c, String context) throws InstantiationException, IllegalAccessException,
-  ClassNotFoundException, IOException {
-	  this();
-    
+      ClassNotFoundException, IOException {
+    this();
+
     for (Entry<String,String> entry : objectStrings.entrySet()) {
       String column = entry.getKey();
       String className = entry.getValue();
-      
+
       Pair<Text,Text> pcic = ColumnSet.decodeColumns(column);
-      
+
       Class<?> clazz;
       if (context != null && !context.equals(""))
         clazz = AccumuloVFSClassLoader.getContextManager().getClassLoader(context).loadClass(className);
       else
         clazz = AccumuloVFSClassLoader.loadClass(className, c);
-      
+
       @SuppressWarnings("unchecked")
       K inst = (K) clazz.newInstance();
       if (pcic.getSecond() == null) {
@@ -71,18 +71,18 @@ public class ColumnToClassMapping<K> {
       }
     }
   }
-  
+
   protected void addObject(Text colf, K obj) {
     objectsCF.put(new ColFamHashKey(new Text(colf)), obj);
   }
-  
+
   protected void addObject(Text colf, Text colq, K obj) {
     objectsCol.put(new ColHashKey(colf, colq), obj);
   }
-  
+
   public K getObject(Key key) {
     K obj = null;
-    
+
     // lookup column family and column qualifier
     if (objectsCol.size() > 0) {
       lookupCol.set(key);
@@ -91,16 +91,16 @@ public class ColumnToClassMapping<K> {
         return obj;
       }
     }
-    
+
     // lookup just column family
     if (objectsCF.size() > 0) {
       lookupCF.set(key);
       obj = objectsCF.get(lookupCF);
     }
-    
+
     return obj;
   }
-  
+
   public boolean isEmpty() {
     return objectsCol.size() == 0 && objectsCF.size() == 0;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnUtil.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnUtil.java b/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnUtil.java
index 984f069..a6a3e65 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnUtil.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/conf/ColumnUtil.java
@@ -24,89 +24,89 @@ public class ColumnUtil {
   private static int hash(byte[] bytes, int offset, int len) {
     int hash = 1;
     int end = offset + len;
-    
+
     for (int i = offset; i < end; i++)
       hash = (31 * hash) + bytes[i];
-    
+
     return hash;
   }
-  
+
   private static int hash(ByteSequence bs) {
     return hash(bs.getBackingArray(), bs.offset(), bs.length());
   }
-  
+
   public static class ColFamHashKey {
     Text columnFamily;
-    
+
     Key key;
-    
+
     private int hashCode;
-    
+
     ColFamHashKey() {
       columnFamily = null;
     }
-    
+
     ColFamHashKey(Text cf) {
       columnFamily = cf;
       hashCode = hash(columnFamily.getBytes(), 0, columnFamily.getLength());
     }
-    
+
     void set(Key key) {
       this.key = key;
       hashCode = hash(key.getColumnFamilyData());
     }
-    
+
     public int hashCode() {
       return hashCode;
     }
-    
+
     public boolean equals(Object o) {
       if (o instanceof ColFamHashKey)
         return equals((ColFamHashKey) o);
       return false;
     }
-    
+
     public boolean equals(ColFamHashKey ohk) {
       if (columnFamily == null)
         return key.compareColumnFamily(ohk.columnFamily) == 0;
       return ohk.key.compareColumnFamily(columnFamily) == 0;
     }
   }
-  
+
   public static class ColHashKey {
     Text columnFamily;
     Text columnQualifier;
-    
+
     Key key;
-    
+
     private int hashValue;
-    
+
     ColHashKey() {
       columnFamily = null;
       columnQualifier = null;
     }
-    
+
     ColHashKey(Text cf, Text cq) {
       columnFamily = cf;
       columnQualifier = cq;
       hashValue = hash(columnFamily.getBytes(), 0, columnFamily.getLength()) + hash(columnQualifier.getBytes(), 0, columnQualifier.getLength());
     }
-    
+
     void set(Key key) {
       this.key = key;
       hashValue = hash(key.getColumnFamilyData()) + hash(key.getColumnQualifierData());
     }
-    
+
     public int hashCode() {
       return hashValue;
     }
-    
+
     public boolean equals(Object o) {
       if (o instanceof ColHashKey)
         return equals((ColHashKey) o);
       return false;
     }
-    
+
     public boolean equals(ColHashKey ohk) {
       if (columnFamily == null)
         return key.compareColumnFamily(ohk.columnFamily) == 0 && key.compareColumnQualifier(ohk.columnQualifier) == 0;


[25/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/LocalityGroupUtilTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/LocalityGroupUtilTest.java b/core/src/test/java/org/apache/accumulo/core/util/LocalityGroupUtilTest.java
index 333ab83..32d5f62 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/LocalityGroupUtilTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/LocalityGroupUtilTest.java
@@ -16,12 +16,15 @@
  */
 package org.apache.accumulo.core.util;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNotNull;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
 import java.util.HashSet;
 import java.util.Map;
 import java.util.Set;
 
-import static org.junit.Assert.*;
-
 import org.apache.accumulo.core.conf.ConfigurationCopy;
 import org.apache.accumulo.core.data.ArrayByteSequence;
 import org.apache.accumulo.core.data.ByteSequence;
@@ -30,10 +33,10 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class LocalityGroupUtilTest {
-  
+
   @Test
   public void testColumnFamilySet() {
-    
+
     ConfigurationCopy conf = new ConfigurationCopy();
     conf.set("table.group.lg1", "cf1,cf2");
     conf.set("table.groups.enabled", "lg1");
@@ -53,7 +56,7 @@ public class LocalityGroupUtilTest {
       fail();
     } catch (LocalityGroupConfigurationError err) {}
   }
-  
+
   @Test
   public void testEncoding() throws Exception {
     byte test1[] = new byte[256];
@@ -62,18 +65,18 @@ public class LocalityGroupUtilTest {
       test1[i] = (byte) (0xff & i);
       test2[i] = (byte) (0xff & (255 - i));
     }
-    
+
     ArrayByteSequence bs1 = new ArrayByteSequence(test1);
-    
+
     String ecf = LocalityGroupUtil.encodeColumnFamily(bs1);
-    
+
     // System.out.println(ecf);
-    
+
     ByteSequence bs2 = LocalityGroupUtil.decodeColumnFamily(ecf);
-    
+
     assertEquals(bs1, bs2);
     assertEquals(ecf, LocalityGroupUtil.encodeColumnFamily(bs2));
-    
+
     // test encoding multiple column fams containing binary data
     HashSet<Text> in = new HashSet<Text>();
     HashSet<ByteSequence> in2 = new HashSet<ByteSequence>();
@@ -82,8 +85,8 @@ public class LocalityGroupUtilTest {
     in.add(new Text(test2));
     in2.add(new ArrayByteSequence(test2));
     Set<ByteSequence> out = LocalityGroupUtil.decodeColumnFamilies(LocalityGroupUtil.encodeColumnFamilies(in));
-    
+
     assertEquals(in2, out);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/MergeTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/MergeTest.java b/core/src/test/java/org/apache/accumulo/core/util/MergeTest.java
index 544e9e4..95d7b66 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/MergeTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/MergeTest.java
@@ -30,11 +30,11 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class MergeTest {
-  
+
   static class MergeTester extends Merge {
     public List<List<Size>> merges = new ArrayList<List<Size>>();
     public List<Size> tablets = new ArrayList<Size>();
-    
+
     MergeTester(Integer... sizes) {
       Text start = null;
       for (Integer size : sizes) {
@@ -48,21 +48,21 @@ public class MergeTest {
         tablets.add(new Size(extent, size));
       }
     }
-    
+
     @Override
     protected void message(String format, Object... args) {}
-    
+
     @Override
     protected Iterator<Size> getSizeIterator(Connector conn, String tablename, final Text start, final Text end) throws MergeException {
       final Iterator<Size> impl = tablets.iterator();
       return new Iterator<Size>() {
         Size next = skip();
-        
+
         @Override
         public boolean hasNext() {
           return next != null;
         }
-        
+
         private Size skip() {
           while (impl.hasNext()) {
             Size candidate = impl.next();
@@ -78,21 +78,21 @@ public class MergeTest {
           }
           return null;
         }
-        
+
         @Override
         public Size next() {
           Size result = next;
           next = skip();
           return result;
         }
-        
+
         @Override
         public void remove() {
           impl.remove();
         }
       };
     }
-    
+
     @Override
     protected void merge(Connector conn, String table, List<Size> sizes, int numToMerge) throws MergeException {
       List<Size> merge = new ArrayList<Size>();
@@ -102,7 +102,7 @@ public class MergeTest {
       merges.add(merge);
     }
   }
-  
+
   static private int[] sizes(List<Size> sizes) {
     int result[] = new int[sizes.size()];
     int i = 0;
@@ -111,7 +111,7 @@ public class MergeTest {
     }
     return result;
   }
-  
+
   @Test
   public void testMergomatic() throws Exception {
     // Merge everything to the last tablet
@@ -120,14 +120,14 @@ public class MergeTest {
     test.mergomatic(null, "table", null, null, 1000, false);
     assertEquals(1, test.merges.size());
     assertArrayEquals(new int[] {10, 20, 30}, sizes(test.merges.get(i = 0)));
-    
+
     // Merge ranges around tablets that are big enough
     test = new MergeTester(1, 2, 100, 1000, 17, 1000, 4, 5, 6, 900);
     test.mergomatic(null, "table", null, null, 1000, false);
     assertEquals(2, test.merges.size());
     assertArrayEquals(new int[] {1, 2, 100}, sizes(test.merges.get(i = 0)));
     assertArrayEquals(new int[] {4, 5, 6, 900}, sizes(test.merges.get(++i)));
-    
+
     // Test the force option
     test = new MergeTester(1, 2, 100, 1000, 17, 1000, 4, 5, 6, 900);
     test.mergomatic(null, "table", null, null, 1000, true);
@@ -135,25 +135,25 @@ public class MergeTest {
     assertArrayEquals(new int[] {1, 2, 100}, sizes(test.merges.get(i = 0)));
     assertArrayEquals(new int[] {17, 1000}, sizes(test.merges.get(++i)));
     assertArrayEquals(new int[] {4, 5, 6, 900}, sizes(test.merges.get(++i)));
-    
+
     // Limit the low-end of the merges
     test = new MergeTester(1, 2, 1000, 17, 1000, 4, 5, 6, 900);
     test.mergomatic(null, "table", new Text("00004"), null, 1000, false);
     assertEquals(1, test.merges.size());
     assertArrayEquals(new int[] {4, 5, 6, 900}, sizes(test.merges.get(i = 0)));
-    
+
     // Limit the upper end of the merges
     test = new MergeTester(1, 2, 1000, 17, 1000, 4, 5, 6, 900);
     test.mergomatic(null, "table", null, new Text("00004"), 1000, false);
     assertEquals(1, test.merges.size());
     assertArrayEquals(new int[] {1, 2}, sizes(test.merges.get(i = 0)));
-    
+
     // Limit both ends
     test = new MergeTester(1, 2, 1000, 17, 1000, 4, 5, 6, 900);
     test.mergomatic(null, "table", new Text("00002"), new Text("00004"), 1000, true);
     assertEquals(1, test.merges.size());
     assertArrayEquals(new int[] {17, 1000}, sizes(test.merges.get(i = 0)));
-    
+
     // Clump up tablets into larger values
     test = new MergeTester(100, 250, 500, 600, 100, 200, 500, 200);
     test.mergomatic(null, "table", null, null, 1000, false);
@@ -162,5 +162,5 @@ public class MergeTest {
     assertArrayEquals(new int[] {600, 100, 200}, sizes(test.merges.get(++i)));
     assertArrayEquals(new int[] {500, 200}, sizes(test.merges.get(++i)));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/PairTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/PairTest.java b/core/src/test/java/org/apache/accumulo/core/util/PairTest.java
index b837f05..60af90e 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/PairTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/PairTest.java
@@ -102,15 +102,15 @@ public class PairTest {
   public void testFromEntry() {
     Entry<Integer,String> entry = new SimpleImmutableEntry<Integer,String>(10, "IO");
 
-    Pair<Integer, String> pair0 = Pair.fromEntry(entry);
+    Pair<Integer,String> pair0 = Pair.fromEntry(entry);
     assertEquals(entry.getKey(), pair0.getFirst());
     assertEquals(entry.getValue(), pair0.getSecond());
 
-    Pair<Object,Object> pair = Pair.<Object, Object, Integer, String>fromEntry(entry);
+    Pair<Object,Object> pair = Pair.<Object,Object,Integer,String> fromEntry(entry);
     assertEquals(entry.getKey(), pair.getFirst());
     assertEquals(entry.getValue(), pair.getSecond());
 
-    Pair<Number,CharSequence> pair2 = Pair.<Number, CharSequence, Integer, String>fromEntry(entry);
+    Pair<Number,CharSequence> pair2 = Pair.<Number,CharSequence,Integer,String> fromEntry(entry);
     assertEquals(entry.getKey(), pair2.getFirst());
     assertEquals(entry.getValue(), pair2.getSecond());
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/PartitionerTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/PartitionerTest.java b/core/src/test/java/org/apache/accumulo/core/util/PartitionerTest.java
index 276720c..c4538ab 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/PartitionerTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/PartitionerTest.java
@@ -37,88 +37,88 @@ import org.junit.Test;
 public class PartitionerTest {
   @Test
   public void test1() {
-    
+
     @SuppressWarnings("unchecked")
     Map<ByteSequence,MutableLong>[] groups = new Map[2];
-    
+
     groups[0] = new HashMap<ByteSequence,MutableLong>();
     groups[0].put(new ArrayByteSequence("cf1"), new MutableLong(1));
     groups[0].put(new ArrayByteSequence("cf2"), new MutableLong(1));
-    
+
     groups[1] = new HashMap<ByteSequence,MutableLong>();
     groups[1].put(new ArrayByteSequence("cf3"), new MutableLong(1));
-    
+
     Partitioner p1 = new Partitioner(groups);
-    
+
     Mutation m1 = new Mutation("r1");
     m1.put("cf1", "cq1", "v1");
-    
+
     Mutation m2 = new Mutation("r2");
     m2.put("cf1", "cq1", "v2");
     m2.put("cf2", "cq2", "v3");
-    
+
     Mutation m3 = new Mutation("r3");
     m3.put("cf1", "cq1", "v4");
     m3.put("cf3", "cq2", "v5");
-    
+
     Mutation m4 = new Mutation("r4");
     m4.put("cf1", "cq1", "v6");
     m4.put("cf3", "cq2", "v7");
     m4.put("cf5", "cq3", "v8");
-    
+
     Mutation m5 = new Mutation("r5");
     m5.put("cf5", "cq3", "v9");
-    
+
     List<Mutation> mutations = Arrays.asList(m1, m2, m3, m4, m5);
     @SuppressWarnings("unchecked")
     List<Mutation>[] partitioned = new List[3];
-    
+
     for (int i = 0; i < partitioned.length; i++) {
       partitioned[i] = new ArrayList<Mutation>();
     }
-    
+
     p1.partition(mutations, partitioned);
-    
+
     m1 = new Mutation("r1");
     m1.put("cf1", "cq1", "v1");
-    
+
     m2 = new Mutation("r2");
     m2.put("cf1", "cq1", "v2");
     m2.put("cf2", "cq2", "v3");
-    
+
     m3 = new Mutation("r3");
     m3.put("cf1", "cq1", "v4");
-    
+
     m4 = new Mutation("r4");
     m4.put("cf1", "cq1", "v6");
-    
-    Assert.assertEquals(toKeySet(m1,m2,m3,m4), toKeySet(partitioned[0]));
-    
+
+    Assert.assertEquals(toKeySet(m1, m2, m3, m4), toKeySet(partitioned[0]));
+
     m3 = new Mutation("r3");
     m3.put("cf3", "cq2", "v5");
-    
+
     m4 = new Mutation("r4");
     m4.put("cf3", "cq2", "v7");
-    
-    Assert.assertEquals(toKeySet(m3,m4),  toKeySet(partitioned[1]));
-    
+
+    Assert.assertEquals(toKeySet(m3, m4), toKeySet(partitioned[1]));
+
     m4 = new Mutation("r4");
     m4.put("cf5", "cq3", "v8");
-    
-    Assert.assertEquals(toKeySet(m4,m5),  toKeySet(partitioned[2]));
-    
+
+    Assert.assertEquals(toKeySet(m4, m5), toKeySet(partitioned[2]));
+
   }
 
-  private Set<Key> toKeySet(List<Mutation> mutations){
+  private Set<Key> toKeySet(List<Mutation> mutations) {
     return toKeySet(mutations.toArray(new Mutation[0]));
   }
-  
-  private Set<Key> toKeySet(Mutation ... expected) {
+
+  private Set<Key> toKeySet(Mutation... expected) {
     HashSet<Key> ret = new HashSet<Key>();
-    for (Mutation mutation : expected) 
-      for(ColumnUpdate cu : mutation.getUpdates())
-       ret.add(new Key(mutation.getRow(), cu.getColumnFamily(), cu.getColumnQualifier(), cu.getColumnVisibility(), cu.getTimestamp()));
-    
+    for (Mutation mutation : expected)
+      for (ColumnUpdate cu : mutation.getUpdates())
+        ret.add(new Key(mutation.getRow(), cu.getColumnFamily(), cu.getColumnQualifier(), cu.getColumnVisibility(), cu.getTimestamp()));
+
     return ret;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/TestVersion.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/TestVersion.java b/core/src/test/java/org/apache/accumulo/core/util/TestVersion.java
index 6783548..af3f391 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/TestVersion.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/TestVersion.java
@@ -22,10 +22,10 @@ public class TestVersion extends TestCase {
   Version make(String version) {
     return new Version(version);
   }
-  
+
   public void testOne() {
     Version v;
-    
+
     v = make("abc-1.2.3-ugly");
     assertTrue(v != null);
     assertTrue(v.getPackage().equals("abc"));
@@ -33,32 +33,32 @@ public class TestVersion extends TestCase {
     assertTrue(v.getMinorVersion() == 2);
     assertTrue(v.getReleaseVersion() == 3);
     assertTrue(v.getEtcetera().equals("ugly"));
-    
+
     v = make("3.2.1");
     assertTrue(v.getPackage() == null);
     assertTrue(v.getMajorVersion() == 3);
     assertTrue(v.getMinorVersion() == 2);
     assertTrue(v.getReleaseVersion() == 1);
     assertTrue(v.getEtcetera() == null);
-    
+
     v = make("55");
     assertTrue(v.getPackage() == null);
     assertTrue(v.getMajorVersion() == 55);
     assertTrue(v.getMinorVersion() == 0);
     assertTrue(v.getReleaseVersion() == 0);
     assertTrue(v.getEtcetera() == null);
-    
+
     v = make("7.1-beta");
     assertTrue(v.getPackage() == null);
     assertTrue(v.getMajorVersion() == 7);
     assertTrue(v.getMinorVersion() == 1);
     assertTrue(v.getReleaseVersion() == 0);
     assertTrue(v.getEtcetera().equals("beta"));
-    
+
     try {
       make("beta");
       fail("Should have thrown an error");
     } catch (IllegalArgumentException t) {}
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/TextUtilTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/TextUtilTest.java b/core/src/test/java/org/apache/accumulo/core/util/TextUtilTest.java
index b592626..cc45526 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/TextUtilTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/TextUtilTest.java
@@ -22,7 +22,7 @@ import org.apache.hadoop.io.Text;
 
 /**
  * Test the TextUtil class.
- * 
+ *
  */
 public class TextUtilTest extends TestCase {
   /**
@@ -40,5 +40,5 @@ public class TextUtilTest extends TestCase {
     assertTrue(TextUtil.getBytes(someText).length == smallerMessage.length());
     assertTrue((new Text(TextUtil.getBytes(someText))).equals(smallerMessageText));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/format/DefaultFormatterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/format/DefaultFormatterTest.java b/core/src/test/java/org/apache/accumulo/core/util/format/DefaultFormatterTest.java
index dd4d93f..7b654d0 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/format/DefaultFormatterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/format/DefaultFormatterTest.java
@@ -51,7 +51,7 @@ public class DefaultFormatterTest {
   @Test
   public void testAppendBytes() {
     StringBuilder sb = new StringBuilder();
-    byte[] data = new byte[] { 0, '\\', 'x', -0x01 };
+    byte[] data = new byte[] {0, '\\', 'x', -0x01};
 
     DefaultFormatter.appendValue(sb, new Value());
     assertEquals("", sb.toString());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/util/format/HexFormatterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/util/format/HexFormatterTest.java b/core/src/test/java/org/apache/accumulo/core/util/format/HexFormatterTest.java
index 7f4f3e5..64a3f1b 100644
--- a/core/src/test/java/org/apache/accumulo/core/util/format/HexFormatterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/util/format/HexFormatterTest.java
@@ -54,7 +54,6 @@ public class HexFormatterTest {
     assertEquals(new Text("\0"), formatter.interpretRow(new Text("0")));
   }
 
-
   @Test
   public void testRoundTripRows() {
     Text bytes = new Text(new byte[] {0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15});

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
index 893ed3f..fb460f6 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/Flush.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.core.client.Connector;
  * Simple example for using tableOperations() (like create, delete, flush, etc).
  */
 public class Flush {
-  
+
   public static void main(String[] args) {
     ClientOnRequiredTable opts = new ClientOnRequiredTable();
     opts.parseArgs(Flush.class.getName(), args);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchScanner.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchScanner.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchScanner.java
index b34261c..e984f1e 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchScanner.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchScanner.java
@@ -45,33 +45,33 @@ import com.beust.jcommander.Parameter;
  */
 class CountingVerifyingReceiver {
   private static final Logger log = Logger.getLogger(CountingVerifyingReceiver.class);
-  
+
   long count = 0;
   int expectedValueSize = 0;
   HashMap<Text,Boolean> expectedRows;
-  
+
   CountingVerifyingReceiver(HashMap<Text,Boolean> expectedRows, int expectedValueSize) {
     this.expectedRows = expectedRows;
     this.expectedValueSize = expectedValueSize;
   }
-  
+
   public void receive(Key key, Value value) {
-    
+
     String row = key.getRow().toString();
     long rowid = Integer.parseInt(row.split("_")[1]);
-    
+
     byte expectedValue[] = RandomBatchWriter.createValue(rowid, expectedValueSize);
-    
+
     if (!Arrays.equals(expectedValue, value.get())) {
       log.error("Got unexpected value for " + key + " expected : " + new String(expectedValue) + " got : " + new String(value.get()));
     }
-    
+
     if (!expectedRows.containsKey(key.getRow())) {
       log.error("Got unexpected key " + key);
     } else {
       expectedRows.put(key.getRow(), true);
     }
-    
+
     count++;
   }
 }
@@ -81,10 +81,10 @@ class CountingVerifyingReceiver {
  */
 public class RandomBatchScanner {
   private static final Logger log = Logger.getLogger(CountingVerifyingReceiver.class);
-  
+
   /**
    * Generate a number of ranges, each covering a single random row.
-   * 
+   *
    * @param num
    *          the number of ranges to generate
    * @param min
@@ -102,20 +102,20 @@ public class RandomBatchScanner {
     log.info(String.format("Generating %,d random queries...", num));
     while (ranges.size() < num) {
       long rowid = (abs(r.nextLong()) % (max - min)) + min;
-      
+
       Text row1 = new Text(String.format("row_%010d", rowid));
-      
+
       Range range = new Range(new Text(row1));
       ranges.add(range);
       expectedRows.put(row1, false);
     }
-    
+
     log.info("finished");
   }
-  
+
   /**
    * Prints a count of the number of rows mapped to false.
-   * 
+   *
    * @return boolean indicating "were all the rows found?"
    */
   private static boolean checkAllRowsFound(HashMap<Text,Boolean> expectedRows) {
@@ -124,18 +124,18 @@ public class RandomBatchScanner {
     for (Entry<Text,Boolean> entry : expectedRows.entrySet())
       if (!entry.getValue())
         count++;
-    
+
     if (count > 0) {
       log.warn("Did not find " + count + " rows");
       allFound = false;
     }
     return allFound;
   }
-  
+
   /**
    * Generates a number of random queries, verifies that the key/value pairs returned were in the queried ranges and that the values were generated by
    * {@link RandomBatchWriter#createValue(long, int)}. Prints information about the results.
-   * 
+   *
    * @param num
    *          the number of queries to generate
    * @param min
@@ -151,43 +151,43 @@ public class RandomBatchScanner {
    * @return boolean indicating "did the queries go fine?"
    */
   static boolean doRandomQueries(int num, long min, long max, int evs, Random r, BatchScanner tsbr) {
-    
+
     HashSet<Range> ranges = new HashSet<Range>(num);
     HashMap<Text,Boolean> expectedRows = new java.util.HashMap<Text,Boolean>();
-    
+
     generateRandomQueries(num, min, max, r, ranges, expectedRows);
-    
+
     tsbr.setRanges(ranges);
-    
+
     CountingVerifyingReceiver receiver = new CountingVerifyingReceiver(expectedRows, evs);
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     for (Entry<Key,Value> entry : tsbr) {
       receiver.receive(entry.getKey(), entry.getValue());
     }
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     log.info(String.format("%6.2f lookups/sec %6.2f secs%n", num / ((t2 - t1) / 1000.0), ((t2 - t1) / 1000.0)));
     log.info(String.format("num results : %,d%n", receiver.count));
-    
+
     return checkAllRowsFound(expectedRows);
   }
-  
-  public static class Opts  extends ClientOnRequiredTable {
-    @Parameter(names="--min", description="miniumum row that will be generated")
+
+  public static class Opts extends ClientOnRequiredTable {
+    @Parameter(names = "--min", description = "miniumum row that will be generated")
     long min = 0;
-    @Parameter(names="--max", description="maximum ow that will be generated")
+    @Parameter(names = "--max", description = "maximum ow that will be generated")
     long max = 0;
-    @Parameter(names="--num", required=true, description="number of ranges to generate")
+    @Parameter(names = "--num", required = true, description = "number of ranges to generate")
     int num = 0;
-    @Parameter(names="--size", required=true, description="size of the value to write")
+    @Parameter(names = "--size", required = true, description = "size of the value to write")
     int size = 0;
-    @Parameter(names="--seed", description="seed for pseudo-random number generator")
+    @Parameter(names = "--seed", description = "seed for pseudo-random number generator")
     Long seed = null;
   }
-  
+
   /**
    * Scans over a specified number of entries to Accumulo using a {@link BatchScanner}. Completes scans twice to compare times for a fresh query with those for
    * a repeated query which has cached metadata and connections already established.
@@ -196,32 +196,32 @@ public class RandomBatchScanner {
     Opts opts = new Opts();
     BatchScannerOpts bsOpts = new BatchScannerOpts();
     opts.parseArgs(RandomBatchScanner.class.getName(), args, bsOpts);
-    
+
     Connector connector = opts.getConnector();
     BatchScanner batchReader = connector.createBatchScanner(opts.getTableName(), opts.auths, bsOpts.scanThreads);
     batchReader.setTimeout(bsOpts.scanTimeout, TimeUnit.MILLISECONDS);
-    
+
     Random r;
     if (opts.seed == null)
       r = new Random();
     else
       r = new Random(opts.seed);
-    
+
     // do one cold
     boolean status = doRandomQueries(opts.num, opts.min, opts.max, opts.size, r, batchReader);
-    
+
     System.gc();
     System.gc();
     System.gc();
-    
+
     if (opts.seed == null)
       r = new Random();
     else
       r = new Random(opts.seed);
-    
+
     // do one hot (connections already established, metadata table cached)
     status = status && doRandomQueries(opts.num, opts.min, opts.max, opts.size, r, batchReader);
-    
+
     batchReader.close();
     if (!status) {
       System.exit(1);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchWriter.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchWriter.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchWriter.java
index fd56a46..b4d92c8 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchWriter.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RandomBatchWriter.java
@@ -41,7 +41,7 @@ import com.beust.jcommander.Parameter;
 
 /**
  * Simple example for writing random data to Accumulo. See docs/examples/README.batch for instructions.
- * 
+ *
  * The rows of the entries will be randomly generated numbers between a specified min and max (prefixed by "row_"). The column families will be "foo" and column
  * qualifiers will be "1". The values will be random byte arrays of a specified size.
  */
@@ -49,7 +49,7 @@ public class RandomBatchWriter {
 
   /**
    * Creates a random byte array of specified size using the specified seed.
-   * 
+   *
    * @param rowid
    *          the seed to use for the random number generator
    * @param dataSize
@@ -72,7 +72,7 @@ public class RandomBatchWriter {
 
   /**
    * Creates a mutation on a specified row with column family "foo", column qualifier "1", specified visibility, and a random value of specified size.
-   * 
+   *
    * @param rowid
    *          the row of the mutation
    * @param dataSize
@@ -111,7 +111,7 @@ public class RandomBatchWriter {
   }
 
   public static long abs(long l) {
-    l = Math.abs(l);  // abs(Long.MIN_VALUE) == Long.MIN_VALUE... 
+    l = Math.abs(l); // abs(Long.MIN_VALUE) == Long.MIN_VALUE...
     if (l < 0)
       return 0;
     return l;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/ReadWriteExample.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/ReadWriteExample.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/ReadWriteExample.java
index ccc924b..b270a6f 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/ReadWriteExample.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/ReadWriteExample.java
@@ -64,7 +64,7 @@ public class ReadWriteExample {
     boolean readEntries = false;
     @Parameter(names = {"-d", "--delete"}, description = "delete entries after any creates")
     boolean deleteEntries = false;
-    @Parameter(names = {"--durability"}, description = "durability used for writes (none, log, flush or sync)", converter=DurabilityConverter.class)
+    @Parameter(names = {"--durability"}, description = "durability used for writes (none, log, flush or sync)", converter = DurabilityConverter.class)
     Durability durability = Durability.DEFAULT;
 
     public Opts() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RowOperations.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RowOperations.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RowOperations.java
index 5276109..ba7f9ad 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RowOperations.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/RowOperations.java
@@ -41,135 +41,135 @@ import org.apache.log4j.Logger;
  * A demonstration of reading entire rows and deleting entire rows.
  */
 public class RowOperations {
-  
+
   private static final Logger log = Logger.getLogger(RowOperations.class);
-  
+
   private static Connector connector;
   private static String table = "example";
   private static BatchWriter bw;
-  
+
   public static void main(String[] args) throws AccumuloException, AccumuloSecurityException, TableExistsException, TableNotFoundException,
       MutationsRejectedException {
-    
+
     ClientOpts opts = new ClientOpts();
     ScannerOpts scanOpts = new ScannerOpts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(RowOperations.class.getName(), args, scanOpts, bwOpts);
-    
+
     // First the setup work
     connector = opts.getConnector();
-    
+
     // lets create an example table
     connector.tableOperations().create(table);
-    
+
     // lets create 3 rows of information
     Text row1 = new Text("row1");
     Text row2 = new Text("row2");
     Text row3 = new Text("row3");
-    
+
     // Which means 3 different mutations
     Mutation mut1 = new Mutation(row1);
     Mutation mut2 = new Mutation(row2);
     Mutation mut3 = new Mutation(row3);
-    
+
     // And we'll put 4 columns in each row
     Text col1 = new Text("1");
     Text col2 = new Text("2");
     Text col3 = new Text("3");
     Text col4 = new Text("4");
-    
+
     // Now we'll add them to the mutations
     mut1.put(new Text("column"), col1, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut1.put(new Text("column"), col2, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut1.put(new Text("column"), col3, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut1.put(new Text("column"), col4, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
-    
+
     mut2.put(new Text("column"), col1, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut2.put(new Text("column"), col2, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut2.put(new Text("column"), col3, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut2.put(new Text("column"), col4, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
-    
+
     mut3.put(new Text("column"), col1, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut3.put(new Text("column"), col2, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut3.put(new Text("column"), col3, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
     mut3.put(new Text("column"), col4, System.currentTimeMillis(), new Value("This is the value for this key".getBytes()));
-    
+
     // Now we'll make a Batch Writer
     bw = connector.createBatchWriter(table, bwOpts.getBatchWriterConfig());
-    
+
     // And add the mutations
     bw.addMutation(mut1);
     bw.addMutation(mut2);
     bw.addMutation(mut3);
-    
+
     // Force a send
     bw.flush();
-    
+
     // Now lets look at the rows
     Scanner rowThree = getRow(scanOpts, new Text("row3"));
     Scanner rowTwo = getRow(scanOpts, new Text("row2"));
     Scanner rowOne = getRow(scanOpts, new Text("row1"));
-    
+
     // And print them
     log.info("This is everything");
     printRow(rowOne);
     printRow(rowTwo);
     printRow(rowThree);
     System.out.flush();
-    
+
     // Now lets delete rowTwo with the iterator
     rowTwo = getRow(scanOpts, new Text("row2"));
     deleteRow(rowTwo);
-    
+
     // Now lets look at the rows again
     rowThree = getRow(scanOpts, new Text("row3"));
     rowTwo = getRow(scanOpts, new Text("row2"));
     rowOne = getRow(scanOpts, new Text("row1"));
-    
+
     // And print them
     log.info("This is row1 and row3");
     printRow(rowOne);
     printRow(rowTwo);
     printRow(rowThree);
     System.out.flush();
-    
+
     // Should only see the two rows
     // Now lets delete rowOne without passing in the iterator
-    
+
     deleteRow(scanOpts, row1);
-    
+
     // Now lets look at the rows one last time
     rowThree = getRow(scanOpts, new Text("row3"));
     rowTwo = getRow(scanOpts, new Text("row2"));
     rowOne = getRow(scanOpts, new Text("row1"));
-    
+
     // And print them
     log.info("This is just row3");
     printRow(rowOne);
     printRow(rowTwo);
     printRow(rowThree);
     System.out.flush();
-    
+
     // Should only see rowThree
-    
+
     // Always close your batchwriter
-    
+
     bw.close();
-    
+
     // and lets clean up our mess
     connector.tableOperations().delete(table);
-    
+
     // fin~
-    
+
   }
-  
+
   /**
    * Deletes a row given a text object
    */
   private static void deleteRow(ScannerOpts scanOpts, Text row) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     deleteRow(getRow(scanOpts, row));
   }
-  
+
   /**
    * Deletes a row, given a Scanner of JUST that row
    */
@@ -186,7 +186,7 @@ public class RowOperations {
     bw.addMutation(deleter);
     bw.flush();
   }
-  
+
   /**
    * Just a generic print function given an iterator. Not necessarily just for printing a single row
    */
@@ -195,7 +195,7 @@ public class RowOperations {
     for (Entry<Key,Value> entry : scanner)
       log.info("Key: " + entry.getKey().toString() + " Value: " + entry.getValue().toString());
   }
-  
+
   /**
    * Gets a scanner over one row
    */
@@ -208,5 +208,5 @@ public class RowOperations {
     scanner.setRange(new Range(row));
     return scanner;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/SequentialBatchWriter.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/SequentialBatchWriter.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/SequentialBatchWriter.java
index 3ae21e9..f2bd4d7 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/SequentialBatchWriter.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/SequentialBatchWriter.java
@@ -33,18 +33,18 @@ import com.beust.jcommander.Parameter;
  * Simple example for writing random data in sequential order to Accumulo. See docs/examples/README.batch for instructions.
  */
 public class SequentialBatchWriter {
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--start")
+    @Parameter(names = "--start")
     long start = 0;
-    @Parameter(names="--num", required=true)
+    @Parameter(names = "--num", required = true)
     long num = 0;
-    @Parameter(names="--size", required=true, description="size of the value to write")
+    @Parameter(names = "--size", required = true, description = "size of the value to write")
     int valueSize = 0;
-    @Parameter(names="--vis", converter=VisibilityConverter.class)
+    @Parameter(names = "--vis", converter = VisibilityConverter.class)
     ColumnVisibility vis = new ColumnVisibility();
   }
-  
+
   /**
    * Writes a specified number of entries to Accumulo using a {@link BatchWriter}. The rows of the entries will be sequential starting at a specified number.
    * The column families will be "foo" and column qualifiers will be "1". The values will be random byte arrays of a specified size.
@@ -55,14 +55,14 @@ public class SequentialBatchWriter {
     opts.parseArgs(SequentialBatchWriter.class.getName(), args, bwOpts);
     Connector connector = opts.getConnector();
     BatchWriter bw = connector.createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig());
-    
+
     long end = opts.start + opts.num;
-    
+
     for (long i = opts.start; i < end; i++) {
       Mutation m = RandomBatchWriter.createMutation(i, opts.valueSize, opts.vis);
       bw.addMutation(m);
     }
-    
+
     bw.close();
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TraceDumpExample.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TraceDumpExample.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TraceDumpExample.java
index 8fbf5e9..46e6a67 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TraceDumpExample.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TraceDumpExample.java
@@ -34,39 +34,39 @@ import com.beust.jcommander.Parameter;
  *
  */
 public class TraceDumpExample {
-	
-	static class Opts extends ClientOnDefaultTable {
-		public Opts() {
-			super("trace");
-		}
 
-		@Parameter(names = {"--traceid"}, description = "The hex string id of a given trace, for example 16cfbbd7beec4ae3")
-		public String traceId = "";
-	}
-	
-	public void dump(Opts opts) throws TableNotFoundException, AccumuloException, AccumuloSecurityException {
-	
-		if (opts.traceId.isEmpty()) {
-			throw new IllegalArgumentException("--traceid option is required");
-		}
-		
-		Scanner scanner = opts.getConnector().createScanner(opts.getTableName(), opts.auths);
-		scanner.setRange(new Range(new Text(opts.traceId)));
-		TraceDump.printTrace(scanner, new Printer() {
-			@Override
+  static class Opts extends ClientOnDefaultTable {
+    public Opts() {
+      super("trace");
+    }
+
+    @Parameter(names = {"--traceid"}, description = "The hex string id of a given trace, for example 16cfbbd7beec4ae3")
+    public String traceId = "";
+  }
+
+  public void dump(Opts opts) throws TableNotFoundException, AccumuloException, AccumuloSecurityException {
+
+    if (opts.traceId.isEmpty()) {
+      throw new IllegalArgumentException("--traceid option is required");
+    }
+
+    Scanner scanner = opts.getConnector().createScanner(opts.getTableName(), opts.auths);
+    scanner.setRange(new Range(new Text(opts.traceId)));
+    TraceDump.printTrace(scanner, new Printer() {
+      @Override
       public void print(String line) {
-				System.out.println(line);
-			}
-		});
-	}
-	
-	public static void main(String[] args) throws TableNotFoundException, AccumuloException, AccumuloSecurityException {
-		TraceDumpExample traceDumpExample = new TraceDumpExample();
-		Opts opts = new Opts();
-		ScannerOpts scannerOpts = new ScannerOpts();
-		opts.parseArgs(TraceDumpExample.class.getName(), args, scannerOpts);
+        System.out.println(line);
+      }
+    });
+  }
+
+  public static void main(String[] args) throws TableNotFoundException, AccumuloException, AccumuloSecurityException {
+    TraceDumpExample traceDumpExample = new TraceDumpExample();
+    Opts opts = new Opts();
+    ScannerOpts scannerOpts = new ScannerOpts();
+    opts.parseArgs(TraceDumpExample.class.getName(), args, scannerOpts);
 
-		traceDumpExample.dump(opts);
-	}
+    traceDumpExample.dump(opts);
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TracingExample.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TracingExample.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TracingExample.java
index 881819b..2c4a8a9 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TracingExample.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/client/TracingExample.java
@@ -18,6 +18,7 @@
 package org.apache.accumulo.examples.simple.client;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
+
 import java.util.Map.Entry;
 
 import org.apache.accumulo.core.cli.ClientOnDefaultTable;
@@ -43,7 +44,7 @@ import com.beust.jcommander.Parameter;
 
 /**
  * A simple example showing how to use the distributed tracing API in client code
- * 
+ *
  */
 public class TracingExample {
   private static final Logger log = Logger.getLogger(TracingExample.class);
@@ -124,8 +125,7 @@ public class TracingExample {
       ++numberOfEntriesRead;
     }
     // You can add additional metadata (key, values) to Spans which will be able to be viewed in the Monitor
-    readScope.getSpan().addKVAnnotation("Number of Entries Read".getBytes(UTF_8),
-        String.valueOf(numberOfEntriesRead).getBytes(UTF_8));
+    readScope.getSpan().addKVAnnotation("Number of Entries Read".getBytes(UTF_8), String.valueOf(numberOfEntriesRead).getBytes(UTF_8));
 
     readScope.close();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/combiner/StatsCombiner.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/combiner/StatsCombiner.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/combiner/StatsCombiner.java
index 41a531b..7dad89c 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/combiner/StatsCombiner.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/combiner/StatsCombiner.java
@@ -33,22 +33,22 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
  * and count. See {@link Combiner} for more information on which values are combined together. See docs/examples/README.combiner for instructions.
  */
 public class StatsCombiner extends Combiner {
-  
+
   public static final String RADIX_OPTION = "radix";
-  
+
   private int radix = 10;
-  
+
   @Override
   public Value reduce(Key key, Iterator<Value> iter) {
-    
+
     long min = Long.MAX_VALUE;
     long max = Long.MIN_VALUE;
     long sum = 0;
     long count = 0;
-    
+
     while (iter.hasNext()) {
       String stats[] = iter.next().toString().split(",");
-      
+
       if (stats.length == 1) {
         long val = Long.parseLong(stats[0], radix);
         min = Math.min(val, min);
@@ -62,21 +62,21 @@ public class StatsCombiner extends Combiner {
         count += Long.parseLong(stats[3], radix);
       }
     }
-    
+
     String ret = Long.toString(min, radix) + "," + Long.toString(max, radix) + "," + Long.toString(sum, radix) + "," + Long.toString(count, radix);
     return new Value(ret.getBytes());
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
-    
+
     if (options.containsKey(RADIX_OPTION))
       radix = Integer.parseInt(options.get(RADIX_OPTION));
     else
       radix = 10;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -85,21 +85,21 @@ public class StatsCombiner extends Combiner {
     io.addNamedOption(RADIX_OPTION, "radix/base of the numbers");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (!super.validateOptions(options))
       return false;
-    
+
     if (options.containsKey(RADIX_OPTION) && !options.get(RADIX_OPTION).matches("\\d+"))
       throw new IllegalArgumentException("invalid option " + RADIX_OPTION + ":" + options.get(RADIX_OPTION));
-    
+
     return true;
   }
-  
+
   /**
    * A convenience method for setting the expected base/radix of the numbers
-   * 
+   *
    * @param iterConfig
    *          Iterator settings to configure
    * @param base

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/AlphaNumKeyConstraint.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/AlphaNumKeyConstraint.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/AlphaNumKeyConstraint.java
index b23a258..8099b7e 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/AlphaNumKeyConstraint.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/AlphaNumKeyConstraint.java
@@ -26,27 +26,27 @@ import org.apache.accumulo.core.data.Mutation;
 
 /**
  * This class is an accumulo constraint that ensures all fields of a key are alpha numeric.
- * 
+ *
  * See docs/examples/README.constraint for instructions.
- * 
+ *
  */
 
 public class AlphaNumKeyConstraint implements Constraint {
-  
+
   private static final short NON_ALPHA_NUM_ROW = 1;
   private static final short NON_ALPHA_NUM_COLF = 2;
   private static final short NON_ALPHA_NUM_COLQ = 3;
-  
+
   private boolean isAlphaNum(byte bytes[]) {
     for (byte b : bytes) {
       boolean ok = ((b >= 'a' && b <= 'z') || (b >= 'A' && b <= 'Z') || (b >= '0' && b <= '9'));
       if (!ok)
         return false;
     }
-    
+
     return true;
   }
-  
+
   private List<Short> addViolation(List<Short> violations, short violation) {
     if (violations == null) {
       violations = new ArrayList<Short>();
@@ -56,29 +56,29 @@ public class AlphaNumKeyConstraint implements Constraint {
     }
     return violations;
   }
-  
+
   @Override
   public List<Short> check(Environment env, Mutation mutation) {
     List<Short> violations = null;
-    
+
     if (!isAlphaNum(mutation.getRow()))
       violations = addViolation(violations, NON_ALPHA_NUM_ROW);
-    
+
     Collection<ColumnUpdate> updates = mutation.getUpdates();
     for (ColumnUpdate columnUpdate : updates) {
       if (!isAlphaNum(columnUpdate.getColumnFamily()))
         violations = addViolation(violations, NON_ALPHA_NUM_COLF);
-      
+
       if (!isAlphaNum(columnUpdate.getColumnQualifier()))
         violations = addViolation(violations, NON_ALPHA_NUM_COLQ);
     }
-    
+
     return violations;
   }
-  
+
   @Override
   public String getViolationDescription(short violationCode) {
-    
+
     switch (violationCode) {
       case NON_ALPHA_NUM_ROW:
         return "Row was not alpha numeric";
@@ -87,8 +87,8 @@ public class AlphaNumKeyConstraint implements Constraint {
       case NON_ALPHA_NUM_COLQ:
         return "Column qualifier was not alpha numeric";
     }
-    
+
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/MaxMutationSize.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/MaxMutationSize.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/MaxMutationSize.java
index 54a863e..1192549 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/MaxMutationSize.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/MaxMutationSize.java
@@ -24,18 +24,18 @@ import org.apache.accumulo.core.data.Mutation;
 
 /**
  * Ensure that mutations are a reasonable size: we must be able to fit several in memory at a time.
- * 
+ *
  */
 public class MaxMutationSize implements Constraint {
   static final long MAX_SIZE = Runtime.getRuntime().maxMemory() >> 8;
   static final List<Short> empty = Collections.emptyList();
   static final List<Short> violations = Collections.singletonList(new Short((short) 0));
-  
+
   @Override
   public String getViolationDescription(short violationCode) {
     return String.format("mutation exceeded maximum size of %d", MAX_SIZE);
   }
-  
+
   @Override
   public List<Short> check(Environment env, Mutation mutation) {
     if (mutation.estimatedMemoryUsed() < MAX_SIZE)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/NumericValueConstraint.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/NumericValueConstraint.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/NumericValueConstraint.java
index b7e1527..f1e6d5a 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/NumericValueConstraint.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/constraints/NumericValueConstraint.java
@@ -28,19 +28,19 @@ import org.apache.accumulo.core.data.Mutation;
  * This class is an accumulo constraint that ensures values are numeric strings. See docs/examples/README.constraint for instructions.
  */
 public class NumericValueConstraint implements Constraint {
-  
+
   private static final short NON_NUMERIC_VALUE = 1;
-  
+
   private boolean isNumeric(byte bytes[]) {
     for (byte b : bytes) {
       boolean ok = (b >= '0' && b <= '9');
       if (!ok)
         return false;
     }
-    
+
     return true;
   }
-  
+
   private List<Short> addViolation(List<Short> violations, short violation) {
     if (violations == null) {
       violations = new ArrayList<Short>();
@@ -50,30 +50,30 @@ public class NumericValueConstraint implements Constraint {
     }
     return violations;
   }
-  
+
   @Override
   public List<Short> check(Environment env, Mutation mutation) {
     List<Short> violations = null;
-    
+
     Collection<ColumnUpdate> updates = mutation.getUpdates();
-    
+
     for (ColumnUpdate columnUpdate : updates) {
       if (!isNumeric(columnUpdate.getValue()))
         violations = addViolation(violations, NON_NUMERIC_VALUE);
     }
-    
+
     return violations;
   }
-  
+
   @Override
   public String getViolationDescription(short violationCode) {
-    
+
     switch (violationCode) {
       case NON_NUMERIC_VALUE:
         return "Value is not numeric";
     }
-    
+
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/FileCount.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/FileCount.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/FileCount.java
index cb6d350..dabb4c1 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/FileCount.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/FileCount.java
@@ -38,21 +38,20 @@ import com.beust.jcommander.Parameter;
  * Computes recursive counts over file system information and stores them back into the same Accumulo table. See docs/examples/README.dirlist for instructions.
  */
 public class FileCount {
-  
+
   private int entriesScanned;
   private int inserts;
-  
+
   private Opts opts;
   private ScannerOpts scanOpts;
   private BatchWriterOpts bwOpts;
-  
-  
+
   private static class CountValue {
     int dirCount = 0;
     int fileCount = 0;
     int recursiveDirCount = 0;
     int recusiveFileCount = 0;
-    
+
     void set(Value val) {
       String sa[] = val.toString().split(",");
       dirCount = Integer.parseInt(sa[0]);
@@ -60,46 +59,46 @@ public class FileCount {
       recursiveDirCount = Integer.parseInt(sa[2]);
       recusiveFileCount = Integer.parseInt(sa[3]);
     }
-    
+
     Value toValue() {
       return new Value((dirCount + "," + fileCount + "," + recursiveDirCount + "," + recusiveFileCount).getBytes());
     }
-    
+
     void incrementFiles() {
       fileCount++;
       recusiveFileCount++;
     }
-    
+
     void incrementDirs() {
       dirCount++;
       recursiveDirCount++;
     }
-    
+
     public void clear() {
       dirCount = 0;
       fileCount = 0;
       recursiveDirCount = 0;
       recusiveFileCount = 0;
     }
-    
+
     public void incrementRecursive(CountValue other) {
       recursiveDirCount += other.recursiveDirCount;
       recusiveFileCount += other.recusiveFileCount;
     }
   }
-  
+
   private int findMaxDepth(Scanner scanner, int min, int max) {
     int mid = min + (max - min) / 2;
     return findMaxDepth(scanner, min, mid, max);
   }
-  
+
   private int findMaxDepth(Scanner scanner, int min, int mid, int max) {
     // check to see if the mid point exist
     if (max < min)
       return -1;
-    
+
     scanner.setRange(new Range(String.format("%03d", mid), true, String.format("%03d", mid + 1), false));
-    
+
     if (scanner.iterator().hasNext()) {
       // this depth exist, check to see if a larger depth exist
       int ret = findMaxDepth(scanner, mid + 1, max);
@@ -111,9 +110,9 @@ public class FileCount {
       // this depth does not exist, look lower
       return findMaxDepth(scanner, min, mid - 1);
     }
-    
+
   }
-  
+
   private int findMaxDepth(Scanner scanner) {
     // do binary search to find max depth
     int origBatchSize = scanner.getBatchSize();
@@ -122,81 +121,81 @@ public class FileCount {
     scanner.setBatchSize(origBatchSize);
     return depth;
   }
-  
+
   // find the count column and consume a row
   private Entry<Key,Value> findCount(Entry<Key,Value> entry, Iterator<Entry<Key,Value>> iterator, CountValue cv) {
-    
+
     Key key = entry.getKey();
     Text currentRow = key.getRow();
-    
+
     if (key.compareColumnQualifier(QueryUtil.COUNTS_COLQ) == 0)
       cv.set(entry.getValue());
-    
+
     while (iterator.hasNext()) {
       entry = iterator.next();
       entriesScanned++;
       key = entry.getKey();
-      
+
       if (key.compareRow(currentRow) != 0)
         return entry;
-      
+
       if (key.compareColumnFamily(QueryUtil.DIR_COLF) == 0 && key.compareColumnQualifier(QueryUtil.COUNTS_COLQ) == 0) {
         cv.set(entry.getValue());
       }
-      
+
     }
-    
+
     return null;
   }
-  
+
   private Entry<Key,Value> consumeRow(Entry<Key,Value> entry, Iterator<Entry<Key,Value>> iterator) {
     Key key = entry.getKey();
     Text currentRow = key.getRow();
-    
+
     while (iterator.hasNext()) {
       entry = iterator.next();
       entriesScanned++;
       key = entry.getKey();
-      
+
       if (key.compareRow(currentRow) != 0)
         return entry;
     }
-    
+
     return null;
   }
-  
+
   private String extractDir(Key key) {
     String row = key.getRowData().toString();
     return row.substring(3, row.lastIndexOf('/'));
   }
-  
+
   private Mutation createMutation(int depth, String dir, CountValue countVal) {
     Mutation m = new Mutation(String.format("%03d%s", depth, dir));
     m.put(QueryUtil.DIR_COLF, QueryUtil.COUNTS_COLQ, opts.visibility, countVal.toValue());
     return m;
   }
-  
+
   private void calculateCounts(Scanner scanner, int depth, BatchWriter batchWriter) throws Exception {
-    
+
     scanner.setRange(new Range(String.format("%03d", depth), true, String.format("%03d", depth + 1), false));
-    
+
     CountValue countVal = new CountValue();
-    
+
     Iterator<Entry<Key,Value>> iterator = scanner.iterator();
-    
+
     String currentDir = null;
-    
+
     Entry<Key,Value> entry = null;
     if (iterator.hasNext()) {
       entry = iterator.next();
       entriesScanned++;
     }
-    
+
     while (entry != null) {
       Key key = entry.getKey();
-      
+
       String dir = extractDir(key);
-      
+
       if (currentDir == null) {
         currentDir = dir;
       } else if (!currentDir.equals(dir)) {
@@ -205,12 +204,12 @@ public class FileCount {
         currentDir = dir;
         countVal.clear();
       }
-      
+
       // process a whole row
       if (key.compareColumnFamily(QueryUtil.DIR_COLF) == 0) {
         CountValue tmpCount = new CountValue();
         entry = findCount(entry, iterator, tmpCount);
-        
+
         if (tmpCount.dirCount == 0 && tmpCount.fileCount == 0) {
           // in this case the higher depth will not insert anything if the
           // dir has no children, so insert something here
@@ -219,7 +218,7 @@ public class FileCount {
           batchWriter.addMutation(m);
           inserts++;
         }
-        
+
         countVal.incrementRecursive(tmpCount);
         countVal.incrementDirs();
       } else {
@@ -227,57 +226,57 @@ public class FileCount {
         countVal.incrementFiles();
       }
     }
-    
+
     if (currentDir != null) {
       batchWriter.addMutation(createMutation(depth - 1, currentDir, countVal));
       inserts++;
     }
   }
-  
+
   FileCount(Opts opts, ScannerOpts scanOpts, BatchWriterOpts bwOpts) throws Exception {
     this.opts = opts;
     this.scanOpts = scanOpts;
     this.bwOpts = bwOpts;
   }
-  
+
   public void run() throws Exception {
-    
+
     entriesScanned = 0;
     inserts = 0;
-    
+
     Connector conn = opts.getConnector();
     Scanner scanner = conn.createScanner(opts.getTableName(), opts.auths);
     scanner.setBatchSize(scanOpts.scanBatchSize);
     BatchWriter bw = conn.createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig());
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     int depth = findMaxDepth(scanner);
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     for (int d = depth; d > 0; d--) {
       calculateCounts(scanner, d, bw);
       // must flush so next depth can read what prev depth wrote
       bw.flush();
     }
-    
+
     bw.close();
-    
+
     long t3 = System.currentTimeMillis();
-    
+
     System.out.printf("Max depth              : %d%n", depth);
     System.out.printf("Time to find max depth : %,d ms%n", (t2 - t1));
     System.out.printf("Time to compute counts : %,d ms%n", (t3 - t2));
     System.out.printf("Entries scanned        : %,d %n", entriesScanned);
     System.out.printf("Counts inserted        : %,d %n", inserts);
   }
-  
+
   public static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--vis", description="use a given visibility for the new counts", converter=VisibilityConverter.class)
+    @Parameter(names = "--vis", description = "use a given visibility for the new counts", converter = VisibilityConverter.class)
     ColumnVisibility visibility = new ColumnVisibility();
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Ingest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Ingest.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Ingest.java
index 0ec10fb..17c9ee8 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Ingest.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Ingest.java
@@ -49,7 +49,7 @@ public class Ingest {
   public static final String LASTMOD_CQ = "lastmod";
   public static final String HASH_CQ = "md5";
   public static final Encoder<Long> encoder = LongCombiner.FIXED_LEN_ENCODER;
-  
+
   public static Mutation buildMutation(ColumnVisibility cv, String path, boolean isDir, boolean isHidden, boolean canExec, long length, long lastmod,
       String hash) {
     if (path.equals("/"))
@@ -68,7 +68,7 @@ public class Ingest {
       m.put(colf, new Text(HASH_CQ), cv, new Value(hash.getBytes()));
     return m;
   }
-  
+
   private static void ingest(File src, ColumnVisibility cv, BatchWriter dirBW, BatchWriter indexBW, FileDataIngest fdi, BatchWriter data) throws Exception {
     // build main table entry
     String path = null;
@@ -78,7 +78,7 @@ public class Ingest {
       path = src.getAbsolutePath();
     }
     System.out.println(path);
-    
+
     String hash = null;
     if (!src.isDirectory()) {
       try {
@@ -88,9 +88,9 @@ public class Ingest {
         return;
       }
     }
-    
+
     dirBW.addMutation(buildMutation(cv, path, src.isDirectory(), src.isHidden(), src.canExecute(), src.length(), src.lastModified(), hash));
-    
+
     // build index table entries
     Text row = QueryUtil.getForwardIndex(path);
     if (row != null) {
@@ -98,14 +98,14 @@ public class Ingest {
       Mutation m = new Mutation(row);
       m.put(QueryUtil.INDEX_COLF, p, cv, nullValue);
       indexBW.addMutation(m);
-      
+
       row = QueryUtil.getReverseIndex(path);
       m = new Mutation(row);
       m.put(QueryUtil.INDEX_COLF, p, cv, nullValue);
       indexBW.addMutation(m);
     }
   }
-  
+
   private static void recurse(File src, ColumnVisibility cv, BatchWriter dirBW, BatchWriter indexBW, FileDataIngest fdi, BatchWriter data) throws Exception {
     // ingest this File
     ingest(src, cv, dirBW, indexBW, fdi, data);
@@ -119,28 +119,27 @@ public class Ingest {
       }
     }
   }
-  
+
   static class Opts extends ClientOpts {
-    @Parameter(names="--dirTable", description="a table to hold the directory information")
+    @Parameter(names = "--dirTable", description = "a table to hold the directory information")
     String nameTable = "dirTable";
-    @Parameter(names="--indexTable", description="an index over the ingested data")
+    @Parameter(names = "--indexTable", description = "an index over the ingested data")
     String indexTable = "indexTable";
-    @Parameter(names="--dataTable", description="the file data, chunked into parts")
+    @Parameter(names = "--dataTable", description = "the file data, chunked into parts")
     String dataTable = "dataTable";
-    @Parameter(names="--vis", description="the visibility to mark the data", converter=VisibilityConverter.class)
+    @Parameter(names = "--vis", description = "the visibility to mark the data", converter = VisibilityConverter.class)
     ColumnVisibility visibility = new ColumnVisibility();
-    @Parameter(names="--chunkSize", description="the size of chunks when breaking down files")
+    @Parameter(names = "--chunkSize", description = "the size of chunks when breaking down files")
     int chunkSize = 100000;
-    @Parameter(description="<dir> { <dir> ... }")
+    @Parameter(description = "<dir> { <dir> ... }")
     List<String> directories = new ArrayList<String>();
   }
-  
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(Ingest.class.getName(), args, bwOpts);
-    
+
     Connector conn = opts.getConnector();
     if (!conn.tableOperations().exists(opts.nameTable))
       conn.tableOperations().create(opts.nameTable);
@@ -150,14 +149,14 @@ public class Ingest {
       conn.tableOperations().create(opts.dataTable);
       conn.tableOperations().attachIterator(opts.dataTable, new IteratorSetting(1, ChunkCombiner.class));
     }
-    
+
     BatchWriter dirBW = conn.createBatchWriter(opts.nameTable, bwOpts.getBatchWriterConfig());
     BatchWriter indexBW = conn.createBatchWriter(opts.indexTable, bwOpts.getBatchWriterConfig());
     BatchWriter dataBW = conn.createBatchWriter(opts.dataTable, bwOpts.getBatchWriterConfig());
     FileDataIngest fdi = new FileDataIngest(opts.chunkSize, opts.visibility);
     for (String dir : opts.directories) {
       recurse(new File(dir), opts.visibility, dirBW, indexBW, fdi, dataBW);
-      
+
       // fill in parent directory info
       int slashIndex = -1;
       while ((slashIndex = dir.lastIndexOf("/")) > 0) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/QueryUtil.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/QueryUtil.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/QueryUtil.java
index 09fb40c..a79b9d2 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/QueryUtil.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/QueryUtil.java
@@ -49,17 +49,16 @@ public class QueryUtil {
   public static final Text REVERSE_PREFIX = new Text("r");
   public static final Text INDEX_COLF = new Text("i");
   public static final Text COUNTS_COLQ = new Text("counts");
-  
-  public QueryUtil(Opts opts) throws AccumuloException,
-      AccumuloSecurityException {
+
+  public QueryUtil(Opts opts) throws AccumuloException, AccumuloSecurityException {
     conn = opts.getConnector();
     this.tableName = opts.getTableName();
     this.auths = opts.auths;
   }
-  
+
   /**
    * Calculates the depth of a path, i.e. the number of forward slashes in the path name.
-   * 
+   *
    * @param path
    *          the full path of a file or directory
    * @return the depth of the path
@@ -71,10 +70,10 @@ public class QueryUtil {
       numSlashes++;
     return numSlashes;
   }
-  
+
   /**
    * Given a path, construct an accumulo row prepended with the path's depth for the directory table.
-   * 
+   *
    * @param path
    *          the full path of a file or directory
    * @return the accumulo row associated with this path
@@ -84,10 +83,10 @@ public class QueryUtil {
     row.append(path.getBytes(), 0, path.length());
     return row;
   }
-  
+
   /**
    * Given a path, construct an accumulo row prepended with the {@link #FORWARD_PREFIX} for the index table.
-   * 
+   *
    * @param path
    *          the full path of a file or directory
    * @return the accumulo row associated with this path
@@ -100,10 +99,10 @@ public class QueryUtil {
     row.append(part.getBytes(), 0, part.length());
     return row;
   }
-  
+
   /**
    * Given a path, construct an accumulo row prepended with the {@link #REVERSE_PREFIX} with the path reversed for the index table.
-   * 
+   *
    * @param path
    *          the full path of a file or directory
    * @return the accumulo row associated with this path
@@ -120,10 +119,10 @@ public class QueryUtil {
     row.append(rev, 0, rev.length);
     return row;
   }
-  
+
   /**
    * Returns either the {@link #DIR_COLF} or a decoded string version of the colf.
-   * 
+   *
    * @param colf
    *          the column family
    */
@@ -132,10 +131,10 @@ public class QueryUtil {
       return colf.toString() + ":";
     return Long.toString(Ingest.encoder.decode(colf.getBytes())) + ":";
   }
-  
+
   /**
    * Scans over the directory table and pulls out stat information about a path.
-   * 
+   *
    * @param path
    *          the full path of a file or directory
    */
@@ -152,10 +151,10 @@ public class QueryUtil {
     }
     return data;
   }
-  
+
   /**
    * Uses the directory table to list the contents of a directory.
-   * 
+   *
    * @param path
    *          the full path of a directory
    */
@@ -177,10 +176,10 @@ public class QueryUtil {
     }
     return fim;
   }
-  
+
   /**
    * Scans over the index table for files or directories with a given name.
-   * 
+   *
    * @param term
    *          the name a file or directory to search for
    */
@@ -190,17 +189,17 @@ public class QueryUtil {
     scanner.setRange(new Range(getForwardIndex(term)));
     return scanner;
   }
-  
+
   /**
    * Scans over the index table for files or directories with a given name, prefix, or suffix (indicated by a wildcard '*' at the beginning or end of the term.
-   * 
+   *
    * @param exp
    *          the name a file or directory to search for with an optional wildcard '*' at the beginning or end
    */
   public Iterable<Entry<Key,Value>> singleRestrictedWildCardSearch(String exp) throws Exception {
     if (exp.indexOf("/") >= 0)
       throw new Exception("this method only works with unqualified names");
-    
+
     Scanner scanner = conn.createScanner(tableName, auths);
     if (exp.startsWith("*")) {
       System.out.println("executing beginning wildcard search for " + exp);
@@ -217,10 +216,10 @@ public class QueryUtil {
     }
     return scanner;
   }
-  
+
   /**
    * Scans over the index table for files or directories with a given name that can contain a single wildcard '*' anywhere in the term.
-   * 
+   *
    * @param exp
    *          the name a file or directory to search for with one optional wildcard '*'
    */
@@ -228,17 +227,17 @@ public class QueryUtil {
     int starIndex = exp.indexOf("*");
     if (exp.indexOf("*", starIndex + 1) >= 0)
       throw new Exception("only one wild card for search");
-    
+
     if (starIndex < 0) {
       return exactTermSearch(exp);
     } else if (starIndex == 0 || starIndex == exp.length() - 1) {
       return singleRestrictedWildCardSearch(exp);
     }
-    
+
     String firstPart = exp.substring(0, starIndex);
     String lastPart = exp.substring(starIndex + 1);
     String regexString = ".*/" + exp.replace("*", "[^/]*");
-    
+
     Scanner scanner = conn.createScanner(tableName, auths);
     if (firstPart.length() >= lastPart.length()) {
       System.out.println("executing middle wildcard search for " + regexString + " from entries starting with " + firstPart);
@@ -252,14 +251,14 @@ public class QueryUtil {
     scanner.addScanIterator(regex);
     return scanner;
   }
-  
+
   public static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--path", description="the directory to list")
+    @Parameter(names = "--path", description = "the directory to list")
     String path = "/";
-    @Parameter(names="--search", description="find a file or directory with the given name")
+    @Parameter(names = "--search", description = "find a file or directory with the given name")
     boolean search = false;
   }
-  
+
   /**
    * Lists the contents of a directory using the directory table, or searches for file or directory names (if the -search flag is included).
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Viewer.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Viewer.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Viewer.java
index 172ba58..70d989f 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Viewer.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/dirlist/Viewer.java
@@ -47,7 +47,7 @@ import com.beust.jcommander.Parameter;
 @SuppressWarnings("serial")
 public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansionListener {
   private static final Logger log = Logger.getLogger(Viewer.class);
-  
+
   JTree tree;
   DefaultTreeModel treeModel;
   QueryUtil q;
@@ -57,35 +57,35 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
   JTextArea text;
   JTextArea data;
   JScrollPane dataPane;
-  
+
   public static class NodeInfo {
     private String name;
     private Map<String,String> data;
-    
+
     public NodeInfo(String name, Map<String,String> data) {
       this.name = name;
       this.data = data;
     }
-    
+
     public String getName() {
       return name;
     }
-    
+
     public String getFullName() {
       String fn = data.get("fullname");
       if (fn == null)
         return name;
       return fn;
     }
-    
+
     public Map<String,String> getData() {
       return data;
     }
-    
+
     public String toString() {
       return getName();
     }
-    
+
     public String getHash() {
       for (String k : data.keySet()) {
         String[] parts = k.split(":");
@@ -96,9 +96,8 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
       return null;
     }
   }
-  
-  public Viewer(Opts opts)
-      throws Exception {
+
+  public Viewer(Opts opts) throws Exception {
     super("File Viewer");
     setSize(1000, 800);
     setDefaultCloseOperation(EXIT_ON_CLOSE);
@@ -106,7 +105,7 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
     fdq = new FileDataQuery(opts.instance, opts.zookeepers, opts.principal, opts.getToken(), opts.dataTable, opts.auths);
     this.topPath = opts.path;
   }
-  
+
   public void populate(DefaultMutableTreeNode node) throws TableNotFoundException {
     String path = ((NodeInfo) node.getUserObject()).getFullName();
     log.debug("listing " + path);
@@ -115,7 +114,7 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
       node.add(new DefaultMutableTreeNode(new NodeInfo(e.getKey(), e.getValue())));
     }
   }
-  
+
   public void populateChildren(DefaultMutableTreeNode node) throws TableNotFoundException {
     @SuppressWarnings("unchecked")
     Enumeration<DefaultMutableTreeNode> children = node.children();
@@ -123,12 +122,12 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
       populate(children.nextElement());
     }
   }
-  
+
   public void init() throws TableNotFoundException {
     DefaultMutableTreeNode root = new DefaultMutableTreeNode(new NodeInfo(topPath, q.getData(topPath)));
     populate(root);
     populateChildren(root);
-    
+
     treeModel = new DefaultTreeModel(root);
     tree = new JTree(treeModel);
     tree.addTreeExpansionListener(this);
@@ -144,11 +143,11 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
     infoSplitPane.setDividerLocation(150);
     getContentPane().add(mainSplitPane, BorderLayout.CENTER);
   }
-  
+
   public static String getText(DefaultMutableTreeNode node) {
     return getText(((NodeInfo) node.getUserObject()).getData());
   }
-  
+
   public static String getText(Map<String,String> data) {
     StringBuilder sb = new StringBuilder();
     for (String name : data.keySet()) {
@@ -159,7 +158,7 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
     }
     return sb.toString();
   }
-  
+
   @Override
   public void treeExpanded(TreeExpansionEvent event) {
     try {
@@ -168,7 +167,7 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
       log.error("Could not find table.", e);
     }
   }
-  
+
   @Override
   public void treeCollapsed(TreeExpansionEvent event) {
     DefaultMutableTreeNode node = (DefaultMutableTreeNode) event.getPath().getLastPathComponent();
@@ -180,7 +179,7 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
       child.removeAllChildren();
     }
   }
-  
+
   @Override
   public void valueChanged(TreeSelectionEvent e) {
     TreePath selected = e.getNewLeadSelectionPath();
@@ -199,16 +198,16 @@ public class Viewer extends JFrame implements TreeSelectionListener, TreeExpansi
       log.error("Could not get data from FileDataQuery.", e1);
     }
   }
-  
+
   static class Opts extends QueryUtil.Opts {
-    @Parameter(names="--dataTable")
+    @Parameter(names = "--dataTable")
     String dataTable = "dataTable";
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(Viewer.class.getName(), args);
-    
+
     Viewer v = new Viewer(opts);
     v.init();
     v.setVisible(true);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkCombiner.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkCombiner.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkCombiner.java
index f997f1a..ca77b39 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkCombiner.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkCombiner.java
@@ -32,7 +32,7 @@ import org.apache.hadoop.io.Text;
 
 /**
  * This iterator dedupes chunks and sets their visibilities to the combined visibility of the refs columns. For example, it would combine
- * 
+ *
  * <pre>
  *    row1 refs uid1\0a A&B V0
  *    row1 refs uid2\0b C&D V0
@@ -41,88 +41,88 @@ import org.apache.hadoop.io.Text;
  *    row1 ~chunk 0 E&F V1
  *    row1 ~chunk 0 G&H V1
  * </pre>
- * 
+ *
  * into the following
- * 
+ *
  * <pre>
  *    row1 refs uid1\0a A&B V0
  *    row1 refs uid2\0b C&D V0
  *    row1 ~chunk 0 (A&B)|(C&D) V1
  * </pre>
- * 
+ *
  * {@link VisibilityCombiner} is used to combie the visibilities.
  */
 
 public class ChunkCombiner implements SortedKeyValueIterator<Key,Value> {
-  
+
   private SortedKeyValueIterator<Key,Value> source;
   private SortedKeyValueIterator<Key,Value> refsSource;
   private static final Collection<ByteSequence> refsColf = Collections.singleton(FileDataIngest.REFS_CF_BS);
   private Map<Text,byte[]> lastRowVC = Collections.emptyMap();
-  
+
   private Key topKey = null;
   private Value topValue = null;
-  
+
   public ChunkCombiner() {}
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     this.source = source;
     this.refsSource = source.deepCopy(env);
   }
-  
+
   @Override
   public boolean hasTop() {
     return topKey != null;
   }
-  
+
   @Override
   public void next() throws IOException {
     findTop();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     source.seek(range, columnFamilies, inclusive);
     findTop();
   }
-  
+
   private void findTop() throws IOException {
     do {
       topKey = null;
       topValue = null;
     } while (source.hasTop() && _findTop() == null);
   }
-  
+
   private byte[] _findTop() throws IOException {
     long maxTS;
-    
+
     topKey = new Key(source.getTopKey());
     topValue = new Value(source.getTopValue());
     source.next();
-    
+
     if (!topKey.getColumnFamilyData().equals(FileDataIngest.CHUNK_CF_BS))
       return topKey.getColumnVisibility().getBytes();
-    
+
     maxTS = topKey.getTimestamp();
-    
+
     while (source.hasTop() && source.getTopKey().equals(topKey, PartialKey.ROW_COLFAM_COLQUAL)) {
       if (source.getTopKey().getTimestamp() > maxTS)
         maxTS = source.getTopKey().getTimestamp();
-      
+
       if (!topValue.equals(source.getTopValue()))
         throw new RuntimeException("values not equals " + topKey + " " + source.getTopKey() + " : " + diffInfo(topValue, source.getTopValue()));
-      
+
       source.next();
     }
-    
+
     byte[] vis = getVisFromRefs();
     if (vis != null) {
       topKey = new Key(topKey.getRowData().toArray(), topKey.getColumnFamilyData().toArray(), topKey.getColumnQualifierData().toArray(), vis, maxTS);
     }
     return vis;
   }
-  
+
   private byte[] getVisFromRefs() throws IOException {
     Text row = topKey.getRow();
     if (lastRowVC.containsKey(row))
@@ -143,34 +143,34 @@ public class ChunkCombiner implements SortedKeyValueIterator<Key,Value> {
     lastRowVC = Collections.singletonMap(row, vc.get());
     return vc.get();
   }
-  
+
   private String diffInfo(Value v1, Value v2) {
     if (v1.getSize() != v2.getSize()) {
       return "val len not equal " + v1.getSize() + "!=" + v2.getSize();
     }
-    
+
     byte[] vb1 = v1.get();
     byte[] vb2 = v2.get();
-    
+
     for (int i = 0; i < vb1.length; i++) {
       if (vb1[i] != vb2[i]) {
         return String.format("first diff at offset %,d 0x%02x != 0x%02x", i, 0xff & vb1[i], 0xff & vb2[i]);
       }
     }
-    
+
     return null;
   }
-  
+
   @Override
   public Key getTopKey() {
     return topKey;
   }
-  
+
   @Override
   public Value getTopValue() {
     return topValue;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     ChunkCombiner cc = new ChunkCombiner();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputFormat.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputFormat.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputFormat.java
index 732b03b..f5da4e5 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputFormat.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputFormat.java
@@ -41,7 +41,7 @@ public class ChunkInputFormat extends InputFormatBase<List<Entry<Key,Value>>,Inp
       InterruptedException {
     return new RecordReaderBase<List<Entry<Key,Value>>,InputStream>() {
       private PeekingIterator<Entry<Key,Value>> peekingScannerIterator;
-      
+
       @Override
       public void initialize(InputSplit inSplit, TaskAttemptContext attempt) throws IOException {
         super.initialize(inSplit, attempt);
@@ -49,7 +49,7 @@ public class ChunkInputFormat extends InputFormatBase<List<Entry<Key,Value>>,Inp
         currentK = new ArrayList<Entry<Key,Value>>();
         currentV = new ChunkInputStream();
       }
-      
+
       @Override
       public boolean nextKeyValue() throws IOException, InterruptedException {
         currentK.clear();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStream.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStream.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStream.java
index 30b8521..c902271 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStream.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStream.java
@@ -34,26 +34,26 @@ import org.apache.log4j.Logger;
  */
 public class ChunkInputStream extends InputStream {
   private static final Logger log = Logger.getLogger(ChunkInputStream.class);
-  
+
   protected PeekingIterator<Entry<Key,Value>> source;
   protected Key currentKey;
   protected Set<Text> currentVis;
   protected int currentChunk;
   protected int currentChunkSize;
   protected boolean gotEndMarker;
-  
+
   protected byte buf[];
   protected int count;
   protected int pos;
-  
+
   public ChunkInputStream() {
     source = null;
   }
-  
+
   public ChunkInputStream(PeekingIterator<Entry<Key,Value>> in) throws IOException {
     setSource(in);
   }
-  
+
   public void setSource(PeekingIterator<Entry<Key,Value>> in) throws IOException {
     if (source != null)
       throw new IOException("setting new source without closing old one");
@@ -65,7 +65,7 @@ public class ChunkInputStream extends InputStream {
       gotEndMarker = true;
       return;
     }
-    
+
     // read forward until we reach a chunk
     Entry<Key,Value> entry = source.next();
     currentKey = entry.getKey();
@@ -91,7 +91,7 @@ public class ChunkInputStream extends InputStream {
       throw new IOException("starting chunk number isn't 0 for " + currentKey.getRow());
     }
   }
-  
+
   private int fill() throws IOException {
     if (source == null || !source.hasNext()) {
       if (gotEndMarker)
@@ -99,11 +99,11 @@ public class ChunkInputStream extends InputStream {
       else
         throw new IOException("no end chunk marker but source has no data");
     }
-    
+
     Entry<Key,Value> entry = source.peek();
     Key thisKey = entry.getKey();
     log.debug("evaluating key: " + thisKey.toString());
-    
+
     // check that we're still on the same row
     if (!thisKey.equals(currentKey, PartialKey.ROW)) {
       if (gotEndMarker)
@@ -115,39 +115,39 @@ public class ChunkInputStream extends InputStream {
       }
     }
     log.debug("matches current key");
-    
+
     // ok to advance the iterator
     source.next();
-    
+
     // check that this is part of a chunk
     if (!thisKey.getColumnFamily().equals(FileDataIngest.CHUNK_CF)) {
       log.debug("skipping non-chunk key");
       return fill();
     }
     log.debug("is a chunk");
-    
+
     // check that the chunk size is the same as the one being read
     if (currentChunkSize != FileDataIngest.bytesToInt(thisKey.getColumnQualifier().getBytes(), 0)) {
       log.debug("skipping chunk of different size");
       return fill();
     }
-    
+
     // add the visibility to the list if it's not there
     if (!currentVis.contains(thisKey.getColumnVisibility()))
       currentVis.add(thisKey.getColumnVisibility());
-    
+
     // check to see if it is an identical chunk with a different visibility
     if (thisKey.getColumnQualifier().equals(currentKey.getColumnQualifier())) {
       log.debug("skipping identical chunk with different visibility");
       return fill();
     }
-    
+
     if (gotEndMarker) {
       log.debug("got another chunk after end marker: " + currentKey.toString() + " " + thisKey.toString());
       clear();
       throw new IOException("found extra chunk after end marker");
     }
-    
+
     // got new chunk of the same file, check that it's the next chunk
     int thisChunk = FileDataIngest.bytesToInt(thisKey.getColumnQualifier().getBytes(), 4);
     if (thisChunk != currentChunk + 1) {
@@ -155,27 +155,27 @@ public class ChunkInputStream extends InputStream {
       clear();
       throw new IOException("missing chunks between " + currentChunk + " and " + thisChunk);
     }
-    
+
     currentKey = thisKey;
     currentChunk = thisChunk;
     buf = entry.getValue().get();
     pos = 0;
-    
+
     // check to see if it's the last chunk
     if (buf.length == 0) {
       gotEndMarker = true;
       return fill();
     }
-    
+
     return count = buf.length;
   }
-  
+
   public Set<Text> getVisibilities() {
     if (source != null)
       throw new IllegalStateException("don't get visibilities before chunks have been completely read");
     return currentVis;
   }
-  
+
   public int read() throws IOException {
     if (source == null)
       return -1;
@@ -191,7 +191,7 @@ public class ChunkInputStream extends InputStream {
     }
     return buf[pos++] & 0xff;
   }
-  
+
   @Override
   public int read(byte[] b, int off, int len) throws IOException {
     if (b == null) {
@@ -201,7 +201,7 @@ public class ChunkInputStream extends InputStream {
     } else if (len == 0) {
       return 0;
     }
-    
+
     log.debug("filling buffer " + off + " " + len);
     int total = 0;
     while (total < len) {
@@ -218,7 +218,7 @@ public class ChunkInputStream extends InputStream {
         }
         avail = count - pos;
       }
-      
+
       int cnt = (avail < len - total) ? avail : len - total;
       log.debug("copying from local buffer: local pos " + pos + " into pos " + off + " len " + cnt);
       System.arraycopy(buf, pos, b, off, cnt);
@@ -229,7 +229,7 @@ public class ChunkInputStream extends InputStream {
     log.debug("filled " + total + " bytes");
     return total;
   }
-  
+
   public void clear() {
     source = null;
     buf = null;
@@ -237,7 +237,7 @@ public class ChunkInputStream extends InputStream {
     currentChunk = 0;
     pos = count = 0;
   }
-  
+
   @Override
   public void close() throws IOException {
     try {


[50/66] [abbrv] accumulo git commit: Merge branch '1.5' into 1.6 with -sours

Posted by ct...@apache.org.
Merge branch '1.5' into 1.6 with -sours


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/dff686b4
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/dff686b4
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/dff686b4

Branch: refs/heads/1.6
Commit: dff686b454e67e74150d967bdc3eb94dcb5e39a6
Parents: 100ffd1 c2155f4
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:21:16 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:21:16 2015 -0500

----------------------------------------------------------------------

----------------------------------------------------------------------



[59/66] [abbrv] accumulo git commit: Merge branch '1.5' into 1.6

Posted by ct...@apache.org.
Merge branch '1.5' into 1.6

Conflicts:
	pom.xml


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/627e525b
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/627e525b
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/627e525b

Branch: refs/heads/1.6
Commit: 627e525b6b881873a1b31d63ca22df26f3b7776f
Parents: dff686b d2c116f
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:35:48 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:35:48 2015 -0500

----------------------------------------------------------------------
 pom.xml | 108 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 108 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/627e525b/pom.xml
----------------------------------------------------------------------
diff --cc pom.xml
index 22d5450,711a21d..6a90918
--- a/pom.xml
+++ b/pom.xml
@@@ -496,21 -367,14 +496,26 @@@
      <pluginManagement>
        <plugins>
          <plugin>
 +          <groupId>org.codehaus.mojo</groupId>
 +          <artifactId>findbugs-maven-plugin</artifactId>
 +          <version>2.5.5</version>
 +          <configuration>
 +            <xmlOutput>true</xmlOutput>
 +            <effort>Max</effort>
 +            <failOnError>true</failOnError>
 +            <includeTests>true</includeTests>
 +            <maxRank>6</maxRank>
 +          </configuration>
 +        </plugin>
 +        <plugin>
+           <groupId>org.apache.maven.plugins</groupId>
+           <artifactId>maven-checkstyle-plugin</artifactId>
+           <version>2.13</version>
+         </plugin>
+         <plugin>
            <groupId>com.google.code.sortpom</groupId>
            <artifactId>maven-sortpom-plugin</artifactId>
 -          <version>2.1.0</version>
 +          <version>2.3.0</version>
            <configuration>
              <predefinedSortOrder>recommended_2008_06</predefinedSortOrder>
              <lineSeparator>\n</lineSeparator>
@@@ -676,20 -571,19 +681,33 @@@
                  <pluginExecution>
                    <pluginExecutionFilter>
                      <groupId>org.apache.maven.plugins</groupId>
 +                    <artifactId>maven-plugin-plugin</artifactId>
 +                    <versionRange>[3.2,)</versionRange>
 +                    <goals>
 +                      <goal>helpmojo</goal>
 +                      <goal>descriptor</goal>
 +                    </goals>
 +                  </pluginExecutionFilter>
 +                  <action>
 +                    <ignore />
 +                  </action>
 +                </pluginExecution>
 +                <pluginExecution>
 +                  <pluginExecutionFilter>
 +                    <groupId>org.apache.maven.plugins</groupId>
+                     <artifactId>maven-checkstyle-plugin</artifactId>
+                     <versionRange>[2.13,)</versionRange>
+                     <goals>
+                       <goal>check</goal>
+                     </goals>
+                   </pluginExecutionFilter>
+                   <action>
+                     <ignore />
+                   </action>
+                 </pluginExecution>
+                 <pluginExecution>
+                   <pluginExecutionFilter>
+                     <groupId>org.apache.maven.plugins</groupId>
                      <artifactId>maven-dependency-plugin</artifactId>
                      <versionRange>[2.0,)</versionRange>
                      <goals>


[66/66] [abbrv] accumulo git commit: Merge branch '1.6'

Posted by ct...@apache.org.
Merge branch '1.6'

Conflicts:
	server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Compactor.java


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/392d9d6a
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/392d9d6a
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/392d9d6a

Branch: refs/heads/master
Commit: 392d9d6a48b2ddf3c8f9b85f91e8e983f0dfa620
Parents: 83a9623 9ca1ff0
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 21:41:48 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 21:41:48 2015 -0500

----------------------------------------------------------------------
 .../org/apache/accumulo/core/cli/ClientOpts.java     |  6 +++---
 .../core/client/admin/InstanceOperations.java        |  6 +++---
 .../apache/accumulo/core/file/rfile/CreateEmpty.java |  4 ++--
 .../accumulo/core/file/rfile/bcfile/Utils.java       |  1 +
 .../org/apache/accumulo/core/iterators/Combiner.java |  2 +-
 .../apache/accumulo/core/iterators/LongCombiner.java |  4 ++--
 .../accumulo/core/iterators/OptionDescriber.java     |  4 ++--
 .../accumulo/core/iterators/TypedValueCombiner.java  | 10 +++++-----
 .../core/iterators/user/SummingArrayCombiner.java    |  5 +++--
 .../core/security/crypto/CryptoModuleParameters.java |  2 +-
 .../NonCachingSecretKeyEncryptionStrategy.java       |  3 ++-
 .../examples/simple/mapreduce/TableToFile.java       |  4 ++--
 pom.xml                                              |  2 +-
 .../java/org/apache/accumulo/server/Accumulo.java    |  4 ++--
 .../org/apache/accumulo/server/init/Initialize.java  |  3 ++-
 .../accumulo/server/rpc/CustomNonBlockingServer.java |  4 ++--
 .../accumulo/tserver/tablet/CompactionInfo.java      |  5 +++--
 .../org/apache/accumulo/shell/ShellOptionsJC.java    |  6 +++---
 .../apache/accumulo/shell/commands/DUCommand.java    |  3 ++-
 .../apache/accumulo/shell/commands/EGrepCommand.java |  3 ++-
 .../shell/commands/QuotedStringTokenizer.java        | 15 ++++++---------
 .../main/java/org/apache/accumulo/start/Main.java    |  4 ++--
 .../start/classloader/AccumuloClassLoader.java       |  4 ++--
 .../test/randomwalk/security/CreateTable.java        |  5 ++---
 .../test/randomwalk/security/CreateUser.java         |  5 ++---
 25 files changed, 58 insertions(+), 56 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/core/src/main/java/org/apache/accumulo/core/iterators/LongCombiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/core/src/main/java/org/apache/accumulo/core/iterators/user/SummingArrayCombiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/pom.xml
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/server/base/src/main/java/org/apache/accumulo/server/rpc/CustomNonBlockingServer.java
----------------------------------------------------------------------
diff --cc server/base/src/main/java/org/apache/accumulo/server/rpc/CustomNonBlockingServer.java
index 21f55b3,0000000..577b5eb
mode 100644,000000..100644
--- a/server/base/src/main/java/org/apache/accumulo/server/rpc/CustomNonBlockingServer.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/rpc/CustomNonBlockingServer.java
@@@ -1,268 -1,0 +1,268 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.server.rpc;
 +
 +import java.io.IOException;
 +import java.net.Socket;
 +import java.nio.channels.SelectionKey;
 +import java.util.Iterator;
 +
 +import org.apache.log4j.Logger;
 +import org.apache.thrift.server.THsHaServer;
 +import org.apache.thrift.server.TNonblockingServer;
 +import org.apache.thrift.transport.TNonblockingServerTransport;
 +import org.apache.thrift.transport.TNonblockingSocket;
 +import org.apache.thrift.transport.TNonblockingTransport;
 +import org.apache.thrift.transport.TTransportException;
 +
 +/**
 + * This class implements a custom non-blocking thrift server, incorporating the {@link THsHaServer} features, and overriding the underlying
 + * {@link TNonblockingServer} methods, especially {@link org.apache.thrift.server.TNonblockingServer.SelectAcceptThread}, in order to override the
 + * {@link org.apache.thrift.server.AbstractNonblockingServer.FrameBuffer} and {@link org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer} with
 + * one that reveals the client address from its transport.
 + *
 + * <p>
 + * The justification for this is explained in https://issues.apache.org/jira/browse/ACCUMULO-1691, and is needed due to the repeated regressions:
 + * <ul>
 + * <li>https://issues.apache.org/jira/browse/THRIFT-958</li>
 + * <li>https://issues.apache.org/jira/browse/THRIFT-1464</li>
 + * <li>https://issues.apache.org/jira/browse/THRIFT-2173</li>
 + * </ul>
 + *
 + * <p>
 + * This class contains a copy of {@link org.apache.thrift.server.TNonblockingServer.SelectAcceptThread} from Thrift 0.9.1, with the slight modification of
 + * instantiating a custom FrameBuffer, rather than the {@link org.apache.thrift.server.AbstractNonblockingServer.FrameBuffer} and
 + * {@link org.apache.thrift.server.AbstractNonblockingServer.AsyncFrameBuffer}. Because of this, any change in the implementation upstream will require a review
 + * of this implementation here, to ensure any new bugfixes/features in the upstream Thrift class are also applied here, at least until
 + * https://issues.apache.org/jira/browse/THRIFT-2173 is implemented. In the meantime, the maven-enforcer-plugin ensures that Thrift remains at version 0.9.1,
 + * which has been reviewed and tested.
 + */
 +public class CustomNonBlockingServer extends THsHaServer {
 +
 +  private static final Logger LOGGER = Logger.getLogger(CustomNonBlockingServer.class);
 +  private SelectAcceptThread selectAcceptThread_;
 +  private volatile boolean stopped_ = false;
 +
 +  public CustomNonBlockingServer(Args args) {
 +    super(args);
 +  }
 +
 +  @Override
 +  protected Runnable getRunnable(final FrameBuffer frameBuffer) {
 +    return new Runnable() {
 +      @Override
 +      public void run() {
 +        if (frameBuffer instanceof CustomNonblockingFrameBuffer) {
 +          TNonblockingTransport trans = ((CustomNonblockingFrameBuffer) frameBuffer).getTransport();
 +          if (trans instanceof TNonblockingSocket) {
 +            TNonblockingSocket tsock = (TNonblockingSocket) trans;
 +            Socket sock = tsock.getSocketChannel().socket();
 +            TServerUtils.clientAddress.set(sock.getInetAddress().getHostAddress() + ":" + sock.getPort());
 +          }
 +        }
 +        frameBuffer.invoke();
 +      }
 +    };
 +  }
 +
 +  @Override
 +  protected boolean startThreads() {
 +    // start the selector
 +    try {
 +      selectAcceptThread_ = new SelectAcceptThread((TNonblockingServerTransport) serverTransport_);
 +      selectAcceptThread_.start();
 +      return true;
 +    } catch (IOException e) {
 +      LOGGER.error("Failed to start selector thread!", e);
 +      return false;
 +    }
 +  }
 +
 +  @Override
 +  public void stop() {
 +    stopped_ = true;
 +    if (selectAcceptThread_ != null) {
 +      selectAcceptThread_.wakeupSelector();
 +    }
 +  }
 +
 +  @Override
 +  public boolean isStopped() {
 +    return selectAcceptThread_.isStopped();
 +  }
 +
 +  @Override
 +  protected void joinSelector() {
 +    // wait until the selector thread exits
 +    try {
 +      selectAcceptThread_.join();
 +    } catch (InterruptedException e) {
 +      // for now, just silently ignore. technically this means we'll have less of
 +      // a graceful shutdown as a result.
 +    }
 +  }
 +
 +  private interface CustomNonblockingFrameBuffer {
 +    TNonblockingTransport getTransport();
 +  }
 +
 +  private class CustomAsyncFrameBuffer extends AsyncFrameBuffer implements CustomNonblockingFrameBuffer {
 +    private TNonblockingTransport trans;
 +
 +    public CustomAsyncFrameBuffer(TNonblockingTransport trans, SelectionKey selectionKey, AbstractSelectThread selectThread) {
 +      super(trans, selectionKey, selectThread);
 +      this.trans = trans;
 +    }
 +
 +    @Override
 +    public TNonblockingTransport getTransport() {
 +      return trans;
 +    }
 +  }
 +
 +  private class CustomFrameBuffer extends FrameBuffer implements CustomNonblockingFrameBuffer {
 +    private TNonblockingTransport trans;
 +
 +    public CustomFrameBuffer(TNonblockingTransport trans, SelectionKey selectionKey, AbstractSelectThread selectThread) {
 +      super(trans, selectionKey, selectThread);
 +      this.trans = trans;
 +    }
 +
 +    @Override
 +    public TNonblockingTransport getTransport() {
 +      return trans;
 +    }
 +  }
 +
 +  // @formatter:off
 +  private class SelectAcceptThread extends AbstractSelectThread {
 +
 +    // The server transport on which new client transports will be accepted
 +    private final TNonblockingServerTransport serverTransport;
 +
 +    /**
 +     * Set up the thread that will handle the non-blocking accepts, reads, and
 +     * writes.
 +     */
 +    public SelectAcceptThread(final TNonblockingServerTransport serverTransport)
 +    throws IOException {
 +      this.serverTransport = serverTransport;
 +      serverTransport.registerSelector(selector);
 +    }
 +
 +    public boolean isStopped() {
 +      return stopped_;
 +    }
 +
 +    /**
 +     * The work loop. Handles both selecting (all IO operations) and managing
 +     * the selection preferences of all existing connections.
 +     */
 +    @Override
 +    public void run() {
 +      try {
 +        if (eventHandler_ != null) {
 +          eventHandler_.preServe();
 +        }
 +
 +        while (!stopped_) {
 +          select();
 +          processInterestChanges();
 +        }
 +        for (SelectionKey selectionKey : selector.keys()) {
 +          cleanupSelectionKey(selectionKey);
 +        }
 +      } catch (Throwable t) {
 +        LOGGER.error("run() exiting due to uncaught error", t);
 +      } finally {
 +        stopped_ = true;
 +      }
 +    }
 +
 +    /**
 +     * Select and process IO events appropriately:
 +     * If there are connections to be accepted, accept them.
 +     * If there are existing connections with data waiting to be read, read it,
 +     * buffering until a whole frame has been read.
 +     * If there are any pending responses, buffer them until their target client
 +     * is available, and then send the data.
 +     */
 +    private void select() {
 +      try {
 +        // wait for io events.
 +        selector.select();
 +
 +        // process the io events we received
 +        Iterator<SelectionKey> selectedKeys = selector.selectedKeys().iterator();
 +        while (!stopped_ && selectedKeys.hasNext()) {
 +          SelectionKey key = selectedKeys.next();
 +          selectedKeys.remove();
 +
 +          // skip if not valid
 +          if (!key.isValid()) {
 +            cleanupSelectionKey(key);
 +            continue;
 +          }
 +
 +          // if the key is marked Accept, then it has to be the server
 +          // transport.
 +          if (key.isAcceptable()) {
 +            handleAccept();
 +          } else if (key.isReadable()) {
 +            // deal with reads
 +            handleRead(key);
 +          } else if (key.isWritable()) {
 +            // deal with writes
 +            handleWrite(key);
 +          } else {
 +            LOGGER.warn("Unexpected state in select! " + key.interestOps());
 +          }
 +        }
 +      } catch (IOException e) {
 +        LOGGER.warn("Got an IOException while selecting!", e);
 +      }
 +    }
 +
 +    /**
 +     * Accept a new connection.
 +     */
 +    @SuppressWarnings("unused")
 +    private void handleAccept() throws IOException {
 +      SelectionKey clientKey = null;
 +      TNonblockingTransport client = null;
 +      try {
 +        // accept the connection
 +        client = (TNonblockingTransport)serverTransport.accept();
 +        clientKey = client.registerSelector(selector, SelectionKey.OP_READ);
 +
 +        // add this key to the map
-           FrameBuffer frameBuffer = processorFactory_.isAsyncProcessor() ?
-                   new CustomAsyncFrameBuffer(client, clientKey,SelectAcceptThread.this) :
++          FrameBuffer frameBuffer =
++              processorFactory_.isAsyncProcessor() ? new CustomAsyncFrameBuffer(client, clientKey,SelectAcceptThread.this) :
 +                  new CustomFrameBuffer(client, clientKey,SelectAcceptThread.this);
 +
 +          clientKey.attach(frameBuffer);
 +      } catch (TTransportException tte) {
 +        // something went wrong accepting.
 +        LOGGER.warn("Exception trying to accept!", tte);
 +        tte.printStackTrace();
 +        if (clientKey != null) cleanupSelectionKey(clientKey);
 +        if (client != null) client.close();
 +      }
 +    }
 +  } // SelectAcceptThread
 +  // @formatter:on
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
----------------------------------------------------------------------
diff --cc server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
index 918edf6,0000000..2023d2c
mode 100644,000000..100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
@@@ -1,129 -1,0 +1,130 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.tserver.tablet;
 +
 +import java.util.ArrayList;
 +import java.util.HashMap;
 +import java.util.List;
 +import java.util.Map;
 +
 +import org.apache.accumulo.core.client.IteratorSetting;
 +import org.apache.accumulo.core.data.KeyExtent;
 +import org.apache.accumulo.core.data.thrift.IterInfo;
 +import org.apache.accumulo.core.tabletserver.thrift.ActiveCompaction;
 +import org.apache.accumulo.core.tabletserver.thrift.CompactionReason;
 +import org.apache.accumulo.core.tabletserver.thrift.CompactionType;
 +import org.apache.accumulo.server.fs.FileRef;
 +
 +public class CompactionInfo {
 +
 +  private final Compactor compactor;
 +  private final String localityGroup;
 +  private final long entriesRead;
 +  private final long entriesWritten;
 +
 +  CompactionInfo(Compactor compactor) {
 +    this.localityGroup = compactor.getCurrentLocalityGroup();
 +    this.entriesRead = compactor.getEntriesRead();
 +    this.entriesWritten = compactor.getEntriesWritten();
 +    this.compactor = compactor;
 +  }
 +
 +  public long getID() {
 +    return compactor.getCompactorID();
 +  }
 +
 +  public KeyExtent getExtent() {
 +    return compactor.getExtent();
 +  }
 +
 +  public long getEntriesRead() {
 +    return entriesRead;
 +  }
 +
 +  public long getEntriesWritten() {
 +    return entriesWritten;
 +  }
 +
 +  public Thread getThread() {
 +    return compactor.thread;
 +  }
 +
 +  public String getOutputFile() {
 +    return compactor.getOutputFile();
 +  }
 +
 +  public ActiveCompaction toThrift() {
 +
 +    CompactionType type;
 +
 +    if (compactor.hasIMM())
 +      if (compactor.getFilesToCompact().size() > 0)
 +        type = CompactionType.MERGE;
 +      else
 +        type = CompactionType.MINOR;
 +    else if (!compactor.willPropogateDeletes())
 +      type = CompactionType.FULL;
 +    else
 +      type = CompactionType.MAJOR;
 +
 +    CompactionReason reason;
 +
-     if (compactor.hasIMM())
++    if (compactor.hasIMM()) {
 +      switch (compactor.getMinCReason()) {
 +        case USER:
 +          reason = CompactionReason.USER;
 +          break;
 +        case CLOSE:
 +          reason = CompactionReason.CLOSE;
 +          break;
 +        case SYSTEM:
 +        default:
 +          reason = CompactionReason.SYSTEM;
 +          break;
 +      }
-     else
++    } else {
 +      switch (compactor.getMajorCompactionReason()) {
 +        case USER:
 +          reason = CompactionReason.USER;
 +          break;
 +        case CHOP:
 +          reason = CompactionReason.CHOP;
 +          break;
 +        case IDLE:
 +          reason = CompactionReason.IDLE;
 +          break;
 +        case NORMAL:
 +        default:
 +          reason = CompactionReason.SYSTEM;
 +          break;
 +      }
++    }
 +
 +    List<IterInfo> iiList = new ArrayList<IterInfo>();
 +    Map<String,Map<String,String>> iterOptions = new HashMap<String,Map<String,String>>();
 +
 +    for (IteratorSetting iterSetting : compactor.getIterators()) {
 +      iiList.add(new IterInfo(iterSetting.getPriority(), iterSetting.getIteratorClass(), iterSetting.getName()));
 +      iterOptions.put(iterSetting.getName(), iterSetting.getOptions());
 +    }
 +    List<String> filesToCompact = new ArrayList<String>();
 +    for (FileRef ref : compactor.getFilesToCompact())
 +      filesToCompact.add(ref.toString());
 +    return new ActiveCompaction(compactor.extent.toThrift(), System.currentTimeMillis() - compactor.getStartTime(), filesToCompact, compactor.getOutputFile(),
 +        type, reason, localityGroup, entriesRead, entriesWritten, iiList, iterOptions);
 +  }
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
----------------------------------------------------------------------
diff --cc shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
index 67c8c40,0000000..875367d
mode 100644,000000..100644
--- a/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/ShellOptionsJC.java
@@@ -1,281 -1,0 +1,281 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.shell;
 +
 +import java.io.File;
 +import java.io.FileNotFoundException;
 +import java.util.ArrayList;
 +import java.util.List;
 +import java.util.Map;
 +import java.util.Scanner;
 +import java.util.TreeMap;
 +
 +import org.apache.accumulo.core.client.ClientConfiguration;
 +import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
 +import org.apache.accumulo.core.client.security.tokens.AuthenticationToken;
 +import org.apache.commons.configuration.ConfigurationException;
 +import org.apache.commons.configuration.PropertiesConfiguration;
 +import org.slf4j.Logger;
 +import org.slf4j.LoggerFactory;
 +
 +import com.beust.jcommander.DynamicParameter;
 +import com.beust.jcommander.IStringConverter;
 +import com.beust.jcommander.Parameter;
 +import com.beust.jcommander.ParameterException;
 +import com.beust.jcommander.converters.FileConverter;
 +
 +public class ShellOptionsJC {
 +  private static final Logger log = LoggerFactory.getLogger(Shell.class);
 +
 +  @Parameter(names = {"-u", "--user"}, description = "username (defaults to your OS user)")
 +  private String username = System.getProperty("user.name", "root");
 +
 +  public static class PasswordConverter implements IStringConverter<String> {
 +    public static final String STDIN = "stdin";
 +
 +    private enum KeyType {
 +      PASS("pass:"), ENV("env:") {
 +        @Override
 +        String process(String value) {
 +          return System.getenv(value);
 +        }
 +      },
 +      FILE("file:") {
 +        @Override
 +        String process(String value) {
 +          Scanner scanner = null;
 +          try {
 +            scanner = new Scanner(new File(value));
 +            return scanner.nextLine();
 +          } catch (FileNotFoundException e) {
 +            throw new ParameterException(e);
 +          } finally {
 +            if (scanner != null) {
 +              scanner.close();
 +            }
 +          }
 +        }
 +      },
 +      STDIN(PasswordConverter.STDIN) {
 +        @Override
 +        public boolean matches(String value) {
 +          return prefix.equals(value);
 +        }
 +
 +        @Override
 +        public String convert(String value) {
 +          // Will check for this later
 +          return prefix;
 +        }
 +      };
 +
 +      String prefix;
 +
 +      private KeyType(String prefix) {
 +        this.prefix = prefix;
 +      }
 +
 +      public boolean matches(String value) {
 +        return value.startsWith(prefix);
 +      }
 +
 +      public String convert(String value) {
 +        return process(value.substring(prefix.length()));
 +      }
 +
 +      String process(String value) {
 +        return value;
 +      }
 +    };
 +
 +    @Override
 +    public String convert(String value) {
 +      for (KeyType keyType : KeyType.values()) {
 +        if (keyType.matches(value)) {
 +          return keyType.convert(value);
 +        }
 +      }
 +
 +      return value;
 +    }
 +  }
 +
 +  // Note: Don't use "password = true" because then it will prompt even if we have a token
 +  @Parameter(names = {"-p", "--password"}, description = "password (can be specified as 'pass:<password>', 'file:<local file containing the password>', "
 +      + "'env:<variable containing the pass>', or stdin)", converter = PasswordConverter.class)
 +  private String password;
 +
 +  public static class TokenConverter implements IStringConverter<AuthenticationToken> {
 +    @Override
 +    public AuthenticationToken convert(String value) {
 +      try {
 +        return Class.forName(value).asSubclass(AuthenticationToken.class).newInstance();
 +      } catch (Exception e) {
 +        // Catching ClassNotFoundException, ClassCastException, InstantiationException and IllegalAccessException
 +        log.error("Could not instantiate AuthenticationToken " + value, e);
 +        throw new ParameterException(e);
 +      }
 +    }
 +  }
 +
 +  @Parameter(names = {"-tc", "--tokenClass"}, description = "token type to create, use the -l to pass options", converter = TokenConverter.class)
 +  private AuthenticationToken authenticationToken;
 +
 +  @DynamicParameter(names = {"-l", "--tokenProperty"}, description = "login properties in the format key=value. Reuse -l for each property")
 +  private Map<String,String> tokenProperties = new TreeMap<String,String>();
 +
 +  @Parameter(names = "--disable-tab-completion", description = "disables tab completion (for less overhead when scripting)")
 +  private boolean tabCompletionDisabled;
 +
 +  @Parameter(names = "--debug", description = "enables client debugging")
 +  private boolean debugEnabled;
 +
 +  @Parameter(names = "--fake", description = "fake a connection to accumulo")
 +  private boolean fake;
 +
 +  @Parameter(names = {"-?", "--help"}, help = true, description = "display this help")
 +  private boolean helpEnabled;
 +
 +  @Parameter(names = {"-e", "--execute-command"}, description = "executes a command, and then exits")
 +  private String execCommand;
 +
 +  @Parameter(names = {"-f", "--execute-file"}, description = "executes commands from a file at startup", converter = FileConverter.class)
 +  private File execFile;
 +
 +  @Parameter(names = {"-fv", "--execute-file-verbose"}, description = "executes commands from a file at startup, with commands shown",
 +      converter = FileConverter.class)
 +  private File execFileVerbose;
 +
 +  @Parameter(names = {"-h", "--hdfsZooInstance"}, description = "use hdfs zoo instance")
 +  private boolean hdfsZooInstance;
 +
 +  @Parameter(names = {"-z", "--zooKeeperInstance"}, description = "use a zookeeper instance with the given instance name and list of zoo hosts", arity = 2)
 +  private List<String> zooKeeperInstance = new ArrayList<String>();
 +
 +  @Parameter(names = {"--ssl"}, description = "use ssl to connect to accumulo")
 +  private boolean useSsl = false;
 +
-   @Parameter(
-       names = "--config-file",
-       description = "read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
++  @Parameter(names = "--config-file", description = "read the given client config file. "
++      + "If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, "
++      + "which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
 +  private String clientConfigFile = null;
 +
 +  @Parameter(names = {"-zi", "--zooKeeperInstanceName"}, description = "use a zookeeper instance with the given instance name")
 +  private String zooKeeperInstanceName;
 +
 +  @Parameter(names = {"-zh", "--zooKeeperHosts"}, description = "use a zookeeper instance with the given list of zoo hosts")
 +  private String zooKeeperHosts;
 +
 +  @Parameter(names = "--auth-timeout", description = "minutes the shell can be idle without re-entering a password")
 +  private int authTimeout = 60; // TODO Add validator for positive number
 +
 +  @Parameter(names = "--disable-auth-timeout", description = "disables requiring the user to re-type a password after being idle")
 +  private boolean authTimeoutDisabled;
 +
 +  @Parameter(hidden = true)
 +  private List<String> unrecognizedOptions;
 +
 +  public String getUsername() {
 +    return username;
 +  }
 +
 +  public String getPassword() {
 +    return password;
 +  }
 +
 +  public AuthenticationToken getAuthenticationToken() {
 +    return authenticationToken;
 +  }
 +
 +  public Map<String,String> getTokenProperties() {
 +    return tokenProperties;
 +  }
 +
 +  public boolean isTabCompletionDisabled() {
 +    return tabCompletionDisabled;
 +  }
 +
 +  public boolean isDebugEnabled() {
 +    return debugEnabled;
 +  }
 +
 +  public boolean isFake() {
 +    return fake;
 +  }
 +
 +  public boolean isHelpEnabled() {
 +    return helpEnabled;
 +  }
 +
 +  public String getExecCommand() {
 +    return execCommand;
 +  }
 +
 +  public File getExecFile() {
 +    return execFile;
 +  }
 +
 +  public File getExecFileVerbose() {
 +    return execFileVerbose;
 +  }
 +
 +  public boolean isHdfsZooInstance() {
 +    return hdfsZooInstance;
 +  }
 +
 +  public List<String> getZooKeeperInstance() {
 +    return zooKeeperInstance;
 +  }
 +
 +  public String getZooKeeperInstanceName() {
 +    return zooKeeperInstanceName;
 +  }
 +
 +  public String getZooKeeperHosts() {
 +    return zooKeeperHosts;
 +  }
 +
 +  public int getAuthTimeout() {
 +    return authTimeout;
 +  }
 +
 +  public boolean isAuthTimeoutDisabled() {
 +    return authTimeoutDisabled;
 +  }
 +
 +  public List<String> getUnrecognizedOptions() {
 +    return unrecognizedOptions;
 +  }
 +
 +  public boolean useSsl() {
 +    return useSsl;
 +  }
 +
 +  public String getClientConfigFile() {
 +    return clientConfigFile;
 +  }
 +
 +  public ClientConfiguration getClientConfiguration() throws ConfigurationException, FileNotFoundException {
 +    ClientConfiguration clientConfig = clientConfigFile == null ? ClientConfiguration.loadDefault() : new ClientConfiguration(new PropertiesConfiguration(
 +        getClientConfigFile()));
 +    if (useSsl()) {
 +      clientConfig.setProperty(ClientProperty.INSTANCE_RPC_SSL_ENABLED, "true");
 +    }
 +    return clientConfig;
 +  }
 +
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/shell/src/main/java/org/apache/accumulo/shell/commands/DUCommand.java
----------------------------------------------------------------------
diff --cc shell/src/main/java/org/apache/accumulo/shell/commands/DUCommand.java
index 1d9b9f1,0000000..3e851d4
mode 100644,000000..100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/DUCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/DUCommand.java
@@@ -1,125 -1,0 +1,126 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.shell.commands;
 +
 +import java.io.IOException;
 +import java.util.Arrays;
 +import java.util.SortedSet;
 +import java.util.TreeSet;
 +
 +import org.apache.accumulo.core.client.Instance;
 +import org.apache.accumulo.core.client.NamespaceNotFoundException;
 +import org.apache.accumulo.core.client.TableNotFoundException;
 +import org.apache.accumulo.core.client.admin.DiskUsage;
 +import org.apache.accumulo.core.client.impl.Namespaces;
 +import org.apache.accumulo.core.util.NumUtil;
 +import org.apache.accumulo.shell.Shell;
 +import org.apache.accumulo.shell.Shell.Command;
 +import org.apache.accumulo.shell.ShellOptions;
 +import org.apache.commons.cli.CommandLine;
 +import org.apache.commons.cli.Option;
 +import org.apache.commons.cli.Options;
 +
 +public class DUCommand extends Command {
 +
 +  private Option optTablePattern, optHumanReadble, optNamespace;
 +
 +  @Override
 +  public int execute(final String fullCommand, final CommandLine cl, final Shell shellState) throws IOException, TableNotFoundException,
 +      NamespaceNotFoundException {
 +
 +    final SortedSet<String> tables = new TreeSet<String>(Arrays.asList(cl.getArgs()));
 +
 +    if (cl.hasOption(ShellOptions.tableOption)) {
 +      String tableName = cl.getOptionValue(ShellOptions.tableOption);
 +      if (!shellState.getConnector().tableOperations().exists(tableName)) {
 +        throw new TableNotFoundException(tableName, tableName, "specified table that doesn't exist");
 +      }
 +      tables.add(tableName);
 +    }
 +
 +    if (cl.hasOption(optNamespace.getOpt())) {
 +      Instance instance = shellState.getInstance();
 +      String namespaceId = Namespaces.getNamespaceId(instance, cl.getOptionValue(optNamespace.getOpt()));
 +      tables.addAll(Namespaces.getTableNames(instance, namespaceId));
 +    }
 +
 +    boolean prettyPrint = cl.hasOption(optHumanReadble.getOpt()) ? true : false;
 +
 +    // Add any patterns
 +    if (cl.hasOption(optTablePattern.getOpt())) {
 +      for (String table : shellState.getConnector().tableOperations().list()) {
 +        if (table.matches(cl.getOptionValue(optTablePattern.getOpt()))) {
 +          tables.add(table);
 +        }
 +      }
 +    }
 +
 +    // If we didn't get any tables, and we have a table selected, add the current table
 +    if (tables.isEmpty() && !shellState.getTableName().isEmpty()) {
 +      tables.add(shellState.getTableName());
 +    }
 +
 +    try {
 +      String valueFormat = prettyPrint ? "%9s" : "%,24d";
 +      for (DiskUsage usage : shellState.getConnector().tableOperations().getDiskUsage(tables)) {
 +        Object value = prettyPrint ? NumUtil.bigNumberForSize(usage.getUsage()) : usage.getUsage();
 +        shellState.getReader().println(String.format(valueFormat + " %s", value, usage.getTables()));
 +      }
 +    } catch (Exception ex) {
 +      throw new RuntimeException(ex);
 +    }
 +    return 0;
 +  }
 +
 +  @Override
 +  public String description() {
-     return "prints how much space, in bytes, is used by files referenced by a table.  When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
++    return "prints how much space, in bytes, is used by files referenced by a table.  "
++        + "When multiple tables are specified it prints how much space, in bytes, is used by files shared between tables, if any.";
 +  }
 +
 +  @Override
 +  public Options getOptions() {
 +    final Options o = new Options();
 +
 +    optTablePattern = new Option("p", "pattern", true, "regex pattern of table names");
 +    optTablePattern.setArgName("pattern");
 +
 +    optHumanReadble = new Option("h", "human-readable", false, "format large sizes to human readable units");
 +    optHumanReadble.setArgName("human readable output");
 +
 +    optNamespace = new Option(ShellOptions.namespaceOption, "namespace", true, "name of a namespace");
 +    optNamespace.setArgName("namespace");
 +
 +    o.addOption(OptUtil.tableOpt("table to examine"));
 +
 +    o.addOption(optTablePattern);
 +    o.addOption(optHumanReadble);
 +    o.addOption(optNamespace);
 +
 +    return o;
 +  }
 +
 +  @Override
 +  public String usage() {
 +    return getName() + " <table>{ <table>}";
 +  }
 +
 +  @Override
 +  public int numArgs() {
 +    return Shell.NO_FIXED_ARG_LENGTH_CHECK;
 +  }
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
----------------------------------------------------------------------
diff --cc shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
index eeac50c,0000000..5ffdca7
mode 100644,000000..100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/EGrepCommand.java
@@@ -1,59 -1,0 +1,60 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.shell.commands;
 +
 +import java.io.IOException;
 +
 +import org.apache.accumulo.core.client.BatchScanner;
 +import org.apache.accumulo.core.client.IteratorSetting;
 +import org.apache.accumulo.core.iterators.user.RegExFilter;
 +import org.apache.commons.cli.CommandLine;
 +import org.apache.commons.cli.Option;
 +import org.apache.commons.cli.Options;
 +
 +public class EGrepCommand extends GrepCommand {
 +
 +  private Option matchSubstringOption;
 +
 +  @Override
 +  protected void setUpIterator(final int prio, final String name, final String term, final BatchScanner scanner, CommandLine cl) throws IOException {
 +    if (prio < 0) {
 +      throw new IllegalArgumentException("Priority < 0 " + prio);
 +    }
 +    final IteratorSetting si = new IteratorSetting(prio, name, RegExFilter.class);
 +    RegExFilter.setRegexs(si, term, term, term, term, true, cl.hasOption(matchSubstringOption.getOpt()));
 +    scanner.addScanIterator(si);
 +  }
 +
 +  @Override
 +  public String description() {
-     return "searches each row, column family, column qualifier and value, in parallel, on the server side (using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
++    return "searches each row, column family, column qualifier and value, in parallel, on the server side "
++        + "(using a java Matcher, so put .* before and after your term if you're not matching the whole element)";
 +  }
 +
 +  @Override
 +  public String usage() {
 +    return getName() + " <regex>{ <regex>}";
 +  }
 +
 +  @Override
 +  public Options getOptions() {
 +    final Options opts = super.getOptions();
 +    matchSubstringOption = new Option("g", "global", false, "forces the use of the find() expression matcher, causing substring matches to return true");
 +    opts.addOption(matchSubstringOption);
 +    return opts;
 +  }
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
----------------------------------------------------------------------
diff --cc shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
index 1f3a1ae,0000000..74397de
mode 100644,000000..100644
--- a/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/commands/QuotedStringTokenizer.java
@@@ -1,142 -1,0 +1,139 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.shell.commands;
 +
 +import static java.nio.charset.StandardCharsets.UTF_8;
 +
 +import java.io.UnsupportedEncodingException;
 +import java.util.ArrayList;
 +import java.util.Iterator;
 +
 +import org.apache.accumulo.core.util.BadArgumentException;
 +import org.apache.accumulo.shell.Shell;
 +
 +/**
 + * A basic tokenizer for generating tokens from a string. It understands quoted strings and escaped quote characters.
 + *
 + * You can use the escape sequence '\' to escape single quotes, double quotes, and spaces only, in addition to the escape character itself.
 + *
 + * The behavior is the same for single and double quoted strings. (i.e. '\'' is the same as "\'")
 + */
 +
 +public class QuotedStringTokenizer implements Iterable<String> {
 +  private ArrayList<String> tokens;
 +  private String input;
 +
 +  public QuotedStringTokenizer(final String t) throws BadArgumentException {
 +    tokens = new ArrayList<String>();
 +    this.input = t;
 +    try {
 +      createTokens();
 +    } catch (UnsupportedEncodingException e) {
 +      throw new IllegalArgumentException(e.getMessage());
 +    }
 +  }
 +
 +  public String[] getTokens() {
 +    return tokens.toArray(new String[tokens.size()]);
 +  }
 +
 +  private void createTokens() throws BadArgumentException, UnsupportedEncodingException {
 +    boolean inQuote = false;
 +    boolean inEscapeSequence = false;
 +    String hexChars = null;
 +    char inQuoteChar = '"';
 +
 +    final byte[] token = new byte[input.length()];
 +    int tokenLength = 0;
 +    final byte[] inputBytes = input.getBytes(UTF_8);
 +    for (int i = 0; i < input.length(); ++i) {
 +      final char ch = input.charAt(i);
 +
 +      // if I ended up in an escape sequence, check for valid escapable character, and add it as a literal
 +      if (inEscapeSequence) {
 +        inEscapeSequence = false;
 +        if (ch == 'x') {
 +          hexChars = "";
 +        } else if (ch == ' ' || ch == '\'' || ch == '"' || ch == '\\') {
 +          token[tokenLength++] = inputBytes[i];
 +        } else {
 +          throw new BadArgumentException("can only escape single quotes, double quotes, the space character, the backslash, and hex input", input, i);
 +        }
-       }
-       // in a hex escape sequence
-       else if (hexChars != null) {
++      } else if (hexChars != null) {
++        // in a hex escape sequence
 +        final int digit = Character.digit(ch, 16);
 +        if (digit < 0) {
 +          throw new BadArgumentException("expected hex character", input, i);
 +        }
 +        hexChars += ch;
 +        if (hexChars.length() == 2) {
 +          byte b;
 +          try {
 +            b = (byte) (0xff & Short.parseShort(hexChars, 16));
 +            if (!Character.isValidCodePoint(0xff & b))
 +              throw new NumberFormatException();
 +          } catch (NumberFormatException e) {
 +            throw new BadArgumentException("unsupported non-ascii character", input, i);
 +          }
 +          token[tokenLength++] = b;
 +          hexChars = null;
 +        }
-       }
-       // in a quote, either end the quote, start escape, or continue a token
-       else if (inQuote) {
++      } else if (inQuote) {
++        // in a quote, either end the quote, start escape, or continue a token
 +        if (ch == inQuoteChar) {
 +          inQuote = false;
 +          tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
 +          tokenLength = 0;
 +        } else if (ch == '\\') {
 +          inEscapeSequence = true;
 +        } else {
 +          token[tokenLength++] = inputBytes[i];
 +        }
-       }
-       // not in a quote, either enter a quote, end a token, start escape, or continue a token
-       else {
++      } else {
++        // not in a quote, either enter a quote, end a token, start escape, or continue a token
 +        if (ch == '\'' || ch == '"') {
 +          if (tokenLength > 0) {
 +            tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
 +            tokenLength = 0;
 +          }
 +          inQuote = true;
 +          inQuoteChar = ch;
 +        } else if (ch == ' ' && tokenLength > 0) {
 +          tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
 +          tokenLength = 0;
 +        } else if (ch == '\\') {
 +          inEscapeSequence = true;
 +        } else if (ch != ' ') {
 +          token[tokenLength++] = inputBytes[i];
 +        }
 +      }
 +    }
 +    if (inQuote) {
 +      throw new BadArgumentException("missing terminating quote", input, input.length());
 +    } else if (inEscapeSequence || hexChars != null) {
 +      throw new BadArgumentException("escape sequence not complete", input, input.length());
 +    }
 +    if (tokenLength > 0) {
 +      tokens.add(new String(token, 0, tokenLength, Shell.CHARSET));
 +    }
 +  }
 +
 +  @Override
 +  public Iterator<String> iterator() {
 +    return tokens.iterator();
 +  }
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------
diff --cc test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
index fb4f309,d3870a5..3392e1d
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
@@@ -45,12 -44,11 +45,11 @@@ public class CreateTable extends Test 
        if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
          if (hasPermission)
            throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-         else
-         // create table anyway for sake of state
-         {
+         else {
+           // create table anyway for sake of state
            try {
 -            state.getConnector().tableOperations().create(tableName);
 -            WalkingSecurity.get(state).initTable(tableName);
 +            env.getConnector().tableOperations().create(tableName);
 +            WalkingSecurity.get(state, env).initTable(tableName);
            } catch (TableExistsException tee) {
              if (exists)
                return;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/392d9d6a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
----------------------------------------------------------------------
diff --cc test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
index eb07c43,1f539ff..5ddd441
--- a/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
+++ b/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateUser.java
@@@ -44,12 -43,11 +44,11 @@@ public class CreateUser extends Test 
          case PERMISSION_DENIED:
            if (hasPermission)
              throw new AccumuloException("Got a security exception when I should have had permission.", ae);
-           else
-           // create user anyway for sake of state
-           {
+           else {
+             // create user anyway for sake of state
              if (!exists) {
 -              state.getConnector().securityOperations().createLocalUser(tableUserName, tabUserPass);
 -              WalkingSecurity.get(state).createUser(tableUserName, tabUserPass);
 +              env.getConnector().securityOperations().createLocalUser(tableUserName, tabUserPass);
 +              WalkingSecurity.get(state, env).createUser(tableUserName, tabUserPass);
                Thread.sleep(1000);
              }
              return;


[16/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/CloneTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/CloneTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/CloneTest.java
index e2d1ecb..ebb83bf 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/CloneTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/CloneTest.java
@@ -38,115 +38,115 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.hadoop.io.Text;
 
 public class CloneTest extends TestCase {
-  
+
   public void testNoFiles() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     KeyExtent ke = new KeyExtent(new Text("0"), null, null);
     Mutation mut = ke.getPrevRowUpdateMutation();
-    
+
     TabletsSection.ServerColumnFamily.TIME_COLUMN.put(mut, new Value("M0".getBytes()));
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(mut, new Value("/default_tablet".getBytes()));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(mut);
-    
+
     bw1.close();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     int rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(0, rc);
-    
+
     // scan tables metadata entries and confirm the same
-    
+
   }
-  
+
   public void testFilesChange() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     KeyExtent ke = new KeyExtent(new Text("0"), null, null);
     Mutation mut = ke.getPrevRowUpdateMutation();
-    
+
     TabletsSection.ServerColumnFamily.TIME_COLUMN.put(mut, new Value("M0".getBytes()));
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(mut, new Value("/default_tablet".getBytes()));
     mut.put(DataFileColumnFamily.NAME.toString(), "/default_tablet/0_0.rf", "1,200");
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(mut);
-    
+
     bw1.flush();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     Mutation mut2 = new Mutation(ke.getMetadataEntry());
     mut2.putDelete(DataFileColumnFamily.NAME.toString(), "/default_tablet/0_0.rf");
     mut2.put(DataFileColumnFamily.NAME.toString(), "/default_tablet/1_0.rf", "2,300");
-    
+
     bw1.addMutation(mut2);
     bw1.flush();
-    
+
     int rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(1, rc);
-    
+
     rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(0, rc);
-    
+
     Scanner scanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.setRange(new KeyExtent(new Text("1"), null, null).toMetadataRange());
-    
+
     HashSet<String> files = new HashSet<String>();
-    
+
     for (Entry<Key,Value> entry : scanner) {
       if (entry.getKey().getColumnFamily().equals(DataFileColumnFamily.NAME))
         files.add(entry.getKey().getColumnQualifier().toString());
     }
-    
+
     assertEquals(1, files.size());
     assertTrue(files.contains("../0/default_tablet/1_0.rf"));
-    
+
   }
-  
+
   // test split where files of children are the same
   public void testSplit1() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(createTablet("0", null, null, "/default_tablet", "/default_tablet/0_0.rf"));
-    
+
     bw1.flush();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     bw1.addMutation(createTablet("0", "m", null, "/default_tablet", "/default_tablet/0_0.rf"));
     bw1.addMutation(createTablet("0", null, "m", "/t-1", "/default_tablet/0_0.rf"));
-    
+
     bw1.flush();
-    
+
     int rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(0, rc);
-    
+
     Scanner scanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.setRange(new KeyExtent(new Text("1"), null, null).toMetadataRange());
-    
+
     HashSet<String> files = new HashSet<String>();
-    
+
     int count = 0;
     for (Entry<Key,Value> entry : scanner) {
       if (entry.getKey().getColumnFamily().equals(DataFileColumnFamily.NAME)) {
@@ -154,61 +154,61 @@ public class CloneTest extends TestCase {
         count++;
       }
     }
-    
+
     assertEquals(1, count);
     assertEquals(1, files.size());
     assertTrue(files.contains("../0/default_tablet/0_0.rf"));
   }
-  
+
   // test split where files of children differ... like majc and split occurred
   public void testSplit2() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(createTablet("0", null, null, "/default_tablet", "/default_tablet/0_0.rf"));
-    
+
     bw1.flush();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     bw1.addMutation(createTablet("0", "m", null, "/default_tablet", "/default_tablet/1_0.rf"));
     Mutation mut3 = createTablet("0", null, "m", "/t-1", "/default_tablet/1_0.rf");
     mut3.putDelete(DataFileColumnFamily.NAME.toString(), "/default_tablet/0_0.rf");
     bw1.addMutation(mut3);
-    
+
     bw1.flush();
-    
+
     int rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(1, rc);
-    
+
     rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(0, rc);
-    
+
     Scanner scanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.setRange(new KeyExtent(new Text("1"), null, null).toMetadataRange());
-    
+
     HashSet<String> files = new HashSet<String>();
-    
+
     int count = 0;
-    
+
     for (Entry<Key,Value> entry : scanner) {
       if (entry.getKey().getColumnFamily().equals(DataFileColumnFamily.NAME)) {
         files.add(entry.getKey().getColumnQualifier().toString());
         count++;
       }
     }
-    
+
     assertEquals(1, files.size());
     assertEquals(2, count);
     assertTrue(files.contains("../0/default_tablet/1_0.rf"));
   }
-  
+
   private static Mutation deleteTablet(String tid, String endRow, String prevRow, String dir, String file) throws Exception {
     KeyExtent ke = new KeyExtent(new Text(tid), endRow == null ? null : new Text(endRow), prevRow == null ? null : new Text(prevRow));
     Mutation mut = new Mutation(ke.getMetadataEntry());
@@ -216,53 +216,53 @@ public class CloneTest extends TestCase {
     TabletsSection.ServerColumnFamily.TIME_COLUMN.putDelete(mut);
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.putDelete(mut);
     mut.putDelete(DataFileColumnFamily.NAME.toString(), file);
-    
+
     return mut;
   }
-  
+
   private static Mutation createTablet(String tid, String endRow, String prevRow, String dir, String file) throws Exception {
     KeyExtent ke = new KeyExtent(new Text(tid), endRow == null ? null : new Text(endRow), prevRow == null ? null : new Text(prevRow));
     Mutation mut = ke.getPrevRowUpdateMutation();
-    
+
     TabletsSection.ServerColumnFamily.TIME_COLUMN.put(mut, new Value("M0".getBytes()));
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(mut, new Value(dir.getBytes()));
     mut.put(DataFileColumnFamily.NAME.toString(), file, "10,200");
-    
+
     return mut;
   }
-  
+
   // test two tablets splitting into four
   public void testSplit3() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(createTablet("0", "m", null, "/d1", "/d1/file1"));
     bw1.addMutation(createTablet("0", null, "m", "/d2", "/d2/file2"));
-    
+
     bw1.flush();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     bw1.addMutation(createTablet("0", "f", null, "/d1", "/d1/file3"));
     bw1.addMutation(createTablet("0", "m", "f", "/d3", "/d1/file1"));
     bw1.addMutation(createTablet("0", "s", "m", "/d2", "/d2/file2"));
     bw1.addMutation(createTablet("0", null, "s", "/d4", "/d2/file2"));
-    
+
     bw1.flush();
-    
+
     int rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(0, rc);
-    
+
     Scanner scanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.setRange(new KeyExtent(new Text("1"), null, null).toMetadataRange());
-    
+
     HashSet<String> files = new HashSet<String>();
-    
+
     int count = 0;
     for (Entry<Key,Value> entry : scanner) {
       if (entry.getKey().getColumnFamily().equals(DataFileColumnFamily.NAME)) {
@@ -270,63 +270,63 @@ public class CloneTest extends TestCase {
         count++;
       }
     }
-    
+
     assertEquals(2, count);
     assertEquals(2, files.size());
     assertTrue(files.contains("../0/d1/file1"));
     assertTrue(files.contains("../0/d2/file2"));
   }
-  
+
   // test cloned marker
   public void testClonedMarker() throws Exception {
-    
+
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(createTablet("0", "m", null, "/d1", "/d1/file1"));
     bw1.addMutation(createTablet("0", null, "m", "/d2", "/d2/file2"));
-    
+
     bw1.flush();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     bw1.addMutation(deleteTablet("0", "m", null, "/d1", "/d1/file1"));
     bw1.addMutation(deleteTablet("0", null, "m", "/d2", "/d2/file2"));
-    
+
     bw1.flush();
-    
+
     bw1.addMutation(createTablet("0", "f", null, "/d1", "/d1/file3"));
     bw1.addMutation(createTablet("0", "m", "f", "/d3", "/d1/file1"));
     bw1.addMutation(createTablet("0", "s", "m", "/d2", "/d2/file3"));
     bw1.addMutation(createTablet("0", null, "s", "/d4", "/d4/file3"));
-    
+
     bw1.flush();
-    
+
     int rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(1, rc);
-    
+
     bw1.addMutation(deleteTablet("0", "m", "f", "/d3", "/d1/file1"));
-    
+
     bw1.flush();
-    
+
     bw1.addMutation(createTablet("0", "m", "f", "/d3", "/d1/file3"));
-    
+
     bw1.flush();
-    
+
     rc = MetadataTableUtil.checkClone("0", "1", conn, bw2);
-    
+
     assertEquals(0, rc);
-    
+
     Scanner scanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.setRange(new KeyExtent(new Text("1"), null, null).toMetadataRange());
-    
+
     HashSet<String> files = new HashSet<String>();
-    
+
     int count = 0;
     for (Entry<Key,Value> entry : scanner) {
       if (entry.getKey().getColumnFamily().equals(DataFileColumnFamily.NAME)) {
@@ -334,42 +334,42 @@ public class CloneTest extends TestCase {
         count++;
       }
     }
-    
+
     assertEquals(3, count);
     assertEquals(3, files.size());
     assertTrue(files.contains("../0/d1/file1"));
     assertTrue(files.contains("../0/d2/file3"));
     assertTrue(files.contains("../0/d4/file3"));
   }
-  
+
   // test two tablets splitting into four
   public void testMerge() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     bw1.addMutation(createTablet("0", "m", null, "/d1", "/d1/file1"));
     bw1.addMutation(createTablet("0", null, "m", "/d2", "/d2/file2"));
-    
+
     bw1.flush();
-    
+
     BatchWriter bw2 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
-    
+
     MetadataTableUtil.initializeClone("0", "1", conn, bw2);
-    
+
     bw1.addMutation(deleteTablet("0", "m", null, "/d1", "/d1/file1"));
     Mutation mut = createTablet("0", null, null, "/d2", "/d2/file2");
     mut.put(DataFileColumnFamily.NAME.toString(), "/d1/file1", "10,200");
     bw1.addMutation(mut);
-    
+
     bw1.flush();
-    
+
     try {
       MetadataTableUtil.checkClone("0", "1", conn, bw2);
       assertTrue(false);
     } catch (TabletIterator.TabletDeletedException tde) {}
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/DefaultMapTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/DefaultMapTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/DefaultMapTest.java
index b68d412..3303d8a 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/DefaultMapTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/DefaultMapTest.java
@@ -16,15 +16,15 @@
  */
 package org.apache.accumulo.server.util;
 
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNotNull;
 
 import java.util.concurrent.atomic.AtomicInteger;
 
-import org.apache.accumulo.server.util.DefaultMap;
 import org.junit.Test;
 
 public class DefaultMapTest {
-  
+
   @Test
   public void testDefaultMap() {
     Integer value = new DefaultMap<String,Integer>(0).get("test");
@@ -33,15 +33,15 @@ public class DefaultMapTest {
     value = new DefaultMap<String,Integer>(1).get("test");
     assertNotNull(value);
     assertEquals(new Integer(1), value);
-    
+
     AtomicInteger canConstruct = new DefaultMap<String,AtomicInteger>(new AtomicInteger(1)).get("test");
     assertNotNull(canConstruct);
     assertEquals(new AtomicInteger(0).get(), canConstruct.get());
-    
+
     DefaultMap<String,String> map = new DefaultMap<String,String>("");
     assertEquals(map.get("foo"), "");
     map.put("foo", "bar");
     assertEquals(map.get("foo"), "bar");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/FileInfoTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/FileInfoTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/FileInfoTest.java
index cd568cf..77fe8b9 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/FileInfoTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/FileInfoTest.java
@@ -16,11 +16,12 @@
  */
 package org.apache.accumulo.server.util;
 
+import static org.junit.Assert.assertEquals;
+
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.server.util.FileUtil.FileInfo;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class FileInfoTest {
   private Key key1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
index 90c6300..65db4b8 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
@@ -83,140 +83,140 @@ public class FileUtilTest {
   @SuppressWarnings("deprecation")
   @Test
   public void testCleanupIndexOpWithDfsDir() throws IOException {
-      // And a "unique" tmp directory for each volume
+    // And a "unique" tmp directory for each volume
     File tmp1 = new File(accumuloDir, "tmp");
-      tmp1.mkdirs();
-      Path tmpPath1 = new Path(tmp1.toURI());
+    tmp1.mkdirs();
+    Path tmpPath1 = new Path(tmp1.toURI());
 
-      HashMap<Property,String> testProps = new HashMap<Property,String>();
+    HashMap<Property,String> testProps = new HashMap<Property,String>();
     testProps.put(Property.INSTANCE_DFS_DIR, accumuloDir.getAbsolutePath());
 
-      AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+    AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
     VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+    Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
   }
 
   @Test
   public void testCleanupIndexOpWithCommonParentVolume() throws IOException {
-      File volumeDir = new File(accumuloDir, "volumes");
-      volumeDir.mkdirs();
+    File volumeDir = new File(accumuloDir, "volumes");
+    volumeDir.mkdirs();
 
-      // Make some directories to simulate multiple volumes
-      File v1 = new File(volumeDir, "v1"), v2 = new File(volumeDir, "v2");
-      v1.mkdirs();
-      v2.mkdirs();
+    // Make some directories to simulate multiple volumes
+    File v1 = new File(volumeDir, "v1"), v2 = new File(volumeDir, "v2");
+    v1.mkdirs();
+    v2.mkdirs();
 
-      // And a "unique" tmp directory for each volume
-      File tmp1 = new File(v1, "tmp"), tmp2 = new File(v2, "tmp");
-      tmp1.mkdirs();
-      tmp2.mkdirs();
-      Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
+    // And a "unique" tmp directory for each volume
+    File tmp1 = new File(v1, "tmp"), tmp2 = new File(v2, "tmp");
+    tmp1.mkdirs();
+    tmp2.mkdirs();
+    Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
 
-      HashMap<Property,String> testProps = new HashMap<Property,String>();
-      testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
+    HashMap<Property,String> testProps = new HashMap<Property,String>();
+    testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
 
-      AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
-      VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
+    AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+    VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+    Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
+    Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
   }
 
   @Test
   public void testCleanupIndexOpWithCommonParentVolumeWithDepth() throws IOException {
-      File volumeDir = new File(accumuloDir, "volumes");
-      volumeDir.mkdirs();
+    File volumeDir = new File(accumuloDir, "volumes");
+    volumeDir.mkdirs();
 
-      // Make some directories to simulate multiple volumes
-      File v1 = new File(volumeDir, "v1"), v2 = new File(volumeDir, "v2");
-      v1.mkdirs();
-      v2.mkdirs();
+    // Make some directories to simulate multiple volumes
+    File v1 = new File(volumeDir, "v1"), v2 = new File(volumeDir, "v2");
+    v1.mkdirs();
+    v2.mkdirs();
 
-      // And a "unique" tmp directory for each volume
-      // Make sure we can handle nested directories (a single tmpdir with potentially multiple unique dirs)
-      File tmp1 = new File(new File(v1, "tmp"), "tmp_1"), tmp2 = new File(new File(v2, "tmp"), "tmp_1");
-      tmp1.mkdirs();
-      tmp2.mkdirs();
-      Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
+    // And a "unique" tmp directory for each volume
+    // Make sure we can handle nested directories (a single tmpdir with potentially multiple unique dirs)
+    File tmp1 = new File(new File(v1, "tmp"), "tmp_1"), tmp2 = new File(new File(v2, "tmp"), "tmp_1");
+    tmp1.mkdirs();
+    tmp2.mkdirs();
+    Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
 
-      HashMap<Property,String> testProps = new HashMap<Property,String>();
-      testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
+    HashMap<Property,String> testProps = new HashMap<Property,String>();
+    testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
 
-      AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
-      VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
+    AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+    VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+    Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
+    Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
   }
 
   @Test
   public void testCleanupIndexOpWithoutCommonParentVolume() throws IOException {
-      // Make some directories to simulate multiple volumes
-      File v1 = new File(accumuloDir, "v1"), v2 = new File(accumuloDir, "v2");
-      v1.mkdirs();
-      v2.mkdirs();
+    // Make some directories to simulate multiple volumes
+    File v1 = new File(accumuloDir, "v1"), v2 = new File(accumuloDir, "v2");
+    v1.mkdirs();
+    v2.mkdirs();
 
-      // And a "unique" tmp directory for each volume
-      File tmp1 = new File(v1, "tmp"), tmp2 = new File(v2, "tmp");
-      tmp1.mkdirs();
-      tmp2.mkdirs();
-      Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
+    // And a "unique" tmp directory for each volume
+    File tmp1 = new File(v1, "tmp"), tmp2 = new File(v2, "tmp");
+    tmp1.mkdirs();
+    tmp2.mkdirs();
+    Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
 
-      HashMap<Property,String> testProps = new HashMap<Property,String>();
-      testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
+    HashMap<Property,String> testProps = new HashMap<Property,String>();
+    testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
 
-      AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
-      VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
+    AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+    VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+    Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
+    Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
   }
 
   @Test
   public void testCleanupIndexOpWithoutCommonParentVolumeWithDepth() throws IOException {
-      // Make some directories to simulate multiple volumes
-      File v1 = new File(accumuloDir, "v1"), v2 = new File(accumuloDir, "v2");
-      v1.mkdirs();
-      v2.mkdirs();
-
-      // And a "unique" tmp directory for each volume
-      // Make sure we can handle nested directories (a single tmpdir with potentially multiple unique dirs)
-      File tmp1 = new File(new File(v1, "tmp"), "tmp_1"), tmp2 = new File(new File(v2, "tmp"), "tmp_1");
-      tmp1.mkdirs();
-      tmp2.mkdirs();
-      Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
-
-      HashMap<Property,String> testProps = new HashMap<Property,String>();
-      testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
-
-      AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
-      VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
+    // Make some directories to simulate multiple volumes
+    File v1 = new File(accumuloDir, "v1"), v2 = new File(accumuloDir, "v2");
+    v1.mkdirs();
+    v2.mkdirs();
+
+    // And a "unique" tmp directory for each volume
+    // Make sure we can handle nested directories (a single tmpdir with potentially multiple unique dirs)
+    File tmp1 = new File(new File(v1, "tmp"), "tmp_1"), tmp2 = new File(new File(v2, "tmp"), "tmp_1");
+    tmp1.mkdirs();
+    tmp2.mkdirs();
+    Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
+
+    HashMap<Property,String> testProps = new HashMap<Property,String>();
+    testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
+
+    AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+    VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+    Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
 
-      FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
+    FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
 
-      Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
+    Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
   }
 
   private static class FileUtilTestConfiguration extends AccumuloConfiguration {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/ReplicationTableUtilTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/ReplicationTableUtilTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/ReplicationTableUtilTest.java
index 4db171e..355fa42 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/ReplicationTableUtilTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/ReplicationTableUtilTest.java
@@ -62,7 +62,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ReplicationTableUtilTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/TServerUtilsTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/TServerUtilsTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/TServerUtilsTest.java
index 75bf953..218d82c 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/TServerUtilsTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/TServerUtilsTest.java
@@ -16,6 +16,13 @@
  */
 package org.apache.accumulo.server.util;
 
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.createNiceMock;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertTrue;
+
 import java.util.concurrent.ExecutorService;
 
 import org.apache.accumulo.server.rpc.TServerUtils;
@@ -23,9 +30,6 @@ import org.apache.thrift.server.TServer;
 import org.apache.thrift.transport.TServerSocket;
 import org.junit.Test;
 
-import static org.junit.Assert.*;
-import static org.easymock.EasyMock.*;
-
 public class TServerUtilsTest {
   private static class TServerWithoutES extends TServer {
     boolean stopCalled;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/TabletIteratorTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/TabletIteratorTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/TabletIteratorTest.java
index 72ce334..0e81b79 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/TabletIteratorTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/TabletIteratorTest.java
@@ -39,64 +39,64 @@ import org.apache.accumulo.server.util.TabletIterator.TabletDeletedException;
 import org.apache.hadoop.io.Text;
 
 public class TabletIteratorTest extends TestCase {
-  
+
   class TestTabletIterator extends TabletIterator {
-    
+
     private Connector conn;
-    
+
     public TestTabletIterator(Connector conn) throws Exception {
       super(conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY), MetadataSchema.TabletsSection.getRange(), true, true);
       this.conn = conn;
     }
-    
+
     @Override
     protected void resetScanner() {
       try {
         Scanner ds = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
         Text tablet = new KeyExtent(new Text("0"), new Text("m"), null).getMetadataEntry();
         ds.setRange(new Range(tablet, true, tablet, true));
-        
+
         Mutation m = new Mutation(tablet);
-        
+
         BatchWriter bw = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
         for (Entry<Key,Value> entry : ds) {
           Key k = entry.getKey();
           m.putDelete(k.getColumnFamily(), k.getColumnQualifier(), k.getTimestamp());
         }
-        
+
         bw.addMutation(m);
-        
+
         bw.close();
-        
+
       } catch (Exception e) {
         throw new RuntimeException(e);
       }
-      
+
       super.resetScanner();
     }
-    
+
   }
-  
+
   // simulate a merge happening while iterating over tablets
   public void testMerge() throws Exception {
     MockInstance mi = new MockInstance();
     Connector conn = mi.getConnector("", new PasswordToken(""));
-    
+
     KeyExtent ke1 = new KeyExtent(new Text("0"), new Text("m"), null);
     Mutation mut1 = ke1.getPrevRowUpdateMutation();
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(mut1, new Value("/d1".getBytes()));
-    
+
     KeyExtent ke2 = new KeyExtent(new Text("0"), null, null);
     Mutation mut2 = ke2.getPrevRowUpdateMutation();
     TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN.put(mut2, new Value("/d2".getBytes()));
-    
+
     BatchWriter bw1 = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
     bw1.addMutation(mut1);
     bw1.addMutation(mut2);
     bw1.close();
-    
+
     TestTabletIterator tabIter = new TestTabletIterator(conn);
-    
+
     try {
       while (tabIter.hasNext()) {
         tabIter.next();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/time/BaseRelativeTimeTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/time/BaseRelativeTimeTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/time/BaseRelativeTimeTest.java
index fdedd84..58760a4 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/time/BaseRelativeTimeTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/time/BaseRelativeTimeTest.java
@@ -16,34 +16,33 @@
  */
 package org.apache.accumulo.server.util.time;
 
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
 
-import org.apache.accumulo.server.util.time.BaseRelativeTime;
-import org.apache.accumulo.server.util.time.ProvidesTime;
 import org.junit.Test;
 
 public class BaseRelativeTimeTest {
-  
+
   static class BogusTime implements ProvidesTime {
     public long value = 0;
-    
+
     public long currentTime() {
       return value;
     }
   }
-  
+
   @Test
   public void testMatchesTime() {
     BogusTime bt = new BogusTime();
     BogusTime now = new BogusTime();
     now.value = bt.value = System.currentTimeMillis();
-    
+
     BaseRelativeTime brt = new BaseRelativeTime(now);
     assertEquals(brt.currentTime(), now.value);
     brt.updateTime(now.value);
     assertEquals(brt.currentTime(), now.value);
   }
-  
+
   @Test
   public void testFutureTime() {
     BogusTime advice = new BogusTime();
@@ -51,14 +50,14 @@ public class BaseRelativeTimeTest {
     local.value = advice.value = System.currentTimeMillis();
     // Ten seconds into the future
     advice.value += 10000;
-    
+
     BaseRelativeTime brt = new BaseRelativeTime(local);
     assertEquals(brt.currentTime(), local.value);
     brt.updateTime(advice.value);
     long once = brt.currentTime();
     assertTrue(once < advice.value);
     assertTrue(once > local.value);
-    
+
     for (int i = 0; i < 100; i++) {
       brt.updateTime(advice.value);
     }
@@ -66,7 +65,7 @@ public class BaseRelativeTimeTest {
     assertTrue(many > once);
     assertTrue("after much advice, relative time is still closer to local time", (advice.value - many) < (once - local.value));
   }
-  
+
   @Test
   public void testPastTime() {
     BogusTime advice = new BogusTime();
@@ -74,7 +73,7 @@ public class BaseRelativeTimeTest {
     local.value = advice.value = System.currentTimeMillis();
     // Ten seconds into the past
     advice.value -= 10000;
-    
+
     BaseRelativeTime brt = new BaseRelativeTime(local);
     brt.updateTime(advice.value);
     long once = brt.currentTime();
@@ -85,5 +84,5 @@ public class BaseRelativeTimeTest {
     brt.updateTime(advice.value - 10000);
     assertTrue("Time cannot go backwards", once <= twice);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/test/java/org/apache/accumulo/server/util/time/SimpleTimerTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/time/SimpleTimerTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/time/SimpleTimerTest.java
index 0a59812..9bde842 100644
--- a/server/base/src/test/java/org/apache/accumulo/server/util/time/SimpleTimerTest.java
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/time/SimpleTimerTest.java
@@ -16,10 +16,14 @@
  */
 package org.apache.accumulo.server.util.time;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertSame;
+import static org.junit.Assert.assertTrue;
+
 import java.util.concurrent.atomic.AtomicInteger;
+
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class SimpleTimerTest {
   private static final long DELAY = 1000L;
@@ -55,6 +59,7 @@ public class SimpleTimerTest {
 
   private static class Thrower implements Runnable {
     boolean wasRun = false;
+
     public void run() {
       wasRun = true;
       throw new IllegalStateException("You shall not pass");
@@ -94,6 +99,6 @@ public class SimpleTimerTest {
     assertEquals(1, SimpleTimer.getInstanceThreadPoolSize());
     SimpleTimer t2 = SimpleTimer.getInstance(2);
     assertSame(t, t2);
-    assertEquals(1, SimpleTimer.getInstanceThreadPoolSize());  // unchanged
+    assertEquals(1, SimpleTimer.getInstanceThreadPoolSize()); // unchanged
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionAlgorithm.java
----------------------------------------------------------------------
diff --git a/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionAlgorithm.java b/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionAlgorithm.java
index 56fbefe..f375328 100644
--- a/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionAlgorithm.java
+++ b/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionAlgorithm.java
@@ -50,7 +50,7 @@ import com.google.common.collect.Iterators;
 import com.google.common.collect.PeekingIterator;
 
 /**
- * 
+ *
  */
 public class GarbageCollectionAlgorithm {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionEnvironment.java
----------------------------------------------------------------------
diff --git a/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionEnvironment.java b/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionEnvironment.java
index 7f208d8..41ac204 100644
--- a/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionEnvironment.java
+++ b/server/gc/src/main/java/org/apache/accumulo/gc/GarbageCollectionEnvironment.java
@@ -36,13 +36,13 @@ import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection.Sc
 import org.apache.accumulo.core.replication.proto.Replication.Status;
 
 /**
- * 
+ *
  */
 public interface GarbageCollectionEnvironment {
 
   /**
    * Return a list of paths to files and dirs which are candidates for deletion from a given table, {@link RootTable#NAME} or {@link MetadataTable#NAME}
-   * 
+   *
    * @param continuePoint
    *          A row to resume from if a previous invocation was stopped due to finding an extremely large number of candidates to remove which would have
    *          exceeded memory limitations
@@ -52,28 +52,28 @@ public interface GarbageCollectionEnvironment {
 
   /**
    * Fetch a list of paths for all bulk loads in progress (blip) from a given table, {@link RootTable#NAME} or {@link MetadataTable#NAME}
-   * 
+   *
    * @return The list of files for each bulk load currently in progress.
    */
   Iterator<String> getBlipIterator() throws TableNotFoundException, AccumuloException, AccumuloSecurityException;
 
   /**
    * Fetches the references to files, {@link DataFileColumnFamily#NAME} or {@link ScanFileColumnFamily#NAME}, from tablets
-   * 
+   *
    * @return An {@link Iterator} of {@link Entry}&lt;{@link Key}, {@link Value}&gt; which constitute a reference to a file.
    */
   Iterator<Entry<Key,Value>> getReferenceIterator() throws TableNotFoundException, AccumuloException, AccumuloSecurityException;
 
   /**
    * Return the set of tableIDs for the given instance this GarbageCollector is running over
-   * 
+   *
    * @return The valueSet for the table name to table id map.
    */
   Set<String> getTableIDs();
 
   /**
    * Delete the given files from the provided {@link Map} of relative path to absolute path for each file that should be deleted
-   * 
+   *
    * @param candidateMap
    *          A Map from relative path to absolute path for files to be deleted.
    */
@@ -81,7 +81,7 @@ public interface GarbageCollectionEnvironment {
 
   /**
    * Delete a table's directory if it is empty.
-   * 
+   *
    * @param tableID
    *          The id of the table whose directory we are to operate on
    */
@@ -89,7 +89,7 @@ public interface GarbageCollectionEnvironment {
 
   /**
    * Increment the number of candidates for deletion for the current garbage collection run
-   * 
+   *
    * @param i
    *          Value to increment the deletion candidates by
    */
@@ -97,7 +97,7 @@ public interface GarbageCollectionEnvironment {
 
   /**
    * Increment the number of files still in use for the current garbage collection run
-   * 
+   *
    * @param i
    *          Value to increment the still-in-use count by.
    */
@@ -105,6 +105,7 @@ public interface GarbageCollectionEnvironment {
 
   /**
    * Determine if the given absolute file is still pending replication
+   *
    * @return True if the file still needs to be replicated
    */
   Iterator<Entry<String,Status>> getReplicationNeededIterator() throws AccumuloException, AccumuloSecurityException;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/gc/src/main/java/org/apache/accumulo/gc/SimpleGarbageCollector.java
----------------------------------------------------------------------
diff --git a/server/gc/src/main/java/org/apache/accumulo/gc/SimpleGarbageCollector.java b/server/gc/src/main/java/org/apache/accumulo/gc/SimpleGarbageCollector.java
index 93a9a49..db37c8b 100644
--- a/server/gc/src/main/java/org/apache/accumulo/gc/SimpleGarbageCollector.java
+++ b/server/gc/src/main/java/org/apache/accumulo/gc/SimpleGarbageCollector.java
@@ -716,8 +716,8 @@ public class SimpleGarbageCollector extends AccumuloServerContext implements Ifa
     HostAndPort result = HostAndPort.fromParts(opts.getAddress(), port);
     log.debug("Starting garbage collector listening on " + result);
     try {
-      return TServerUtils.startTServer(getConfiguration(), result, processor, this.getClass().getSimpleName(), "GC Monitor Service", 2,
-          getConfiguration().getCount(Property.GENERAL_SIMPLETIMER_THREADPOOL_SIZE), 1000, maxMessageSize, getServerSslParams(), 0).address;
+      return TServerUtils.startTServer(getConfiguration(), result, processor, this.getClass().getSimpleName(), "GC Monitor Service", 2, getConfiguration()
+          .getCount(Property.GENERAL_SIMPLETIMER_THREADPOOL_SIZE), 1000, maxMessageSize, getServerSslParams(), 0).address;
     } catch (Exception ex) {
       log.fatal(ex, ex);
       throw new RuntimeException(ex);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/gc/src/test/java/org/apache/accumulo/gc/GarbageCollectionTest.java
----------------------------------------------------------------------
diff --git a/server/gc/src/test/java/org/apache/accumulo/gc/GarbageCollectionTest.java b/server/gc/src/test/java/org/apache/accumulo/gc/GarbageCollectionTest.java
index 51b9596..286723d 100644
--- a/server/gc/src/test/java/org/apache/accumulo/gc/GarbageCollectionTest.java
+++ b/server/gc/src/test/java/org/apache/accumulo/gc/GarbageCollectionTest.java
@@ -42,7 +42,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class GarbageCollectionTest {
   static class TestGCE implements GarbageCollectionEnvironment {
@@ -345,7 +345,6 @@ public class GarbageCollectionTest {
     assertRemoved(gce);
   }
 
-
   @Test
   public void testCustomDirectories() throws Exception {
     TestGCE gce = new TestGCE();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/gc/src/test/java/org/apache/accumulo/gc/SimpleGarbageCollectorOptsTest.java
----------------------------------------------------------------------
diff --git a/server/gc/src/test/java/org/apache/accumulo/gc/SimpleGarbageCollectorOptsTest.java b/server/gc/src/test/java/org/apache/accumulo/gc/SimpleGarbageCollectorOptsTest.java
index d484741..b91784d 100644
--- a/server/gc/src/test/java/org/apache/accumulo/gc/SimpleGarbageCollectorOptsTest.java
+++ b/server/gc/src/test/java/org/apache/accumulo/gc/SimpleGarbageCollectorOptsTest.java
@@ -16,10 +16,11 @@
  */
 package org.apache.accumulo.gc;
 
+import static org.junit.Assert.assertFalse;
+
 import org.apache.accumulo.gc.SimpleGarbageCollector.Opts;
 import org.junit.Before;
 import org.junit.Test;
-import static org.junit.Assert.assertFalse;
 
 public class SimpleGarbageCollectorOptsTest {
   private Opts opts;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/EventCoordinator.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/EventCoordinator.java b/server/master/src/main/java/org/apache/accumulo/master/EventCoordinator.java
index e2f32c4..ebff7ab 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/EventCoordinator.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/EventCoordinator.java
@@ -19,10 +19,10 @@ package org.apache.accumulo.master;
 import org.apache.log4j.Logger;
 
 public class EventCoordinator {
-  
+
   private static final Logger log = Logger.getLogger(EventCoordinator.class);
   long eventCounter = 0;
-  
+
   synchronized long waitForEvents(long millis, long lastEvent) {
     // Did something happen since the last time we waited?
     if (lastEvent == eventCounter) {
@@ -37,27 +37,27 @@ public class EventCoordinator {
     }
     return eventCounter;
   }
-  
+
   synchronized public void event(String msg, Object... args) {
     log.info(String.format(msg, args));
     eventCounter++;
     notifyAll();
   }
-  
+
   public Listener getListener() {
     return new Listener();
   }
-  
+
   public class Listener {
     long lastEvent;
-    
+
     Listener() {
       lastEvent = eventCounter;
     }
-    
+
     public void waitForEvents(long millis) {
       lastEvent = EventCoordinator.this.waitForEvents(millis, lastEvent);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/FateServiceHandler.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/FateServiceHandler.java b/server/master/src/main/java/org/apache/accumulo/master/FateServiceHandler.java
index 5207745..d10a7ad 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/FateServiceHandler.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/FateServiceHandler.java
@@ -31,12 +31,12 @@ import java.util.Map.Entry;
 import java.util.Set;
 
 import org.apache.accumulo.core.client.AccumuloSecurityException;
-import org.apache.accumulo.core.client.impl.CompactionStrategyConfigUtil;
-import org.apache.accumulo.core.client.admin.CompactionStrategyConfig;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.NamespaceNotFoundException;
 import org.apache.accumulo.core.client.TableNotFoundException;
+import org.apache.accumulo.core.client.admin.CompactionStrategyConfig;
 import org.apache.accumulo.core.client.admin.TimeType;
+import org.apache.accumulo.core.client.impl.CompactionStrategyConfigUtil;
 import org.apache.accumulo.core.client.impl.Namespaces;
 import org.apache.accumulo.core.client.impl.TableOperationsImpl;
 import org.apache.accumulo.core.client.impl.Tables;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/TabletGroupWatcher.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/TabletGroupWatcher.java b/server/master/src/main/java/org/apache/accumulo/master/TabletGroupWatcher.java
index 802c967..35a2d10 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/TabletGroupWatcher.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/TabletGroupWatcher.java
@@ -94,30 +94,30 @@ class TabletGroupWatcher extends Daemon {
   private final Master master;
   final TabletStateStore store;
   final TabletGroupWatcher dependentWatcher;
-  
+
   final TableStats stats = new TableStats();
-  
+
   TabletGroupWatcher(Master master, TabletStateStore store, TabletGroupWatcher dependentWatcher) {
     this.master = master;
     this.store = store;
     this.dependentWatcher = dependentWatcher;
   }
-  
+
   Map<Text,TableCounts> getStats() {
     return stats.getLast();
   }
-  
+
   TableCounts getStats(Text tableId) {
     return stats.getLast(tableId);
   }
-  
+
   @Override
   public void run() {
-    
+
     Thread.currentThread().setName("Watching " + store.name());
     int[] oldCounts = new int[TabletState.values().length];
     EventCoordinator.Listener eventListener = this.master.nextEvent.getListener();
-    
+
     while (this.master.stillMaster()) {
       // slow things down a little, otherwise we spam the logs when there are many wake-up events
       UtilWaitThread.sleep(100);
@@ -133,27 +133,27 @@ class TabletGroupWatcher extends Daemon {
             currentMerges.put(merge.getExtent().getTableId(), new MergeStats(merge));
           }
         }
-        
+
         // Get the current status for the current list of tservers
         SortedMap<TServerInstance,TabletServerStatus> currentTServers = new TreeMap<TServerInstance,TabletServerStatus>();
         for (TServerInstance entry : this.master.tserverSet.getCurrentServers()) {
           currentTServers.put(entry, this.master.tserverStatus.get(entry));
         }
-        
+
         if (currentTServers.size() == 0) {
           eventListener.waitForEvents(Master.TIME_TO_WAIT_BETWEEN_SCANS);
           continue;
         }
-        
+
         // Don't move tablets to servers that are shutting down
         SortedMap<TServerInstance,TabletServerStatus> destinations = new TreeMap<TServerInstance,TabletServerStatus>(currentTServers);
         destinations.keySet().removeAll(this.master.serversToShutdown);
-        
+
         List<Assignment> assignments = new ArrayList<Assignment>();
         List<Assignment> assigned = new ArrayList<Assignment>();
         List<TabletLocationState> assignedToDeadServers = new ArrayList<TabletLocationState>();
         Map<KeyExtent,TServerInstance> unassigned = new HashMap<KeyExtent,TServerInstance>();
-        
+
         int[] counts = new int[TabletState.values().length];
         stats.begin();
         // Walk through the tablets in our store, and work tablets
@@ -167,10 +167,10 @@ class TabletGroupWatcher extends Daemon {
           // ignore entries for tables that do not exist in zookeeper
           if (TableManager.getInstance().getTableState(tls.extent.getTableId().toString()) == null)
             continue;
-          
+
           if (Master.log.isTraceEnabled())
             Master.log.trace(tls + " walogs " + tls.walogs.size());
-                    
+
           // Don't overwhelm the tablet servers with work
           if (unassigned.size() + unloaded > Master.MAX_TSERVER_WORK_CHUNK * currentTServers.size()) {
             flushChanges(destinations, assignments, assigned, assignedToDeadServers, unassigned);
@@ -199,12 +199,12 @@ class TabletGroupWatcher extends Daemon {
           mergeStats.update(tls.extent, state, tls.chopped, !tls.walogs.isEmpty());
           sendChopRequest(mergeStats.getMergeInfo(), state, tls);
           sendSplitRequest(mergeStats.getMergeInfo(), state, tls);
-          
+
           // Always follow through with assignments
           if (state == TabletState.ASSIGNED) {
             goal = TabletGoalState.HOSTED;
           }
-          
+
           // if we are shutting down all the tabletservers, we have to do it in order
           if (goal == TabletGoalState.UNASSIGNED && state == TabletState.HOSTED) {
             if (this.master.serversToShutdown.equals(currentTServers.keySet())) {
@@ -213,7 +213,7 @@ class TabletGroupWatcher extends Daemon {
               }
             }
           }
-          
+
           if (goal == TabletGoalState.HOSTED) {
             if (state != TabletState.HOSTED && !tls.walogs.isEmpty()) {
               if (this.master.recoveryManager.recoverLogs(tls.extent, tls.walogs))
@@ -275,12 +275,12 @@ class TabletGroupWatcher extends Daemon {
           }
           counts[state.ordinal()]++;
         }
-        
+
         flushChanges(destinations, assignments, assigned, assignedToDeadServers, unassigned);
-        
+
         // provide stats after flushing changes to avoid race conditions w/ delete table
         stats.end();
-        
+
         // Report changes
         for (TabletState state : TabletState.values()) {
           int i = state.ordinal();
@@ -293,14 +293,14 @@ class TabletGroupWatcher extends Daemon {
         if (totalUnloaded > 0) {
           this.master.nextEvent.event("[%s]: %d tablets unloaded", store.name(), totalUnloaded);
         }
-        
+
         updateMergeState(mergeStatsCache);
-        
+
         Master.log.debug(String.format("[%s] sleeping for %.2f seconds", store.name(), Master.TIME_TO_WAIT_BETWEEN_SCANS / 1000.));
         eventListener.waitForEvents(Master.TIME_TO_WAIT_BETWEEN_SCANS);
       } catch (Exception ex) {
         Master.log.error("Error processing table state for store " + store.name(), ex);
-        if (ex.getCause() != null && ex.getCause() instanceof BadLocationStateException) { 
+        if (ex.getCause() != null && ex.getCause() instanceof BadLocationStateException) {
           repairMetadata(((BadLocationStateException) ex.getCause()).getEncodedEndRow());
         } else {
           UtilWaitThread.sleep(Master.WAIT_BETWEEN_ERRORS);
@@ -316,15 +316,15 @@ class TabletGroupWatcher extends Daemon {
       }
     }
   }
-  
+
   private void repairMetadata(Text row) {
     Master.log.debug("Attempting repair on " + row);
     // ACCUMULO-2261 if a dying tserver writes a location before its lock information propagates, it may cause duplicate assignment.
     // Attempt to find the dead server entry and remove it.
     try {
-      Map<Key, Value> future = new HashMap<Key, Value>();
-      Map<Key, Value> assigned = new HashMap<Key, Value>();
-      KeyExtent extent = new KeyExtent(row, new Value(new byte[]{0}));
+      Map<Key,Value> future = new HashMap<Key,Value>();
+      Map<Key,Value> assigned = new HashMap<Key,Value>();
+      KeyExtent extent = new KeyExtent(row, new Value(new byte[] {0}));
       String table = MetadataTable.NAME;
       if (extent.isMeta())
         table = RootTable.NAME;
@@ -349,9 +349,9 @@ class TabletGroupWatcher extends Daemon {
         Master.log.info("Attempted a repair, but nothing seems to be obviously wrong. " + assigned + " " + future);
         return;
       }
-      Iterator<Entry<Key, Value>> iter = Iterators.concat(future.entrySet().iterator(), assigned.entrySet().iterator());
+      Iterator<Entry<Key,Value>> iter = Iterators.concat(future.entrySet().iterator(), assigned.entrySet().iterator());
       while (iter.hasNext()) {
-        Entry<Key, Value> entry = iter.next();
+        Entry<Key,Value> entry = iter.next();
         TServerInstance alive = master.tserverSet.find(entry.getValue().toString());
         if (alive == null) {
           Master.log.info("Removing entry " + entry);
@@ -376,7 +376,7 @@ class TabletGroupWatcher extends Daemon {
     }
     return result;
   }
-  
+
   private void sendSplitRequest(MergeInfo info, TabletState state, TabletLocationState tls) {
     // Already split?
     if (!info.getState().equals(MergeState.SPLITTING))
@@ -416,7 +416,7 @@ class TabletGroupWatcher extends Daemon {
       }
     }
   }
-  
+
   private void sendChopRequest(MergeInfo info, TabletState state, TabletLocationState tls) {
     // Don't bother if we're in the wrong state
     if (!info.getState().equals(MergeState.WAITING_FOR_CHOPPED))
@@ -443,7 +443,7 @@ class TabletGroupWatcher extends Daemon {
       }
     }
   }
-  
+
   private void updateMergeState(Map<Text,MergeStats> mergeStatsCache) {
     for (MergeStats stats : mergeStatsCache.values()) {
       try {
@@ -457,7 +457,7 @@ class TabletGroupWatcher extends Daemon {
         if (update != stats.getMergeInfo().getState()) {
           this.master.setMergeState(stats.getMergeInfo(), update);
         }
-        
+
         if (update == MergeState.MERGING) {
           try {
             if (stats.getMergeInfo().isDelete()) {
@@ -475,7 +475,7 @@ class TabletGroupWatcher extends Daemon {
       }
     }
   }
-  
+
   private void deleteTablets(MergeInfo info) throws AccumuloException {
     KeyExtent extent = info.getExtent();
     String targetSystemTable = extent.isMeta() ? RootTable.NAME : MetadataTable.NAME;
@@ -537,7 +537,7 @@ class TabletGroupWatcher extends Daemon {
       } finally {
         bw.close();
       }
-      
+
       if (followingTablet != null) {
         Master.log.debug("Updating prevRow of " + followingTablet + " to " + extent.getPrevEndRow());
         bw = conn.createBatchWriter(targetSystemTable, new BatchWriterConfig());
@@ -553,15 +553,15 @@ class TabletGroupWatcher extends Daemon {
       } else {
         // Recreate the default tablet to hold the end of the table
         Master.log.debug("Recreating the last tablet to point to " + extent.getPrevEndRow());
-        String tdir = master.getFileSystem().choose(Optional.of(extent.getTableId().toString()), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR + extent.getTableId()
-            + Constants.DEFAULT_TABLET_LOCATION;
+        String tdir = master.getFileSystem().choose(Optional.of(extent.getTableId().toString()), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR
+            + Path.SEPARATOR + extent.getTableId() + Constants.DEFAULT_TABLET_LOCATION;
         MetadataTableUtil.addTablet(new KeyExtent(extent.getTableId(), null, extent.getPrevEndRow()), tdir, master, timeType, this.master.masterLock);
       }
     } catch (Exception ex) {
       throw new AccumuloException(ex);
     }
   }
-  
+
   private void mergeMetadataRecords(MergeInfo info) throws AccumuloException {
     KeyExtent range = info.getExtent();
     Master.log.debug("Merging metadata for " + range);
@@ -578,7 +578,7 @@ class TabletGroupWatcher extends Daemon {
     if (range.isMeta()) {
       targetSystemTable = RootTable.NAME;
     }
-    
+
     BatchWriter bw = null;
     try {
       long fileCount = 0;
@@ -608,7 +608,7 @@ class TabletGroupWatcher extends Daemon {
           bw.addMutation(MetadataTableUtil.createDeleteMutation(range.getTableId().toString(), entry.getValue().toString()));
         }
       }
-      
+
       // read the logical time from the last tablet in the merge range, it is not included in
       // the loop above
       scanner = conn.createScanner(targetSystemTable, Authorizations.EMPTY);
@@ -619,37 +619,37 @@ class TabletGroupWatcher extends Daemon {
           maxLogicalTime = TabletTime.maxMetadataTime(maxLogicalTime, entry.getValue().toString());
         }
       }
-      
+
       if (maxLogicalTime != null)
         TabletsSection.ServerColumnFamily.TIME_COLUMN.put(m, new Value(maxLogicalTime.getBytes()));
-      
+
       if (!m.getUpdates().isEmpty()) {
         bw.addMutation(m);
       }
-      
+
       bw.flush();
-      
+
       Master.log.debug("Moved " + fileCount + " files to " + stop);
-      
+
       if (firstPrevRowValue == null) {
         Master.log.debug("tablet already merged");
         return;
       }
-      
+
       stop.setPrevEndRow(KeyExtent.decodePrevEndRow(firstPrevRowValue));
       Mutation updatePrevRow = stop.getPrevRowUpdateMutation();
       Master.log.debug("Setting the prevRow for last tablet: " + stop);
       bw.addMutation(updatePrevRow);
       bw.flush();
-      
+
       deleteTablets(info, scanRange, bw, conn);
-      
+
       // Clean-up the last chopped marker
       m = new Mutation(stopRow);
       ChoppedColumnFamily.CHOPPED_COLUMN.putDelete(m);
       bw.addMutation(m);
       bw.flush();
-      
+
     } catch (Exception ex) {
       throw new AccumuloException(ex);
     } finally {
@@ -661,7 +661,7 @@ class TabletGroupWatcher extends Daemon {
         }
     }
   }
-  
+
   private void deleteTablets(MergeInfo info, Range scanRange, BatchWriter bw, Connector conn) throws TableNotFoundException, MutationsRejectedException {
     Scanner scanner;
     Mutation m;
@@ -679,19 +679,19 @@ class TabletGroupWatcher extends Daemon {
       while (row.hasNext()) {
         Entry<Key,Value> entry = row.next();
         Key key = entry.getKey();
-        
+
         if (m == null)
           m = new Mutation(key.getRow());
-        
+
         m.putDelete(key.getColumnFamily(), key.getColumnQualifier());
         Master.log.debug("deleting entry " + key);
       }
       bw.addMutation(m);
     }
-    
+
     bw.flush();
   }
-  
+
   private KeyExtent getHighTablet(KeyExtent range) throws AccumuloException {
     try {
       Connector conn = this.master.getConnector();
@@ -713,7 +713,7 @@ class TabletGroupWatcher extends Daemon {
       throw new AccumuloException("Unexpected failure finding the last tablet for a merge " + range, ex);
     }
   }
-  
+
   private void flushChanges(SortedMap<TServerInstance,TabletServerStatus> currentTServers, List<Assignment> assignments, List<Assignment> assigned,
       List<TabletLocationState> assignedToDeadServers, Map<KeyExtent,TServerInstance> unassigned) throws DistributedStoreException, TException {
     if (!assignedToDeadServers.isEmpty()) {
@@ -722,7 +722,7 @@ class TabletGroupWatcher extends Daemon {
       store.unassign(assignedToDeadServers);
       this.master.nextEvent.event("Marked %d tablets as unassigned because they don't have current servers", assignedToDeadServers.size());
     }
-    
+
     if (!currentTServers.isEmpty()) {
       Map<KeyExtent,TServerInstance> assignedOut = new HashMap<KeyExtent,TServerInstance>();
       final StringBuilder builder = new StringBuilder(64);
@@ -764,7 +764,7 @@ class TabletGroupWatcher extends Daemon {
       if (!unassigned.isEmpty() && assignedOut.isEmpty())
         Master.log.warn("Load balancer failed to assign any tablets");
     }
-    
+
     if (assignments.size() > 0) {
       Master.log.info(String.format("Assigning %d tablets", assignments.size()));
       store.setFutureLocations(assignments);
@@ -780,5 +780,5 @@ class TabletGroupWatcher extends Daemon {
       master.assignedTablet(a.tablet);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/metrics/ReplicationMetricsMBean.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/metrics/ReplicationMetricsMBean.java b/server/master/src/main/java/org/apache/accumulo/master/metrics/ReplicationMetricsMBean.java
index 84f8142..4d19126 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/metrics/ReplicationMetricsMBean.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/metrics/ReplicationMetricsMBean.java
@@ -17,32 +17,32 @@
 package org.apache.accumulo.master.metrics;
 
 /**
- * 
+ *
  */
 public interface ReplicationMetricsMBean {
-  
+
   /**
    * A system may have multiple Replication targets, each of which have a queue of files to be replicated. This returns the sum across all targets, not
    * de-duplicating files.
-   * 
+   *
    * @return The number of files pending replication across all targets
    */
   public int getNumFilesPendingReplication();
-  
+
   /**
    * The total number of threads available to replicate data to peers. Each TabletServer has a number of threads devoted to replication, so this value is
    * affected by the number of currently active TabletServers.
-   * 
+   *
    * @return The number of threads available to replicate data across the instance
    */
   public int getMaxReplicationThreads();
-  
+
   /**
    * Peers are systems which data can be replicated to. This is the number of peers that are defined, but this is not necessarily the number of peers which are
    * actively being replicated to.
-   * 
+   *
    * @return The number of peers/targets which are defined for data to be replicated to.
    */
   public int getNumConfiguredPeers();
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/recovery/RecoveryManager.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/recovery/RecoveryManager.java b/server/master/src/main/java/org/apache/accumulo/master/recovery/RecoveryManager.java
index c636dbb..b9211d2 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/recovery/RecoveryManager.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/recovery/RecoveryManager.java
@@ -115,8 +115,8 @@ public class RecoveryManager {
 
   }
 
-  private void initiateSort(String sortId, String source, final String destination, AccumuloConfiguration aconf)
-    throws KeeperException, InterruptedException, IOException {
+  private void initiateSort(String sortId, String source, final String destination, AccumuloConfiguration aconf) throws KeeperException, InterruptedException,
+      IOException {
     String work = source + "|" + destination;
     new DistributedWorkQueue(ZooUtil.getRoot(master.getInstance()) + Constants.ZRECOVERY, aconf).addWork(sortId, work.getBytes(UTF_8));
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/replication/MasterReplicationCoordinator.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/replication/MasterReplicationCoordinator.java b/server/master/src/main/java/org/apache/accumulo/master/replication/MasterReplicationCoordinator.java
index 73131c7..be8a264 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/replication/MasterReplicationCoordinator.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/replication/MasterReplicationCoordinator.java
@@ -65,10 +65,9 @@ public class MasterReplicationCoordinator implements ReplicationCoordinator.Ifac
     this.security = SecurityOperation.getInstance(master, false);
   }
 
-
   @Override
   public String getServicerAddress(String remoteTableId, TCredentials creds) throws ReplicationCoordinatorException, TException {
-    try { 
+    try {
       security.authenticateUser(master.rpcCreds(), creds);
     } catch (ThriftSecurityException e) {
       log.error("{} failed to authenticate for replication to {}", creds.getPrincipal(), remoteTableId);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/replication/SequentialWorkAssigner.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/replication/SequentialWorkAssigner.java b/server/master/src/main/java/org/apache/accumulo/master/replication/SequentialWorkAssigner.java
index 4b2936c..e30e9ac 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/replication/SequentialWorkAssigner.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/replication/SequentialWorkAssigner.java
@@ -36,11 +36,9 @@ import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
 /**
- * Creates work in ZK which is <code>filename.serialized_ReplicationTarget => filename</code>, but replicates
- * files in the order in which they were created.
+ * Creates work in ZK which is <code>filename.serialized_ReplicationTarget => filename</code>, but replicates files in the order in which they were created.
  * <p>
- * The intent is to ensure that WALs are replayed in the same order on the peer in which
- * they were applied on the primary.
+ * The intent is to ensure that WALs are replayed in the same order on the peer in which they were applied on the primary.
  */
 public class SequentialWorkAssigner extends DistributedWorkQueueWorkAssigner {
   private static final Logger log = LoggerFactory.getLogger(SequentialWorkAssigner.class);
@@ -48,7 +46,7 @@ public class SequentialWorkAssigner extends DistributedWorkQueueWorkAssigner {
 
   // @formatter:off
   /*
-   * { 
+   * {
    *    peer1 => {sourceTableId1 => work_queue_key1, sourceTableId2 => work_queue_key2, ...}
    *    peer2 => {sourceTableId1 => work_queue_key1, sourceTableId3 => work_queue_key4, ...}
    *    ...

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/replication/WorkDriver.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/replication/WorkDriver.java b/server/master/src/main/java/org/apache/accumulo/master/replication/WorkDriver.java
index fbc9c80..3558d2d 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/replication/WorkDriver.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/replication/WorkDriver.java
@@ -63,7 +63,7 @@ public class WorkDriver extends Daemon {
         log.error("Could not instantiate configured work assigner {}", workAssignerClass, e);
         throw new RuntimeException(e);
       }
-  
+
       this.assigner.configure(conf, conn);
       this.assignerImplName = assigner.getClass().getName();
       this.setName(assigner.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/state/MergeStats.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/state/MergeStats.java b/server/master/src/main/java/org/apache/accumulo/master/state/MergeStats.java
index 4737b6e..8cdaf9f 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/state/MergeStats.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/state/MergeStats.java
@@ -56,7 +56,7 @@ public class MergeStats {
   int total = 0;
   boolean lowerSplit = false;
   boolean upperSplit = false;
-  
+
   public MergeStats(MergeInfo info) {
     this.info = info;
     if (info.getState().equals(MergeState.NONE))
@@ -66,11 +66,11 @@ public class MergeStats {
     if (info.getExtent().getPrevEndRow() == null)
       lowerSplit = true;
   }
-  
+
   public MergeInfo getMergeInfo() {
     return info;
   }
-  
+
   public void update(KeyExtent ke, TabletState state, boolean chopped, boolean hasWALs) {
     if (info.getState().equals(MergeState.NONE))
       return;
@@ -100,7 +100,7 @@ public class MergeStats {
     if (state.equals(TabletState.UNASSIGNED))
       this.unassigned++;
   }
-  
+
   public MergeState nextMergeState(Connector connector, CurrentState master) throws Exception {
     MergeState state = info.getState();
     if (state == MergeState.NONE)
@@ -173,7 +173,7 @@ public class MergeStats {
     }
     return state;
   }
-  
+
   private boolean verifyMergeConsistency(Connector connector, CurrentState master) throws TableNotFoundException, IOException {
     MergeStats verify = new MergeStats(info);
     KeyExtent extent = info.getExtent();
@@ -188,7 +188,7 @@ public class MergeStats {
     Range range = new Range(first, false, null, true);
     scanner.setRange(range);
     KeyExtent prevExtent = null;
-    
+
     log.debug("Scanning range " + range);
     for (Entry<Key,Value> entry : scanner) {
       TabletLocationState tls;
@@ -202,31 +202,31 @@ public class MergeStats {
       if (!tls.extent.getTableId().equals(tableId)) {
         break;
       }
-      
+
       if (!tls.walogs.isEmpty() && verify.getMergeInfo().needsToBeChopped(tls.extent)) {
         log.debug("failing consistency: needs to be chopped" + tls.extent);
         return false;
       }
-      
+
       if (prevExtent == null) {
         // this is the first tablet observed, it must be offline and its prev row must be less than the start of the merge range
         if (tls.extent.getPrevEndRow() != null && tls.extent.getPrevEndRow().compareTo(start) > 0) {
           log.debug("failing consistency: prev row is too high " + start);
           return false;
         }
-        
+
         if (tls.getState(master.onlineTabletServers()) != TabletState.UNASSIGNED) {
           log.debug("failing consistency: assigned or hosted " + tls);
           return false;
         }
-        
+
       } else if (!tls.extent.isPreviousExtent(prevExtent)) {
         log.debug("hole in " + MetadataTable.NAME);
         return false;
       }
-      
+
       prevExtent = tls.extent;
-      
+
       verify.update(tls.extent, tls.getState(master.onlineTabletServers()), tls.chopped, !tls.walogs.isEmpty());
       // stop when we've seen the tablet just beyond our range
       if (tls.extent.getPrevEndRow() != null && extent.getEndRow() != null && tls.extent.getPrevEndRow().compareTo(extent.getEndRow()) > 0) {
@@ -237,11 +237,11 @@ public class MergeStats {
         + verify.total);
     return chopped == verify.chopped && unassigned == verify.unassigned && unassigned == verify.total;
   }
-  
+
   public static void main(String[] args) throws Exception {
     ClientOpts opts = new ClientOpts();
     opts.parseArgs(MergeStats.class.getName(), args);
-    
+
     Connector conn = opts.getConnector();
     Map<String,String> tableIdMap = conn.tableOperations().tableIdMap();
     for (Entry<String,String> entry : tableIdMap.entrySet()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/state/SetGoalState.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/state/SetGoalState.java b/server/master/src/main/java/org/apache/accumulo/master/state/SetGoalState.java
index bd65163..3442171 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/state/SetGoalState.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/state/SetGoalState.java
@@ -31,7 +31,7 @@ import org.apache.accumulo.server.security.SecurityUtil;
 import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 
 public class SetGoalState {
-  
+
   /**
    * Utility program that will change the goal state for the master from the command line.
    */
@@ -47,5 +47,5 @@ public class SetGoalState {
     ZooReaderWriter.getInstance().putPersistentData(ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZMASTER_GOAL_STATE, args[0].getBytes(UTF_8),
         NodeExistsPolicy.OVERWRITE);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/state/TableCounts.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/state/TableCounts.java b/server/master/src/main/java/org/apache/accumulo/master/state/TableCounts.java
index 4ebd745..73395ea 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/state/TableCounts.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/state/TableCounts.java
@@ -20,19 +20,19 @@ import org.apache.accumulo.server.master.state.TabletState;
 
 public class TableCounts {
   int counts[] = new int[TabletState.values().length];
-  
+
   public int unassigned() {
     return counts[TabletState.UNASSIGNED.ordinal()];
   }
-  
+
   public int assigned() {
     return counts[TabletState.ASSIGNED.ordinal()];
   }
-  
+
   public int assignedToDeadServers() {
     return counts[TabletState.ASSIGNED_TO_DEAD_SERVER.ordinal()];
   }
-  
+
   public int hosted() {
     return counts[TabletState.HOSTED.ordinal()];
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/state/TableStats.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/state/TableStats.java b/server/master/src/main/java/org/apache/accumulo/master/state/TableStats.java
index f088a5d..127406c 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/state/TableStats.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/state/TableStats.java
@@ -27,12 +27,12 @@ public class TableStats {
   private Map<Text,TableCounts> next;
   private long startScan = 0;
   private long endScan = 0;
-  
+
   public synchronized void begin() {
     next = new HashMap<Text,TableCounts>();
     startScan = System.currentTimeMillis();
   }
-  
+
   public synchronized void update(Text tableId, TabletState state) {
     TableCounts counts = next.get(tableId);
     if (counts == null) {
@@ -41,30 +41,30 @@ public class TableStats {
     }
     counts.counts[state.ordinal()]++;
   }
-  
+
   public synchronized void end() {
     last = next;
     next = null;
     endScan = System.currentTimeMillis();
   }
-  
+
   public synchronized Map<Text,TableCounts> getLast() {
     return last;
   }
-  
+
   public synchronized TableCounts getLast(Text tableId) {
     TableCounts result = last.get(tableId);
     if (result == null)
       return new TableCounts();
     return result;
   }
-  
+
   public synchronized long getScanTime() {
     if (endScan <= startScan)
       return System.currentTimeMillis() - startScan;
     return endScan - startScan;
   }
-  
+
   public synchronized long lastScanFinished() {
     return endScan;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/BulkImport.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/BulkImport.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/BulkImport.java
index 049c9b3..c663686 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/BulkImport.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/BulkImport.java
@@ -141,7 +141,7 @@ public class BulkImport extends MasterRepo {
   }
 
   @Override
-  //TODO Remove deprecation warning suppression when Hadoop1 support is dropped
+  // TODO Remove deprecation warning suppression when Hadoop1 support is dropped
   @SuppressWarnings("deprecation")
   public Repo<Master> call(long tid, Master master) throws Exception {
     log.debug(" tid " + tid + " sourceDir " + sourceDir);
@@ -213,7 +213,7 @@ public class BulkImport extends MasterRepo {
     }
   }
 
-  //TODO Remove deprecation warning suppression when Hadoop1 support is dropped
+  // TODO Remove deprecation warning suppression when Hadoop1 support is dropped
   @SuppressWarnings("deprecation")
   private String prepareBulkImport(Master master, final VolumeManager fs, String dir, String tableId) throws Exception {
     final Path bulkDir = createNewBulkDir(fs, tableId);
@@ -288,7 +288,7 @@ public class BulkImport extends MasterRepo {
       }));
     }
     workers.shutdown();
-    while (!workers.awaitTermination(1000L, TimeUnit.MILLISECONDS)) { }
+    while (!workers.awaitTermination(1000L, TimeUnit.MILLISECONDS)) {}
 
     for (Future<Exception> ex : results) {
       if (ex.get() != null) {
@@ -456,8 +456,8 @@ class CopyFailed extends MasterRepo {
     }
 
     if (loadedFailures.size() > 0) {
-      DistributedWorkQueue bifCopyQueue = new DistributedWorkQueue(Constants.ZROOT + "/" + master.getInstance().getInstanceID()
-          + Constants.ZBULK_FAILED_COPYQ, master.getConfiguration());
+      DistributedWorkQueue bifCopyQueue = new DistributedWorkQueue(Constants.ZROOT + "/" + master.getInstance().getInstanceID() + Constants.ZBULK_FAILED_COPYQ,
+          master.getConfiguration());
 
       HashSet<String> workIds = new HashSet<String>();
 
@@ -575,8 +575,7 @@ class LoadFiles extends MasterRepo {
               server = pair.getFirst();
               List<String> attempt = Collections.singletonList(file);
               log.debug("Asking " + pair.getFirst() + " to bulk import " + file);
-              List<String> fail = client.bulkImportFiles(Tracer.traceInfo(), master.rpcCreds(), tid, tableId, attempt,
-                  errorDir, setTime);
+              List<String> fail = client.bulkImportFiles(Tracer.traceInfo(), master.rpcCreds(), tid, tableId, attempt, errorDir, setTime);
               if (fail.isEmpty()) {
                 loaded.add(file);
               } else {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/CloneTable.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/CloneTable.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/CloneTable.java
index 7034e39..f1cf35c 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/CloneTable.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/CloneTable.java
@@ -195,8 +195,8 @@ class ClonePermissions extends MasterRepo {
     // give all table permissions to the creator
     for (TablePermission permission : TablePermission.values()) {
       try {
-        AuditedSecurityOperation.getInstance(environment).grantTablePermission(environment.rpcCreds(), cloneInfo.user,
-            cloneInfo.tableId, permission, cloneInfo.namespaceId);
+        AuditedSecurityOperation.getInstance(environment).grantTablePermission(environment.rpcCreds(), cloneInfo.user, cloneInfo.tableId, permission,
+            cloneInfo.namespaceId);
       } catch (ThriftSecurityException e) {
         Logger.getLogger(FinishCloneTable.class).error(e.getMessage(), e);
         throw e;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/CompactRange.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/CompactRange.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/CompactRange.java
index db8bbfe..fd7decf 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/CompactRange.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/CompactRange.java
@@ -81,8 +81,7 @@ class CompactionDriver extends MasterRepo {
   @Override
   public long isReady(long tid, Master master) throws Exception {
 
-    String zCancelID = Constants.ZROOT + "/" + master.getInstance().getInstanceID() + Constants.ZTABLES + "/" + tableId
-        + Constants.ZTABLE_COMPACT_CANCEL_ID;
+    String zCancelID = Constants.ZROOT + "/" + master.getInstance().getInstanceID() + Constants.ZTABLES + "/" + tableId + Constants.ZTABLE_COMPACT_CANCEL_ID;
 
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
 
@@ -206,7 +205,6 @@ public class CompactRange extends MasterRepo {
   private byte[] endRow;
   private byte[] config;
 
-
   public CompactRange(String tableId, byte[] startRow, byte[] endRow, List<IteratorSetting> iterators, CompactionStrategyConfig compactionStrategy)
       throws ThriftTableOperationException {
     this.tableId = tableId;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/CreateTable.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/CreateTable.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/CreateTable.java
index 95c9f79..103eef8 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/CreateTable.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/CreateTable.java
@@ -177,8 +177,8 @@ class ChooseDir extends MasterRepo {
   @Override
   public Repo<Master> call(long tid, Master master) throws Exception {
     // Constants.DEFAULT_TABLET_LOCATION has a leading slash prepended to it so we don't need to add one here
-    tableInfo.dir = master.getFileSystem().choose(Optional.of(tableInfo.tableId), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR + tableInfo.tableId
-        + Constants.DEFAULT_TABLET_LOCATION;
+    tableInfo.dir = master.getFileSystem().choose(Optional.of(tableInfo.tableId), ServerConstants.getBaseUris()) + Constants.HDFS_TABLES_DIR + Path.SEPARATOR
+        + tableInfo.tableId + Constants.DEFAULT_TABLET_LOCATION;
     return new CreateDir(tableInfo);
   }
 


[19/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/SecurityUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/SecurityUtil.java b/server/base/src/main/java/org/apache/accumulo/server/security/SecurityUtil.java
index 29e4939..42d1313 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/SecurityUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/SecurityUtil.java
@@ -25,7 +25,7 @@ import org.apache.hadoop.security.UserGroupInformation;
 import org.apache.log4j.Logger;
 
 /**
- * 
+ *
  */
 public class SecurityUtil {
   private static final Logger log = Logger.getLogger(SecurityUtil.class);
@@ -39,13 +39,13 @@ public class SecurityUtil {
     String keyTab = acuConf.getPath(Property.GENERAL_KERBEROS_KEYTAB);
     if (keyTab == null || keyTab.length() == 0)
       return;
-    
+
     usingKerberos = true;
-    
+
     String principalConfig = acuConf.get(Property.GENERAL_KERBEROS_PRINCIPAL);
     if (principalConfig == null || principalConfig.length() == 0)
       return;
-    
+
     if (login(principalConfig, keyTab)) {
       try {
         // This spawns a thread to periodically renew the logged in (accumulo) user
@@ -58,10 +58,10 @@ public class SecurityUtil {
 
     throw new RuntimeException("Failed to perform Kerberos login for " + principalConfig + " using  " + keyTab);
   }
-  
+
   /**
    * This will log in the given user in kerberos.
-   * 
+   *
    * @param principalConfig
    *          This is the principals name in the format NAME/HOST@REALM. {@link org.apache.hadoop.security.SecurityUtil#HOSTNAME_PATTERN} will automatically be
    *          replaced by the systems host name.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/SystemCredentials.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/SystemCredentials.java b/server/base/src/main/java/org/apache/accumulo/server/security/SystemCredentials.java
index a59d57c..79201b1 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/SystemCredentials.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/SystemCredentials.java
@@ -40,7 +40,7 @@ import org.apache.hadoop.io.Writable;
 
 /**
  * Credentials for the system services.
- * 
+ *
  * @since 1.6.0
  */
 public final class SystemCredentials extends Credentials {
@@ -77,7 +77,7 @@ public final class SystemCredentials extends Credentials {
 
   /**
    * An {@link AuthenticationToken} type for Accumulo servers for inter-server communication.
-   * 
+   *
    * @since 1.6.0
    */
   public static final class SystemToken extends PasswordToken {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authenticator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authenticator.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authenticator.java
index 5f0da82..092b92d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authenticator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authenticator.java
@@ -29,27 +29,27 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
  */
 
 public interface Authenticator {
-  
+
   void initialize(String instanceId, boolean initialize);
-  
+
   boolean validSecurityHandlers(Authorizor auth, PermissionHandler pm);
-  
+
   void initializeSecurity(TCredentials credentials, String principal, byte[] token) throws AccumuloSecurityException, ThriftSecurityException;
-  
+
   boolean authenticateUser(String principal, AuthenticationToken token) throws AccumuloSecurityException;
-  
+
   Set<String> listUsers() throws AccumuloSecurityException;
-  
+
   void createUser(String principal, AuthenticationToken token) throws AccumuloSecurityException;
-  
+
   void dropUser(String user) throws AccumuloSecurityException;
-  
+
   void changePassword(String principal, AuthenticationToken token) throws AccumuloSecurityException;
-  
+
   boolean userExists(String user) throws AccumuloSecurityException;
-  
+
   Set<Class<? extends AuthenticationToken>> getSupportedTokenTypes();
-  
+
   /**
    * Returns true if the given token is appropriate for this Authenticator
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authorizor.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authorizor.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authorizor.java
index 5131bd3..abb579d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authorizor.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/Authorizor.java
@@ -29,27 +29,27 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
  * Accumulo, it should throw an AccumuloSecurityException with the error code UNSUPPORTED_OPERATION
  */
 public interface Authorizor {
-  
+
   /**
    * Sets up the authorizor for a new instance of Accumulo
    */
   void initialize(String instanceId, boolean initialize);
-  
+
   /**
    * Used to validate that the Authorizor, Authenticator, and permission handler can coexist
    */
   boolean validSecurityHandlers(Authenticator auth, PermissionHandler pm);
-  
+
   /**
    * Used to initialize security for the root user
    */
   void initializeSecurity(TCredentials credentials, String rootuser) throws AccumuloSecurityException, ThriftSecurityException;
-  
+
   /**
    * Used to change the authorizations for the user
    */
   void changeAuthorizations(String user, Authorizations authorizations) throws AccumuloSecurityException;
-  
+
   /**
    * Used to get the authorizations for the user
    */
@@ -59,12 +59,12 @@ public interface Authorizor {
    * Used to check if a user has valid auths.
    */
   boolean isValidAuthorizations(String user, List<ByteBuffer> list) throws AccumuloSecurityException;
-  
+
   /**
    * Initializes a new user
    */
   void initUser(String user) throws AccumuloSecurityException;
-  
+
   /**
    * Deletes a user
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecureAuthenticator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecureAuthenticator.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecureAuthenticator.java
index 5b374d9..a57608c 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecureAuthenticator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecureAuthenticator.java
@@ -30,47 +30,42 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
  * primarily for testing, but can also be used for any system where user space management is not a concern.
  */
 public class InsecureAuthenticator implements Authenticator {
-  
+
   @Override
-  public void initialize(String instanceId, boolean initialize) {
-  }
-  
+  public void initialize(String instanceId, boolean initialize) {}
+
   @Override
   public boolean validSecurityHandlers(Authorizor auth, PermissionHandler pm) {
     return true;
   }
-  
+
   @Override
-  public void initializeSecurity(TCredentials credentials, String principal, byte[] token) throws AccumuloSecurityException {
-  }
-  
+  public void initializeSecurity(TCredentials credentials, String principal, byte[] token) throws AccumuloSecurityException {}
+
   @Override
   public boolean authenticateUser(String principal, AuthenticationToken token) {
     return token instanceof NullToken;
   }
-  
+
   @Override
   public Set<String> listUsers() throws AccumuloSecurityException {
     return Collections.emptySet();
   }
-  
+
   @Override
-  public void createUser(String principal, AuthenticationToken token) throws AccumuloSecurityException {
-  }
-  
+  public void createUser(String principal, AuthenticationToken token) throws AccumuloSecurityException {}
+
   @Override
-  public void dropUser(String user) throws AccumuloSecurityException {
-  }
-  
+  public void dropUser(String user) throws AccumuloSecurityException {}
+
   @Override
-  public void changePassword(String user, AuthenticationToken token) throws AccumuloSecurityException {
-  }
-  
+  public void changePassword(String user, AuthenticationToken token) throws AccumuloSecurityException {}
+
   @Override
   public boolean userExists(String user) {
     return true;
   }
-  
+
   @Override
   public boolean validTokenClass(String tokenClass) {
     return tokenClass.equals(NullToken.class.getName());
@@ -82,5 +77,5 @@ public class InsecureAuthenticator implements Authenticator {
     cs.add(NullToken.class);
     return cs;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecurePermHandler.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecurePermHandler.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecurePermHandler.java
index dfe3c88..b4e5e25 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecurePermHandler.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/InsecurePermHandler.java
@@ -19,8 +19,8 @@ package org.apache.accumulo.server.security.handler;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.NamespaceNotFoundException;
 import org.apache.accumulo.core.client.TableNotFoundException;
-import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.NamespacePermission;
+import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
 import org.apache.accumulo.core.security.thrift.TCredentials;
 
@@ -28,68 +28,59 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
  * This is a Permission Handler implementation that doesn't actually do any security. Use at your own risk.
  */
 public class InsecurePermHandler implements PermissionHandler {
-  
+
   @Override
-  public void initialize(String instanceId, boolean initialize) {
-  }
-  
+  public void initialize(String instanceId, boolean initialize) {}
+
   @Override
   public boolean validSecurityHandlers(Authenticator authent, Authorizor author) {
     return true;
   }
-  
+
   @Override
-  public void initializeSecurity(TCredentials token, String rootuser) throws AccumuloSecurityException {
-  }
-  
+  public void initializeSecurity(TCredentials token, String rootuser) throws AccumuloSecurityException {}
+
   @Override
   public boolean hasSystemPermission(String user, SystemPermission permission) throws AccumuloSecurityException {
     return true;
   }
-  
+
   @Override
   public boolean hasCachedSystemPermission(String user, SystemPermission permission) throws AccumuloSecurityException {
     return true;
   }
-  
+
   @Override
   public boolean hasTablePermission(String user, String table, TablePermission permission) throws AccumuloSecurityException, TableNotFoundException {
     return true;
   }
-  
+
   @Override
   public boolean hasCachedTablePermission(String user, String table, TablePermission permission) throws AccumuloSecurityException, TableNotFoundException {
     return true;
   }
-  
+
   @Override
-  public void grantSystemPermission(String user, SystemPermission permission) throws AccumuloSecurityException {
-  }
-  
+  public void grantSystemPermission(String user, SystemPermission permission) throws AccumuloSecurityException {}
+
   @Override
-  public void revokeSystemPermission(String user, SystemPermission permission) throws AccumuloSecurityException {
-  }
-  
+  public void revokeSystemPermission(String user, SystemPermission permission) throws AccumuloSecurityException {}
+
   @Override
-  public void grantTablePermission(String user, String table, TablePermission permission) throws AccumuloSecurityException, TableNotFoundException {
-  }
-  
+  public void grantTablePermission(String user, String table, TablePermission permission) throws AccumuloSecurityException, TableNotFoundException {}
+
   @Override
-  public void revokeTablePermission(String user, String table, TablePermission permission) throws AccumuloSecurityException, TableNotFoundException {
-  }
-  
+  public void revokeTablePermission(String user, String table, TablePermission permission) throws AccumuloSecurityException, TableNotFoundException {}
+
   @Override
-  public void cleanTablePermissions(String table) throws AccumuloSecurityException, TableNotFoundException {
-  }
-  
+  public void cleanTablePermissions(String table) throws AccumuloSecurityException, TableNotFoundException {}
+
   @Override
-  public void initUser(String user) throws AccumuloSecurityException {
-  }
-  
+  public void initUser(String user) throws AccumuloSecurityException {}
+
   @Override
-  public void cleanUser(String user) throws AccumuloSecurityException {
-  }
-  
+  public void cleanUser(String user) throws AccumuloSecurityException {}
+
   @Override
   public void initTable(String table) throws AccumuloSecurityException {}
 
@@ -107,16 +98,13 @@ public class InsecurePermHandler implements PermissionHandler {
 
   @Override
   public void grantNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException,
-      NamespaceNotFoundException {
-  }
+      NamespaceNotFoundException {}
 
   @Override
   public void revokeNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException,
-      NamespaceNotFoundException {
-  }
+      NamespaceNotFoundException {}
 
   @Override
-  public void cleanNamespacePermissions(String namespace) throws AccumuloSecurityException, NamespaceNotFoundException {
-  }
-  
+  public void cleanNamespacePermissions(String namespace) throws AccumuloSecurityException, NamespaceNotFoundException {}
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/PermissionHandler.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/PermissionHandler.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/PermissionHandler.java
index 914bab3..117aa77 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/PermissionHandler.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/PermissionHandler.java
@@ -70,8 +70,7 @@ public interface PermissionHandler {
   /**
    * Used to get the namespace permission of a user for a namespace
    */
-  boolean hasNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException,
-      NamespaceNotFoundException;
+  boolean hasNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException, NamespaceNotFoundException;
 
   /**
    * Used to get the namespace permission of a user for a namespace, with caching. This method is for high frequency operations
@@ -102,14 +101,12 @@ public interface PermissionHandler {
   /**
    * Gives the user the given namespace permission
    */
-  void grantNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException,
-      NamespaceNotFoundException;
+  void grantNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException, NamespaceNotFoundException;
 
   /**
    * Denies the user the given namespace permission.
    */
-  void revokeNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException,
-      NamespaceNotFoundException;
+  void revokeNamespacePermission(String user, String namespace, NamespacePermission permission) throws AccumuloSecurityException, NamespaceNotFoundException;
 
   /**
    * Cleans up the permissions for a table. Used when a table gets deleted.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKPermHandler.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKPermHandler.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKPermHandler.java
index c03c40b..f032aa2 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKPermHandler.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKPermHandler.java
@@ -217,8 +217,7 @@ public class ZKPermHandler implements PermissionHandler {
         synchronized (zooCache) {
           zooCache.clear(ZKUserPath + "/" + user + ZKUserTablePerms + "/" + table);
           ZooReaderWriter.getInstance().putPersistentData(ZKUserPath + "/" + user + ZKUserTablePerms + "/" + table,
-              ZKSecurityTool.convertTablePermissions(tablePerms),
-              NodeExistsPolicy.OVERWRITE);
+              ZKSecurityTool.convertTablePermissions(tablePerms), NodeExistsPolicy.OVERWRITE);
         }
       }
     } catch (KeeperException e) {
@@ -244,8 +243,7 @@ public class ZKPermHandler implements PermissionHandler {
         synchronized (zooCache) {
           zooCache.clear(ZKUserPath + "/" + user + ZKUserNamespacePerms + "/" + namespace);
           ZooReaderWriter.getInstance().putPersistentData(ZKUserPath + "/" + user + ZKUserNamespacePerms + "/" + namespace,
-              ZKSecurityTool.convertNamespacePermissions(namespacePerms),
-              NodeExistsPolicy.OVERWRITE);
+              ZKSecurityTool.convertNamespacePermissions(namespacePerms), NodeExistsPolicy.OVERWRITE);
         }
       }
     } catch (KeeperException e) {
@@ -435,8 +433,8 @@ public class ZKPermHandler implements PermissionHandler {
   private void createTablePerm(String user, String table, Set<TablePermission> perms) throws KeeperException, InterruptedException {
     synchronized (zooCache) {
       zooCache.clear();
-      ZooReaderWriter.getInstance().putPersistentData(ZKUserPath + "/" + user + ZKUserTablePerms + "/" + table,
-          ZKSecurityTool.convertTablePermissions(perms), NodeExistsPolicy.FAIL);
+      ZooReaderWriter.getInstance().putPersistentData(ZKUserPath + "/" + user + ZKUserTablePerms + "/" + table, ZKSecurityTool.convertTablePermissions(perms),
+          NodeExistsPolicy.FAIL);
     }
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKSecurityTool.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKSecurityTool.java b/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKSecurityTool.java
index 2ed430c..aec3078 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKSecurityTool.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/handler/ZKSecurityTool.java
@@ -30,8 +30,8 @@ import java.util.Set;
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.security.Authorizations;
-import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.NamespacePermission;
+import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
 import org.apache.log4j.Logger;
 
@@ -43,7 +43,7 @@ import org.apache.log4j.Logger;
 class ZKSecurityTool {
   private static final Logger log = Logger.getLogger(ZKSecurityTool.class);
   private static final int SALT_LENGTH = 8;
-  
+
   // Generates a byte array salt of length SALT_LENGTH
   private static byte[] generateSalt() {
     final SecureRandom random = new SecureRandom();
@@ -51,17 +51,17 @@ class ZKSecurityTool {
     random.nextBytes(salt);
     return salt;
   }
-  
+
   private static byte[] hash(byte[] raw) throws NoSuchAlgorithmException {
     MessageDigest md = MessageDigest.getInstance(Constants.PW_HASH_ALGORITHM);
     md.update(raw);
     return md.digest();
   }
-  
+
   public static boolean checkPass(byte[] password, byte[] zkData) {
     if (zkData == null)
       return false;
-    
+
     byte[] salt = new byte[SALT_LENGTH];
     System.arraycopy(zkData, 0, salt, 0, SALT_LENGTH);
     byte[] passwordToCheck;
@@ -73,7 +73,7 @@ class ZKSecurityTool {
     }
     return java.util.Arrays.equals(passwordToCheck, zkData);
   }
-  
+
   public static byte[] createPass(byte[] password) throws AccumuloException {
     byte[] salt = generateSalt();
     try {
@@ -83,7 +83,7 @@ class ZKSecurityTool {
       throw new AccumuloException("Count not create hashed password", e);
     }
   }
-  
+
   private static byte[] convertPass(byte[] password, byte[] salt) throws NoSuchAlgorithmException {
     byte[] plainSalt = new byte[password.length + SALT_LENGTH];
     System.arraycopy(password, 0, plainSalt, 0, password.length);
@@ -94,15 +94,15 @@ class ZKSecurityTool {
     System.arraycopy(hashed, 0, saltedHash, SALT_LENGTH, hashed.length);
     return saltedHash; // contains salt+hash(password+salt)
   }
-  
+
   public static Authorizations convertAuthorizations(byte[] authorizations) {
     return new Authorizations(authorizations);
   }
-  
+
   public static byte[] convertAuthorizations(Authorizations authorizations) {
     return authorizations.getAuthorizationsArray();
   }
-  
+
   public static byte[] convertSystemPermissions(Set<SystemPermission> systempermissions) {
     ByteArrayOutputStream bytes = new ByteArrayOutputStream(systempermissions.size());
     DataOutputStream out = new DataOutputStream(bytes);
@@ -115,7 +115,7 @@ class ZKSecurityTool {
     }
     return bytes.toByteArray();
   }
-  
+
   public static Set<SystemPermission> convertSystemPermissions(byte[] systempermissions) {
     ByteArrayInputStream bytes = new ByteArrayInputStream(systempermissions);
     DataInputStream in = new DataInputStream(bytes);
@@ -129,7 +129,7 @@ class ZKSecurityTool {
     }
     return toReturn;
   }
-  
+
   public static byte[] convertTablePermissions(Set<TablePermission> tablepermissions) {
     ByteArrayOutputStream bytes = new ByteArrayOutputStream(tablepermissions.size());
     DataOutputStream out = new DataOutputStream(bytes);
@@ -142,14 +142,14 @@ class ZKSecurityTool {
     }
     return bytes.toByteArray();
   }
-  
+
   public static Set<TablePermission> convertTablePermissions(byte[] tablepermissions) {
     Set<TablePermission> toReturn = new HashSet<TablePermission>();
     for (byte b : tablepermissions)
       toReturn.add(TablePermission.getPermissionById(b));
     return toReturn;
   }
-  
+
   public static byte[] convertNamespacePermissions(Set<NamespacePermission> namespacepermissions) {
     ByteArrayOutputStream bytes = new ByteArrayOutputStream(namespacepermissions.size());
     DataOutputStream out = new DataOutputStream(bytes);
@@ -162,14 +162,14 @@ class ZKSecurityTool {
     }
     return bytes.toByteArray();
   }
-  
+
   public static Set<NamespacePermission> convertNamespacePermissions(byte[] namespacepermissions) {
     Set<NamespacePermission> toReturn = new HashSet<NamespacePermission>();
     for (byte b : namespacepermissions)
       toReturn.add(NamespacePermission.getPermissionById(b));
     return toReturn;
   }
-  
+
   public static String getInstancePath(String instanceId) {
     return Constants.ZROOT + "/" + instanceId;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tables/TableManager.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tables/TableManager.java b/server/base/src/main/java/org/apache/accumulo/server/tables/TableManager.java
index 886c28f..7e5f54d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tables/TableManager.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tables/TableManager.java
@@ -54,7 +54,7 @@ public class TableManager {
   private static final Set<TableObserver> observers = Collections.synchronizedSet(new HashSet<TableObserver>());
   private static final Map<String,TableState> tableStateCache = Collections.synchronizedMap(new HashMap<String,TableState>());
   private static final byte[] ZERO_BYTE = new byte[] {'0'};
-  
+
   private static TableManager tableManager = null;
 
   private final Instance instance;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tables/TableObserver.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tables/TableObserver.java b/server/base/src/main/java/org/apache/accumulo/server/tables/TableObserver.java
index 80aec3a..de88b96 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tables/TableObserver.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tables/TableObserver.java
@@ -22,8 +22,8 @@ import org.apache.accumulo.core.master.state.tables.TableState;
 
 public interface TableObserver {
   void initialize(Map<String,TableState> tableIdToStateMap);
-  
+
   void stateChanged(String tableId, TableState tState);
-  
+
   void sessionExpired();
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tablets/TabletTime.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tablets/TabletTime.java b/server/base/src/main/java/org/apache/accumulo/server/tablets/TabletTime.java
index e3fd8f3..3dd9e7b 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tablets/TabletTime.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tablets/TabletTime.java
@@ -27,7 +27,7 @@ import org.apache.accumulo.server.util.time.RelativeTime;
 public abstract class TabletTime {
   public static final char LOGICAL_TIME_ID = 'L';
   public static final char MILLIS_TIME_ID = 'M';
-  
+
   public static char getTimeID(TimeType timeType) {
     switch (timeType) {
       case LOGICAL:
@@ -35,115 +35,115 @@ public abstract class TabletTime {
       case MILLIS:
         return MILLIS_TIME_ID;
     }
-    
+
     throw new IllegalArgumentException("Unknown time type " + timeType);
   }
-  
+
   public abstract void useMaxTimeFromWALog(long time);
-  
+
   public abstract String getMetadataValue(long time);
-  
+
   public abstract String getMetadataValue();
-  
+
   public abstract long setUpdateTimes(List<Mutation> mutations);
-  
+
   public abstract long getTime();
-  
+
   public abstract long getAndUpdateTime();
-  
+
   protected void setSystemTimes(Mutation mutation, long lastCommitTime) {
     ServerMutation m = (ServerMutation) mutation;
     m.setSystemTimestamp(lastCommitTime);
   }
-  
+
   public static TabletTime getInstance(String metadataValue) {
     if (metadataValue.charAt(0) == LOGICAL_TIME_ID) {
       return new LogicalTime(Long.parseLong(metadataValue.substring(1)));
     } else if (metadataValue.charAt(0) == MILLIS_TIME_ID) {
       return new MillisTime(Long.parseLong(metadataValue.substring(1)));
     }
-    
+
     throw new IllegalArgumentException("Time type unknown : " + metadataValue);
-    
+
   }
-  
+
   public static String maxMetadataTime(String mv1, String mv2) {
     if (mv1 == null && mv2 == null) {
       return null;
     }
-    
+
     if (mv1 == null) {
       checkType(mv2);
       return mv2;
     }
-    
+
     if (mv2 == null) {
       checkType(mv1);
       return mv1;
     }
-    
+
     if (mv1.charAt(0) != mv2.charAt(0))
       throw new IllegalArgumentException("Time types differ " + mv1 + " " + mv2);
     checkType(mv1);
-    
+
     long t1 = Long.parseLong(mv1.substring(1));
     long t2 = Long.parseLong(mv2.substring(1));
-    
+
     if (t1 < t2)
       return mv2;
     else
       return mv1;
-    
+
   }
-  
+
   private static void checkType(String mv1) {
     if (mv1.charAt(0) != LOGICAL_TIME_ID && mv1.charAt(0) != MILLIS_TIME_ID)
       throw new IllegalArgumentException("Invalid time type " + mv1);
   }
-  
+
   static class MillisTime extends TabletTime {
-    
+
     private long lastTime;
     private long lastUpdateTime = 0;
-    
+
     public MillisTime(long time) {
       this.lastTime = time;
     }
-    
+
     @Override
     public String getMetadataValue(long time) {
       return MILLIS_TIME_ID + "" + time;
     }
-    
+
     @Override
     public String getMetadataValue() {
       return getMetadataValue(lastTime);
     }
-    
+
     @Override
     public void useMaxTimeFromWALog(long time) {
       if (time > lastTime)
         lastTime = time;
     }
-    
+
     @Override
     public long setUpdateTimes(List<Mutation> mutations) {
-      
+
       long currTime = RelativeTime.currentTimeMillis();
-      
+
       synchronized (this) {
         if (mutations.size() == 0)
           return lastTime;
-        
+
         currTime = updateTime(currTime);
       }
-      
+
       for (Mutation mutation : mutations)
         setSystemTimes(mutation, currTime);
-      
+
       return currTime;
     }
-    
+
     private long updateTime(long currTime) {
       if (currTime < lastTime) {
         if (currTime - lastUpdateTime > 0) {
@@ -151,81 +151,81 @@ public abstract class TabletTime {
           // to this method so move ahead slowly
           lastTime++;
         }
-        
+
         lastUpdateTime = currTime;
-        
+
         currTime = lastTime;
       } else {
         lastTime = currTime;
       }
       return currTime;
     }
-    
+
     @Override
     public long getTime() {
       return lastTime;
     }
-    
+
     @Override
     public long getAndUpdateTime() {
       long currTime = RelativeTime.currentTimeMillis();
-      
+
       synchronized (this) {
         currTime = updateTime(currTime);
       }
-      
+
       return currTime;
     }
-    
+
   }
-  
+
   static class LogicalTime extends TabletTime {
     AtomicLong nextTime;
-    
+
     private LogicalTime(Long time) {
       this.nextTime = new AtomicLong(time.longValue() + 1);
     }
-    
+
     @Override
     public void useMaxTimeFromWALog(long time) {
       time++;
-      
+
       if (this.nextTime.get() < time) {
         this.nextTime.set(time);
       }
     }
-    
+
     @Override
     public String getMetadataValue() {
       return getMetadataValue(getTime());
     }
-    
+
     @Override
     public String getMetadataValue(long time) {
       return LOGICAL_TIME_ID + "" + time;
     }
-    
+
     @Override
     public long setUpdateTimes(List<Mutation> mutations) {
       if (mutations.size() == 0)
         return getTime();
-      
+
       long time = nextTime.getAndAdd(mutations.size());
       for (Mutation mutation : mutations)
         setSystemTimes(mutation, time++);
-      
+
       return time - 1;
     }
-    
+
     @Override
     public long getTime() {
       return nextTime.get() - 1;
     }
-    
+
     @Override
     public long getAndUpdateTime() {
       return nextTime.getAndIncrement();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tablets/UniqueNameAllocator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tablets/UniqueNameAllocator.java b/server/base/src/main/java/org/apache/accumulo/server/tablets/UniqueNameAllocator.java
index 8bf4dc8..c2eac02 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tablets/UniqueNameAllocator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tablets/UniqueNameAllocator.java
@@ -28,9 +28,9 @@ import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 
 /**
  * Allocates unique names for an accumulo instance. The names are unique for the lifetime of the instance.
- * 
+ *
  * This is useful for filenames because it makes caching easy.
- * 
+ *
  */
 
 public class UniqueNameAllocator {
@@ -38,17 +38,17 @@ public class UniqueNameAllocator {
   private long maxAllocated = 0;
   private String nextNamePath;
   private Random rand;
-  
+
   private UniqueNameAllocator() {
     nextNamePath = Constants.ZROOT + "/" + HdfsZooInstance.getInstance().getInstanceID() + Constants.ZNEXT_FILE;
     rand = new Random();
   }
-  
+
   public synchronized String getNextName() {
-    
+
     while (next >= maxAllocated) {
       final int allocate = 100 + rand.nextInt(100);
-      
+
       try {
         byte[] max = ZooReaderWriter.getInstance().mutate(nextNamePath, null, ZooUtil.PRIVATE, new ZooReaderWriter.Mutator() {
           public byte[] mutate(byte[] currentValue) throws Exception {
@@ -57,25 +57,25 @@ public class UniqueNameAllocator {
             return Long.toString(l, Character.MAX_RADIX).getBytes(UTF_8);
           }
         });
-        
+
         maxAllocated = Long.parseLong(new String(max, UTF_8), Character.MAX_RADIX);
         next = maxAllocated - allocate;
-        
+
       } catch (Exception e) {
         throw new RuntimeException(e);
       }
     }
-    
+
     return new String(FastFormat.toZeroPaddedString(next++, 7, Character.MAX_RADIX, new byte[0]), UTF_8);
   }
-  
+
   private static UniqueNameAllocator instance = null;
-  
+
   public static synchronized UniqueNameAllocator getInstance() {
     if (instance == null)
       instance = new UniqueNameAllocator();
-    
+
     return instance;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManager.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManager.java b/server/base/src/main/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManager.java
index 41a1edb..5557f03 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManager.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManager.java
@@ -32,16 +32,15 @@ import org.apache.log4j.Logger;
  * The LargestFirstMemoryManager attempts to keep memory between 80% and 90% full. It adapts over time the point at which it should start a compaction based on
  * how full memory gets between successive calls. It will also flush idle tablets based on a per-table configurable idle time. It will only attempt to flush
  * tablets up to 20% of all memory. And, as the name of the class would suggest, it flushes the tablet with the highest memory footprint. However, it actually
- * chooses the tablet as a function of its size doubled for every 15 minutes of idle time. 
+ * chooses the tablet as a function of its size doubled for every 15 minutes of idle time.
  */
 public class LargestFirstMemoryManager implements MemoryManager {
-  
+
   private static final Logger log = Logger.getLogger(LargestFirstMemoryManager.class);
   private static final long ZERO_TIME = System.currentTimeMillis();
   private static final int TSERV_MINC_MAXCONCURRENT_NUMWAITING_MULTIPLIER = 2;
   private static final double MAX_FLUSH_AT_ONCE_PERCENT = 0.20;
 
-
   private long maxMemory = -1;
   private int maxConcurrentMincs;
   private int numWaitingMultiplier;
@@ -51,13 +50,13 @@ public class LargestFirstMemoryManager implements MemoryManager {
   private long maxObserved;
   private final HashMap<Text,Long> mincIdleThresholds = new HashMap<Text,Long>();
   private ServerConfiguration config = null;
-  
+
   private static class TabletInfo {
     final KeyExtent extent;
     final long memTableSize;
     final long idleTime;
     final long load;
-    
+
     public TabletInfo(KeyExtent extent, long memTableSize, long idleTime, long load) {
       this.extent = extent;
       this.memTableSize = memTableSize;
@@ -65,16 +64,16 @@ public class LargestFirstMemoryManager implements MemoryManager {
       this.load = load;
     }
   }
-  
+
   // A little map that will hold the "largest" N tablets, where largest is a result of the timeMemoryLoad function
   private static class LargestMap {
     final int max;
-    final TreeMap<Long, List<TabletInfo>> map = new TreeMap<Long, List<TabletInfo>>(); 
-    
+    final TreeMap<Long,List<TabletInfo>> map = new TreeMap<Long,List<TabletInfo>>();
+
     LargestMap(int n) {
       max = n;
     }
-    
+
     public boolean put(Long key, TabletInfo value) {
       if (map.size() == max) {
         if (key.compareTo(map.firstKey()) < 0)
@@ -114,14 +113,14 @@ public class LargestFirstMemoryManager implements MemoryManager {
       map.remove(key);
     }
   }
-  
+
   LargestFirstMemoryManager(long maxMemory, int maxConcurrentMincs, int numWaitingMultiplier) {
     this();
     this.maxMemory = maxMemory;
     this.maxConcurrentMincs = maxConcurrentMincs;
     this.numWaitingMultiplier = numWaitingMultiplier;
   }
-  
+
   @Override
   public void init(ServerConfiguration conf) {
     this.config = conf;
@@ -129,39 +128,39 @@ public class LargestFirstMemoryManager implements MemoryManager {
     maxConcurrentMincs = conf.getConfiguration().getCount(Property.TSERV_MINC_MAXCONCURRENT);
     numWaitingMultiplier = TSERV_MINC_MAXCONCURRENT_NUMWAITING_MULTIPLIER;
   }
-  
+
   public LargestFirstMemoryManager() {
     prevIngestMemory = 0;
     compactionThreshold = 0.5;
     maxObserved = 0;
   }
-  
+
   protected long getMinCIdleThreshold(KeyExtent extent) {
     Text tableId = extent.getTableId();
     if (!mincIdleThresholds.containsKey(tableId))
       mincIdleThresholds.put(tableId, config.getTableConfiguration(tableId.toString()).getTimeInMillis(Property.TABLE_MINC_COMPACT_IDLETIME));
     return mincIdleThresholds.get(tableId);
   }
-  
+
   @Override
   public MemoryManagementActions getMemoryManagementActions(List<TabletState> tablets) {
     if (maxMemory < 0)
       throw new IllegalStateException("need to initialize " + LargestFirstMemoryManager.class.getName());
-    
+
     final int maxMinCs = maxConcurrentMincs * numWaitingMultiplier;
-    
+
     mincIdleThresholds.clear();
     final MemoryManagementActions result = new MemoryManagementActions();
     result.tabletsToMinorCompact = new ArrayList<KeyExtent>();
-    
+
     LargestMap largestMemTablets = new LargestMap(maxMinCs);
     final LargestMap largestIdleMemTablets = new LargestMap(maxConcurrentMincs);
     final long now = currentTimeMillis();
-    
+
     long ingestMemory = 0;
     long compactionMemory = 0;
     int numWaitingMincs = 0;
-    
+
     // find the largest and most idle tablets
     for (TabletState ts : tablets) {
       final long memTabletSize = ts.getMemTableSize();
@@ -176,21 +175,21 @@ public class LargestFirstMemoryManager implements MemoryManager {
           largestIdleMemTablets.put(timeMemoryLoad, tabletInfo);
         }
       }
-      
+
       compactionMemory += minorCompactingSize;
       if (minorCompactingSize > 0)
         numWaitingMincs++;
     }
-    
+
     if (ingestMemory + compactionMemory > maxObserved) {
       maxObserved = ingestMemory + compactionMemory;
     }
-    
+
     final long memoryChange = ingestMemory - prevIngestMemory;
     prevIngestMemory = ingestMemory;
-    
+
     boolean startMinC = false;
-    
+
     if (numWaitingMincs < maxMinCs) {
       // based on previous ingest memory increase, if we think that the next increase will
       // take us over the threshold for non-compacting memory, then start a minor compaction
@@ -204,23 +203,22 @@ public class LargestFirstMemoryManager implements MemoryManager {
         log.debug("IDLE minor compaction chosen");
       }
     }
-    
+
     if (startMinC) {
       long toBeCompacted = compactionMemory;
-      outer:
-        for (int i = numWaitingMincs; i < maxMinCs && !largestMemTablets.isEmpty(); /* empty */) {
-          Entry<Long,List<TabletInfo>> lastEntry = largestMemTablets.lastEntry();
-          for (TabletInfo largest : lastEntry.getValue()) {
-            toBeCompacted += largest.memTableSize;
-            result.tabletsToMinorCompact.add(largest.extent);
-            log.debug(String.format("COMPACTING %s  total = %,d ingestMemory = %,d", largest.extent.toString(), (ingestMemory + compactionMemory), ingestMemory));
-            log.debug(String.format("chosenMem = %,d chosenIT = %.2f load %,d", largest.memTableSize, largest.idleTime / 1000.0, largest.load));
-            if (toBeCompacted > ingestMemory * MAX_FLUSH_AT_ONCE_PERCENT)
-              break outer;
-            i++;
-          }
-          largestMemTablets.remove(lastEntry.getKey());
+      outer: for (int i = numWaitingMincs; i < maxMinCs && !largestMemTablets.isEmpty(); /* empty */) {
+        Entry<Long,List<TabletInfo>> lastEntry = largestMemTablets.lastEntry();
+        for (TabletInfo largest : lastEntry.getValue()) {
+          toBeCompacted += largest.memTableSize;
+          result.tabletsToMinorCompact.add(largest.extent);
+          log.debug(String.format("COMPACTING %s  total = %,d ingestMemory = %,d", largest.extent.toString(), (ingestMemory + compactionMemory), ingestMemory));
+          log.debug(String.format("chosenMem = %,d chosenIT = %.2f load %,d", largest.memTableSize, largest.idleTime / 1000.0, largest.load));
+          if (toBeCompacted > ingestMemory * MAX_FLUSH_AT_ONCE_PERCENT)
+            break outer;
+          i++;
         }
+        largestMemTablets.remove(lastEntry.getKey());
+      }
     } else if (memoryChange < 0) {
       // before idle mincs, starting a minor compaction meant that memoryChange >= 0.
       // we thought we might want to remove the "else" if that changed,
@@ -228,12 +226,12 @@ public class LargestFirstMemoryManager implements MemoryManager {
       // change more often, so it is staying for now.
       // also, now we have the case where memoryChange < 0 due to an idle compaction, yet
       // we are still adjusting the threshold. should this be tracked and prevented?
-      
+
       // memory change < 0 means a minor compaction occurred
       // we want to see how full the memory got during the compaction
       // (the goal is for it to have between 80% and 90% memory utilization)
       // and adjust the compactionThreshold accordingly
-      
+
       log.debug(String.format("BEFORE compactionThreshold = %.3f maxObserved = %,d", compactionThreshold, maxObserved));
       if (compactionThreshold < 0.82 && maxObserved < 0.8 * maxMemory) {
         // 0.82 * 1.1 is about 0.9, which is our desired max threshold
@@ -243,24 +241,24 @@ public class LargestFirstMemoryManager implements MemoryManager {
         compactionThreshold *= 0.9;
       }
       maxObserved = 0;
-      
+
       log.debug(String.format("AFTER compactionThreshold = %.3f", compactionThreshold));
     }
-    
+
     return result;
   }
-  
+
   protected long currentTimeMillis() {
     return System.currentTimeMillis();
   }
 
   @Override
   public void tabletClosed(KeyExtent extent) {}
-  
+
   // The load function: memory times the idle time, doubling every 15 mins
   static long timeMemoryLoad(long mem, long time) {
     double minutesIdle = time / 60000.0;
-    
+
     return (long) (mem * Math.pow(2, minutesIdle / 15.0));
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tabletserver/MemoryManager.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tabletserver/MemoryManager.java b/server/base/src/main/java/org/apache/accumulo/server/tabletserver/MemoryManager.java
index 5e6fecc..11327f6 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tabletserver/MemoryManager.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tabletserver/MemoryManager.java
@@ -24,13 +24,13 @@ import org.apache.accumulo.server.conf.ServerConfiguration;
 /**
  * A MemoryManager in accumulo currently determines when minor compactions should occur and when ingest should be put on hold. The goal of a memory manager
  * implementation is to maximize ingest throughput and minimize the number of minor compactions.
- * 
- * 
- * 
+ *
+ *
+ *
  */
 
 public interface MemoryManager {
-  
+
   /**
    * Initialize the memory manager.
    */
@@ -38,14 +38,14 @@ public interface MemoryManager {
 
   /**
    * An implementation of this function will be called periodically by accumulo and should return a list of tablets to minor compact.
-   * 
+   *
    * Instructing a tablet that is already minor compacting (this can be inferred from the TabletState) to minor compact has no effect.
-   * 
+   *
    * Holding all ingest does not affect metadata tablets.
    */
-  
+
   MemoryManagementActions getMemoryManagementActions(List<TabletState> tablets);
-  
+
   /**
    * This method is called when a tablet is closed. A memory manger can clean up any per tablet state it is keeping when this is called.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/tabletserver/TabletState.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/tabletserver/TabletState.java b/server/base/src/main/java/org/apache/accumulo/server/tabletserver/TabletState.java
index aeacb8d..78301e8 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/tabletserver/TabletState.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/tabletserver/TabletState.java
@@ -20,10 +20,10 @@ import org.apache.accumulo.core.data.KeyExtent;
 
 public interface TabletState {
   KeyExtent getExtent();
-  
+
   long getLastCommitTime();
-  
+
   long getMemTableSize();
-  
+
   long getMinorCompactingMemTableSize();
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/ActionStatsUpdator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/ActionStatsUpdator.java b/server/base/src/main/java/org/apache/accumulo/server/util/ActionStatsUpdator.java
index dd4573b..de28ad9 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/ActionStatsUpdator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/ActionStatsUpdator.java
@@ -19,10 +19,10 @@ package org.apache.accumulo.server.util;
 import org.apache.accumulo.core.tabletserver.thrift.ActionStats;
 
 /**
- * 
+ *
  */
 public class ActionStatsUpdator {
-  
+
   public static void update(ActionStats summary, ActionStats td) {
     summary.status += td.status;
     summary.elapsed += td.elapsed;
@@ -33,5 +33,5 @@ public class ActionStatsUpdator {
     summary.queueSumDev += td.queueSumDev;
     summary.fail += td.fail;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/Admin.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/Admin.java b/server/base/src/main/java/org/apache/accumulo/server/util/Admin.java
index ae36f1f..7d247f7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/Admin.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/Admin.java
@@ -129,10 +129,10 @@ public class Admin {
     @Parameter(names = {"-u", "--users"}, description = "print users and their authorizations and permissions")
     boolean users = false;
   }
-  
+
   @Parameters(commandDescription = "redistribute tablet directories across the current volume list")
   static class RandomizeVolumesCommand {
-    @Parameter(names={"-t"}, description = "table to update", required=true)
+    @Parameter(names = {"-t"}, description = "table to update", required = true)
     String table = null;
   }
 
@@ -164,10 +164,10 @@ public class Admin {
     cl.addCommand("stopAll", stopAllOpts);
     StopMasterCommand stopMasterOpts = new StopMasterCommand();
     cl.addCommand("stopMaster", stopMasterOpts);
-    
+
     RandomizeVolumesCommand randomizeVolumesOpts = new RandomizeVolumesCommand();
     cl.addCommand("randomizeVolumes", randomizeVolumesOpts);
-    
+
     cl.parse(args);
 
     if (opts.help || cl.getParsedCommand() == null) {
@@ -269,7 +269,7 @@ public class Admin {
   /**
    * flushing during shutdown is a performance optimization, its not required. The method will make an attempt to initiate flushes of all tables and give up if
    * it takes too long.
-   * 
+   *
    */
   private static void flushAll(final ClientContext context) throws AccumuloException, AccumuloSecurityException {
 
@@ -385,21 +385,21 @@ public class Admin {
     }
 
     if (opts.allConfiguration) {
-      //print accumulo site
+      // print accumulo site
       printSystemConfiguration(connector, outputDirectory);
-      //print namespaces
+      // print namespaces
       for (String namespace : connector.namespaceOperations().list()) {
         printNameSpaceConfiguration(connector, namespace, outputDirectory);
       }
-      //print tables
+      // print tables
       SortedSet<String> tableNames = connector.tableOperations().list();
       for (String tableName : tableNames) {
         printTableConfiguration(connector, tableName, outputDirectory);
       }
-      //print users
+      // print users
       for (String user : localUsers) {
-		printUserConfiguration(connector, user, outputDirectory);
-	  }
+        printUserConfiguration(connector, user, outputDirectory);
+      }
     } else {
       if (opts.systemConfiguration) {
         printSystemConfiguration(connector, outputDirectory);
@@ -436,56 +436,57 @@ public class Admin {
     return defaultValue;
   }
 
-  private static void printNameSpaceConfiguration(Connector connector, String namespace, File outputDirectory) throws IOException, AccumuloException, AccumuloSecurityException, NamespaceNotFoundException {
-	  File namespaceScript = new File(outputDirectory, namespace + NS_FILE_SUFFIX);
-	  FileWriter nsWriter = new FileWriter(namespaceScript);
-	  nsWriter.write(createNsFormat.format(new String[] {namespace}));
-	  TreeMap<String,String> props = new TreeMap<String,String>();
-	  for (Entry<String,String> p : connector.namespaceOperations().getProperties(namespace)) {
-		  props.put(p.getKey(), p.getValue());
-	  }
-	  for (Entry<String,String> entry : props.entrySet()) {
-        String defaultValue = getDefaultConfigValue(entry.getKey());
-        if (defaultValue == null || !defaultValue.equals(entry.getValue())) {
-          if (!entry.getValue().equals(siteConfig.get(entry.getKey())) && !entry.getValue().equals(systemConfig.get(entry.getKey()))) {
-            nsWriter.write(nsConfigFormat.format(new String[] {namespace, entry.getKey()+"="+entry.getValue()}));
-          }
+  private static void printNameSpaceConfiguration(Connector connector, String namespace, File outputDirectory) throws IOException, AccumuloException,
+      AccumuloSecurityException, NamespaceNotFoundException {
+    File namespaceScript = new File(outputDirectory, namespace + NS_FILE_SUFFIX);
+    FileWriter nsWriter = new FileWriter(namespaceScript);
+    nsWriter.write(createNsFormat.format(new String[] {namespace}));
+    TreeMap<String,String> props = new TreeMap<String,String>();
+    for (Entry<String,String> p : connector.namespaceOperations().getProperties(namespace)) {
+      props.put(p.getKey(), p.getValue());
+    }
+    for (Entry<String,String> entry : props.entrySet()) {
+      String defaultValue = getDefaultConfigValue(entry.getKey());
+      if (defaultValue == null || !defaultValue.equals(entry.getValue())) {
+        if (!entry.getValue().equals(siteConfig.get(entry.getKey())) && !entry.getValue().equals(systemConfig.get(entry.getKey()))) {
+          nsWriter.write(nsConfigFormat.format(new String[] {namespace, entry.getKey() + "=" + entry.getValue()}));
         }
-	  }
-	  nsWriter.close();
+      }
+    }
+    nsWriter.close();
   }
 
-  private static void printUserConfiguration(Connector connector, String user, File outputDirectory) throws IOException, AccumuloException, AccumuloSecurityException {
-      File userScript = new File(outputDirectory, user + USER_FILE_SUFFIX);
-      FileWriter userWriter = new FileWriter(userScript);
-      userWriter.write(createUserFormat.format(new String[] {user}));
-      Authorizations auths = connector.securityOperations().getUserAuthorizations(user);
-      userWriter.write(userAuthsFormat.format(new String[] {user, auths.toString()}));
-      for (SystemPermission sp : SystemPermission.values()) {
-        if (connector.securityOperations().hasSystemPermission(user, sp)) {
-          userWriter.write(sysPermFormat.format(new String[] {sp.name(), user}));
-        }
+  private static void printUserConfiguration(Connector connector, String user, File outputDirectory) throws IOException, AccumuloException,
+      AccumuloSecurityException {
+    File userScript = new File(outputDirectory, user + USER_FILE_SUFFIX);
+    FileWriter userWriter = new FileWriter(userScript);
+    userWriter.write(createUserFormat.format(new String[] {user}));
+    Authorizations auths = connector.securityOperations().getUserAuthorizations(user);
+    userWriter.write(userAuthsFormat.format(new String[] {user, auths.toString()}));
+    for (SystemPermission sp : SystemPermission.values()) {
+      if (connector.securityOperations().hasSystemPermission(user, sp)) {
+        userWriter.write(sysPermFormat.format(new String[] {sp.name(), user}));
       }
-      for (String namespace : connector.namespaceOperations().list()) {
-        for (NamespacePermission np : NamespacePermission.values()) {
-          if (connector.securityOperations().hasNamespacePermission(user, namespace, np)) {
-            userWriter.write(nsPermFormat.format(new String[] {np.name(), namespace, user}));
-          }
+    }
+    for (String namespace : connector.namespaceOperations().list()) {
+      for (NamespacePermission np : NamespacePermission.values()) {
+        if (connector.securityOperations().hasNamespacePermission(user, namespace, np)) {
+          userWriter.write(nsPermFormat.format(new String[] {np.name(), namespace, user}));
         }
       }
-      for (String tableName : connector.tableOperations().list()) {
-        for (TablePermission perm : TablePermission.values()) {
-          if (connector.securityOperations().hasTablePermission(user, tableName, perm)) {
-            userWriter.write(tablePermFormat.format(new String[] {perm.name(), tableName, user}));
-          }
+    }
+    for (String tableName : connector.tableOperations().list()) {
+      for (TablePermission perm : TablePermission.values()) {
+        if (connector.securityOperations().hasTablePermission(user, tableName, perm)) {
+          userWriter.write(tablePermFormat.format(new String[] {perm.name(), tableName, user}));
         }
       }
+    }
 
-      userWriter.close();
+    userWriter.close();
   }
 
-  private static void printSystemConfiguration(Connector connector, File outputDirectory) throws IOException, AccumuloException,
-      AccumuloSecurityException {
+  private static void printSystemConfiguration(Connector connector, File outputDirectory) throws IOException, AccumuloException, AccumuloSecurityException {
     Configuration conf = new Configuration(false);
     TreeMap<String,String> site = new TreeMap<String,String>(siteConfig);
     for (Entry<String,String> prop : site.entrySet()) {
@@ -510,8 +511,8 @@ public class Admin {
     }
   }
 
-  private static void printTableConfiguration(Connector connector, String tableName, File outputDirectory) throws AccumuloException,
-      TableNotFoundException, IOException, AccumuloSecurityException {
+  private static void printTableConfiguration(Connector connector, String tableName, File outputDirectory) throws AccumuloException, TableNotFoundException,
+      IOException, AccumuloSecurityException {
     File tableBackup = new File(outputDirectory, tableName + ".cfg");
     FileWriter writer = new FileWriter(tableBackup);
     writer.write(createTableFormat.format(new String[] {tableName}));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/ChangeSecret.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/ChangeSecret.java b/server/base/src/main/java/org/apache/accumulo/server/util/ChangeSecret.java
index c93f680..18ca881 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/ChangeSecret.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/ChangeSecret.java
@@ -44,14 +44,14 @@ import org.apache.zookeeper.data.Stat;
 import com.beust.jcommander.Parameter;
 
 public class ChangeSecret {
-  
+
   static class Opts extends ClientOpts {
-    @Parameter(names="--old", description="old zookeeper password", password=true, hidden=true)
+    @Parameter(names = "--old", description = "old zookeeper password", password = true, hidden = true)
     String oldPass;
-    @Parameter(names="--new", description="new zookeeper password", password=true, hidden=true)
+    @Parameter(names = "--new", description = "new zookeeper password", password = true, hidden = true)
     String newPass;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     List<String> argsList = new ArrayList<String>(args.length + 2);
@@ -71,11 +71,11 @@ public class ChangeSecret {
     System.out.println("New instance id is " + instanceId);
     System.out.println("Be sure to put your new secret in accumulo-site.xml");
   }
-  
+
   interface Visitor {
     void visit(ZooReader zoo, String path) throws Exception;
   }
-  
+
   private static void recurse(ZooReader zoo, String root, Visitor v) {
     try {
       v.visit(zoo, root);
@@ -86,7 +86,7 @@ public class ChangeSecret {
       throw new RuntimeException(ex);
     }
   }
-  
+
   private static boolean verifyAccumuloIsDown(Instance inst, String oldPassword) {
     ZooReader zooReader = new ZooReaderWriter(inst.getZooKeepers(), inst.getZooKeepersSessionTimeOut(), oldPassword);
     String root = ZooUtil.getRoot(inst);
@@ -101,14 +101,14 @@ public class ChangeSecret {
     if (ephemerals.size() == 0) {
       return true;
     }
-    
+
     System.err.println("The following ephemeral nodes exist, something is still running:");
     for (String path : ephemerals) {
       System.err.println(path);
     }
     return false;
   }
-  
+
   private static String rewriteZooKeeperInstance(final Instance inst, String oldPass, String newPass) throws Exception {
     final ZooReaderWriter orig = new ZooReaderWriter(inst.getZooKeepers(), inst.getZooKeepersSessionTimeOut(), oldPass);
     final IZooReaderWriter new_ = new ZooReaderWriter(inst.getZooKeepers(), inst.getZooKeepersSessionTimeOut(), newPass);
@@ -143,7 +143,7 @@ public class ChangeSecret {
     new_.putPersistentData(path, newInstanceId.getBytes(UTF_8), NodeExistsPolicy.OVERWRITE);
     return newInstanceId;
   }
-  
+
   private static void updateHdfs(VolumeManager fs, Instance inst, String newInstanceId) throws IOException {
     // Need to recreate the instanceId on all of them to keep consistency
     for (Volume v : fs.getVolumes()) {
@@ -159,7 +159,7 @@ public class ChangeSecret {
       v.getFileSystem().create(new Path(instanceId, newInstanceId)).close();
     }
   }
-  
+
   private static void deleteInstance(Instance origInstance, String oldPass) throws Exception {
     IZooReaderWriter orig = new ZooReaderWriter(origInstance.getZooKeepers(), origInstance.getZooKeepersSessionTimeOut(), oldPass);
     orig.recursiveDelete("/accumulo/" + origInstance.getInstanceID(), NodeMissingPolicy.SKIP);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/CleanZookeeper.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/CleanZookeeper.java b/server/base/src/main/java/org/apache/accumulo/server/util/CleanZookeeper.java
index 0652cf1..3cfb1a7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/CleanZookeeper.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/CleanZookeeper.java
@@ -19,6 +19,7 @@ package org.apache.accumulo.server.util;
 import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.io.IOException;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.cli.Help;
 import org.apache.accumulo.fate.zookeeper.IZooReaderWriter;
@@ -31,16 +32,16 @@ import org.apache.zookeeper.KeeperException;
 import com.beust.jcommander.Parameter;
 
 public class CleanZookeeper {
-  
+
   private static final Logger log = Logger.getLogger(CleanZookeeper.class);
-  
+
   static class Opts extends Help {
-    @Parameter(names={"-z", "--keepers"}, description="comma separated list of zookeeper hosts")
+    @Parameter(names = {"-z", "--keepers"}, description = "comma separated list of zookeeper hosts")
     String keepers = "localhost:2181";
-    @Parameter(names={"--password"}, description="the system secret", password=true)
+    @Parameter(names = {"--password"}, description = "the system secret", password = true)
     String auth;
   }
-  
+
   /**
    * @param args
    *          must contain one element: the address of a zookeeper node a second parameter provides an additional authentication value
@@ -50,13 +51,13 @@ public class CleanZookeeper {
   public static void main(String[] args) throws IOException {
     Opts opts = new Opts();
     opts.parseArgs(CleanZookeeper.class.getName(), args);
-    
+
     String root = Constants.ZROOT;
     IZooReaderWriter zk = ZooReaderWriter.getInstance();
     if (opts.auth != null) {
-      zk.getZooKeeper().addAuthInfo("digest", ("accumulo:"+opts.auth).getBytes(UTF_8));
+      zk.getZooKeeper().addAuthInfo("digest", ("accumulo:" + opts.auth).getBytes(UTF_8));
     }
-    
+
     try {
       for (String child : zk.getChildren(root)) {
         if (Constants.ZINSTANCES.equals("/" + child)) {
@@ -84,5 +85,5 @@ public class CleanZookeeper {
       System.out.println("Error Occurred: " + ex);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/DefaultMap.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/DefaultMap.java b/server/base/src/main/java/org/apache/accumulo/server/util/DefaultMap.java
index 7038be7..66c5d8e 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/DefaultMap.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/DefaultMap.java
@@ -20,17 +20,17 @@ import java.util.HashMap;
 
 /**
  * A HashMap that returns a default value if the key is not stored in the map.
- * 
+ *
  * A zero-argument constructor of the default object's class is used, otherwise the default object is used.
  */
 public class DefaultMap<K,V> extends HashMap<K,V> {
   private static final long serialVersionUID = 1L;
   V dfault;
-  
+
   public DefaultMap(V dfault) {
     this.dfault = dfault;
   }
-  
+
   @SuppressWarnings("unchecked")
   @Override
   public V get(Object key) {
@@ -44,7 +44,7 @@ public class DefaultMap<K,V> extends HashMap<K,V> {
     }
     return result;
   }
-  
+
   @SuppressWarnings("unchecked")
   private V construct() {
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/DeleteZooInstance.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/DeleteZooInstance.java b/server/base/src/main/java/org/apache/accumulo/server/util/DeleteZooInstance.java
index 0793c6b..ba27733 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/DeleteZooInstance.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/DeleteZooInstance.java
@@ -31,14 +31,14 @@ import org.apache.zookeeper.KeeperException;
 import com.beust.jcommander.Parameter;
 
 public class DeleteZooInstance {
-  
+
   static class Opts extends Help {
-    @Parameter(names={"-i", "--instance"}, description="the instance name or id to delete")
+    @Parameter(names = {"-i", "--instance"}, description = "the instance name or id to delete")
     String instance;
   }
-  
+
   static void deleteRetry(IZooReaderWriter zk, String path) throws Exception {
-    for (int i = 0; i < 10; i++){
+    for (int i = 0; i < 10; i++) {
       try {
         zk.recursiveDelete(path, NodeMissingPolicy.SKIP);
         return;
@@ -49,7 +49,7 @@ public class DeleteZooInstance {
       }
     }
   }
-  
+
   /**
    * @param args
    *          : the name or UUID of the instance to be deleted
@@ -57,7 +57,7 @@ public class DeleteZooInstance {
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(DeleteZooInstance.class.getName(), args);
-    
+
     IZooReaderWriter zk = ZooReaderWriter.getInstance();
     // try instance name:
     Set<String> instances = new HashSet<String>(zk.getChildren(Constants.ZROOT + Constants.ZINSTANCES));
@@ -79,5 +79,5 @@ public class DeleteZooInstance {
       deleteRetry(zk, Constants.ZROOT + "/" + opts.instance);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/DumpZookeeper.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/DumpZookeeper.java b/server/base/src/main/java/org/apache/accumulo/server/util/DumpZookeeper.java
index 478b3b5..e900202 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/DumpZookeeper.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/DumpZookeeper.java
@@ -20,6 +20,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.io.PrintStream;
 import java.io.UnsupportedEncodingException;
+
 import org.apache.accumulo.core.cli.Help;
 import org.apache.accumulo.core.util.Base64;
 import org.apache.accumulo.fate.zookeeper.IZooReaderWriter;
@@ -32,35 +33,35 @@ import org.apache.zookeeper.data.Stat;
 import com.beust.jcommander.Parameter;
 
 public class DumpZookeeper {
-  
+
   static IZooReaderWriter zk = null;
-  
+
   private static final Logger log = Logger.getLogger(DumpZookeeper.class);
-  
+
   private static class Encoded {
     public String encoding;
     public String value;
-    
+
     Encoded(String e, String v) {
       encoding = e;
       value = v;
     }
   }
-  
+
   static class Opts extends Help {
     @Parameter(names = "--root", description = "the root of the znode tree to dump")
     String root = "/";
   }
-  
+
   public static void main(String[] args) {
     Opts opts = new Opts();
     opts.parseArgs(DumpZookeeper.class.getName(), args);
-    
+
     Logger.getRootLogger().setLevel(Level.WARN);
     PrintStream out = System.out;
     try {
       zk = ZooReaderWriter.getInstance();
-      
+
       write(out, 0, "<dump root='%s'>", opts.root);
       for (String child : zk.getChildren(opts.root, null))
         if (!child.equals("zookeeper"))
@@ -70,7 +71,7 @@ public class DumpZookeeper {
       log.error(ex, ex);
     }
   }
-  
+
   private static void dump(PrintStream out, String root, String child, int indent) throws KeeperException, InterruptedException, UnsupportedEncodingException {
     String path = root + "/" + child;
     if (root.endsWith("/"))
@@ -102,7 +103,7 @@ public class DumpZookeeper {
       write(out, indent, "</node>");
     }
   }
-  
+
   private static Encoded value(String path) throws KeeperException, InterruptedException, UnsupportedEncodingException {
     byte[] data = zk.getData(path, null);
     for (int i = 0; i < data.length; i++) {
@@ -112,7 +113,7 @@ public class DumpZookeeper {
     }
     return new Encoded(UTF_8.name(), new String(data, UTF_8));
   }
-  
+
   private static void write(PrintStream out, int indent, String fmt, Object... args) {
     for (int i = 0; i < indent; i++)
       out.print(" ");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/util/FileSystemMonitor.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/FileSystemMonitor.java b/server/base/src/main/java/org/apache/accumulo/server/util/FileSystemMonitor.java
index 665bf25..94a80d9 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/FileSystemMonitor.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/FileSystemMonitor.java
@@ -38,43 +38,43 @@ import org.apache.log4j.Logger;
 public class FileSystemMonitor {
   private static final String PROC_MOUNTS = "/proc/mounts";
   private static final Logger log = Logger.getLogger(FileSystemMonitor.class);
-  
+
   private static class Mount {
     String mountPoint;
     Set<String> options;
-    
+
     Mount(String line) {
       String tokens[] = line.split("\\s+");
-      
+
       mountPoint = tokens[1];
-      
+
       options = new HashSet<String>(Arrays.asList(tokens[3].split(",")));
     }
   }
-  
+
   static List<Mount> parse(String procFile) throws IOException {
-    
+
     List<Mount> mounts = new ArrayList<Mount>();
-    
+
     FileReader fr = new FileReader(procFile);
     BufferedReader br = new BufferedReader(fr);
-    
+
     String line;
     try {
-    while ((line = br.readLine()) != null)
-      mounts.add(new Mount(line));
+      while ((line = br.readLine()) != null)
+        mounts.add(new Mount(line));
     } finally {
       br.close();
     }
-    
+
     return mounts;
   }
-  
+
   private Map<String,Boolean> readWriteFilesystems = new HashMap<String,Boolean>();
-  
+
   public FileSystemMonitor(final String procFile, long period) throws IOException {
     List<Mount> mounts = parse(procFile);
-    
+
     for (Mount mount : mounts) {
       if (mount.options.contains("rw"))
         readWriteFilesystems.put(mount.mountPoint, true);
@@ -83,7 +83,7 @@ public class FileSystemMonitor {
       else
         throw new IOException("Filesystem " + mount + " does not have ro or rw option");
     }
-    
+
     TimerTask tt = new TimerTask() {
       @Override
       public void run() {
@@ -98,16 +98,16 @@ public class FileSystemMonitor {
         }
       }
     };
-    
+
     // use a new Timer object instead of a shared one.
     // trying to avoid the case where one the timers other
     // task gets stuck because a FS went read only, and this task
     // does not execute
     Timer timer = new Timer("filesystem monitor timer", true);
     timer.schedule(tt, period, period);
-    
+
   }
-  
+
   protected void logAsync(final Level level, final String msg, final Exception e) {
     Runnable r = new Runnable() {
       @Override
@@ -115,13 +115,13 @@ public class FileSystemMonitor {
         log.log(level, msg, e);
       }
     };
-    
+
     new Thread(r).start();
   }
-  
+
   protected void checkMounts(String procFile) throws Exception {
     List<Mount> mounts = parse(procFile);
-    
+
     for (Mount mount : mounts) {
       if (!readWriteFilesystems.containsKey(mount.mountPoint))
         if (mount.options.contains("rw"))
@@ -134,7 +134,7 @@ public class FileSystemMonitor {
         throw new Exception("Filesystem " + mount.mountPoint + " switched to read only");
     }
   }
-  
+
   public static void start(AccumuloConfiguration conf, Property prop) {
     if (conf.getBoolean(prop)) {
       if (new File(PROC_MOUNTS).exists()) {


[03/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/scalability/Ingest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/scalability/Ingest.java b/test/src/main/java/org/apache/accumulo/test/scalability/Ingest.java
index 238a88d..dfb5547 100644
--- a/test/src/main/java/org/apache/accumulo/test/scalability/Ingest.java
+++ b/test/src/main/java/org/apache/accumulo/test/scalability/Ingest.java
@@ -21,6 +21,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 import java.util.Random;
 import java.util.UUID;
 import java.util.concurrent.TimeUnit;
+
 import org.apache.accumulo.core.client.BatchWriter;
 import org.apache.accumulo.core.client.BatchWriterConfig;
 import org.apache.accumulo.core.client.Connector;
@@ -32,56 +33,56 @@ import org.apache.accumulo.test.continuous.ContinuousIngest;
 import org.apache.log4j.Logger;
 
 public class Ingest extends ScaleTest {
-  
+
   private static final Logger log = Logger.getLogger(Ingest.class);
-  
+
   @Override
   public void setup() {
-    
+
     Connector conn = getConnector();
     String tableName = getTestProperty("TABLE");
-    
+
     // delete existing table
     if (conn.tableOperations().exists(tableName)) {
       System.out.println("Deleting existing table: " + tableName);
       try {
         conn.tableOperations().delete(tableName);
       } catch (Exception e) {
-        log.error("Failed to delete table '"+tableName+"'.", e);
+        log.error("Failed to delete table '" + tableName + "'.", e);
       }
     }
-    
+
     // create table
     try {
       conn.tableOperations().create(tableName);
       conn.tableOperations().addSplits(tableName, calculateSplits());
       conn.tableOperations().setProperty(tableName, "table.split.threshold", "256M");
     } catch (Exception e) {
-      log.error("Failed to create table '"+tableName+"'.", e);
+      log.error("Failed to create table '" + tableName + "'.", e);
     }
-    
+
   }
-  
+
   @Override
   public void client() {
-    
+
     Connector conn = getConnector();
     String tableName = getTestProperty("TABLE");
-    
+
     // get batch writer configuration
     long maxMemory = Long.parseLong(getTestProperty("MAX_MEMORY"));
     long maxLatency = Long.parseLong(getTestProperty("MAX_LATENCY"));
     int maxWriteThreads = Integer.parseInt(getTestProperty("NUM_THREADS"));
-    
+
     // create batch writer
     BatchWriter bw = null;
     try {
       bw = conn.createBatchWriter(tableName, new BatchWriterConfig().setMaxMemory(maxMemory).setMaxLatency(maxLatency, TimeUnit.MILLISECONDS)
           .setMaxWriteThreads(maxWriteThreads));
     } catch (TableNotFoundException e) {
-      log.error("Table '"+tableName+"' not found.", e);
+      log.error("Table '" + tableName + "' not found.", e);
     }
-    
+
     // configure writing
     Random r = new Random();
     String ingestInstanceId = UUID.randomUUID().toString();
@@ -92,12 +93,12 @@ public class Ingest extends ScaleTest {
     int maxColQ = 32767;
     long count = 0;
     long totalBytes = 0;
-    
+
     ColumnVisibility cv = new ColumnVisibility();
 
     // start timer
     startTimer();
-    
+
     // write specified number of entries
     while (count < numIngestEntries) {
       count++;
@@ -111,7 +112,7 @@ public class Ingest extends ScaleTest {
         System.exit(-1);
       }
     }
-    
+
     // close writer
     try {
       bw.close();
@@ -119,22 +120,22 @@ public class Ingest extends ScaleTest {
       log.error("Could not close BatchWriter due to mutations being rejected.", e);
       System.exit(-1);
     }
-    
+
     // stop timer
     stopTimer(count, totalBytes);
   }
-  
+
   @Override
   public void teardown() {
-    
+
     Connector conn = getConnector();
     String tableName = getTestProperty("TABLE");
-    
+
     try {
       conn.tableOperations().delete(tableName);
     } catch (Exception e) {
-      log.error("Failed to delete table '"+tableName+"'", e);
+      log.error("Failed to delete table '" + tableName + "'", e);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/scalability/Run.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/scalability/Run.java b/test/src/main/java/org/apache/accumulo/test/scalability/Run.java
index 42c5ec3..f7af2ff 100644
--- a/test/src/main/java/org/apache/accumulo/test/scalability/Run.java
+++ b/test/src/main/java/org/apache/accumulo/test/scalability/Run.java
@@ -16,10 +16,10 @@
  */
 package org.apache.accumulo.test.scalability;
 
-import com.beust.jcommander.Parameter;
 import java.io.FileInputStream;
 import java.net.InetAddress;
 import java.util.Properties;
+
 import org.apache.accumulo.core.cli.Help;
 import org.apache.accumulo.core.util.CachedConfiguration;
 import org.apache.hadoop.conf.Configuration;
@@ -27,33 +27,35 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
+import com.beust.jcommander.Parameter;
+
 public class Run {
-  
+
   private static final Logger log = Logger.getLogger(Run.class);
-  
+
   static class Opts extends Help {
-    @Parameter(names="--testId", required=true)
+    @Parameter(names = "--testId", required = true)
     String testId;
-    @Parameter(names="--action", required=true, description="one of 'setup', 'teardown' or 'client'")
+    @Parameter(names = "--action", required = true, description = "one of 'setup', 'teardown' or 'client'")
     String action;
-    @Parameter(names="--count", description="number of tablet servers", required=true)
-    int numTabletServers; 
+    @Parameter(names = "--count", description = "number of tablet servers", required = true)
+    int numTabletServers;
   }
-  
+
   public static void main(String[] args) throws Exception {
-    
+
     final String sitePath = "/tmp/scale-site.conf";
     final String testPath = "/tmp/scale-test.conf";
     Opts opts = new Opts();
     opts.parseArgs(Run.class.getName(), args);
-    
+
     Configuration conf = CachedConfiguration.getInstance();
     FileSystem fs;
     fs = FileSystem.get(conf);
-    
+
     fs.copyToLocalFile(new Path("/accumulo-scale/conf/site.conf"), new Path(sitePath));
     fs.copyToLocalFile(new Path(String.format("/accumulo-scale/conf/%s.conf", opts.testId)), new Path(testPath));
-    
+
     // load configuration file properties
     Properties scaleProps = new Properties();
     Properties testProps = new Properties();
@@ -69,11 +71,11 @@ public class Run {
     } catch (Exception e) {
       log.error("Error loading config file.", e);
     }
-    
+
     ScaleTest test = (ScaleTest) Class.forName(String.format("org.apache.accumulo.test.scalability.%s", opts.testId)).newInstance();
-    
+
     test.init(scaleProps, testProps, opts.numTabletServers);
-    
+
     if (opts.action.equalsIgnoreCase("setup")) {
       test.setup();
     } else if (opts.action.equalsIgnoreCase("client")) {
@@ -86,5 +88,5 @@ public class Run {
       test.teardown();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/scalability/ScaleTest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/scalability/ScaleTest.java b/test/src/main/java/org/apache/accumulo/test/scalability/ScaleTest.java
index 46377d6..f908296 100644
--- a/test/src/main/java/org/apache/accumulo/test/scalability/ScaleTest.java
+++ b/test/src/main/java/org/apache/accumulo/test/scalability/ScaleTest.java
@@ -28,44 +28,44 @@ import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.apache.hadoop.io.Text;
 
 public abstract class ScaleTest {
-  
+
   private Connector conn;
   private Properties scaleProps;
   private Properties testProps;
   private int numTabletServers;
   private long startTime;
-  
+
   public void init(Properties scaleProps, Properties testProps, int numTabletServers) throws AccumuloException, AccumuloSecurityException {
-    
+
     this.scaleProps = scaleProps;
     this.testProps = testProps;
     this.numTabletServers = numTabletServers;
-    
+
     // get properties to create connector
     String instanceName = this.scaleProps.getProperty("INSTANCE_NAME");
     String zookeepers = this.scaleProps.getProperty("ZOOKEEPERS");
     String user = this.scaleProps.getProperty("USER");
     String password = this.scaleProps.getProperty("PASSWORD");
     System.out.println(password);
-    
+
     conn = new ZooKeeperInstance(new ClientConfiguration().withInstance(instanceName).withZkHosts(zookeepers)).getConnector(user, new PasswordToken(password));
   }
-  
+
   protected void startTimer() {
     startTime = System.currentTimeMillis();
   }
-  
+
   protected void stopTimer(long numEntries, long numBytes) {
     long endTime = System.currentTimeMillis();
     System.out.printf("ELAPSEDMS %d %d %d%n", endTime - startTime, numEntries, numBytes);
   }
-  
+
   public abstract void setup();
-  
+
   public abstract void client();
-  
+
   public abstract void teardown();
-  
+
   public TreeSet<Text> calculateSplits() {
     int numSplits = numTabletServers - 1;
     long distance = (Long.MAX_VALUE / numTabletServers) + 1;
@@ -77,11 +77,11 @@ public abstract class ScaleTest {
     }
     return keys;
   }
-  
+
   public Connector getConnector() {
     return conn;
   }
-  
+
   public String getTestProperty(String key) {
     return testProps.getProperty(key);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/DataWriter.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/DataWriter.java b/test/src/main/java/org/apache/accumulo/test/stress/random/DataWriter.java
index 33a3984..e7158e2 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/DataWriter.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/DataWriter.java
@@ -22,12 +22,12 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
 public class DataWriter extends Stream<Void> {
   private final BatchWriter writer;
   private final RandomMutations mutations;
-  
+
   public DataWriter(BatchWriter writer, RandomMutations mutations) {
     this.writer = writer;
     this.mutations = mutations;
   }
-  
+
   @Override
   public Void next() {
     try {
@@ -37,7 +37,7 @@ public class DataWriter extends Stream<Void> {
     }
     return null;
   }
-  
+
   @Override
   public void finalize() {
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/IntArgValidator.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/IntArgValidator.java b/test/src/main/java/org/apache/accumulo/test/stress/random/IntArgValidator.java
index 6ba6ca9..1582f0d 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/IntArgValidator.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/IntArgValidator.java
@@ -22,11 +22,11 @@ import com.beust.jcommander.ParameterException;
 import com.google.common.base.Preconditions;
 
 public class IntArgValidator implements IValueValidator<Integer> {
-  
+
   @Override
   public void validate(String name, Integer value) throws ParameterException {
     Preconditions.checkNotNull(value);
     Preconditions.checkArgument(value > 0);
   }
-  
-}
\ No newline at end of file
+
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/RandomByteArrays.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/RandomByteArrays.java b/test/src/main/java/org/apache/accumulo/test/stress/random/RandomByteArrays.java
index 0b6b36a..a3bdd43 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/RandomByteArrays.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/RandomByteArrays.java
@@ -21,11 +21,11 @@ package org.apache.accumulo.test.stress.random;
  */
 public class RandomByteArrays extends Stream<byte[]> {
   private final RandomWithinRange random_arrays;
-  
+
   public RandomByteArrays(RandomWithinRange random_arrays) {
     this.random_arrays = random_arrays;
   }
-  
+
   public byte[] next() {
     return random_arrays.next_bytes();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/RandomMutations.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/RandomMutations.java b/test/src/main/java/org/apache/accumulo/test/stress/random/RandomMutations.java
index 679b983..db5da55 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/RandomMutations.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/RandomMutations.java
@@ -24,10 +24,9 @@ public class RandomMutations extends Stream<Mutation> {
   private final int max_cells_per_mutation;
   private byte[] current_row;
   private int cells_remaining_in_row;
-  
-  public RandomMutations(RandomByteArrays rows, RandomByteArrays column_families,
-      RandomByteArrays column_qualifiers, RandomByteArrays values, RandomWithinRange row_widths,
-      int max_cells_per_mutation) {
+
+  public RandomMutations(RandomByteArrays rows, RandomByteArrays column_families, RandomByteArrays column_qualifiers, RandomByteArrays values,
+      RandomWithinRange row_widths, int max_cells_per_mutation) {
     this.rows = rows;
     this.column_families = column_families;
     this.column_qualifiers = column_qualifiers;
@@ -48,7 +47,7 @@ public class RandomMutations extends Stream<Mutation> {
     }
     Mutation m = new Mutation(current_row);
     final int cells = Math.min(cells_remaining_in_row, max_cells_per_mutation);
-    for(int i = 1; i <= cells; i++) {
+    for (int i = 1; i <= cells; i++) {
       m.put(column_families.next(), column_qualifiers.next(), values.next());
     }
     cells_remaining_in_row -= cells;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/RandomWithinRange.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/RandomWithinRange.java b/test/src/main/java/org/apache/accumulo/test/stress/random/RandomWithinRange.java
index 8eb978b..8da9a37 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/RandomWithinRange.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/RandomWithinRange.java
@@ -21,18 +21,17 @@ import java.util.Random;
 import com.google.common.base.Preconditions;
 
 /**
- * Class that returns positive integers between some minimum
- * and maximum.
+ * Class that returns positive integers between some minimum and maximum.
  *
  */
 public class RandomWithinRange {
   private final Random random;
   private final int min, max;
-  
+
   public RandomWithinRange(int seed, int min, int max) {
     this(new Random(seed), min, max);
   }
-  
+
   public RandomWithinRange(Random random, int min, int max) {
     Preconditions.checkArgument(min > 0, "Min must be positive.");
     Preconditions.checkArgument(max >= min, "Max must be greater than or equal to min.");
@@ -40,7 +39,7 @@ public class RandomWithinRange {
     this.min = min;
     this.max = max;
   }
-  
+
   public int next() {
     if (min == max) {
       return min;
@@ -50,7 +49,7 @@ public class RandomWithinRange {
       return random.nextInt(max - min) + min;
     }
   }
-  
+
   public byte[] next_bytes() {
     byte[] b = new byte[next()];
     random.nextBytes(b);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/Scan.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/Scan.java b/test/src/main/java/org/apache/accumulo/test/stress/random/Scan.java
index 9408770..c59cd1c 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/Scan.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/Scan.java
@@ -33,26 +33,24 @@ import org.apache.hadoop.io.Text;
 import com.google.common.collect.Lists;
 
 public class Scan {
-  
+
   public static void main(String[] args) throws Exception {
     ScanOpts opts = new ScanOpts();
     opts.parseArgs(Scan.class.getName(), args);
-    
+
     Connector connector = opts.getConnector();
     Scanner scanner = connector.createScanner(opts.getTableName(), new Authorizations());
-    
+
     if (opts.isolate) {
       scanner.enableIsolation();
     }
-    
+
     Random tablet_index_generator = new Random(opts.scan_seed);
-    
-    LoopControl scanning_condition = opts.continuous ? new ContinuousLoopControl() :
-                                                       new IterativeLoopControl(opts.scan_iterations);
-    
-    while(scanning_condition.keepScanning()) {
-      Range range = pickRange(connector.tableOperations(), opts.getTableName(),
-          tablet_index_generator);
+
+    LoopControl scanning_condition = opts.continuous ? new ContinuousLoopControl() : new IterativeLoopControl(opts.scan_iterations);
+
+    while (scanning_condition.keepScanning()) {
+      Range range = pickRange(connector.tableOperations(), opts.getTableName(), tablet_index_generator);
       scanner.setRange(range);
       if (opts.batch_size > 0) {
         scanner.setBatchSize(opts.batch_size);
@@ -60,24 +58,20 @@ public class Scan {
       try {
         consume(scanner);
       } catch (Exception e) {
-        System.err.println(
-            String.format(
-                "Exception while scanning range %s. Check the state of Accumulo for errors.",
-                range));
+        System.err.println(String.format("Exception while scanning range %s. Check the state of Accumulo for errors.", range));
         throw e;
       }
     }
   }
-  
+
   public static void consume(Iterable<?> iterable) {
     Iterator<?> itr = iterable.iterator();
     while (itr.hasNext()) {
       itr.next();
     }
   }
-  
-  public static Range pickRange(TableOperations tops, String table, Random r)
-      throws TableNotFoundException, AccumuloSecurityException, AccumuloException {
+
+  public static Range pickRange(TableOperations tops, String table, Random r) throws TableNotFoundException, AccumuloSecurityException, AccumuloException {
     ArrayList<Text> splits = Lists.newArrayList(tops.listSplits(table));
     if (splits.isEmpty()) {
       return new Range();
@@ -88,28 +82,26 @@ public class Scan {
       return new Range(startRow, false, endRow, true);
     }
   }
-  
+
   /*
-   * These interfaces + implementations are used to determine
-   * how many times the scanner should look up a random tablet
-   * and scan it.
+   * These interfaces + implementations are used to determine how many times the scanner should look up a random tablet and scan it.
    */
   static interface LoopControl {
     public boolean keepScanning();
   }
-  
+
   // Does a finite number of iterations
   static class IterativeLoopControl implements LoopControl {
     private final int max;
     private int current;
-    
+
     public IterativeLoopControl(int max) {
       this.max = max;
       this.current = 0;
     }
-    
+
     public boolean keepScanning() {
-      if(current < max) {
+      if (current < max) {
         ++current;
         return true;
       } else {
@@ -117,7 +109,7 @@ public class Scan {
       }
     }
   }
-  
+
   // Does an infinite number of iterations
   static class ContinuousLoopControl implements LoopControl {
     public boolean keepScanning() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/ScanOpts.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/ScanOpts.java b/test/src/main/java/org/apache/accumulo/test/stress/random/ScanOpts.java
index 86e7920..e3f73f7 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/ScanOpts.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/ScanOpts.java
@@ -21,26 +21,25 @@ import org.apache.accumulo.core.cli.ClientOnDefaultTable;
 import com.beust.jcommander.Parameter;
 
 class ScanOpts extends ClientOnDefaultTable {
-  @Parameter(names = "--isolate",
-      description = "true to turn on scan isolation, false to turn off. default is false.")
+  @Parameter(names = "--isolate", description = "true to turn on scan isolation, false to turn off. default is false.")
   boolean isolate = false;
-  
+
   @Parameter(names = "--num-iterations", description = "number of scan iterations")
   int scan_iterations = 1024;
-  
+
   @Parameter(names = "--continuous", description = "continuously scan the table. note that this overrides --num-iterations")
   boolean continuous;
-  
+
   @Parameter(names = "--scan-seed", description = "seed for randomly choosing tablets to scan")
   int scan_seed = 1337;
 
-  @Parameter(names = "--scan-batch-size", description="scanner batch size")
+  @Parameter(names = "--scan-batch-size", description = "scanner batch size")
   int batch_size = -1;
 
   public ScanOpts() {
     this(WriteOptions.DEFAULT_TABLE);
   }
-  
+
   public ScanOpts(String table) {
     super(table);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/Stream.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/Stream.java b/test/src/main/java/org/apache/accumulo/test/stress/random/Stream.java
index adacfb8..72b31e5 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/Stream.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/Stream.java
@@ -20,21 +20,21 @@ import java.util.Iterator;
 
 /**
  * Base class to model an infinite stream of data. A stream implements an iterator whose {{@link #hasNext()} method will always return true.
- * 
+ *
  */
 public abstract class Stream<T> implements Iterator<T> {
-  
+
   @Override
   public final boolean hasNext() {
     return true;
   }
-  
+
   @Override
   public abstract T next();
-  
+
   @Override
   public final void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/Write.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/Write.java b/test/src/main/java/org/apache/accumulo/test/stress/random/Write.java
index bb679ad..ea6f164 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/Write.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/Write.java
@@ -22,25 +22,25 @@ import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 
 public class Write {
-  
+
   public static void main(String[] args) throws Exception {
     WriteOptions opts = new WriteOptions();
     BatchWriterOpts batch_writer_opts = new BatchWriterOpts();
     opts.parseArgs(Write.class.getName(), args, batch_writer_opts);
-    
+
     opts.check();
-    
+
     Connector c = opts.getConnector();
-    
-    if(opts.clear_table && c.tableOperations().exists(opts.getTableName())) {
+
+    if (opts.clear_table && c.tableOperations().exists(opts.getTableName())) {
       try {
-      c.tableOperations().delete(opts.getTableName());
-      } catch(TableNotFoundException e) {
+        c.tableOperations().delete(opts.getTableName());
+      } catch (TableNotFoundException e) {
         System.err.println("Couldn't delete the table because it doesn't exist any more.");
       }
     }
-    
-    if(!c.tableOperations().exists(opts.getTableName())) {
+
+    if (!c.tableOperations().exists(opts.getTableName())) {
       try {
         c.tableOperations().create(opts.getTableName());
       } catch (TableExistsException e) {
@@ -53,42 +53,21 @@ public class Write {
       writeDelay = 0;
     }
 
-    DataWriter dw = new DataWriter(c.createBatchWriter(opts.getTableName(), batch_writer_opts.getBatchWriterConfig()),
-        new RandomMutations(
-            //rows
-            new RandomByteArrays(
-                new RandomWithinRange(
-                    opts.row_seed,
-                    opts.rowMin(),
-                    opts.rowMax())),
-            //cfs
-            new RandomByteArrays(
-                new RandomWithinRange(
-                    opts.cf_seed,
-                    opts.cfMin(),
-                    opts.cfMax())),
-            //cqs
-            new RandomByteArrays(
-                new RandomWithinRange(
-                    opts.cq_seed,
-                    opts.cqMin(),
-                    opts.cqMax())),
-            //vals
-            new RandomByteArrays(
-                new RandomWithinRange(
-                    opts.value_seed,
-                    opts.valueMin(),
-                    opts.valueMax())),
-            //number of cells per row
-            new RandomWithinRange(
-                opts.row_width_seed,
-                opts.rowWidthMin(),
-                opts.rowWidthMax()),
-            // max cells per mutation
-            opts.max_cells_per_mutation)
-        );
-    
-    while(true) {
+    DataWriter dw = new DataWriter(c.createBatchWriter(opts.getTableName(), batch_writer_opts.getBatchWriterConfig()), new RandomMutations(
+    // rows
+        new RandomByteArrays(new RandomWithinRange(opts.row_seed, opts.rowMin(), opts.rowMax())),
+        // cfs
+        new RandomByteArrays(new RandomWithinRange(opts.cf_seed, opts.cfMin(), opts.cfMax())),
+        // cqs
+        new RandomByteArrays(new RandomWithinRange(opts.cq_seed, opts.cqMin(), opts.cqMax())),
+        // vals
+        new RandomByteArrays(new RandomWithinRange(opts.value_seed, opts.valueMin(), opts.valueMax())),
+        // number of cells per row
+        new RandomWithinRange(opts.row_width_seed, opts.rowWidthMin(), opts.rowWidthMax()),
+        // max cells per mutation
+        opts.max_cells_per_mutation));
+
+    while (true) {
       dw.next();
       if (writeDelay > 0) {
         Thread.sleep(writeDelay);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/WriteOptions.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/WriteOptions.java b/test/src/main/java/org/apache/accumulo/test/stress/random/WriteOptions.java
index 3e6e647..f92a9eb 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/WriteOptions.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/WriteOptions.java
@@ -23,52 +23,52 @@ import com.beust.jcommander.Parameter;
 class WriteOptions extends ClientOnDefaultTable {
   static final String DEFAULT_TABLE = "stress_test";
   static final int DEFAULT_MIN = 1, DEFAULT_MAX = 128, DEFAULT_SPREAD = DEFAULT_MAX - DEFAULT_MIN;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--min-row-size", description = "minimum row size")
   Integer row_min;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--max-row-size", description = "maximum row size")
   Integer row_max;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--min-cf-size", description = "minimum column family size")
   Integer cf_min;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--max-cf-size", description = "maximum column family size")
   Integer cf_max;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--min-cq-size", description = "minimum column qualifier size")
   Integer cq_min;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--max-cq-size", description = "maximum column qualifier size")
   Integer cq_max;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--min-value-size", description = "minimum value size")
   Integer value_min;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--max-value-size", description = "maximum value size")
   Integer value_max;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--min-row-width", description = "minimum row width")
   Integer row_width_min;
-  
+
   @Parameter(validateValueWith = IntArgValidator.class, names = "--max-row-width", description = "maximum row width")
   Integer row_width_max;
-  
+
   @Parameter(names = "--clear-table", description = "clears the table before ingesting")
   boolean clear_table;
-  
+
   @Parameter(names = "--row-seed", description = "seed for generating rows")
   int row_seed = 87;
-  
+
   @Parameter(names = "--cf-seed", description = "seed for generating column families")
   int cf_seed = 7;
-  
+
   @Parameter(names = "--cq-seed", description = "seed for generating column qualifiers")
   int cq_seed = 43;
-  
+
   @Parameter(names = "--value-seed", description = "seed for generating values")
   int value_seed = 99;
-  
+
   @Parameter(names = "--row-width-seed", description = "seed for generating the number of cells within a row (a row's \"width\")")
   int row_width_seed = 444;
 
@@ -81,15 +81,15 @@ class WriteOptions extends ClientOnDefaultTable {
   public WriteOptions(String table) {
     super(table);
   }
-  
+
   public WriteOptions() {
     this(DEFAULT_TABLE);
   }
-  
+
   private static int minOrDefault(Integer ref) {
     return ref == null ? DEFAULT_MIN : ref;
   }
-  
+
   private static int calculateMax(Integer min_ref, Integer max_ref) {
     if (max_ref == null) {
       if (min_ref == null) {
@@ -101,74 +101,68 @@ class WriteOptions extends ClientOnDefaultTable {
       return max_ref;
     }
   }
-  
+
   public void check() {
     checkPair("ROW", row_min, row_max);
     checkPair("COLUMN FAMILY", cf_min, cf_max);
     checkPair("COLUMN QUALIFIER", cq_min, cq_max);
     checkPair("VALUE", value_min, value_max);
   }
-  
+
   public void checkPair(String label, Integer min_ref, Integer max_ref) {
     // we've already asserted that the numbers will either be
     // 1) null
     // 2) positive
     // need to verify that they're coherent here
-    
-    if(min_ref == null && max_ref != null) {
+
+    if (min_ref == null && max_ref != null) {
       // we don't support just specifying a max yet
-      throw new IllegalArgumentException(
-          String.format("[%s] Maximum value supplied, but no minimum. Must supply a minimum with a maximum value.",
-                        label));
-    } else if(min_ref != null && max_ref != null) {
+      throw new IllegalArgumentException(String.format("[%s] Maximum value supplied, but no minimum. Must supply a minimum with a maximum value.", label));
+    } else if (min_ref != null && max_ref != null) {
       // if a user supplied lower and upper bounds, we need to verify
       // that min <= max
-      if(min_ref.compareTo(max_ref) > 0) {
-        throw new IllegalArgumentException(
-            String.format("[%s] Min value (%d) is greater than max value (%d)",
-                          label,
-                          min_ref,
-                          max_ref));
+      if (min_ref.compareTo(max_ref) > 0) {
+        throw new IllegalArgumentException(String.format("[%s] Min value (%d) is greater than max value (%d)", label, min_ref, max_ref));
       }
     }
   }
-  
+
   public int rowMin() {
     return minOrDefault(row_min);
   }
-  
+
   public int rowMax() {
     return calculateMax(row_min, row_max);
   }
-  
+
   public int cfMin() {
     return minOrDefault(cf_min);
   }
-  
+
   public int cfMax() {
     return calculateMax(cf_min, cf_max);
   }
-  
+
   public int cqMin() {
     return minOrDefault(cq_min);
   }
-  
+
   public int cqMax() {
     return calculateMax(cq_min, cq_max);
   }
-  
+
   public int valueMin() {
     return minOrDefault(value_min);
   }
-  
+
   public int valueMax() {
     return calculateMax(value_min, value_max);
   }
-  
+
   public int rowWidthMin() {
     return minOrDefault(row_width_min);
   }
-  
+
   public int rowWidthMax() {
     return calculateMax(row_width_min, row_width_max);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/stress/random/package-info.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/stress/random/package-info.java b/test/src/main/java/org/apache/accumulo/test/stress/random/package-info.java
index 488a279..fdbf72e 100644
--- a/test/src/main/java/org/apache/accumulo/test/stress/random/package-info.java
+++ b/test/src/main/java/org/apache/accumulo/test/stress/random/package-info.java
@@ -17,19 +17,20 @@
 /**
  * This package contains utility classes designed to test Accumulo when large cells are being written. This is an attempt to observe the behavior Accumulo
  * displays when compacting and reading these cells.
- * 
+ *
  * There are two components to this package: {@link org.apache.accumulo.test.stress.random.Write} and {@link org.apache.accumulo.test.stress.random.Scan}.
- * 
+ *
  * The {@link org.apache.accumulo.test.stress.random.Write} provides facilities for writing random sized cells. Users can configure minimum and maximum
  * sized portions of a cell. The portions users can configure are the row, column family, column qualifier and value. Note that the sizes are uniformly
  * distributed between the minimum and maximum values. See {@link org.apache.accumulo.test.stress.random.WriteOptions} for available options and default sizing
  * information.
- * 
+ *
  * The Scan provides users with the ability to query tables generated by the Write. It will pick a tablet at random and scan the entire range. The
  * amount of times this process is done is user configurable. By default, it happens 1,024 times. Users can also specify whether or not the scan should be
  * isolated or not.
- * 
+ *
  * There is no shared state intended by either of these services. This allows multiple clients to be run in parallel, either on the same host or distributed
  * across hosts.
  */
 package org.apache.accumulo.test.stress.random;
+

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/fate/zookeeper/ZooLockTest.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/fate/zookeeper/ZooLockTest.java b/test/src/test/java/org/apache/accumulo/fate/zookeeper/ZooLockTest.java
index 4045087..2ad8161 100644
--- a/test/src/test/java/org/apache/accumulo/fate/zookeeper/ZooLockTest.java
+++ b/test/src/test/java/org/apache/accumulo/fate/zookeeper/ZooLockTest.java
@@ -40,7 +40,7 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
 /**
- * 
+ *
  */
 public class ZooLockTest {
 
@@ -155,7 +155,7 @@ public class ZooLockTest {
   @Test(timeout = 10000)
   public void testNoParent() throws Exception {
     String parent = "/zltest-" + this.hashCode() + "-l" + pdCount.incrementAndGet();
- 
+
     ZooLock zl = new ZooLock(accumulo.getZooKeepers(), 30000, "digest", "secret".getBytes(), parent);
 
     Assert.assertFalse(zl.isLocked());
@@ -175,7 +175,7 @@ public class ZooLockTest {
   @Test(timeout = 10000)
   public void testDeleteLock() throws Exception {
     String parent = "/zltest-" + this.hashCode() + "-l" + pdCount.incrementAndGet();
- 
+
     ZooReaderWriter zk = ZooReaderWriter.getInstance(accumulo.getZooKeepers(), 30000, "digest", "secret".getBytes());
     zk.mkdirs(parent);
 
@@ -206,7 +206,7 @@ public class ZooLockTest {
   @Test(timeout = 10000)
   public void testDeleteWaiting() throws Exception {
     String parent = "/zltest-" + this.hashCode() + "-l" + pdCount.incrementAndGet();
- 
+
     ZooReaderWriter zk = ZooReaderWriter.getInstance(accumulo.getZooKeepers(), 30000, "digest", "secret".getBytes());
     zk.mkdirs(parent);
 
@@ -280,7 +280,7 @@ public class ZooLockTest {
     while (!watcher.isConnected()) {
       Thread.sleep(200);
     }
-    
+
     zk.create(parent, new byte[0], ZooDefs.Ids.OPEN_ACL_UNSAFE, CreateMode.PERSISTENT);
 
     ZooLock zl = new ZooLock(accumulo.getZooKeepers(), 30000, "digest", "secret".getBytes(), parent);
@@ -318,7 +318,7 @@ public class ZooLockTest {
     String parent = "/zltest-" + this.hashCode() + "-l" + pdCount.incrementAndGet();
 
     ZooLock zl = new ZooLock(accumulo.getZooKeepers(), 1000, "digest", "secret".getBytes(), parent);
-    
+
     ConnectedWatcher watcher = new ConnectedWatcher();
     ZooKeeper zk = new ZooKeeper(accumulo.getZooKeepers(), 1000, watcher);
     zk.addAuthInfo("digest", "secret".getBytes());
@@ -326,7 +326,7 @@ public class ZooLockTest {
     while (!watcher.isConnected()) {
       Thread.sleep(200);
     }
-    
+
     for (int i = 0; i < 10; i++) {
       zk.create(parent, new byte[0], ZooDefs.Ids.OPEN_ACL_UNSAFE, CreateMode.PERSISTENT);
       zk.delete(parent, -1);
@@ -360,7 +360,7 @@ public class ZooLockTest {
     while (!watcher.isConnected()) {
       Thread.sleep(200);
     }
-    
+
     zk.create(parent, new byte[0], ZooDefs.Ids.OPEN_ACL_UNSAFE, CreateMode.PERSISTENT);
 
     ZooLock zl = new ZooLock(accumulo.getZooKeepers(), 1000, "digest", "secret".getBytes(), parent);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/proxy/ProxyDurabilityIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/proxy/ProxyDurabilityIT.java b/test/src/test/java/org/apache/accumulo/proxy/ProxyDurabilityIT.java
index f8bcbfb..c632b47 100644
--- a/test/src/test/java/org/apache/accumulo/proxy/ProxyDurabilityIT.java
+++ b/test/src/test/java/org/apache/accumulo/proxy/ProxyDurabilityIT.java
@@ -54,18 +54,18 @@ import org.apache.thrift.server.TServer;
 import org.junit.Test;
 
 public class ProxyDurabilityIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     hadoopCoreSite.set("fs.file.impl", RawLocalFileSystem.class.getName());
     cfg.setProperty(Property.INSTANCE_ZK_TIMEOUT, "5s");
     cfg.setNumTservers(1);
   }
-  
+
   private static ByteBuffer bytes(String value) {
     return ByteBuffer.wrap(value.getBytes());
   }
-  
+
   @Test
   public void testDurability() throws Exception {
     Connector c = getConnector();
@@ -77,8 +77,8 @@ public class ProxyDurabilityIT extends ConfigurableMacIT {
     Class<Factory> protocolClass = org.apache.thrift.protocol.TJSONProtocol.Factory.class;
 
     int proxyPort = PortUtils.getRandomFreePort();
-    final TServer proxyServer = Proxy.createProxyServer(org.apache.accumulo.proxy.thrift.AccumuloProxy.class, org.apache.accumulo.proxy.ProxyServer.class, proxyPort,
-        protocolClass, props);
+    final TServer proxyServer = Proxy.createProxyServer(org.apache.accumulo.proxy.thrift.AccumuloProxy.class, org.apache.accumulo.proxy.ProxyServer.class,
+        proxyPort, protocolClass, props);
     Thread thread = new Thread() {
       @Override
       public void run() {
@@ -92,15 +92,15 @@ public class ProxyDurabilityIT extends ConfigurableMacIT {
     Map<String,String> properties = new TreeMap<String,String>();
     properties.put("password", ROOT_PASSWORD);
     ByteBuffer login = client.login("root", properties);
-    
+
     String tableName = getUniqueNames(1)[0];
     client.createTable(login, tableName, true, TimeType.MILLIS);
     assertTrue(c.tableOperations().exists(tableName));
-    
+
     WriterOptions options = new WriterOptions();
     options.setDurability(Durability.NONE);
     String writer = client.createWriter(login, tableName, options);
-    Map<ByteBuffer,List<ColumnUpdate>> cells = new TreeMap<ByteBuffer, List<ColumnUpdate>>();
+    Map<ByteBuffer,List<ColumnUpdate>> cells = new TreeMap<ByteBuffer,List<ColumnUpdate>>();
     ColumnUpdate column = new ColumnUpdate(bytes("cf"), bytes("cq"));
     column.setValue("value".getBytes());
     cells.put(bytes("row"), Collections.singletonList(column));
@@ -109,7 +109,7 @@ public class ProxyDurabilityIT extends ConfigurableMacIT {
     assertEquals(1, count(tableName));
     restartTServer();
     assertEquals(0, count(tableName));
-    
+
     ConditionalWriterOptions cfg = new ConditionalWriterOptions();
     cfg.setDurability(Durability.LOG);
     String cwriter = client.createConditionalWriter(login, tableName, cfg);
@@ -121,7 +121,7 @@ public class ProxyDurabilityIT extends ConfigurableMacIT {
     assertEquals(1, count(tableName));
     restartTServer();
     assertEquals(0, count(tableName));
-    
+
     proxyServer.stop();
     thread.join();
   }
@@ -137,5 +137,5 @@ public class ProxyDurabilityIT extends ConfigurableMacIT {
     Connector c = getConnector();
     return FunctionalTestUtils.count(c.createScanner(tableName, Authorizations.EMPTY));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/proxy/SimpleProxyIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/proxy/SimpleProxyIT.java b/test/src/test/java/org/apache/accumulo/proxy/SimpleProxyIT.java
index c075075..f55e593 100644
--- a/test/src/test/java/org/apache/accumulo/proxy/SimpleProxyIT.java
+++ b/test/src/test/java/org/apache/accumulo/proxy/SimpleProxyIT.java
@@ -186,8 +186,9 @@ public class SimpleProxyIT {
     // wait for accumulo to be up and functional
     ZooKeeperInstance zoo = new ZooKeeperInstance(accumulo.getInstanceName(), accumulo.getZooKeepers());
     Connector c = zoo.getConnector("root", new PasswordToken(secret.getBytes()));
-    for (@SuppressWarnings("unused") Entry<org.apache.accumulo.core.data.Key,Value> entry : c.createScanner(MetadataTable.NAME, Authorizations.EMPTY))
-        ;
+    for (@SuppressWarnings("unused")
+    Entry<org.apache.accumulo.core.data.Key,Value> entry : c.createScanner(MetadataTable.NAME, Authorizations.EMPTY))
+      ;
 
     Properties props = new Properties();
     props.put("instance", accumulo.getConfig().getInstanceName());
@@ -1599,8 +1600,8 @@ public class SimpleProxyIT {
 
     client.createTable(creds, tableName, true, TimeType.MILLIS);
 
-    client.setProperty(creds, Property.VFS_CONTEXT_CLASSPATH_PROPERTY.getKey() + "context1",
-        System.getProperty("user.dir") + "/src/test/resources/TestCompactionStrat.jar");
+    client.setProperty(creds, Property.VFS_CONTEXT_CLASSPATH_PROPERTY.getKey() + "context1", System.getProperty("user.dir")
+        + "/src/test/resources/TestCompactionStrat.jar");
     client.setTableProperty(creds, tableName, Property.TABLE_CLASSPATH.getKey(), "context1");
 
     client.addSplits(creds, tableName, Collections.singleton(s2bb("efg")));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/proxy/TestProxyInstanceOperations.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/proxy/TestProxyInstanceOperations.java b/test/src/test/java/org/apache/accumulo/proxy/TestProxyInstanceOperations.java
index ad21f91..92d065f 100644
--- a/test/src/test/java/org/apache/accumulo/proxy/TestProxyInstanceOperations.java
+++ b/test/src/test/java/org/apache/accumulo/proxy/TestProxyInstanceOperations.java
@@ -38,14 +38,14 @@ public class TestProxyInstanceOperations {
   protected static TestProxyClient tpc;
   protected static ByteBuffer userpass;
   protected static final int port = 10197;
-  
+
   @SuppressWarnings("serial")
   @BeforeClass
   public static void setup() throws Exception {
     Properties prop = new Properties();
     prop.setProperty("useMockInstance", "true");
     prop.put("tokenClass", PasswordToken.class.getName());
-    
+
     proxy = Proxy.createProxyServer(Class.forName("org.apache.accumulo.proxy.thrift.AccumuloProxy"), Class.forName("org.apache.accumulo.proxy.ProxyServer"),
         port, TCompactProtocol.Factory.class, prop);
     thread = new Thread() {
@@ -56,28 +56,32 @@ public class TestProxyInstanceOperations {
     };
     thread.start();
     tpc = new TestProxyClient("localhost", port);
-    userpass = tpc.proxy.login("root", new TreeMap<String, String>() {{ put("password",""); }});
+    userpass = tpc.proxy.login("root", new TreeMap<String,String>() {
+      {
+        put("password", "");
+      }
+    });
   }
-  
+
   @AfterClass
   public static void tearDown() throws InterruptedException {
     proxy.stop();
     thread.join();
   }
-  
+
   @Test
   public void properties() throws TException {
     tpc.proxy().setProperty(userpass, "test.systemprop", "whistletips");
-    
+
     assertEquals(tpc.proxy().getSystemConfiguration(userpass).get("test.systemprop"), "whistletips");
     tpc.proxy().removeProperty(userpass, "test.systemprop");
     assertNull(tpc.proxy().getSystemConfiguration(userpass).get("test.systemprop"));
-    
+
   }
-  
+
   @Test
   public void testClassLoad() throws TException {
     assertTrue(tpc.proxy().testClassLoad(userpass, "org.apache.accumulo.core.iterators.user.RegExFilter", "org.apache.accumulo.core.iterators.Filter"));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/proxy/TestProxyReadWrite.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/proxy/TestProxyReadWrite.java b/test/src/test/java/org/apache/accumulo/proxy/TestProxyReadWrite.java
index c0049a0..4528a0e 100644
--- a/test/src/test/java/org/apache/accumulo/proxy/TestProxyReadWrite.java
+++ b/test/src/test/java/org/apache/accumulo/proxy/TestProxyReadWrite.java
@@ -55,14 +55,14 @@ public class TestProxyReadWrite {
   protected static ByteBuffer userpass;
   protected static final int port = 10194;
   protected static final String testtable = "testtable";
-  
+
   @SuppressWarnings("serial")
   @BeforeClass
   public static void setup() throws Exception {
     Properties prop = new Properties();
     prop.setProperty("useMockInstance", "true");
     prop.put("tokenClass", PasswordToken.class.getName());
-    
+
     proxy = Proxy.createProxyServer(Class.forName("org.apache.accumulo.proxy.thrift.AccumuloProxy"), Class.forName("org.apache.accumulo.proxy.ProxyServer"),
         port, TCompactProtocol.Factory.class, prop);
     thread = new Thread() {
@@ -73,38 +73,42 @@ public class TestProxyReadWrite {
     };
     thread.start();
     tpc = new TestProxyClient("localhost", port);
-    userpass = tpc.proxy().login("root", new TreeMap<String, String>() {{put("password",""); }});
+    userpass = tpc.proxy().login("root", new TreeMap<String,String>() {
+      {
+        put("password", "");
+      }
+    });
   }
-  
+
   @AfterClass
   public static void tearDown() throws InterruptedException {
     proxy.stop();
     thread.join();
   }
-  
+
   @Before
   public void makeTestTable() throws Exception {
     tpc.proxy().createTable(userpass, testtable, true, TimeType.MILLIS);
   }
-  
+
   @After
   public void deleteTestTable() throws Exception {
     tpc.proxy().deleteTable(userpass, testtable);
   }
-  
+
   private static void addMutation(Map<ByteBuffer,List<ColumnUpdate>> mutations, String row, String cf, String cq, String value) {
     ColumnUpdate update = new ColumnUpdate(ByteBuffer.wrap(cf.getBytes()), ByteBuffer.wrap(cq.getBytes()));
     update.setValue(value.getBytes());
     mutations.put(ByteBuffer.wrap(row.getBytes()), Collections.singletonList(update));
   }
-  
+
   private static void addMutation(Map<ByteBuffer,List<ColumnUpdate>> mutations, String row, String cf, String cq, String vis, String value) {
     ColumnUpdate update = new ColumnUpdate(ByteBuffer.wrap(cf.getBytes()), ByteBuffer.wrap(cq.getBytes()));
     update.setValue(value.getBytes());
     update.setColVisibility(vis.getBytes());
     mutations.put(ByteBuffer.wrap(row.getBytes()), Collections.singletonList(update));
   }
-  
+
   /**
    * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Set a range so only the entries between -Inf...5 come back (there should be
    * 50,000)
@@ -116,22 +120,22 @@ public class TestProxyReadWrite {
     String format = "%1$05d";
     for (int i = 0; i < maxInserts; i++) {
       addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().updateAndFlush(userpass, testtable, mutations);
         mutations.clear();
       }
     }
-    
+
     Key stop = new Key();
     stop.setRow("5".getBytes());
     BatchScanOptions options = new BatchScanOptions();
     options.ranges = Collections.singletonList(new Range(null, false, stop, false));
     String cookie = tpc.proxy().createBatchScanner(userpass, testtable, options);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
@@ -142,8 +146,8 @@ public class TestProxyReadWrite {
   }
 
   /**
-   * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Set a columnFamily so only the entries with specified column family come back (there should be
-   * 50,000)
+   * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Set a columnFamily so only the entries with specified column family come back
+   * (there should be 50,000)
    */
   @Test
   public void readWriteBatchOneShotWithColumnFamilyOnly() throws Exception {
@@ -151,26 +155,26 @@ public class TestProxyReadWrite {
     Map<ByteBuffer,List<ColumnUpdate>> mutations = new HashMap<ByteBuffer,List<ColumnUpdate>>();
     String format = "%1$05d";
     for (int i = 0; i < maxInserts; i++) {
-	
-      addMutation(mutations, String.format(format, i), "cf" + (i % 2) , "cq" + (i % 2), Util.randString(10));
-      
+
+      addMutation(mutations, String.format(format, i), "cf" + (i % 2), "cq" + (i % 2), Util.randString(10));
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().updateAndFlush(userpass, testtable, mutations);
         mutations.clear();
       }
     }
-    
+
     BatchScanOptions options = new BatchScanOptions();
 
-	ScanColumn sc = new ScanColumn();
-	sc.colFamily = ByteBuffer.wrap("cf0".getBytes());
+    ScanColumn sc = new ScanColumn();
+    sc.colFamily = ByteBuffer.wrap("cf0".getBytes());
 
     options.columns = Collections.singletonList(sc);
     String cookie = tpc.proxy().createBatchScanner(userpass, testtable, options);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
@@ -180,9 +184,8 @@ public class TestProxyReadWrite {
     assertEquals(i, 50000);
   }
 
-
   /**
-   * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Set a columnFamily + columnQualififer so only the entries with specified column 
+   * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Set a columnFamily + columnQualififer so only the entries with specified column
    * come back (there should be 50,000)
    */
   @Test
@@ -191,27 +194,27 @@ public class TestProxyReadWrite {
     Map<ByteBuffer,List<ColumnUpdate>> mutations = new HashMap<ByteBuffer,List<ColumnUpdate>>();
     String format = "%1$05d";
     for (int i = 0; i < maxInserts; i++) {
-	
-      addMutation(mutations, String.format(format, i), "cf" + (i % 2) , "cq" + (i % 2), Util.randString(10));
-      
+
+      addMutation(mutations, String.format(format, i), "cf" + (i % 2), "cq" + (i % 2), Util.randString(10));
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().updateAndFlush(userpass, testtable, mutations);
         mutations.clear();
       }
     }
-    
+
     BatchScanOptions options = new BatchScanOptions();
 
-	ScanColumn sc = new ScanColumn();
-	sc.colFamily = ByteBuffer.wrap("cf0".getBytes());
-	sc.colQualifier = ByteBuffer.wrap("cq0".getBytes());
+    ScanColumn sc = new ScanColumn();
+    sc.colFamily = ByteBuffer.wrap("cf0".getBytes());
+    sc.colQualifier = ByteBuffer.wrap("cq0".getBytes());
 
     options.columns = Collections.singletonList(sc);
     String cookie = tpc.proxy().createBatchScanner(userpass, testtable, options);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
@@ -221,7 +224,6 @@ public class TestProxyReadWrite {
     assertEquals(i, 50000);
   }
 
-
   /**
    * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Filter the results so only the even numbers come back.
    */
@@ -232,39 +234,39 @@ public class TestProxyReadWrite {
     String format = "%1$05d";
     for (int i = 0; i < maxInserts; i++) {
       addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().updateAndFlush(userpass, testtable, mutations);
         mutations.clear();
       }
-      
+
     }
-    
+
     String regex = ".*[02468]";
-    
+
     org.apache.accumulo.core.client.IteratorSetting is = new org.apache.accumulo.core.client.IteratorSetting(50, regex, RegExFilter.class);
     RegExFilter.setRegexs(is, regex, null, null, null, false);
-    
+
     IteratorSetting pis = Util.iteratorSetting2ProxyIteratorSetting(is);
     ScanOptions opts = new ScanOptions();
     opts.iterators = Collections.singletonList(pis);
     String cookie = tpc.proxy().createScanner(userpass, testtable, opts);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
       for (KeyValue kv : kvList.getResults()) {
         assertEquals(Integer.parseInt(new String(kv.getKey().getRow())), i);
-        
+
         i += 2;
       }
       hasNext = kvList.isMore();
     }
   }
-  
+
   @Test
   public void readWriteOneShotWithRange() throws Exception {
     int maxInserts = 100000;
@@ -272,22 +274,22 @@ public class TestProxyReadWrite {
     String format = "%1$05d";
     for (int i = 0; i < maxInserts; i++) {
       addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().updateAndFlush(userpass, testtable, mutations);
         mutations.clear();
       }
     }
-    
+
     Key stop = new Key();
     stop.setRow("5".getBytes());
     ScanOptions opts = new ScanOptions();
     opts.range = new Range(null, false, stop, false);
     String cookie = tpc.proxy().createScanner(userpass, testtable, opts);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
@@ -296,7 +298,7 @@ public class TestProxyReadWrite {
     }
     assertEquals(i, 50000);
   }
-  
+
   /**
    * Insert 100000 cells which have as the row [0..99999] (padded with zeros). Filter the results so only the even numbers come back.
    */
@@ -307,41 +309,41 @@ public class TestProxyReadWrite {
     String format = "%1$05d";
     for (int i = 0; i < maxInserts; i++) {
       addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
-        
+
         tpc.proxy().updateAndFlush(userpass, testtable, mutations);
         mutations.clear();
-        
+
       }
-      
+
     }
-    
+
     String regex = ".*[02468]";
-    
+
     org.apache.accumulo.core.client.IteratorSetting is = new org.apache.accumulo.core.client.IteratorSetting(50, regex, RegExFilter.class);
     RegExFilter.setRegexs(is, regex, null, null, null, false);
-    
+
     IteratorSetting pis = Util.iteratorSetting2ProxyIteratorSetting(is);
     ScanOptions opts = new ScanOptions();
     opts.iterators = Collections.singletonList(pis);
     String cookie = tpc.proxy().createScanner(userpass, testtable, opts);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
       for (KeyValue kv : kvList.getResults()) {
         assertEquals(Integer.parseInt(new String(kv.getKey().getRow())), i);
-        
+
         i += 2;
       }
       hasNext = kvList.isMore();
     }
   }
-  
+
   // @Test
   // This test takes kind of a long time. Enable it if you think you may have memory issues.
   public void manyWritesAndReads() throws Exception {
@@ -351,24 +353,24 @@ public class TestProxyReadWrite {
     String writer = tpc.proxy().createWriter(userpass, testtable, null);
     for (int i = 0; i < maxInserts; i++) {
       addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
-        
+
         tpc.proxy().update(writer, mutations);
         mutations.clear();
-        
+
       }
-      
+
     }
-    
+
     tpc.proxy().flush(writer);
     tpc.proxy().closeWriter(writer);
-    
+
     String cookie = tpc.proxy().createScanner(userpass, testtable, null);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
@@ -382,7 +384,7 @@ public class TestProxyReadWrite {
     }
     assertEquals(maxInserts, i);
   }
-  
+
   @Test
   public void asynchReadWrite() throws Exception {
     int maxInserts = 10000;
@@ -391,29 +393,29 @@ public class TestProxyReadWrite {
     String writer = tpc.proxy().createWriter(userpass, testtable, null);
     for (int i = 0; i < maxInserts; i++) {
       addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().update(writer, mutations);
         mutations.clear();
       }
     }
-    
+
     tpc.proxy().flush(writer);
     tpc.proxy().closeWriter(writer);
-    
+
     String regex = ".*[02468]";
-    
+
     org.apache.accumulo.core.client.IteratorSetting is = new org.apache.accumulo.core.client.IteratorSetting(50, regex, RegExFilter.class);
     RegExFilter.setRegexs(is, regex, null, null, null, false);
-    
+
     IteratorSetting pis = Util.iteratorSetting2ProxyIteratorSetting(is);
     ScanOptions opts = new ScanOptions();
     opts.iterators = Collections.singletonList(pis);
     String cookie = tpc.proxy().createScanner(userpass, testtable, opts);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     int numRead = 0;
     while (hasNext) {
@@ -427,14 +429,14 @@ public class TestProxyReadWrite {
     }
     assertEquals(maxInserts / 2, numRead);
   }
-  
+
   @Test
   public void testVisibility() throws Exception {
-    
+
     Set<ByteBuffer> auths = new HashSet<ByteBuffer>();
     auths.add(ByteBuffer.wrap("even".getBytes()));
     tpc.proxy().changeUserAuthorizations(userpass, "root", auths);
-    
+
     int maxInserts = 10000;
     Map<ByteBuffer,List<ColumnUpdate>> mutations = new HashMap<ByteBuffer,List<ColumnUpdate>>();
     String format = "%1$05d";
@@ -444,22 +446,22 @@ public class TestProxyReadWrite {
         addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, "even", Util.randString(10));
       else
         addMutation(mutations, String.format(format, i), "cf" + i, "cq" + i, "odd", Util.randString(10));
-      
+
       if (i % 1000 == 0 || i == maxInserts - 1) {
         tpc.proxy().update(writer, mutations);
         mutations.clear();
       }
     }
-    
+
     tpc.proxy().flush(writer);
     tpc.proxy().closeWriter(writer);
     ScanOptions opts = new ScanOptions();
     opts.authorizations = auths;
     String cookie = tpc.proxy().createScanner(userpass, testtable, opts);
-    
+
     int i = 0;
     boolean hasNext = true;
-    
+
     int k = 1000;
     int numRead = 0;
     while (hasNext) {
@@ -470,9 +472,9 @@ public class TestProxyReadWrite {
         numRead++;
       }
       hasNext = kvList.isMore();
-      
+
     }
     assertEquals(maxInserts / 2, numRead);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/proxy/TestProxySecurityOperations.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/proxy/TestProxySecurityOperations.java b/test/src/test/java/org/apache/accumulo/proxy/TestProxySecurityOperations.java
index e0b17ac..1a87200 100644
--- a/test/src/test/java/org/apache/accumulo/proxy/TestProxySecurityOperations.java
+++ b/test/src/test/java/org/apache/accumulo/proxy/TestProxySecurityOperations.java
@@ -50,13 +50,13 @@ public class TestProxySecurityOperations {
   protected static final String testtable = "testtable";
   protected static final String testuser = "VonJines";
   protected static final ByteBuffer testpw = ByteBuffer.wrap("fiveones".getBytes());
-  
+
   @BeforeClass
   public static void setup() throws Exception {
     Properties prop = new Properties();
     prop.setProperty("useMockInstance", "true");
     prop.put("tokenClass", PasswordToken.class.getName());
-    
+
     proxy = Proxy.createProxyServer(Class.forName("org.apache.accumulo.proxy.thrift.AccumuloProxy"), Class.forName("org.apache.accumulo.proxy.ProxyServer"),
         port, TCompactProtocol.Factory.class, prop);
     thread = new Thread() {
@@ -66,35 +66,35 @@ public class TestProxySecurityOperations {
       }
     };
     thread.start();
-    
+
     tpc = new TestProxyClient("localhost", port);
     userpass = tpc.proxy().login("root", new TreeMap<String,String>() {
       private static final long serialVersionUID = 1L;
-      
+
       {
         put("password", "");
       }
     });
   }
-  
+
   @AfterClass
   public static void tearDown() throws InterruptedException {
     proxy.stop();
     thread.join();
   }
-  
+
   @Before
   public void makeTestTableAndUser() throws Exception {
     tpc.proxy().createTable(userpass, testtable, true, TimeType.MILLIS);
     tpc.proxy().createLocalUser(userpass, testuser, testpw);
   }
-  
+
   @After
   public void deleteTestTable() throws Exception {
     tpc.proxy().deleteTable(userpass, testtable);
     tpc.proxy().dropLocalUser(userpass, testuser);
   }
-  
+
   @Test
   public void create() throws TException {
     tpc.proxy().createLocalUser(userpass, testuser + "2", testpw);
@@ -102,38 +102,38 @@ public class TestProxySecurityOperations {
     tpc.proxy().dropLocalUser(userpass, testuser + "2");
     assertTrue(!tpc.proxy().listLocalUsers(userpass).contains(testuser + "2"));
   }
-  
+
   @Test
   public void authenticate() throws TException {
     assertTrue(tpc.proxy().authenticateUser(userpass, testuser, bb2pp(testpw)));
     assertFalse(tpc.proxy().authenticateUser(userpass, "EvilUser", bb2pp(testpw)));
-    
+
     tpc.proxy().changeLocalUserPassword(userpass, testuser, ByteBuffer.wrap("newpass".getBytes()));
     assertFalse(tpc.proxy().authenticateUser(userpass, testuser, bb2pp(testpw)));
     assertTrue(tpc.proxy().authenticateUser(userpass, testuser, bb2pp(ByteBuffer.wrap("newpass".getBytes()))));
-    
+
   }
-  
+
   @Test
   public void tablePermissions() throws TException {
     tpc.proxy().grantTablePermission(userpass, testuser, testtable, TablePermission.ALTER_TABLE);
     assertTrue(tpc.proxy().hasTablePermission(userpass, testuser, testtable, TablePermission.ALTER_TABLE));
-    
+
     tpc.proxy().revokeTablePermission(userpass, testuser, testtable, TablePermission.ALTER_TABLE);
     assertFalse(tpc.proxy().hasTablePermission(userpass, testuser, testtable, TablePermission.ALTER_TABLE));
-    
+
   }
-  
+
   @Test
   public void systemPermissions() throws TException {
     tpc.proxy().grantSystemPermission(userpass, testuser, SystemPermission.ALTER_USER);
     assertTrue(tpc.proxy().hasSystemPermission(userpass, testuser, SystemPermission.ALTER_USER));
-    
+
     tpc.proxy().revokeSystemPermission(userpass, testuser, SystemPermission.ALTER_USER);
     assertFalse(tpc.proxy().hasSystemPermission(userpass, testuser, SystemPermission.ALTER_USER));
-    
+
   }
-  
+
   @Test
   public void auths() throws TException {
     HashSet<ByteBuffer> newauths = new HashSet<ByteBuffer>();
@@ -142,16 +142,16 @@ public class TestProxySecurityOperations {
     tpc.proxy().changeUserAuthorizations(userpass, testuser, newauths);
     List<ByteBuffer> actualauths = tpc.proxy().getUserAuthorizations(userpass, testuser);
     assertEquals(actualauths.size(), newauths.size());
-    
+
     for (ByteBuffer auth : actualauths) {
       assertTrue(newauths.contains(auth));
     }
   }
-  
+
   private Map<String,String> bb2pp(ByteBuffer cf) {
     Map<String,String> toRet = new TreeMap<String,String>();
     toRet.put("password", ByteBufferUtil.toString(cf));
     return toRet;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/proxy/TestProxyTableOperations.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/proxy/TestProxyTableOperations.java b/test/src/test/java/org/apache/accumulo/proxy/TestProxyTableOperations.java
index 87d3454..dd01af9 100644
--- a/test/src/test/java/org/apache/accumulo/proxy/TestProxyTableOperations.java
+++ b/test/src/test/java/org/apache/accumulo/proxy/TestProxyTableOperations.java
@@ -44,21 +44,21 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 
 public class TestProxyTableOperations {
-  
+
   protected static TServer proxy;
   protected static Thread thread;
   protected static TestProxyClient tpc;
   protected static ByteBuffer userpass;
   protected static final int port = 10195;
   protected static final String testtable = "testtable";
-  
+
   @SuppressWarnings("serial")
   @BeforeClass
   public static void setup() throws Exception {
     Properties prop = new Properties();
     prop.setProperty("useMockInstance", "true");
     prop.put("tokenClass", PasswordToken.class.getName());
-    
+
     proxy = Proxy.createProxyServer(Class.forName("org.apache.accumulo.proxy.thrift.AccumuloProxy"), Class.forName("org.apache.accumulo.proxy.ProxyServer"),
         port, TCompactProtocol.Factory.class, prop);
     thread = new Thread() {
@@ -75,23 +75,23 @@ public class TestProxyTableOperations {
       }
     });
   }
-  
+
   @AfterClass
   public static void tearDown() throws InterruptedException {
     proxy.stop();
     thread.join();
   }
-  
+
   @Before
   public void makeTestTable() throws Exception {
     tpc.proxy().createTable(userpass, testtable, true, TimeType.MILLIS);
   }
-  
+
   @After
   public void deleteTestTable() throws Exception {
     tpc.proxy().deleteTable(userpass, testtable);
   }
-  
+
   @Test
   public void createExistsDelete() throws TException {
     assertFalse(tpc.proxy().tableExists(userpass, "testtable2"));
@@ -100,7 +100,7 @@ public class TestProxyTableOperations {
     tpc.proxy().deleteTable(userpass, "testtable2");
     assertFalse(tpc.proxy().tableExists(userpass, "testtable2"));
   }
-  
+
   @Test
   public void listRename() throws TException {
     assertFalse(tpc.proxy().tableExists(userpass, "testtable2"));
@@ -108,9 +108,9 @@ public class TestProxyTableOperations {
     assertTrue(tpc.proxy().tableExists(userpass, "testtable2"));
     tpc.proxy().renameTable(userpass, "testtable2", testtable);
     assertTrue(tpc.proxy().listTables(userpass).contains("testtable"));
-    
+
   }
-  
+
   // This test does not yet function because the backing Mock instance does not yet support merging
   @Test
   public void merge() throws TException {
@@ -119,19 +119,19 @@ public class TestProxyTableOperations {
     splits.add(ByteBuffer.wrap("c".getBytes()));
     splits.add(ByteBuffer.wrap("z".getBytes()));
     tpc.proxy().addSplits(userpass, testtable, splits);
-    
+
     tpc.proxy().mergeTablets(userpass, testtable, ByteBuffer.wrap("b".getBytes()), ByteBuffer.wrap("d".getBytes()));
-    
+
     splits.remove(ByteBuffer.wrap("c".getBytes()));
-    
+
     List<ByteBuffer> tableSplits = tpc.proxy().listSplits(userpass, testtable, 10);
-    
+
     for (ByteBuffer split : tableSplits)
       assertTrue(splits.contains(split));
     assertTrue(tableSplits.size() == splits.size());
-    
+
   }
-  
+
   @Test
   public void splits() throws TException {
     Set<ByteBuffer> splits = new HashSet<ByteBuffer>();
@@ -139,14 +139,14 @@ public class TestProxyTableOperations {
     splits.add(ByteBuffer.wrap("b".getBytes()));
     splits.add(ByteBuffer.wrap("z".getBytes()));
     tpc.proxy().addSplits(userpass, testtable, splits);
-    
+
     List<ByteBuffer> tableSplits = tpc.proxy().listSplits(userpass, testtable, 10);
-    
+
     for (ByteBuffer split : tableSplits)
       assertTrue(splits.contains(split));
     assertTrue(tableSplits.size() == splits.size());
   }
-  
+
   @Test
   public void constraints() throws TException {
     int cid = tpc.proxy().addConstraint(userpass, testtable, "org.apache.accumulo.TestConstraint");
@@ -156,7 +156,7 @@ public class TestProxyTableOperations {
     constraints = tpc.proxy().listConstraints(userpass, testtable);
     assertNull(constraints.get("org.apache.accumulo.TestConstraint"));
   }
-  
+
   @Test
   public void localityGroups() throws TException {
     Map<String,Set<String>> groups = new HashMap<String,Set<String>>();
@@ -168,9 +168,9 @@ public class TestProxyTableOperations {
     group2.add("cf3");
     groups.put("group2", group2);
     tpc.proxy().setLocalityGroups(userpass, testtable, groups);
-    
+
     Map<String,Set<String>> actualGroups = tpc.proxy().getLocalityGroups(userpass, testtable);
-    
+
     assertEquals(groups.size(), actualGroups.size());
     for (String groupName : groups.keySet()) {
       assertTrue(actualGroups.containsKey(groupName));
@@ -180,7 +180,7 @@ public class TestProxyTableOperations {
       }
     }
   }
-  
+
   @Test
   public void tableProperties() throws TException {
     tpc.proxy().setTableProperty(userpass, testtable, "test.property1", "wharrrgarbl");
@@ -188,13 +188,13 @@ public class TestProxyTableOperations {
     tpc.proxy().removeTableProperty(userpass, testtable, "test.property1");
     assertNull(tpc.proxy().getTableProperties(userpass, testtable).get("test.property1"));
   }
-  
+
   private static void addMutation(Map<ByteBuffer,List<ColumnUpdate>> mutations, String row, String cf, String cq, String value) {
     ColumnUpdate update = new ColumnUpdate(ByteBuffer.wrap(cf.getBytes()), ByteBuffer.wrap(cq.getBytes()));
     update.setValue(value.getBytes());
     mutations.put(ByteBuffer.wrap(row.getBytes()), Collections.singletonList(update));
   }
-  
+
   @Test
   public void tableOperationsRowMethods() throws TException {
     Map<ByteBuffer,List<ColumnUpdate>> mutations = new HashMap<ByteBuffer,List<ColumnUpdate>>();
@@ -202,11 +202,11 @@ public class TestProxyTableOperations {
       addMutation(mutations, "" + i, "cf", "cq", "");
     }
     tpc.proxy().updateAndFlush(userpass, testtable, mutations);
-    
+
     assertEquals(tpc.proxy().getMaxRow(userpass, testtable, null, null, true, null, true), ByteBuffer.wrap("9".getBytes()));
-    
+
     tpc.proxy().deleteRows(userpass, testtable, ByteBuffer.wrap("51".getBytes()), ByteBuffer.wrap("99".getBytes()));
     assertEquals(tpc.proxy().getMaxRow(userpass, testtable, null, null, true, null, true), ByteBuffer.wrap("5".getBytes()));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/server/security/SystemCredentialsIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/server/security/SystemCredentialsIT.java b/test/src/test/java/org/apache/accumulo/server/security/SystemCredentialsIT.java
index fb71f5f..abbe5e6 100644
--- a/test/src/test/java/org/apache/accumulo/server/security/SystemCredentialsIT.java
+++ b/test/src/test/java/org/apache/accumulo/server/security/SystemCredentialsIT.java
@@ -127,9 +127,10 @@ public class SystemCredentialsIT extends ConfigurableMacIT {
       }
     } catch (RuntimeException e) {
       // catch the runtime exception from the scanner iterator
-      if (e.getCause() instanceof AccumuloSecurityException && ((AccumuloSecurityException) e.getCause()).getSecurityErrorCode() == SecurityErrorCode.BAD_CREDENTIALS) {
-          e.printStackTrace(System.err);
-          System.exit(FAIL_CODE);
+      if (e.getCause() instanceof AccumuloSecurityException
+          && ((AccumuloSecurityException) e.getCause()).getSecurityErrorCode() == SecurityErrorCode.BAD_CREDENTIALS) {
+        e.printStackTrace(System.err);
+        System.exit(FAIL_CODE);
       }
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/AssignmentThreadsIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/AssignmentThreadsIT.java b/test/src/test/java/org/apache/accumulo/test/AssignmentThreadsIT.java
index d7a5ac2..ccf234d 100644
--- a/test/src/test/java/org/apache/accumulo/test/AssignmentThreadsIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/AssignmentThreadsIT.java
@@ -14,7 +14,9 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
- package org.apache.accumulo.test;
+package org.apache.accumulo.test;
+
+import static org.junit.Assert.assertTrue;
 
 import java.util.Random;
 import java.util.SortedSet;
@@ -29,8 +31,6 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
-import static org.junit.Assert.assertTrue;
-
 // ACCUMULO-1177
 public class AssignmentThreadsIT extends ConfigurableMacIT {
 
@@ -41,7 +41,7 @@ public class AssignmentThreadsIT extends ConfigurableMacIT {
   }
 
   // [0-9a-f]
-  private final static byte[] HEXCHARS = { 0x30, 0x31, 0x32, 0x33, 0x34, 0x35, 0x36, 0x37, 0x38, 0x39, 0x61, 0x62, 0x63, 0x64, 0x65, 0x66 };
+  private final static byte[] HEXCHARS = {0x30, 0x31, 0x32, 0x33, 0x34, 0x35, 0x36, 0x37, 0x38, 0x39, 0x61, 0x62, 0x63, 0x64, 0x65, 0x66};
   private final static Random random = new Random();
 
   public static byte[] randomHex(int n) {
@@ -50,8 +50,8 @@ public class AssignmentThreadsIT extends ConfigurableMacIT {
     random.nextBytes(binary);
     int count = 0;
     for (byte x : binary) {
-      hex[count++] = HEXCHARS[(x >> 4)&0xf];
-      hex[count++] = HEXCHARS[x&0xf];
+      hex[count++] = HEXCHARS[(x >> 4) & 0xf];
+      hex[count++] = HEXCHARS[x & 0xf];
     }
     return hex;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/BalanceFasterIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/BalanceFasterIT.java b/test/src/test/java/org/apache/accumulo/test/BalanceFasterIT.java
index d4de5e7..2cc5d34 100644
--- a/test/src/test/java/org/apache/accumulo/test/BalanceFasterIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/BalanceFasterIT.java
@@ -16,7 +16,7 @@
  */
 package org.apache.accumulo.test;
 
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertTrue;
 
 import java.util.HashMap;
 import java.util.Iterator;
@@ -41,13 +41,13 @@ import org.junit.Test;
 
 // ACCUMULO-2952
 public class BalanceFasterIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setNumTservers(3);
   }
 
-  @Test(timeout=90*1000)
+  @Test(timeout = 90 * 1000)
   public void test() throws Exception {
     // create a table, add a bunch of splits
     String tableName = getUniqueNames(1)[0];
@@ -59,12 +59,12 @@ public class BalanceFasterIT extends ConfigurableMacIT {
     }
     conn.tableOperations().addSplits(tableName, splits);
     // give a short wait for balancing
-    UtilWaitThread.sleep(10*1000);
+    UtilWaitThread.sleep(10 * 1000);
     // find out where the tabets are
     Scanner s = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     s.fetchColumnFamily(MetadataSchema.TabletsSection.CurrentLocationColumnFamily.NAME);
     s.setRange(MetadataSchema.TabletsSection.getRange());
-    Map<String, Integer> counts = new HashMap<String, Integer>();
+    Map<String,Integer> counts = new HashMap<String,Integer>();
     while (true) {
       int total = 0;
       counts.clear();
@@ -90,5 +90,5 @@ public class BalanceFasterIT extends ConfigurableMacIT {
     assertTrue(Math.abs(a - c) < 3);
     assertTrue(a > 330);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/BulkImportVolumeIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/BulkImportVolumeIT.java b/test/src/test/java/org/apache/accumulo/test/BulkImportVolumeIT.java
index 78cdfe6..62031d5 100644
--- a/test/src/test/java/org/apache/accumulo/test/BulkImportVolumeIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/BulkImportVolumeIT.java
@@ -90,5 +90,4 @@ public class BulkImportVolumeIT extends AccumuloClusterIT {
     assertEquals(1, fs.listStatus(err).length);
   }
 
-
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/ConditionalWriterIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/ConditionalWriterIT.java b/test/src/test/java/org/apache/accumulo/test/ConditionalWriterIT.java
index 262acb6..b68870d 100644
--- a/test/src/test/java/org/apache/accumulo/test/ConditionalWriterIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/ConditionalWriterIT.java
@@ -104,7 +104,7 @@ public class ConditionalWriterIT extends AccumuloClusterIT {
   }
 
   public static long abs(long l) {
-    l = Math.abs(l);  // abs(Long.MIN_VALUE) == Long.MIN_VALUE...
+    l = Math.abs(l); // abs(Long.MIN_VALUE) == Long.MIN_VALUE...
     if (l < 0)
       return 0;
     return l;
@@ -1264,8 +1264,7 @@ public class ConditionalWriterIT extends AccumuloClusterIT {
 
     final Scanner scanner = conn.createScanner("trace", Authorizations.EMPTY);
     scanner.setRange(new Range(new Text(Long.toHexString(root.traceId()))));
-    loop:
-    while (true) {
+    loop: while (true) {
       final StringBuffer finalBuffer = new StringBuffer();
       int traceCount = TraceDump.printTrace(scanner, new Printer() {
         @Override
@@ -1281,8 +1280,7 @@ public class ConditionalWriterIT extends AccumuloClusterIT {
       log.info("Trace output:" + traceOutput);
       if (traceCount > 0) {
         int lastPos = 0;
-        for (String part : "traceTest, startScan,startConditionalUpdate,conditionalUpdate,Check conditions,apply conditional mutations".split(","))
-        {
+        for (String part : "traceTest, startScan,startConditionalUpdate,conditionalUpdate,Check conditions,apply conditional mutations".split(",")) {
           log.info("Looking in trace output for '" + part + "'");
           int pos = traceOutput.indexOf(part);
           if (-1 == pos) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/InterruptibleScannersIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/InterruptibleScannersIT.java b/test/src/test/java/org/apache/accumulo/test/InterruptibleScannersIT.java
index 4c8bcba..5f9b97a 100644
--- a/test/src/test/java/org/apache/accumulo/test/InterruptibleScannersIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/InterruptibleScannersIT.java
@@ -93,11 +93,10 @@ public class InterruptibleScannersIT extends AccumuloClusterIT {
     thread.start();
     try {
       // Use the scanner, expect problems
-      for (@SuppressWarnings("unused") Entry<Key,Value> entry : scanner) {
-      }
+      for (@SuppressWarnings("unused")
+      Entry<Key,Value> entry : scanner) {}
       Assert.fail("Scan should not succeed");
-    } catch (Exception ex) {
-    } finally {
+    } catch (Exception ex) {} finally {
       thread.join();
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/KeyValueEqualityIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/KeyValueEqualityIT.java b/test/src/test/java/org/apache/accumulo/test/KeyValueEqualityIT.java
index 7dd7e52..1bcd82c 100644
--- a/test/src/test/java/org/apache/accumulo/test/KeyValueEqualityIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/KeyValueEqualityIT.java
@@ -63,7 +63,8 @@ public class KeyValueEqualityIT extends AccumuloClusterIT {
     bw1.close();
     bw2.close();
 
-    Iterator<Entry<Key,Value>> t1 = conn.createScanner(table1, Authorizations.EMPTY).iterator(), t2 = conn.createScanner(table2, Authorizations.EMPTY).iterator();
+    Iterator<Entry<Key,Value>> t1 = conn.createScanner(table1, Authorizations.EMPTY).iterator(), t2 = conn.createScanner(table2, Authorizations.EMPTY)
+        .iterator();
     while (t1.hasNext() && t2.hasNext()) {
       // KeyValue, the implementation of Entry<Key,Value>, should support equality and hashCode properly
       Entry<Key,Value> e1 = t1.next(), e2 = t2.next();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/MetaGetsReadersIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/MetaGetsReadersIT.java b/test/src/test/java/org/apache/accumulo/test/MetaGetsReadersIT.java
index 7bf88bc..7f9983a 100644
--- a/test/src/test/java/org/apache/accumulo/test/MetaGetsReadersIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/MetaGetsReadersIT.java
@@ -17,6 +17,9 @@
 
 package org.apache.accumulo.test;
 
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
+
 import java.util.Iterator;
 import java.util.Map.Entry;
 import java.util.Random;
@@ -39,9 +42,6 @@ import org.apache.accumulo.test.functional.SlowIterator;
 import org.apache.hadoop.conf.Configuration;
 import org.junit.Test;
 
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
-
 public class MetaGetsReadersIT extends ConfigurableMacIT {
 
   @Override
@@ -50,7 +50,7 @@ public class MetaGetsReadersIT extends ConfigurableMacIT {
     cfg.setProperty(Property.TSERV_SCAN_MAX_OPENFILES, "2");
     cfg.setProperty(Property.TABLE_BLOCKCACHE_ENABLED, "false");
   }
-  
+
   private static Thread slowScan(final Connector c, final String tableName, final AtomicBoolean stop) {
     Thread thread = new Thread() {
       public void run() {
@@ -73,7 +73,7 @@ public class MetaGetsReadersIT extends ConfigurableMacIT {
     };
     return thread;
   }
-  
+
   @Test(timeout = 2 * 60 * 1000)
   public void test() throws Exception {
     final String tableName = getUniqueNames(1)[0];
@@ -98,8 +98,8 @@ public class MetaGetsReadersIT extends ConfigurableMacIT {
     UtilWaitThread.sleep(500);
     long now = System.currentTimeMillis();
     Scanner m = c.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
-    for (@SuppressWarnings("unused") Entry<Key,Value> entry : m) {
-    }
+    for (@SuppressWarnings("unused")
+    Entry<Key,Value> entry : m) {}
     long delay = System.currentTimeMillis() - now;
     System.out.println("Delay = " + delay);
     assertTrue("metadata table scan was slow", delay < 1000);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/NamespacesIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/NamespacesIT.java b/test/src/test/java/org/apache/accumulo/test/NamespacesIT.java
index 5bf68c2..c6b9b23 100644
--- a/test/src/test/java/org/apache/accumulo/test/NamespacesIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/NamespacesIT.java
@@ -116,7 +116,7 @@ public class NamespacesIT extends AccumuloIT {
     cluster.getConfig().setNumTservers(1);
     cluster.start();
   }
-  
+
   @Before
   public void setupConnectorAndNamespace() throws Exception {
     // prepare a unique namespace and get a new root connector for each test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/NoMutationRecoveryIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/NoMutationRecoveryIT.java b/test/src/test/java/org/apache/accumulo/test/NoMutationRecoveryIT.java
index b2b24fa..10b8810 100644
--- a/test/src/test/java/org/apache/accumulo/test/NoMutationRecoveryIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/NoMutationRecoveryIT.java
@@ -59,7 +59,7 @@ public class NoMutationRecoveryIT extends AccumuloClusterIT {
     cfg.setNumTservers(1);
   }
 
-  public boolean equals(Entry<Key, Value> a, Entry<Key, Value> b) {
+  public boolean equals(Entry<Key,Value> a, Entry<Key,Value> b) {
     // comparison, without timestamp
     Key akey = a.getKey();
     Key bkey = b.getKey();
@@ -73,7 +73,7 @@ public class NoMutationRecoveryIT extends AccumuloClusterIT {
     conn.tableOperations().create(table);
     String tableId = conn.tableOperations().tableIdMap().get(table);
     update(conn, table, new Text("row"), new Text("cf"), new Text("cq"), new Value("value".getBytes()));
-    Entry<Key, Value> logRef = getLogRef(conn, MetadataTable.NAME);
+    Entry<Key,Value> logRef = getLogRef(conn, MetadataTable.NAME);
     conn.tableOperations().flush(table, null, null, true);
     assertEquals("should not have any refs", 0, FunctionalTestUtils.count(getLogRefs(conn, MetadataTable.NAME, Range.prefix(tableId))));
     conn.securityOperations().grantTablePermission(conn.whoami(), MetadataTable.NAME, TablePermission.WRITE);
@@ -96,7 +96,7 @@ public class NoMutationRecoveryIT extends AccumuloClusterIT {
       count++;
     }
     assertEquals(1, count);
-    for (Entry<Key, Value> ref : getLogRefs(conn, MetadataTable.NAME)) {
+    for (Entry<Key,Value> ref : getLogRefs(conn, MetadataTable.NAME)) {
       assertFalse(equals(ref, logRef));
     }
   }
@@ -106,7 +106,7 @@ public class NoMutationRecoveryIT extends AccumuloClusterIT {
     update(conn, name, k.getRow(), k.getColumnFamily(), k.getColumnQualifier(), logRef.getValue());
   }
 
-  private Iterable<Entry<Key, Value>> getLogRefs(Connector conn, String table) throws Exception {
+  private Iterable<Entry<Key,Value>> getLogRefs(Connector conn, String table) throws Exception {
     return getLogRefs(conn, table, new Range());
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/RecoveryCompactionsAreFlushesIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/RecoveryCompactionsAreFlushesIT.java b/test/src/test/java/org/apache/accumulo/test/RecoveryCompactionsAreFlushesIT.java
index b78e724..49ed92f 100644
--- a/test/src/test/java/org/apache/accumulo/test/RecoveryCompactionsAreFlushesIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/RecoveryCompactionsAreFlushesIT.java
@@ -84,18 +84,17 @@ public class RecoveryCompactionsAreFlushesIT extends AccumuloClusterIT {
     control.startAllServers(ServerType.TABLET_SERVER);
 
     // ensure the table is readable
-    for (@SuppressWarnings("unused") Entry<Key,Value> entry : c.createScanner(tableName, Authorizations.EMPTY)) {
-    }
+    for (@SuppressWarnings("unused")
+    Entry<Key,Value> entry : c.createScanner(tableName, Authorizations.EMPTY)) {}
 
     // ensure that the recovery was not a merging minor compaction
     Scanner s = c.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     s.fetchColumnFamily(MetadataSchema.TabletsSection.DataFileColumnFamily.NAME);
-    for (Entry<Key, Value> entry : s) {
+    for (Entry<Key,Value> entry : s) {
       String filename = entry.getKey().getColumnQualifier().toString();
       String parts[] = filename.split("/");
-      Assert.assertFalse(parts[parts.length-1].startsWith("M"));
+      Assert.assertFalse(parts[parts.length - 1].startsWith("M"));
     }
   }
 
-
 }


[12/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/LookupTask.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/LookupTask.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/LookupTask.java
index 184dc4e..cb341bb 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/LookupTask.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/LookupTask.java
@@ -48,7 +48,7 @@ import org.apache.log4j.Logger;
 public class LookupTask extends ScanTask<MultiScanResult> {
 
   private static final Logger log = Logger.getLogger(LookupTask.class);
-  
+
   private final long scanID;
 
   public LookupTask(TabletServer server, long scanID) {
@@ -109,8 +109,8 @@ public class LookupTask extends ScanTask<MultiScanResult> {
           if (isCancelled())
             interruptFlag.set(true);
 
-          lookupResult = tablet.lookup(entry.getValue(), session.columnSet, session.auths, results, maxResultsSize - bytesAdded, session.ssiList,
-              session.ssio, interruptFlag);
+          lookupResult = tablet.lookup(entry.getValue(), session.columnSet, session.auths, results, maxResultsSize - bytesAdded, session.ssiList, session.ssio,
+              interruptFlag);
 
           // if the tablet was closed it it possible that the
           // interrupt flag was set.... do not want it set for
@@ -170,4 +170,4 @@ public class LookupTask extends ScanTask<MultiScanResult> {
       runState.set(ScanRunState.FINISHED);
     }
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/NextBatchTask.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/NextBatchTask.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/NextBatchTask.java
index da11f14..22ea384 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/NextBatchTask.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/NextBatchTask.java
@@ -30,7 +30,7 @@ import org.apache.log4j.Logger;
 public class NextBatchTask extends ScanTask<ScanBatch> {
 
   private static final Logger log = Logger.getLogger(TabletServer.class);
-  
+
   private final long scanID;
 
   public NextBatchTask(TabletServer server, long scanID, AtomicBoolean interruptFlag) {
@@ -93,4 +93,4 @@ public class NextBatchTask extends ScanTask<ScanBatch> {
     }
 
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanRunState.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanRunState.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanRunState.java
index f2b804c..f5a699b 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanRunState.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanRunState.java
@@ -18,4 +18,4 @@ package org.apache.accumulo.tserver.scan;
 
 public enum ScanRunState {
   QUEUED, RUNNING, FINISHED
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanTask.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanTask.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanTask.java
index a647c84..dffdf76 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanTask.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/scan/ScanTask.java
@@ -126,4 +126,4 @@ public abstract class ScanTask<T> implements RunnableFuture<T> {
     return runState.get();
   }
 
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ConditionalSession.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ConditionalSession.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ConditionalSession.java
index d2515e6..cd5e617 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ConditionalSession.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ConditionalSession.java
@@ -18,9 +18,9 @@ package org.apache.accumulo.tserver.session;
 
 import java.util.concurrent.atomic.AtomicBoolean;
 
+import org.apache.accumulo.core.client.Durability;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.thrift.TCredentials;
-import org.apache.accumulo.core.client.Durability;
 
 public class ConditionalSession extends Session {
   public final TCredentials credentials;
@@ -28,7 +28,7 @@ public class ConditionalSession extends Session {
   public final String tableId;
   public final AtomicBoolean interruptFlag = new AtomicBoolean();
   public final Durability durability;
-  
+
   public ConditionalSession(TCredentials credentials, Authorizations authorizations, String tableId, Durability durability) {
     super(credentials);
     this.credentials = credentials;
@@ -41,4 +41,4 @@ public class ConditionalSession extends Session {
   public void cleanup() {
     interruptFlag.set(true);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/session/MultiScanSession.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/MultiScanSession.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/MultiScanSession.java
index 110fcac..b91c9e6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/MultiScanSession.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/MultiScanSession.java
@@ -45,7 +45,8 @@ public class MultiScanSession extends Session {
 
   public volatile ScanTask<MultiScanResult> lookupTask;
 
-  public MultiScanSession(TCredentials credentials, KeyExtent threadPoolExtent, Map<KeyExtent,List<Range>> queries, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio, Authorizations authorizations) {
+  public MultiScanSession(TCredentials credentials, KeyExtent threadPoolExtent, Map<KeyExtent,List<Range>> queries, List<IterInfo> ssiList,
+      Map<String,Map<String,String>> ssio, Authorizations authorizations) {
     super(credentials);
     this.queries = queries;
     this.ssiList = ssiList;
@@ -59,4 +60,4 @@ public class MultiScanSession extends Session {
     if (lookupTask != null)
       lookupTask.cancel(true);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ScanSession.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ScanSession.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ScanSession.java
index 3f9b7af..f4d72a0 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ScanSession.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/ScanSession.java
@@ -44,8 +44,9 @@ public class ScanSession extends Session {
   public volatile ScanTask<ScanBatch> nextBatchTask;
   public Scanner scanner;
   public final long readaheadThreshold;
-  
-  public ScanSession(TCredentials credentials, KeyExtent extent, Set<Column> columnSet, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio, Authorizations authorizations, long readaheadThreshold) {
+
+  public ScanSession(TCredentials credentials, KeyExtent extent, Set<Column> columnSet, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio,
+      Authorizations authorizations, long readaheadThreshold) {
     super(credentials);
     this.extent = extent;
     this.columnSet = columnSet;
@@ -66,4 +67,4 @@ public class ScanSession extends Session {
     }
   }
 
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/session/Session.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/Session.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/Session.java
index dfe38e4..9aaa17a 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/Session.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/Session.java
@@ -25,19 +25,19 @@ public class Session {
   public long startTime;
   public boolean reserved;
   private final TCredentials credentials;
-  
+
   public Session(TCredentials credentials) {
     this.credentials = credentials;
     this.client = TServerUtils.clientAddress.get();
   }
-  
+
   public String getUser() {
     return credentials.getPrincipal();
   }
-  
+
   public TCredentials getCredentials() {
     return credentials;
   }
 
   public void cleanup() {}
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/session/SessionManager.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/SessionManager.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/SessionManager.java
index 13049e2..51722d9 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/SessionManager.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/SessionManager.java
@@ -276,8 +276,9 @@ public class SessionManager {
           }
         }
 
-        ActiveScan activeScan = new ActiveScan(ss.client, ss.getUser(), ss.extent.getTableId().toString(), ct - ss.startTime, ct - ss.lastAccessTime, ScanType.SINGLE,
-            state, ss.extent.toThrift(), Translator.translate(ss.columnSet, Translators.CT), ss.ssiList, ss.ssio, ss.auths.getAuthorizationsBB());
+        ActiveScan activeScan = new ActiveScan(ss.client, ss.getUser(), ss.extent.getTableId().toString(), ct - ss.startTime, ct - ss.lastAccessTime,
+            ScanType.SINGLE, state, ss.extent.toThrift(), Translator.translate(ss.columnSet, Translators.CT), ss.ssiList, ss.ssio,
+            ss.auths.getAuthorizationsBB());
 
         // scanId added by ACCUMULO-2641 is an optional thrift argument and not available in ActiveScan constructor
         activeScan.setScanId(entry.getKey());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/session/UpdateSession.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/UpdateSession.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/UpdateSession.java
index 65430ce..08930ac 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/session/UpdateSession.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/session/UpdateSession.java
@@ -20,6 +20,7 @@ import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
 
+import org.apache.accumulo.core.client.Durability;
 import org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode;
 import org.apache.accumulo.core.constraints.Violations;
 import org.apache.accumulo.core.data.KeyExtent;
@@ -28,7 +29,6 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
 import org.apache.accumulo.core.util.MapCounter;
 import org.apache.accumulo.core.util.Stat;
 import org.apache.accumulo.tserver.TservConstraintEnv;
-import org.apache.accumulo.core.client.Durability;
 import org.apache.accumulo.tserver.tablet.Tablet;
 
 public class UpdateSession extends Session {
@@ -42,17 +42,17 @@ public class UpdateSession extends Session {
   public final Stat authTimes = new Stat();
   public final Map<Tablet,List<Mutation>> queuedMutations = new HashMap<Tablet,List<Mutation>>();
   public final Violations violations;
-  
+
   public Tablet currentTablet = null;
   public long totalUpdates = 0;
   public long flushTime = 0;
   public long queuedMutationSize = 0;
   public final Durability durability;
-  
+
   public UpdateSession(TservConstraintEnv env, TCredentials credentials, Durability durability) {
     super(credentials);
     this.cenv = env;
     this.violations = new Violations();
     this.durability = durability;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Batch.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Batch.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Batch.java
index 1a83ba4..4388867 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Batch.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Batch.java
@@ -25,7 +25,7 @@ final class Batch {
   private final List<KVEntry> results;
   private final Key continueKey;
   private final long numBytes;
-  
+
   Batch(boolean skipContinueKey, List<KVEntry> results, Key continueKey, long numBytes) {
     this.skipContinueKey = skipContinueKey;
     this.results = results;
@@ -48,4 +48,4 @@ final class Batch {
   public long getNumBytes() {
     return numBytes;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CommitSession.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CommitSession.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CommitSession.java
index db4100f..17290c0 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CommitSession.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CommitSession.java
@@ -26,7 +26,7 @@ import org.apache.accumulo.tserver.log.DfsLogger;
 import org.apache.log4j.Logger;
 
 public class CommitSession {
-  
+
   private static final Logger log = Logger.getLogger(CommitSession.class);
 
   private final int seq;
@@ -114,4 +114,4 @@ public class CommitSession {
   public void mutate(List<Mutation> mutations) {
     memTable.mutate(mutations);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
index 8e9fb9b..918edf6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionInfo.java
@@ -123,7 +123,7 @@ public class CompactionInfo {
     List<String> filesToCompact = new ArrayList<String>();
     for (FileRef ref : compactor.getFilesToCompact())
       filesToCompact.add(ref.toString());
-    return new ActiveCompaction(compactor.extent.toThrift(), System.currentTimeMillis() - compactor.getStartTime(), filesToCompact,
-        compactor.getOutputFile(), type, reason, localityGroup, entriesRead, entriesWritten, iiList, iterOptions);
+    return new ActiveCompaction(compactor.extent.toThrift(), System.currentTimeMillis() - compactor.getStartTime(), filesToCompact, compactor.getOutputFile(),
+        type, reason, localityGroup, entriesRead, entriesWritten, iiList, iterOptions);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionRunner.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionRunner.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionRunner.java
index 1dee64b..53cc750 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionRunner.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionRunner.java
@@ -73,4 +73,4 @@ final class CompactionRunner implements Runnable, Comparable<CompactionRunner> {
 
     return o.getNumFiles() - this.getNumFiles();
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionStats.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionStats.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionStats.java
index 69832e9..68a2307 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionStats.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionStats.java
@@ -20,39 +20,39 @@ public class CompactionStats {
   private long entriesRead;
   private long entriesWritten;
   private long fileSize;
-  
+
   CompactionStats(long er, long ew) {
     this.setEntriesRead(er);
     this.setEntriesWritten(ew);
   }
-  
+
   public CompactionStats() {}
-  
+
   private void setEntriesRead(long entriesRead) {
     this.entriesRead = entriesRead;
   }
-  
+
   public long getEntriesRead() {
     return entriesRead;
   }
-  
+
   private void setEntriesWritten(long entriesWritten) {
     this.entriesWritten = entriesWritten;
   }
-  
+
   public long getEntriesWritten() {
     return entriesWritten;
   }
-  
+
   public void add(CompactionStats mcs) {
     this.entriesRead += mcs.entriesRead;
     this.entriesWritten += mcs.entriesWritten;
   }
-  
+
   public void setFileSize(long fileSize) {
     this.fileSize = fileSize;
   }
-  
+
   public long getFileSize() {
     return this.fileSize;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionWatcher.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionWatcher.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionWatcher.java
index adc01b2..94fbfdc 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionWatcher.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CompactionWatcher.java
@@ -29,18 +29,18 @@ import org.apache.accumulo.server.util.time.SimpleTimer;
 import org.apache.log4j.Logger;
 
 /**
- * 
+ *
  */
 public class CompactionWatcher implements Runnable {
   private final Map<List<Long>,ObservedCompactionInfo> observedCompactions = new HashMap<List<Long>,ObservedCompactionInfo>();
   private final AccumuloConfiguration config;
   private static boolean watching = false;
-  
+
   private static class ObservedCompactionInfo {
     CompactionInfo compactionInfo;
     long firstSeen;
     boolean loggedWarning;
-    
+
     ObservedCompactionInfo(CompactionInfo ci, long time) {
       this.compactionInfo = ci;
       this.firstSeen = time;
@@ -54,24 +54,24 @@ public class CompactionWatcher implements Runnable {
   @Override
   public void run() {
     List<CompactionInfo> runningCompactions = Compactor.getRunningCompactions();
-    
+
     Set<List<Long>> newKeys = new HashSet<List<Long>>();
-    
+
     long time = System.currentTimeMillis();
 
     for (CompactionInfo ci : runningCompactions) {
       List<Long> compactionKey = Arrays.asList(ci.getID(), ci.getEntriesRead(), ci.getEntriesWritten());
       newKeys.add(compactionKey);
-      
+
       if (!observedCompactions.containsKey(compactionKey)) {
         observedCompactions.put(compactionKey, new ObservedCompactionInfo(ci, time));
       }
     }
-    
+
     // look for compactions that finished or made progress and logged a warning
     HashMap<List<Long>,ObservedCompactionInfo> copy = new HashMap<List<Long>,ObservedCompactionInfo>(observedCompactions);
     copy.keySet().removeAll(newKeys);
-    
+
     for (ObservedCompactionInfo oci : copy.values()) {
       if (oci.loggedWarning) {
         Logger.getLogger(CompactionWatcher.class).info("Compaction of " + oci.compactionInfo.getExtent() + " is no longer stuck");
@@ -80,7 +80,7 @@ public class CompactionWatcher implements Runnable {
 
     // remove any compaction that completed or made progress
     observedCompactions.keySet().retainAll(newKeys);
-    
+
     long warnTime = config.getTimeInMillis(Property.TSERV_COMPACTION_WARN_TIME);
 
     // check for stuck compactions

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Compactor.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Compactor.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Compactor.java
index 869cc33..1c23293 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Compactor.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Compactor.java
@@ -75,7 +75,7 @@ public class Compactor implements Callable<CompactionStats> {
   }
 
   public interface CompactionEnv {
-    
+
     boolean isCompactionEnabled();
 
     IteratorScope getIteratorScope();
@@ -113,7 +113,7 @@ public class Compactor implements Callable<CompactionStats> {
   private synchronized void setLocalityGroup(String name) {
     this.currentLocalityGroup = name;
   }
-  
+
   public synchronized String getCurrentLocalityGroup() {
     return currentLocalityGroup;
   }
@@ -411,7 +411,7 @@ public class Compactor implements Callable<CompactionStats> {
   long getEntriesRead() {
     return entriesRead.get();
   }
-  
+
   long getEntriesWritten() {
     return entriesWritten.get();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CountingIterator.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CountingIterator.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CountingIterator.java
index 44b8460..8716695 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CountingIterator.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/CountingIterator.java
@@ -75,4 +75,4 @@ public class CountingIterator extends WrappingIterator {
 
     return count + sum;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/KVEntry.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/KVEntry.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/KVEntry.java
index 4919be9..4b1cf8c 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/KVEntry.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/KVEntry.java
@@ -36,4 +36,4 @@ public class KVEntry extends KeyValue {
   int estimateMemoryUsed() {
     return getKey().getSize() + getValue().get().length + (9 * 32); // overhead is 32 per object
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactionTask.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactionTask.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactionTask.java
index c5dd4a4..8609a4b 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactionTask.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactionTask.java
@@ -76,8 +76,8 @@ class MinorCompactionTask implements Runnable {
       }
       span.stop();
       span = Trace.start("compact");
-      this.stats = tablet.minorCompact(tablet.getTabletServer().getFileSystem(), tablet.getTabletMemory().getMinCMemTable(), tmpFileRef, newMapfileLocation, mergeFile, true, queued, commitSession, flushId,
-          mincReason);
+      this.stats = tablet.minorCompact(tablet.getTabletServer().getFileSystem(), tablet.getTabletMemory().getMinCMemTable(), tmpFileRef, newMapfileLocation,
+          mergeFile, true, queued, commitSession, flushId, mincReason);
       span.stop();
 
       if (tablet.needsSplit()) {
@@ -96,4 +96,4 @@ class MinorCompactionTask implements Runnable {
       minorCompaction.stop();
     }
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactor.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactor.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactor.java
index b513167..a425aa1 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactor.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/MinorCompactor.java
@@ -39,37 +39,37 @@ import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
 public class MinorCompactor extends Compactor {
-  
+
   private static final Logger log = Logger.getLogger(MinorCompactor.class);
-  
+
   private static final Map<FileRef,DataFileValue> EMPTY_MAP = Collections.emptyMap();
-  
+
   private static Map<FileRef,DataFileValue> toFileMap(FileRef mergeFile, DataFileValue dfv) {
     if (mergeFile == null)
       return EMPTY_MAP;
-    
+
     return Collections.singletonMap(mergeFile, dfv);
   }
-  
+
   private final TabletServer tabletServer;
 
   public MinorCompactor(TabletServer tabletServer, Tablet tablet, InMemoryMap imm, FileRef mergeFile, DataFileValue dfv, FileRef outputFile,
       MinorCompactionReason mincReason, TableConfiguration tableConfig) {
     super(tabletServer, tablet, toFileMap(mergeFile, dfv), imm, outputFile, true, new CompactionEnv() {
-      
+
       @Override
       public boolean isCompactionEnabled() {
         return true;
       }
-      
+
       @Override
       public IteratorScope getIteratorScope() {
         return IteratorScope.minc;
       }
-    }, Collections.<IteratorSetting>emptyList(), mincReason.ordinal(), tableConfig);
+    }, Collections.<IteratorSetting> emptyList(), mincReason.ordinal(), tableConfig);
     this.tabletServer = tabletServer;
   }
-  
+
   private boolean isTableDeleting() {
     try {
       return Tables.getTableState(tabletServer.getInstance(), extent.getTableId().toString()) == TableState.DELETING;
@@ -78,30 +78,30 @@ public class MinorCompactor extends Compactor {
       return false; // can not get positive confirmation that its deleting.
     }
   }
-  
+
   @Override
   public CompactionStats call() {
     log.debug("Begin minor compaction " + getOutputFile() + " " + getExtent());
-    
+
     // output to new MapFile with a temporary name
     int sleepTime = 100;
     double growthFactor = 4;
     int maxSleepTime = 1000 * 60 * 3; // 3 minutes
     boolean reportedProblem = false;
-    
+
     runningCompactions.add(this);
     try {
       do {
         try {
           CompactionStats ret = super.call();
-          
+
           // log.debug(String.format("MinC %,d recs in | %,d recs out | %,d recs/sec | %6.3f secs | %,d bytes ",map.size(), entriesCompacted,
           // (int)(map.size()/((t2 - t1)/1000.0)), (t2 - t1)/1000.0, estimatedSizeInBytes()));
-          
+
           if (reportedProblem) {
             ProblemReports.getInstance(tabletServer).deleteProblemReport(getExtent().getTableId().toString(), ProblemType.FILE_WRITE, getOutputFile());
           }
-          
+
           return ret;
         } catch (IOException e) {
           log.warn("MinC failed (" + e.getMessage() + ") to create " + getOutputFile() + " retrying ...");
@@ -116,14 +116,14 @@ public class MinorCompactor extends Compactor {
         } catch (CompactionCanceledException e) {
           throw new IllegalStateException(e);
         }
-        
+
         Random random = new Random();
-        
+
         int sleep = sleepTime + random.nextInt(sleepTime);
         log.debug("MinC failed sleeping " + sleep + " ms before retrying");
         UtilWaitThread.sleep(sleep);
         sleepTime = (int) Math.round(Math.min(maxSleepTime, sleepTime * growthFactor));
-        
+
         // clean up
         try {
           if (getFileSystem().exists(new Path(getOutputFile()))) {
@@ -132,15 +132,15 @@ public class MinorCompactor extends Compactor {
         } catch (IOException e) {
           log.warn("Failed to delete failed MinC file " + getOutputFile() + " " + e.getMessage());
         }
-        
+
         if (isTableDeleting())
           return new CompactionStats(0, 0);
-        
+
       } while (true);
     } finally {
       thread = null;
       runningCompactions.remove(this);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Rate.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Rate.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Rate.java
index 450fffe..a0ea2d6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Rate.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Rate.java
@@ -21,10 +21,10 @@ public class Rate {
   private long lastTime = -1;
   private double current = 0.0;
   final double ratio;
-  
+
   /**
    * Turn a counter into an exponentially smoothed rate over time.
-   * 
+   *
    * @param ratio
    *          the rate at which each update influences the curve; must be (0., 1.0)
    */
@@ -33,11 +33,11 @@ public class Rate {
       throw new IllegalArgumentException("ratio must be > 0. and < 1.0");
     this.ratio = ratio;
   }
-  
+
   public double update(long counter) {
     return update(System.currentTimeMillis(), counter);
   }
-  
+
   synchronized public double update(long when, long counter) {
     if (lastCounter < 0) {
       lastTime = when;
@@ -53,7 +53,7 @@ public class Rate {
     lastCounter = counter;
     return current;
   }
-  
+
   synchronized public double rate() {
     return this.current;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/RootFiles.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/RootFiles.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/RootFiles.java
index 875643e..d0278f4 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/RootFiles.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/RootFiles.java
@@ -32,7 +32,7 @@ import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
 /**
- * 
+ *
  */
 public class RootFiles {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanBatch.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanBatch.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanBatch.java
index dc932c6..888d6f5 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanBatch.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanBatch.java
@@ -34,4 +34,4 @@ final public class ScanBatch {
   public List<KVEntry> getResults() {
     return results;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanDataSource.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanDataSource.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanDataSource.java
index fe4b16b..00333c6 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanDataSource.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanDataSource.java
@@ -64,8 +64,8 @@ class ScanDataSource implements DataSource {
 
   private final ScanOptions options;
 
-  ScanDataSource(Tablet tablet, Authorizations authorizations, byte[] defaultLabels, HashSet<Column> columnSet, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio,
-      AtomicBoolean interruptFlag) {
+  ScanDataSource(Tablet tablet, Authorizations authorizations, byte[] defaultLabels, HashSet<Column> columnSet, List<IterInfo> ssiList,
+      Map<String,Map<String,String>> ssio, AtomicBoolean interruptFlag) {
     this.tablet = tablet;
     expectedDeletionCount = tablet.getDataSourceDeletions();
     this.options = new ScanOptions(-1, authorizations, defaultLabels, columnSet, ssiList, ssio, interruptFlag, false);
@@ -171,8 +171,8 @@ class ScanDataSource implements DataSource {
 
     VisibilityFilter visFilter = new VisibilityFilter(colFilter, options.getAuthorizations(), options.getDefaultLabels());
 
-    return iterEnv.getTopLevelIterator(IteratorUtil
-        .loadIterators(IteratorScope.scan, visFilter, tablet.getExtent(), tablet.getTableConfiguration(), options.getSsiList(), options.getSsio(), iterEnv));
+    return iterEnv.getTopLevelIterator(IteratorUtil.loadIterators(IteratorScope.scan, visFilter, tablet.getExtent(), tablet.getTableConfiguration(),
+        options.getSsiList(), options.getSsio(), iterEnv));
   }
 
   void close(boolean sawErrors) {
@@ -213,7 +213,7 @@ class ScanDataSource implements DataSource {
     if (fileManager != null)
       fileManager.reattach();
   }
-  
+
   public void detachFileManager() {
     if (fileManager != null)
       fileManager.detach();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanOptions.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanOptions.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanOptions.java
index 07aa8e7..93e8eee 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanOptions.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/ScanOptions.java
@@ -36,8 +36,8 @@ final class ScanOptions {
   private final int num;
   private final boolean isolated;
 
-  ScanOptions(int num, Authorizations authorizations, byte[] defaultLabels, Set<Column> columnSet, List<IterInfo> ssiList,
-      Map<String,Map<String,String>> ssio, AtomicBoolean interruptFlag, boolean isolated) {
+  ScanOptions(int num, Authorizations authorizations, byte[] defaultLabels, Set<Column> columnSet, List<IterInfo> ssiList, Map<String,Map<String,String>> ssio,
+      AtomicBoolean interruptFlag, boolean isolated) {
     this.num = num;
     this.authorizations = authorizations;
     this.defaultLabels = defaultLabels;
@@ -79,4 +79,4 @@ final class ScanOptions {
   public boolean isIsolated() {
     return isolated;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Scanner.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Scanner.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Scanner.java
index ad3fcb2..17ae729 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Scanner.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Scanner.java
@@ -113,10 +113,10 @@ public class Scanner {
       // to return mapfiles, even when exception is thrown
       if (!options.isIsolated()) {
         dataSource.close(false);
-      } else { 
+      } else {
         dataSource.detachFileManager();
       }
-      
+
       if (results != null && results.getResults() != null)
         tablet.updateQueryStats(results.getResults().size(), results.getNumBytes());
     }
@@ -133,4 +133,4 @@ public class Scanner {
         isolatedDataSource.close(false);
     }
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitInfo.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitInfo.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitInfo.java
index ec84aa8..f8f2183 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitInfo.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitInfo.java
@@ -24,12 +24,12 @@ import org.apache.accumulo.server.master.state.TServerInstance;
 
 /**
  * operations are disallowed while we split which is ok since splitting is fast
- * 
+ *
  * a minor compaction should have taken place before calling this so there should be relatively little left to compact
- * 
+ *
  * we just need to make sure major compactions aren't occurring if we have the major compactor thread decide who needs splitting we can avoid synchronization
  * issues with major compactions
- * 
+ *
  */
 
 final public class SplitInfo {
@@ -73,4 +73,4 @@ final public class SplitInfo {
     return lastLocation;
   }
 
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitRowSpec.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitRowSpec.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitRowSpec.java
index 75cf91e..367902f 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitRowSpec.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/SplitRowSpec.java
@@ -26,4 +26,4 @@ class SplitRowSpec {
     this.splitRatio = splitRatio;
     this.row = row;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Tablet.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Tablet.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Tablet.java
index 7279953..17c60ec 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Tablet.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/Tablet.java
@@ -1786,7 +1786,7 @@ public class Tablet implements TabletCommitter {
     CompactionStrategy strategy = null;
     Map<FileRef,Pair<Key,Key>> firstAndLastKeys = null;
 
-    if(reason == MajorCompactionReason.USER){
+    if (reason == MajorCompactionReason.USER) {
       try {
         compactionId = getCompactionID();
         strategy = createCompactionStrategy(compactionId.getSecond().getCompactionStrategy());
@@ -1887,7 +1887,7 @@ public class Tablet implements TabletCommitter {
       log.debug(String.format("MajC initiate lock %.2f secs, wait %.2f secs", (t3 - t2) / 1000.0, (t2 - t1) / 1000.0));
 
       if (updateCompactionID) {
-        MetadataTableUtil.updateTabletCompactID(extent, compactionId.getFirst(),tabletServer, getTabletServer().getLock());
+        MetadataTableUtil.updateTabletCompactID(extent, compactionId.getFirst(), tabletServer, getTabletServer().getLock());
         return majCStats;
       }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletClosedException.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletClosedException.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletClosedException.java
index 827b803..d3ed507 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletClosedException.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletClosedException.java
@@ -26,4 +26,4 @@ public class TabletClosedException extends RuntimeException {
   }
 
   private static final long serialVersionUID = 1L;
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletCommitter.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletCommitter.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletCommitter.java
index 3042267..b56d0af 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletCommitter.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletCommitter.java
@@ -48,5 +48,5 @@ public interface TabletCommitter {
   Durability getDurability();
 
   void updateMemoryUsageStats(long estimatedSizeInBytes, long estimatedSizeInBytes2);
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletMemory.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletMemory.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletMemory.java
index 155d6b5..15cd50d 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletMemory.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/tablet/TabletMemory.java
@@ -29,7 +29,7 @@ import org.apache.log4j.Logger;
 
 class TabletMemory implements Closeable {
   static private final Logger log = Logger.getLogger(TabletMemory.class);
-  
+
   private final TabletCommitter tablet;
   private InMemoryMap memTable;
   private InMemoryMap otherMemTable;
@@ -93,7 +93,7 @@ class TabletMemory implements Closeable {
     if (deletingMemTable != null) {
       throw new IllegalStateException();
     }
-    
+
     if (commitSession == null) {
       throw new IllegalStateException();
     }
@@ -187,4 +187,4 @@ class TabletMemory implements Closeable {
   public boolean isClosed() {
     return commitSession == null;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManagerTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManagerTest.java b/server/tserver/src/test/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManagerTest.java
index 2a85825..e17281a 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManagerTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/server/tabletserver/LargestFirstMemoryManagerTest.java
@@ -34,12 +34,12 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class LargestFirstMemoryManagerTest {
-  
+
   private static final long ZERO = System.currentTimeMillis();
   private static final long LATER = ZERO + 20 * 60 * 1000;
-  private static final long ONE_GIG = 1024*1024*1024;
-  private static final long HALF_GIG = ONE_GIG/2;
-  private static final long QGIG = ONE_GIG/4;
+  private static final long ONE_GIG = 1024 * 1024 * 1024;
+  private static final long HALF_GIG = ONE_GIG / 2;
+  private static final long QGIG = ONE_GIG / 4;
   private static final long ONE_MINUTE = 60 * 1000;
 
   @Test
@@ -94,138 +94,72 @@ public class LargestFirstMemoryManagerTest {
     // lots of work to do
     mgr = new LargestFirstMemoryManagerUnderTest();
     mgr.init(config);
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, HALF_GIG, 0),
-        t(k("b"), ZERO, HALF_GIG+1, 0),
-        t(k("c"), ZERO, HALF_GIG+2, 0),
-        t(k("d"), ZERO, HALF_GIG+3, 0),
-        t(k("e"), ZERO, HALF_GIG+4, 0),
-        t(k("f"), ZERO, HALF_GIG+5, 0),
-        t(k("g"), ZERO, HALF_GIG+6, 0),
-        t(k("h"), ZERO, HALF_GIG+7, 0),
-        t(k("i"), ZERO, HALF_GIG+8, 0))
-        );
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, HALF_GIG, 0), t(k("b"), ZERO, HALF_GIG + 1, 0), t(k("c"), ZERO, HALF_GIG + 2, 0),
+        t(k("d"), ZERO, HALF_GIG + 3, 0), t(k("e"), ZERO, HALF_GIG + 4, 0), t(k("f"), ZERO, HALF_GIG + 5, 0), t(k("g"), ZERO, HALF_GIG + 6, 0),
+        t(k("h"), ZERO, HALF_GIG + 7, 0), t(k("i"), ZERO, HALF_GIG + 8, 0)));
     assertEquals(2, result.tabletsToMinorCompact.size());
     assertEquals(k("i"), result.tabletsToMinorCompact.get(0));
     assertEquals(k("h"), result.tabletsToMinorCompact.get(1));
     // one finished, one in progress, one filled up
     mgr = new LargestFirstMemoryManagerUnderTest();
     mgr.init(config);
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, HALF_GIG, 0),
-        t(k("b"), ZERO, HALF_GIG+1, 0),
-        t(k("c"), ZERO, HALF_GIG+2, 0),
-        t(k("d"), ZERO, HALF_GIG+3, 0),
-        t(k("e"), ZERO, HALF_GIG+4, 0),
-        t(k("f"), ZERO, HALF_GIG+5, 0),
-        t(k("g"), ZERO, ONE_GIG, 0),
-        t(k("h"), ZERO, 0, HALF_GIG+7),
-        t(k("i"), ZERO, 0, 0))
-        );
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, HALF_GIG, 0), t(k("b"), ZERO, HALF_GIG + 1, 0), t(k("c"), ZERO, HALF_GIG + 2, 0),
+        t(k("d"), ZERO, HALF_GIG + 3, 0), t(k("e"), ZERO, HALF_GIG + 4, 0), t(k("f"), ZERO, HALF_GIG + 5, 0), t(k("g"), ZERO, ONE_GIG, 0),
+        t(k("h"), ZERO, 0, HALF_GIG + 7), t(k("i"), ZERO, 0, 0)));
     assertEquals(1, result.tabletsToMinorCompact.size());
     assertEquals(k("g"), result.tabletsToMinorCompact.get(0));
     // memory is very full, lots of candidates
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, HALF_GIG, 0),
-        t(k("b"), ZERO, ONE_GIG+1, 0),
-        t(k("c"), ZERO, ONE_GIG+2, 0),
-        t(k("d"), ZERO, ONE_GIG+3, 0),
-        t(k("e"), ZERO, ONE_GIG+4, 0),
-        t(k("f"), ZERO, ONE_GIG+5, 0),
-        t(k("g"), ZERO, ONE_GIG+6, 0),
-        t(k("h"), ZERO, 0, 0),
-        t(k("i"), ZERO, 0, 0))
-        );
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, HALF_GIG, 0), t(k("b"), ZERO, ONE_GIG + 1, 0), t(k("c"), ZERO, ONE_GIG + 2, 0),
+        t(k("d"), ZERO, ONE_GIG + 3, 0), t(k("e"), ZERO, ONE_GIG + 4, 0), t(k("f"), ZERO, ONE_GIG + 5, 0), t(k("g"), ZERO, ONE_GIG + 6, 0),
+        t(k("h"), ZERO, 0, 0), t(k("i"), ZERO, 0, 0)));
     assertEquals(2, result.tabletsToMinorCompact.size());
     assertEquals(k("g"), result.tabletsToMinorCompact.get(0));
     assertEquals(k("f"), result.tabletsToMinorCompact.get(1));
     // only have two compactors, still busy
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, HALF_GIG, 0),
-        t(k("b"), ZERO, ONE_GIG+1, 0),
-        t(k("c"), ZERO, ONE_GIG+2, 0),
-        t(k("d"), ZERO, ONE_GIG+3, 0),
-        t(k("e"), ZERO, ONE_GIG+4, 0),
-        t(k("f"), ZERO, ONE_GIG, ONE_GIG+5),
-        t(k("g"), ZERO, ONE_GIG, ONE_GIG+6),
-        t(k("h"), ZERO, 0, 0),
-        t(k("i"), ZERO, 0, 0))
-        );
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, HALF_GIG, 0), t(k("b"), ZERO, ONE_GIG + 1, 0), t(k("c"), ZERO, ONE_GIG + 2, 0),
+        t(k("d"), ZERO, ONE_GIG + 3, 0), t(k("e"), ZERO, ONE_GIG + 4, 0), t(k("f"), ZERO, ONE_GIG, ONE_GIG + 5), t(k("g"), ZERO, ONE_GIG, ONE_GIG + 6),
+        t(k("h"), ZERO, 0, 0), t(k("i"), ZERO, 0, 0)));
     assertEquals(0, result.tabletsToMinorCompact.size());
-    // finished one 
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, HALF_GIG, 0),
-        t(k("b"), ZERO, ONE_GIG+1, 0),
-        t(k("c"), ZERO, ONE_GIG+2, 0),
-        t(k("d"), ZERO, ONE_GIG+3, 0),
-        t(k("e"), ZERO, ONE_GIG+4, 0),
-        t(k("f"), ZERO, ONE_GIG, ONE_GIG+5),
-        t(k("g"), ZERO, ONE_GIG, 0),
-        t(k("h"), ZERO, 0, 0),
-        t(k("i"), ZERO, 0, 0))
-        );
+    // finished one
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, HALF_GIG, 0), t(k("b"), ZERO, ONE_GIG + 1, 0), t(k("c"), ZERO, ONE_GIG + 2, 0),
+        t(k("d"), ZERO, ONE_GIG + 3, 0), t(k("e"), ZERO, ONE_GIG + 4, 0), t(k("f"), ZERO, ONE_GIG, ONE_GIG + 5), t(k("g"), ZERO, ONE_GIG, 0),
+        t(k("h"), ZERO, 0, 0), t(k("i"), ZERO, 0, 0)));
     assertEquals(1, result.tabletsToMinorCompact.size());
     assertEquals(k("e"), result.tabletsToMinorCompact.get(0));
 
     // many are running: do nothing
     mgr = new LargestFirstMemoryManagerUnderTest();
     mgr.init(config);
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, HALF_GIG, 0),
-        t(k("b"), ZERO, HALF_GIG+1, 0),
-        t(k("c"), ZERO, HALF_GIG+2, 0),
-        t(k("d"), ZERO, 0, HALF_GIG),
-        t(k("e"), ZERO, 0, HALF_GIG),
-        t(k("f"), ZERO, 0, HALF_GIG),
-        t(k("g"), ZERO, 0, HALF_GIG),
-        t(k("i"), ZERO, 0, HALF_GIG),
-        t(k("j"), ZERO, 0, HALF_GIG),
-        t(k("k"), ZERO, 0, HALF_GIG),
-        t(k("l"), ZERO, 0, HALF_GIG),
-        t(k("m"), ZERO, 0, HALF_GIG)
-        ));
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, HALF_GIG, 0), t(k("b"), ZERO, HALF_GIG + 1, 0), t(k("c"), ZERO, HALF_GIG + 2, 0),
+        t(k("d"), ZERO, 0, HALF_GIG), t(k("e"), ZERO, 0, HALF_GIG), t(k("f"), ZERO, 0, HALF_GIG), t(k("g"), ZERO, 0, HALF_GIG), t(k("i"), ZERO, 0, HALF_GIG),
+        t(k("j"), ZERO, 0, HALF_GIG), t(k("k"), ZERO, 0, HALF_GIG), t(k("l"), ZERO, 0, HALF_GIG), t(k("m"), ZERO, 0, HALF_GIG)));
     assertEquals(0, result.tabletsToMinorCompact.size());
-    
+
     // observe adjustment:
     mgr = new LargestFirstMemoryManagerUnderTest();
     mgr.init(config);
     // compact the largest
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, QGIG, 0),
-        t(k("b"), ZERO, QGIG+1, 0),
-        t(k("c"), ZERO, QGIG+2, 0)));
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, QGIG, 0), t(k("b"), ZERO, QGIG + 1, 0), t(k("c"), ZERO, QGIG + 2, 0)));
     assertEquals(1, result.tabletsToMinorCompact.size());
     assertEquals(k("c"), result.tabletsToMinorCompact.get(0));
     // show that it is compacting... do nothing
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, QGIG, 0),
-        t(k("b"), ZERO, QGIG+1, 0),
-        t(k("c"), ZERO, 0, QGIG+2)));
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, QGIG, 0), t(k("b"), ZERO, QGIG + 1, 0), t(k("c"), ZERO, 0, QGIG + 2)));
     assertEquals(0, result.tabletsToMinorCompact.size());
     // not going to bother compacting any more
     mgr.currentTime += ONE_MINUTE;
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, QGIG, 0),
-        t(k("b"), ZERO, QGIG+1, 0),
-        t(k("c"), ZERO, 0, QGIG+2)));
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, QGIG, 0), t(k("b"), ZERO, QGIG + 1, 0), t(k("c"), ZERO, 0, QGIG + 2)));
     assertEquals(0, result.tabletsToMinorCompact.size());
     // now do nothing
     mgr.currentTime += ONE_MINUTE;
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, QGIG, 0),
-        t(k("b"), ZERO, 0, 0),
-        t(k("c"), ZERO, 0, 0)));
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, QGIG, 0), t(k("b"), ZERO, 0, 0), t(k("c"), ZERO, 0, 0)));
     assertEquals(0, result.tabletsToMinorCompact.size());
     // on no! more data, this time we compact because we've adjusted
     mgr.currentTime += ONE_MINUTE;
-    result = mgr.getMemoryManagementActions(tablets(
-        t(k("a"), ZERO, QGIG, 0),
-        t(k("b"), ZERO, QGIG+1, 0),
-        t(k("c"), ZERO, 0, 0)));
+    result = mgr.getMemoryManagementActions(tablets(t(k("a"), ZERO, QGIG, 0), t(k("b"), ZERO, QGIG + 1, 0), t(k("c"), ZERO, 0, 0)));
     assertEquals(1, result.tabletsToMinorCompact.size());
     assertEquals(k("b"), result.tabletsToMinorCompact.get(0));
   }
-  
+
   private static class LargestFirstMemoryManagerUnderTest extends LargestFirstMemoryManager {
 
     public long currentTime = ZERO;
@@ -239,20 +173,20 @@ public class LargestFirstMemoryManagerTest {
     protected long getMinCIdleThreshold(KeyExtent extent) {
       return 15 * 60 * 1000;
     }
-    
+
   }
-  
+
   private static KeyExtent k(String endRow) {
     return new KeyExtent(new Text("1"), new Text(endRow), null);
   }
-  
+
   private static class TestTabletState implements TabletState {
-    
+
     private final KeyExtent extent;
     private final long lastCommit;
     private final long memSize;
     private final long compactingSize;
-    
+
     TestTabletState(KeyExtent extent, long commit, long memsize, long compactingTableSize) {
       this.extent = extent;
       this.lastCommit = commit;
@@ -279,14 +213,14 @@ public class LargestFirstMemoryManagerTest {
     public long getMinorCompactingMemTableSize() {
       return compactingSize;
     }
-    
+
   }
 
   private TabletState t(KeyExtent ke, long lastCommit, long memSize, long compactingSize) {
     return new TestTabletState(ke, lastCommit, memSize, compactingSize);
   }
-  
-  private static List<TabletState> tablets(TabletState ... states) {
+
+  private static List<TabletState> tablets(TabletState... states) {
     return Arrays.asList(states);
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/CheckTabletMetadataTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/CheckTabletMetadataTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/CheckTabletMetadataTest.java
index 0209278..54be395 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/CheckTabletMetadataTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/CheckTabletMetadataTest.java
@@ -25,99 +25,98 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection;
 import org.apache.accumulo.core.util.ColumnFQ;
 import org.apache.accumulo.server.master.state.TServerInstance;
-import org.apache.accumulo.tserver.TabletServer;
 import org.apache.hadoop.io.Text;
 import org.junit.Assert;
 import org.junit.Test;
 
 public class CheckTabletMetadataTest {
-  
+
   private static Key nk(String row, ColumnFQ cfq) {
     return new Key(new Text(row), cfq.getColumnFamily(), cfq.getColumnQualifier());
   }
-  
+
   private static Key nk(String row, Text cf, String cq) {
     return new Key(row, cf.toString(), cq);
   }
-  
+
   private static void put(TreeMap<Key,Value> tabletMeta, String row, ColumnFQ cfq, byte[] val) {
     Key k = new Key(new Text(row), cfq.getColumnFamily(), cfq.getColumnQualifier());
     tabletMeta.put(k, new Value(val));
   }
-  
+
   private static void put(TreeMap<Key,Value> tabletMeta, String row, Text cf, String cq, String val) {
     Key k = new Key(new Text(row), cf, new Text(cq));
     tabletMeta.put(k, new Value(val.getBytes()));
   }
-  
+
   private static void assertFail(TreeMap<Key,Value> tabletMeta, KeyExtent ke, TServerInstance tsi) {
     try {
       Assert.assertNull(TabletServer.checkTabletMetadata(ke, tsi, tabletMeta, ke.getMetadataEntry()));
     } catch (Exception e) {
-      
+
     }
   }
-  
+
   private static void assertFail(TreeMap<Key,Value> tabletMeta, KeyExtent ke, TServerInstance tsi, Key keyToDelete) {
     TreeMap<Key,Value> copy = new TreeMap<Key,Value>(tabletMeta);
     Assert.assertNotNull(copy.remove(keyToDelete));
     try {
       Assert.assertNull(TabletServer.checkTabletMetadata(ke, tsi, copy, ke.getMetadataEntry()));
     } catch (Exception e) {
-      
+
     }
   }
-  
+
   @Test
   public void testBadTabletMetadata() throws Exception {
-    
+
     KeyExtent ke = new KeyExtent(new Text("1"), null, null);
-    
+
     TreeMap<Key,Value> tabletMeta = new TreeMap<Key,Value>();
-    
+
     put(tabletMeta, "1<", TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN, KeyExtent.encodePrevEndRow(null).get());
     put(tabletMeta, "1<", TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN, "/t1".getBytes());
     put(tabletMeta, "1<", TabletsSection.ServerColumnFamily.TIME_COLUMN, "M0".getBytes());
     put(tabletMeta, "1<", TabletsSection.FutureLocationColumnFamily.NAME, "4", "127.0.0.1:9997");
-    
+
     TServerInstance tsi = new TServerInstance("127.0.0.1:9997", 4);
-    
+
     Assert.assertNotNull(TabletServer.checkTabletMetadata(ke, tsi, tabletMeta, ke.getMetadataEntry()));
-    
+
     assertFail(tabletMeta, ke, new TServerInstance("127.0.0.1:9998", 4));
     assertFail(tabletMeta, ke, new TServerInstance("127.0.0.1:9998", 5));
     assertFail(tabletMeta, ke, new TServerInstance("127.0.0.1:9997", 5));
     assertFail(tabletMeta, ke, new TServerInstance("127.0.0.2:9997", 4));
     assertFail(tabletMeta, ke, new TServerInstance("127.0.0.2:9997", 5));
-    
+
     assertFail(tabletMeta, new KeyExtent(new Text("1"), null, new Text("m")), tsi);
-    
+
     assertFail(tabletMeta, new KeyExtent(new Text("1"), new Text("r"), new Text("m")), tsi);
-    
+
     assertFail(tabletMeta, ke, tsi, nk("1<", TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN));
-    
+
     assertFail(tabletMeta, ke, tsi, nk("1<", TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN));
-    
+
     assertFail(tabletMeta, ke, tsi, nk("1<", TabletsSection.ServerColumnFamily.TIME_COLUMN));
-    
+
     assertFail(tabletMeta, ke, tsi, nk("1<", TabletsSection.FutureLocationColumnFamily.NAME, "4"));
-    
+
     TreeMap<Key,Value> copy = new TreeMap<Key,Value>(tabletMeta);
     put(copy, "1<", TabletsSection.CurrentLocationColumnFamily.NAME, "4", "127.0.0.1:9997");
     assertFail(copy, ke, tsi);
     assertFail(copy, ke, tsi, nk("1<", TabletsSection.FutureLocationColumnFamily.NAME, "4"));
-    
+
     copy = new TreeMap<Key,Value>(tabletMeta);
     put(copy, "1<", TabletsSection.CurrentLocationColumnFamily.NAME, "5", "127.0.0.1:9998");
     assertFail(copy, ke, tsi);
     put(copy, "1<", TabletsSection.CurrentLocationColumnFamily.NAME, "6", "127.0.0.1:9999");
     assertFail(copy, ke, tsi);
-    
+
     copy = new TreeMap<Key,Value>(tabletMeta);
     put(copy, "1<", TabletsSection.FutureLocationColumnFamily.NAME, "5", "127.0.0.1:9998");
     assertFail(copy, ke, tsi);
-    
+
     assertFail(new TreeMap<Key,Value>(), ke, tsi);
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/CountingIteratorTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/CountingIteratorTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/CountingIteratorTest.java
index 154b121..8a1dca5 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/CountingIteratorTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/CountingIteratorTest.java
@@ -31,7 +31,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class CountingIteratorTest {
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/InMemoryMapTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/InMemoryMapTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/InMemoryMapTest.java
index 39c4c39..da7157a 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/InMemoryMapTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/InMemoryMapTest.java
@@ -267,17 +267,17 @@ public class InMemoryMapTest {
 
     ski1.close();
   }
-  
+
   private void deepCopyAndDelete(int interleaving, boolean interrupt) throws Exception {
     // interleaving == 0 intentionally omitted, this runs the test w/o deleting in mem map
 
     InMemoryMap imm = new InMemoryMap(false, tempFolder.newFolder().getAbsolutePath());
-    
+
     mutate(imm, "r1", "foo:cq1", 3, "bar1");
     mutate(imm, "r1", "foo:cq2", 3, "bar2");
-    
+
     MemoryIterator ski1 = imm.skvIterator();
-    
+
     AtomicBoolean iflag = new AtomicBoolean(false);
     ski1.setInterruptFlag(iflag);
 
@@ -286,7 +286,7 @@ public class InMemoryMapTest {
       if (interrupt)
         iflag.set(true);
     }
-    
+
     SortedKeyValueIterator<Key,Value> dc = ski1.deepCopy(null);
 
     if (interleaving == 2) {
@@ -335,7 +335,7 @@ public class InMemoryMapTest {
         fail("i = " + i);
       } catch (IterationInterruptedException iie) {}
   }
-   
+
   @Test
   public void testBug1() throws Exception {
     InMemoryMap imm = new InMemoryMap(false, tempFolder.newFolder().getAbsolutePath());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/TabletServerSyncCheckTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/TabletServerSyncCheckTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/TabletServerSyncCheckTest.java
index d35f07f..1a3f9fc 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/TabletServerSyncCheckTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/TabletServerSyncCheckTest.java
@@ -91,7 +91,6 @@ public class TabletServerSyncCheckTest {
 
   private class TestVolumeManagerImpl extends VolumeManagerImpl {
 
-   
     public TestVolumeManagerImpl(Map<String,Volume> volumes) {
       super(volumes, volumes.values().iterator().next(), new ConfigurationCopy(Collections.<String,String> emptyMap()));
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/TservConstraintEnvTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/TservConstraintEnvTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/TservConstraintEnvTest.java
index 2676aca..a84e890 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/TservConstraintEnvTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/TservConstraintEnvTest.java
@@ -16,9 +16,11 @@
  */
 package org.apache.accumulo.tserver;
 
-import static org.junit.Assert.*;
-import static org.powermock.api.easymock.PowerMock.*;
 import static org.easymock.EasyMock.expect;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
+import static org.powermock.api.easymock.PowerMock.createMock;
+import static org.powermock.api.easymock.PowerMock.replay;
 
 import java.nio.ByteBuffer;
 import java.util.Collections;
@@ -29,7 +31,6 @@ import org.apache.accumulo.core.data.ArrayByteSequence;
 import org.apache.accumulo.core.data.ByteSequence;
 import org.apache.accumulo.core.security.thrift.TCredentials;
 import org.apache.accumulo.server.security.SecurityOperation;
-import org.apache.accumulo.tserver.TservConstraintEnv;
 import org.junit.Test;
 
 public class TservConstraintEnvTest {
@@ -41,8 +42,7 @@ public class TservConstraintEnvTest {
     TCredentials badCred = createMock(TCredentials.class);
 
     ByteSequence bs = new ArrayByteSequence("foo".getBytes());
-    List<ByteBuffer> bbList = Collections.<ByteBuffer> singletonList(ByteBuffer.wrap(
-        bs.getBackingArray(), bs.offset(), bs.length()));
+    List<ByteBuffer> bbList = Collections.<ByteBuffer> singletonList(ByteBuffer.wrap(bs.getBackingArray(), bs.offset(), bs.length()));
 
     expect(security.userHasAuthorizations(goodCred, bbList)).andReturn(true);
     expect(security.userHasAuthorizations(badCred, bbList)).andReturn(false);
@@ -51,4 +51,4 @@ public class TservConstraintEnvTest {
     assertTrue(new TservConstraintEnv(security, goodCred).getAuthorizationsContainer().contains(bs));
     assertFalse(new TservConstraintEnv(security, badCred).getAuthorizationsContainer().contains(bs));
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/CompactionPlanTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/CompactionPlanTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/CompactionPlanTest.java
index 988d87f..1b7d00d 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/CompactionPlanTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/CompactionPlanTest.java
@@ -19,12 +19,13 @@ package org.apache.accumulo.tserver.compaction;
 
 import java.util.Set;
 
-import com.google.common.collect.ImmutableSet;
 import org.apache.accumulo.server.fs.FileRef;
 import org.junit.Rule;
 import org.junit.Test;
 import org.junit.rules.ExpectedException;
 
+import com.google.common.collect.ImmutableSet;
+
 public class CompactionPlanTest {
 
   @Rule

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategyTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategyTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategyTest.java
index abab34f..ed75a60 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategyTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/compaction/SizeLimitCompactionStrategyTest.java
@@ -30,7 +30,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class SizeLimitCompactionStrategyTest {
   private Map<FileRef,DataFileValue> nfl(String... sa) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/constraints/ConstraintCheckerTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/constraints/ConstraintCheckerTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/constraints/ConstraintCheckerTest.java
index 490c096..dd09f69 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/constraints/ConstraintCheckerTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/constraints/ConstraintCheckerTest.java
@@ -16,8 +16,13 @@
  */
 package org.apache.accumulo.tserver.constraints;
 
-import static org.easymock.EasyMock.*;
-import static org.junit.Assert.*;
+import static org.easymock.EasyMock.anyObject;
+import static org.easymock.EasyMock.createMock;
+import static org.easymock.EasyMock.createMockBuilder;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.replay;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNull;
 
 import java.util.ArrayList;
 import java.util.List;
@@ -47,9 +52,7 @@ public class ConstraintCheckerTest {
 
   @Before
   public void setup() throws NoSuchMethodException, SecurityException {
-    cc = createMockBuilder(ConstraintChecker.class)
-           .addMockedMethod("getConstraints")
-           .createMock();
+    cc = createMockBuilder(ConstraintChecker.class).addMockedMethod("getConstraints").createMock();
     constraints = new ArrayList<Constraint>();
     expect(cc.getConstraints()).andReturn(constraints);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/log/DfsLoggerTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/log/DfsLoggerTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/log/DfsLoggerTest.java
index 2ae37ed..cd652e4 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/log/DfsLoggerTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/log/DfsLoggerTest.java
@@ -16,7 +16,7 @@
  */
 package org.apache.accumulo.tserver.log;
 
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
 
 import java.util.ArrayList;
 import java.util.Collections;
@@ -28,32 +28,32 @@ import org.apache.accumulo.tserver.TabletMutations;
 import org.junit.Test;
 
 public class DfsLoggerTest {
-  
+
   @Test
   public void testDurabilityForGroupCommit() {
     List<TabletMutations> lst = new ArrayList<TabletMutations>();
     assertEquals(Durability.NONE, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m1 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.NONE);
+    TabletMutations m1 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.NONE);
     lst.add(m1);
     assertEquals(Durability.NONE, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m2 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.LOG);
+    TabletMutations m2 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.LOG);
     lst.add(m2);
     assertEquals(Durability.LOG, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m3 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.NONE);
+    TabletMutations m3 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.NONE);
     lst.add(m3);
     assertEquals(Durability.LOG, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m4 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.FLUSH);
+    TabletMutations m4 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.FLUSH);
     lst.add(m4);
     assertEquals(Durability.FLUSH, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m5 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.LOG);
+    TabletMutations m5 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.LOG);
     lst.add(m5);
     assertEquals(Durability.FLUSH, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m6 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.SYNC);
+    TabletMutations m6 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.SYNC);
     lst.add(m6);
     assertEquals(Durability.SYNC, DfsLogger.chooseDurabilityForGroupCommit(lst));
-    TabletMutations m7 = new TabletMutations(0, 1, Collections.<Mutation>emptyList(), Durability.FLUSH);
+    TabletMutations m7 = new TabletMutations(0, 1, Collections.<Mutation> emptyList(), Durability.FLUSH);
     lst.add(m7);
     assertEquals(Durability.SYNC, DfsLogger.chooseDurabilityForGroupCommit(lst));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/log/SortedLogRecoveryTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/log/SortedLogRecoveryTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/log/SortedLogRecoveryTest.java
index 03361d1..67cdb04 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/log/SortedLogRecoveryTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/log/SortedLogRecoveryTest.java
@@ -611,8 +611,8 @@ public class SortedLogRecoveryTest {
     Mutation m2 = new ServerMutation(new Text("row1"));
     m2.put("foo", "bar", "v2");
 
-    KeyValue entries[] = new KeyValue[] {createKeyValue(OPEN, 0, -1, "1"), createKeyValue(DEFINE_TABLET, 1, 2, extent),
-        createKeyValue(MUTATION, 2, 2, m1), createKeyValue(COMPACTION_START, 3, 2, compactionStartFile), createKeyValue(MUTATION, 4, 2, m2),};
+    KeyValue entries[] = new KeyValue[] {createKeyValue(OPEN, 0, -1, "1"), createKeyValue(DEFINE_TABLET, 1, 2, extent), createKeyValue(MUTATION, 2, 2, m1),
+        createKeyValue(COMPACTION_START, 3, 2, compactionStartFile), createKeyValue(MUTATION, 4, 2, m2),};
 
     Arrays.sort(entries);
     Map<String,KeyValue[]> logs = new TreeMap<String,KeyValue[]>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/log/TestUpgradePathForWALogs.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/log/TestUpgradePathForWALogs.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/log/TestUpgradePathForWALogs.java
index 1014306..45d3e73 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/log/TestUpgradePathForWALogs.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/log/TestUpgradePathForWALogs.java
@@ -50,7 +50,7 @@ public class TestUpgradePathForWALogs {
   public static void createTestDirectory() {
     File baseDir = new File(System.getProperty("user.dir") + "/target/upgrade-tests");
     baseDir.mkdirs();
-    testDir = new File(baseDir,  TestUpgradePathForWALogs.class.getName());
+    testDir = new File(baseDir, TestUpgradePathForWALogs.class.getName());
     FileUtils.deleteQuietly(testDir);
     testDir.mkdir();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/logger/LogFileTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/logger/LogFileTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/logger/LogFileTest.java
index f91c195..cf2266f 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/logger/LogFileTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/logger/LogFileTest.java
@@ -33,16 +33,13 @@ import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.security.ColumnVisibility;
 import org.apache.accumulo.server.data.ServerMutation;
-import org.apache.accumulo.tserver.logger.LogEvents;
-import org.apache.accumulo.tserver.logger.LogFileKey;
-import org.apache.accumulo.tserver.logger.LogFileValue;
 import org.apache.hadoop.io.DataInputBuffer;
 import org.apache.hadoop.io.DataOutputBuffer;
 import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class LogFileTest {
-  
+
   static private void readWrite(LogEvents event, long seq, int tid, String filename, KeyExtent tablet, Mutation[] mutations, LogFileKey keyResult,
       LogFileValue valueResult) throws IOException {
     LogFileKey key = new LogFileKey();
@@ -66,7 +63,7 @@ public class LogFileTest {
     assertEquals(value.mutations, valueResult.mutations);
     assertTrue(in.read() == -1);
   }
-  
+
   @Test
   public void testReadFields() throws IOException {
     LogFileKey key = new LogFileKey();
@@ -111,14 +108,14 @@ public class LogFileTest {
     assertEquals(key.tid, 10);
     assertEquals(value.mutations, Arrays.asList(m, m));
   }
-  
+
   @Test
   public void testEventType() {
     assertEquals(LogFileKey.eventType(MUTATION), LogFileKey.eventType(MANY_MUTATIONS));
     assertEquals(LogFileKey.eventType(COMPACTION_START), LogFileKey.eventType(COMPACTION_FINISH));
     assertTrue(LogFileKey.eventType(DEFINE_TABLET) < LogFileKey.eventType(COMPACTION_FINISH));
     assertTrue(LogFileKey.eventType(COMPACTION_FINISH) < LogFileKey.eventType(MUTATION));
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystemTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystemTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystemTest.java
index 90a2aa4..1a32ea4 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystemTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/AccumuloReplicaSystemTest.java
@@ -59,7 +59,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class AccumuloReplicaSystemTest {
 
@@ -159,7 +159,8 @@ public class AccumuloReplicaSystemTest {
 
     Status status = Status.newBuilder().setBegin(0).setEnd(0).setInfiniteEnd(true).setClosed(false).build();
     DataInputStream dis = new DataInputStream(new ByteArrayInputStream(baos.toByteArray()));
-    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE, new HashSet<Integer>());
+    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE,
+        new HashSet<Integer>());
 
     // We stopped because we got to the end of the file
     Assert.assertEquals(9, repl.entriesConsumed);
@@ -266,7 +267,8 @@ public class AccumuloReplicaSystemTest {
     // If it were still open, more data could be appended that we need to process
     Status status = Status.newBuilder().setBegin(0).setEnd(0).setInfiniteEnd(true).setClosed(true).build();
     DataInputStream dis = new DataInputStream(new ByteArrayInputStream(baos.toByteArray()));
-    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE, new HashSet<Integer>());
+    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE,
+        new HashSet<Integer>());
 
     // We stopped because we got to the end of the file
     Assert.assertEquals(Long.MAX_VALUE, repl.entriesConsumed);
@@ -331,7 +333,8 @@ public class AccumuloReplicaSystemTest {
     // If it were still open, more data could be appended that we need to process
     Status status = Status.newBuilder().setBegin(100).setEnd(0).setInfiniteEnd(true).setClosed(true).build();
     DataInputStream dis = new DataInputStream(new ByteArrayInputStream(new byte[0]));
-    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE, new HashSet<Integer>());
+    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE,
+        new HashSet<Integer>());
 
     // We stopped because we got to the end of the file
     Assert.assertEquals(Long.MAX_VALUE, repl.entriesConsumed);
@@ -353,7 +356,8 @@ public class AccumuloReplicaSystemTest {
     // If it were still open, more data could be appended that we need to process
     Status status = Status.newBuilder().setBegin(100).setEnd(0).setInfiniteEnd(true).setClosed(false).build();
     DataInputStream dis = new DataInputStream(new ByteArrayInputStream(new byte[0]));
-    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE, new HashSet<Integer>());
+    WalReplication repl = ars.getWalEdits(new ReplicationTarget("peer", "1", "1"), dis, new Path("/accumulo/wals/tserver+port/wal"), status, Long.MAX_VALUE,
+        new HashSet<Integer>());
 
     // We stopped because we got to the end of the file
     Assert.assertEquals(0, repl.entriesConsumed);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/BatchWriterReplicationReplayerTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/BatchWriterReplicationReplayerTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/BatchWriterReplicationReplayerTest.java
index aba0b1c..4584f56 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/BatchWriterReplicationReplayerTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/BatchWriterReplicationReplayerTest.java
@@ -47,7 +47,7 @@ import org.junit.Test;
 import com.google.common.collect.Lists;
 
 /**
- * 
+ *
  */
 public class BatchWriterReplicationReplayerTest {
 
@@ -80,7 +80,6 @@ public class BatchWriterReplicationReplayerTest {
     final BatchWriterConfig bwCfg = new BatchWriterConfig();
     bwCfg.setMaxMemory(1l);
 
-
     LogFileKey key = new LogFileKey();
     key.event = LogEvents.MANY_MUTATIONS;
     key.seq = 1;
@@ -133,7 +132,7 @@ public class BatchWriterReplicationReplayerTest {
     expectLastCall().once();
 
     replay(conn, conf, bw);
-    
+
     replayer.replicateLog(context, tableName, edits);
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/ReplicationProcessorTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/ReplicationProcessorTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/ReplicationProcessorTest.java
index d389c0b..15306a8 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/ReplicationProcessorTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/replication/ReplicationProcessorTest.java
@@ -38,7 +38,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ReplicationProcessorTest {
 
@@ -78,7 +78,8 @@ public class ReplicationProcessorTest {
   public void filesWhichMakeNoProgressArentReplicatedAgain() throws Exception {
     ReplicaSystem replica = EasyMock.createMock(ReplicaSystem.class);
     ReplicaSystemHelper helper = EasyMock.createMock(ReplicaSystemHelper.class);
-    ReplicationProcessor proc = EasyMock.createMockBuilder(ReplicationProcessor.class).addMockedMethods("getReplicaSystem", "doesFileExist", "getStatus", "getHelper").createMock();
+    ReplicationProcessor proc = EasyMock.createMockBuilder(ReplicationProcessor.class)
+        .addMockedMethods("getReplicaSystem", "doesFileExist", "getStatus", "getHelper").createMock();
 
     ReplicationTarget target = new ReplicationTarget("peer", "1", "1");
     Status status = Status.newBuilder().setBegin(0).setEnd(0).setInfiniteEnd(true).setClosed(true).build();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/RootFilesTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/RootFilesTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/RootFilesTest.java
index 9c75a66..743cfc2 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/RootFilesTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/RootFilesTest.java
@@ -29,7 +29,6 @@ import org.apache.accumulo.core.conf.Property;
 import org.apache.accumulo.server.fs.FileRef;
 import org.apache.accumulo.server.fs.VolumeManager;
 import org.apache.accumulo.server.fs.VolumeManagerImpl;
-import org.apache.accumulo.tserver.tablet.RootFiles;
 import org.apache.hadoop.fs.Path;
 import org.junit.Assert;
 import org.junit.Rule;
@@ -37,7 +36,7 @@ import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
 /**
- * 
+ *
  */
 public class RootFilesTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/TabletTest.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/TabletTest.java b/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/TabletTest.java
index a40a871..c4969e6 100644
--- a/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/TabletTest.java
+++ b/server/tserver/src/test/java/org/apache/accumulo/tserver/tablet/TabletTest.java
@@ -36,7 +36,7 @@ import org.junit.Test;
 import com.google.common.collect.Iterators;
 
 /**
- * 
+ *
  */
 public class TabletTest {
 
@@ -55,7 +55,7 @@ public class TabletTest {
     ConfigurationObserver obs = EasyMock.createMock(ConfigurationObserver.class);
 
     Tablet tablet = new Tablet(time, "", 0, new Path("/foo"), dfm, tserver, tserverResourceManager, tabletMemory, tableConf, extent, obs);
-    
+
     long hdfsBlockSize = 10000l, blockSize = 5000l, indexBlockSize = 500l;
     int replication = 5;
     String compressType = "snappy";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/shell/src/main/java/org/apache/accumulo/shell/Shell.java
----------------------------------------------------------------------
diff --git a/shell/src/main/java/org/apache/accumulo/shell/Shell.java b/shell/src/main/java/org/apache/accumulo/shell/Shell.java
index 58308ff..9697a85 100644
--- a/shell/src/main/java/org/apache/accumulo/shell/Shell.java
+++ b/shell/src/main/java/org/apache/accumulo/shell/Shell.java
@@ -229,9 +229,8 @@ public class Shell extends ShellOptions {
   private boolean masking = false;
 
   public Shell() throws IOException {
-    this(new ConsoleReader(), new PrintWriter(
-        new OutputStreamWriter(System.out,
-        System.getProperty("jline.WindowsTerminal.output.encoding", System.getProperty("file.encoding")))));
+    this(new ConsoleReader(), new PrintWriter(new OutputStreamWriter(System.out, System.getProperty("jline.WindowsTerminal.output.encoding",
+        System.getProperty("file.encoding")))));
   }
 
   public Shell(ConsoleReader reader, PrintWriter writer) {


[44/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ThriftScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ThriftScanner.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ThriftScanner.java
index 7b6284d..1963532 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ThriftScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ThriftScanner.java
@@ -69,15 +69,15 @@ import com.google.common.net.HostAndPort;
 
 public class ThriftScanner {
   private static final Logger log = Logger.getLogger(ThriftScanner.class);
-  
+
   public static final Map<TabletType,Set<String>> serversWaitedForWrites = new EnumMap<TabletType,Set<String>>(TabletType.class);
-  
+
   static {
     for (TabletType ttype : TabletType.values()) {
       serversWaitedForWrites.put(ttype, Collections.synchronizedSet(new HashSet<String>()));
     }
   }
-  
+
   public static boolean getBatchFromServer(ClientContext context, Range range, KeyExtent extent, String server, SortedMap<Key,Value> results,
       SortedSet<Column> fetchedColumns, List<IterInfo> serverSideIteratorList, Map<String,Map<String,String>> serverSideIteratorOptions, int size,
       Authorizations authorizations, boolean retry) throws AccumuloException, AccumuloSecurityException, NotServingTabletException {
@@ -92,7 +92,7 @@ public class ThriftScanner {
         // not reading whole rows (or stopping on row boundries) so there is no need to enable isolation below
         ScanState scanState = new ScanState(context, extent.getTableId(), authorizations, range, fetchedColumns, size, serverSideIteratorList,
             serverSideIteratorOptions, false);
-        
+
         TabletType ttype = TabletType.type(extent);
         boolean waitForWrites = !serversWaitedForWrites.get(ttype).contains(server);
         InitialScan isr = client.startScan(tinfo, scanState.context.rpcCreds(), extent.toThrift(), scanState.range.toThrift(),
@@ -100,14 +100,14 @@ public class ThriftScanner {
             scanState.authorizations.getAuthorizationsBB(), waitForWrites, scanState.isolated, scanState.readaheadThreshold);
         if (waitForWrites)
           serversWaitedForWrites.get(ttype).add(server);
-        
+
         Key.decompress(isr.result.results);
-        
+
         for (TKeyValue kv : isr.result.results)
           results.put(new Key(kv.key), new Value(kv.value));
-        
+
         client.closeScan(tinfo, isr.scanID);
-        
+
         return isr.result.more;
       } finally {
         ThriftUtil.returnClient(client);
@@ -122,33 +122,33 @@ public class ThriftScanner {
     } catch (TException e) {
       log.debug("Error getting transport to " + server + " : " + e);
     }
-    
+
     throw new AccumuloException("getBatchFromServer: failed");
   }
-  
+
   public static class ScanState {
-    
+
     boolean isolated;
     Text tableId;
     Text startRow;
     boolean skipStartRow;
     long readaheadThreshold;
-    
+
     Range range;
-    
+
     int size;
-    
+
     ClientContext context;
     Authorizations authorizations;
     List<Column> columns;
-    
+
     TabletLocation prevLoc;
     Long scanID;
-    
+
     boolean finished = false;
-    
+
     List<IterInfo> serverSideIteratorList;
-    
+
     Map<String,Map<String,String>> serverSideIteratorOptions;
 
     public ScanState(ClientContext context, Text tableId, Authorizations authorizations, Range range, SortedSet<Column> fetchedColumns, int size,
@@ -162,40 +162,40 @@ public class ThriftScanner {
       this.context = context;
       ;
       this.authorizations = authorizations;
-      
+
       columns = new ArrayList<Column>(fetchedColumns.size());
       for (Column column : fetchedColumns) {
         columns.add(column);
       }
-      
+
       this.tableId = tableId;
       this.range = range;
-      
+
       Key startKey = range.getStartKey();
       if (startKey == null) {
         startKey = new Key();
       }
       this.startRow = startKey.getRow();
-      
+
       this.skipStartRow = false;
-      
+
       this.size = size;
-      
+
       this.serverSideIteratorList = serverSideIteratorList;
       this.serverSideIteratorOptions = serverSideIteratorOptions;
-      
+
       this.isolated = isolated;
       this.readaheadThreshold = readaheadThreshold;
-      
+
     }
   }
-  
+
   public static class ScanTimedOutException extends IOException {
-    
+
     private static final long serialVersionUID = 1L;
-    
+
   }
-  
+
   public static List<KeyValue> scan(ClientContext context, ScanState scanState, int timeOut) throws ScanTimedOutException, AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
     TabletLocation loc = null;
@@ -204,34 +204,34 @@ public class ThriftScanner {
     String lastError = null;
     String error = null;
     int tooManyFilesCount = 0;
-    
+
     List<KeyValue> results = null;
-    
+
     Span span = Trace.start("scan");
     try {
       while (results == null && !scanState.finished) {
         if (Thread.currentThread().isInterrupted()) {
           throw new AccumuloException("Thread interrupted");
         }
-        
+
         if ((System.currentTimeMillis() - startTime) / 1000.0 > timeOut)
           throw new ScanTimedOutException();
-        
+
         while (loc == null) {
           long currentTime = System.currentTimeMillis();
           if ((currentTime - startTime) / 1000.0 > timeOut)
             throw new ScanTimedOutException();
-          
+
           Span locateSpan = Trace.start("scan:locateTablet");
           try {
             loc = TabletLocator.getLocator(context, scanState.tableId).locateTablet(context, scanState.startRow, scanState.skipStartRow, false);
-            
+
             if (loc == null) {
               if (!Tables.exists(instance, scanState.tableId.toString()))
                 throw new TableDeletedException(scanState.tableId.toString());
               else if (Tables.getTableState(instance, scanState.tableId.toString()) == TableState.OFFLINE)
                 throw new TableOfflineException(instance, scanState.tableId.toString());
-              
+
               error = "Failed to locate tablet for table : " + scanState.tableId + " row : " + scanState.startRow;
               if (!error.equals(lastError))
                 log.debug(error);
@@ -243,7 +243,7 @@ public class ThriftScanner {
               // when a tablet splits we do want to continue scanning the low child
               // of the split if we are already passed it
               Range dataRange = loc.tablet_extent.toDataRange();
-              
+
               if (scanState.range.getStartKey() != null && dataRange.afterEndKey(scanState.range.getStartKey())) {
                 // go to the next tablet
                 scanState.startRow = loc.tablet_extent.getEndRow();
@@ -264,14 +264,14 @@ public class ThriftScanner {
               log.debug(error);
             else if (log.isTraceEnabled())
               log.trace(error);
-            
+
             lastError = error;
             Thread.sleep(100);
           } finally {
             locateSpan.stop();
           }
         }
-        
+
         Span scanLocation = Trace.start("scan:location");
         scanLocation.data("tserver", loc.tablet_location);
         try {
@@ -291,16 +291,16 @@ public class ThriftScanner {
           else if (log.isTraceEnabled())
             log.trace(error);
           lastError = error;
-          
+
           TabletLocator.getLocator(context, scanState.tableId).invalidateCache(loc.tablet_extent);
           loc = null;
-          
+
           // no need to try the current scan id somewhere else
           scanState.scanID = null;
-          
+
           if (scanState.isolated)
             throw new IsolationException();
-          
+
           Thread.sleep(100);
         } catch (NoSuchScanIDException e) {
           error = "Scan failed, no such scan id " + scanState.scanID + " " + loc;
@@ -309,10 +309,10 @@ public class ThriftScanner {
           else if (log.isTraceEnabled())
             log.trace(error);
           lastError = error;
-          
+
           if (scanState.isolated)
             throw new IsolationException();
-          
+
           scanState.scanID = null;
         } catch (TooManyFilesException e) {
           error = "Tablet has too many files " + loc + " retrying...";
@@ -327,15 +327,15 @@ public class ThriftScanner {
               log.trace(error);
           }
           lastError = error;
-          
+
           // not sure what state the scan session on the server side is
           // in after this occurs, so lets be cautious and start a new
           // scan session
           scanState.scanID = null;
-          
+
           if (scanState.isolated)
             throw new IsolationException();
-          
+
           Thread.sleep(100);
         } catch (TException e) {
           TabletLocator.getLocator(context, scanState.tableId).invalidateCache(context.getInstance(), loc.tablet_location);
@@ -346,24 +346,24 @@ public class ThriftScanner {
             log.trace(error);
           lastError = error;
           loc = null;
-          
+
           // do not want to continue using the same scan id, if a timeout occurred could cause a batch to be skipped
           // because a thread on the server side may still be processing the timed out continue scan
           scanState.scanID = null;
-          
+
           if (scanState.isolated)
             throw new IsolationException();
-          
+
           Thread.sleep(100);
         } finally {
           scanLocation.stop();
         }
       }
-      
+
       if (results != null && results.size() == 0 && scanState.finished) {
         results = null;
       }
-      
+
       return results;
     } catch (InterruptedException ex) {
       throw new AccumuloException(ex);
@@ -371,12 +371,12 @@ public class ThriftScanner {
       span.stop();
     }
   }
-  
+
   private static List<KeyValue> scan(TabletLocation loc, ScanState scanState, ClientContext context) throws AccumuloSecurityException,
       NotServingTabletException, TException, NoSuchScanIDException, TooManyFilesException {
     if (scanState.finished)
       return null;
-    
+
     OpTimer opTimer = new OpTimer(log, Level.TRACE);
 
     final TInfo tinfo = Tracer.traceInfo();
@@ -386,18 +386,18 @@ public class ThriftScanner {
     String old = Thread.currentThread().getName();
     try {
       ScanResult sr;
-      
+
       if (scanState.prevLoc != null && !scanState.prevLoc.equals(loc))
         scanState.scanID = null;
-      
+
       scanState.prevLoc = loc;
-      
+
       if (scanState.scanID == null) {
         String msg = "Starting scan tserver=" + loc.tablet_location + " tablet=" + loc.tablet_extent + " range=" + scanState.range + " ssil="
             + scanState.serverSideIteratorList + " ssio=" + scanState.serverSideIteratorOptions;
         Thread.currentThread().setName(msg);
         opTimer.start(msg);
-        
+
         TabletType ttype = TabletType.type(loc.tablet_extent);
         boolean waitForWrites = !serversWaitedForWrites.get(ttype).contains(loc.tablet_location);
         InitialScan is = client.startScan(tinfo, scanState.context.rpcCreds(), loc.tablet_extent.toThrift(), scanState.range.toThrift(),
@@ -405,27 +405,27 @@ public class ThriftScanner {
             scanState.authorizations.getAuthorizationsBB(), waitForWrites, scanState.isolated, scanState.readaheadThreshold);
         if (waitForWrites)
           serversWaitedForWrites.get(ttype).add(loc.tablet_location);
-        
+
         sr = is.result;
-        
+
         if (sr.more)
           scanState.scanID = is.scanID;
         else
           client.closeScan(tinfo, is.scanID);
-        
+
       } else {
         // log.debug("Calling continue scan : "+scanState.range+"  loc = "+loc);
         String msg = "Continuing scan tserver=" + loc.tablet_location + " scanid=" + scanState.scanID;
         Thread.currentThread().setName(msg);
         opTimer.start(msg);
-        
+
         sr = client.continueScan(tinfo, scanState.scanID);
         if (!sr.more) {
           client.closeScan(tinfo, scanState.scanID);
           scanState.scanID = null;
         }
       }
-      
+
       if (!sr.more) {
         // log.debug("No more : tab end row = "+loc.tablet_extent.getEndRow()+" range = "+scanState.range);
         if (loc.tablet_extent.getEndRow() == null) {
@@ -442,18 +442,18 @@ public class ThriftScanner {
       } else {
         opTimer.stop("Finished scan in %DURATION% #results=" + sr.results.size() + " scanid=" + scanState.scanID);
       }
-      
+
       Key.decompress(sr.results);
-      
+
       if (sr.results.size() > 0 && !scanState.finished)
         scanState.range = new Range(new Key(sr.results.get(sr.results.size() - 1).key), false, scanState.range.getEndKey(), scanState.range.isEndKeyInclusive());
-      
+
       List<KeyValue> results = new ArrayList<KeyValue>(sr.results.size());
       for (TKeyValue tkv : sr.results)
         results.add(new KeyValue(new Key(tkv.key), tkv.value));
-      
+
       return results;
-      
+
     } catch (ThriftSecurityException e) {
       throw new AccumuloSecurityException(e.user, e.code, e);
     } finally {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TimeoutTabletLocator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TimeoutTabletLocator.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TimeoutTabletLocator.java
index 644ba31..8a6777d 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TimeoutTabletLocator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TimeoutTabletLocator.java
@@ -31,14 +31,14 @@ import org.apache.accumulo.core.data.Range;
 import org.apache.hadoop.io.Text;
 
 /**
- * 
+ *
  */
 public class TimeoutTabletLocator extends TabletLocator {
-  
+
   private TabletLocator locator;
   private long timeout;
   private Long firstFailTime = null;
-  
+
   private void failed() {
     if (firstFailTime == null) {
       firstFailTime = System.currentTimeMillis();
@@ -46,93 +46,93 @@ public class TimeoutTabletLocator extends TabletLocator {
       throw new TimedOutException("Failed to obtain metadata");
     }
   }
-  
+
   private void succeeded() {
     firstFailTime = null;
   }
-  
+
   public TimeoutTabletLocator(TabletLocator locator, long timeout) {
     this.locator = locator;
     this.timeout = timeout;
   }
-  
+
   @Override
   public TabletLocation locateTablet(ClientContext context, Text row, boolean skipRow, boolean retry) throws AccumuloException, AccumuloSecurityException,
       TableNotFoundException {
-    
+
     try {
       TabletLocation ret = locator.locateTablet(context, row, skipRow, retry);
-      
+
       if (ret == null)
         failed();
       else
         succeeded();
-      
+
       return ret;
     } catch (AccumuloException ae) {
       failed();
       throw ae;
     }
   }
-  
+
   @Override
   public <T extends Mutation> void binMutations(ClientContext context, List<T> mutations, Map<String,TabletServerMutations<T>> binnedMutations, List<T> failures)
       throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     try {
       locator.binMutations(context, mutations, binnedMutations, failures);
-      
+
       if (failures.size() == mutations.size())
         failed();
       else
         succeeded();
-      
+
     } catch (AccumuloException ae) {
       failed();
       throw ae;
     }
   }
-  
+
   /**
-   * 
+   *
    */
-  
+
   @Override
   public List<Range> binRanges(ClientContext context, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
-    
+
     try {
       List<Range> ret = locator.binRanges(context, ranges, binnedRanges);
-      
+
       if (ranges.size() == ret.size())
         failed();
       else
         succeeded();
-      
+
       return ret;
     } catch (AccumuloException ae) {
       failed();
       throw ae;
     }
   }
-  
+
   @Override
   public void invalidateCache(KeyExtent failedExtent) {
     locator.invalidateCache(failedExtent);
   }
-  
+
   @Override
   public void invalidateCache(Collection<KeyExtent> keySet) {
     locator.invalidateCache(keySet);
   }
-  
+
   @Override
   public void invalidateCache() {
     locator.invalidateCache();
   }
-  
+
   @Override
   public void invalidateCache(Instance instance, String server) {
     locator.invalidateCache(instance, server);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/Translator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/Translator.java b/core/src/main/java/org/apache/accumulo/core/client/impl/Translator.java
index 3ce542d..2503d95 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/Translator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/Translator.java
@@ -33,31 +33,31 @@ import org.apache.accumulo.core.data.thrift.TKeyExtent;
 import org.apache.accumulo.core.data.thrift.TRange;
 
 public abstract class Translator<IT,OT> {
-  
+
   public abstract OT translate(IT input);
-  
+
   public static class TKeyExtentTranslator extends Translator<TKeyExtent,KeyExtent> {
     @Override
     public KeyExtent translate(TKeyExtent input) {
       return new KeyExtent(input);
     }
-    
+
   }
-  
+
   public static class KeyExtentTranslator extends Translator<KeyExtent,TKeyExtent> {
     @Override
     public TKeyExtent translate(KeyExtent input) {
       return input.toThrift();
     }
   }
-  
+
   public static class TCVSTranslator extends Translator<TConstraintViolationSummary,ConstraintViolationSummary> {
     @Override
     public ConstraintViolationSummary translate(TConstraintViolationSummary input) {
       return new ConstraintViolationSummary(input);
     }
   }
-  
+
   public static class CVSTranslator extends Translator<ConstraintViolationSummary,TConstraintViolationSummary> {
     @Override
     public TConstraintViolationSummary translate(ConstraintViolationSummary input) {
@@ -71,69 +71,69 @@ public abstract class Translator<IT,OT> {
       return new Column(input);
     }
   }
-  
+
   public static class ColumnTranslator extends Translator<Column,TColumn> {
     @Override
     public TColumn translate(Column input) {
       return input.toThrift();
     }
   }
-  
+
   public static class TRangeTranslator extends Translator<TRange,Range> {
-    
+
     @Override
     public Range translate(TRange input) {
       return new Range(input);
     }
-    
+
   }
-  
+
   public static class RangeTranslator extends Translator<Range,TRange> {
     @Override
     public TRange translate(Range input) {
       return input.toThrift();
     }
   }
-  
+
   public static class ListTranslator<IT,OT> extends Translator<List<IT>,List<OT>> {
-    
+
     private Translator<IT,OT> translator;
-    
+
     public ListTranslator(Translator<IT,OT> translator) {
       this.translator = translator;
     }
-    
+
     @Override
     public List<OT> translate(List<IT> input) {
       return translate(input, this.translator);
     }
-    
+
   }
-  
+
   public static <IKT,OKT,T> Map<OKT,T> translate(Map<IKT,T> input, Translator<IKT,OKT> keyTranslator) {
     HashMap<OKT,T> output = new HashMap<OKT,T>();
-    
+
     for (Entry<IKT,T> entry : input.entrySet())
       output.put(keyTranslator.translate(entry.getKey()), entry.getValue());
-    
+
     return output;
   }
-  
+
   public static <IKT,OKT,IVT,OVT> Map<OKT,OVT> translate(Map<IKT,IVT> input, Translator<IKT,OKT> keyTranslator, Translator<IVT,OVT> valueTranslator) {
     HashMap<OKT,OVT> output = new HashMap<OKT,OVT>();
-    
+
     for (Entry<IKT,IVT> entry : input.entrySet())
       output.put(keyTranslator.translate(entry.getKey()), valueTranslator.translate(entry.getValue()));
-    
+
     return output;
   }
-  
+
   public static <IT,OT> List<OT> translate(Collection<IT> input, Translator<IT,OT> translator) {
     ArrayList<OT> output = new ArrayList<OT>(input.size());
-    
+
     for (IT in : input)
       output.add(translator.translate(in));
-    
+
     return output;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/Writer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/Writer.java b/core/src/main/java/org/apache/accumulo/core/client/impl/Writer.java
index 552ddae..4f325be 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/Writer.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/Writer.java
@@ -41,19 +41,19 @@ import org.apache.thrift.TServiceClient;
 import com.google.common.net.HostAndPort;
 
 public class Writer {
-  
+
   private static final Logger log = Logger.getLogger(Writer.class);
-  
+
   private ClientContext context;
   private Text table;
-  
+
   public Writer(ClientContext context, Text table) {
     checkArgument(context != null, "context is null");
     checkArgument(table != null, "table is null");
     this.context = context;
     this.table = table;
   }
-  
+
   public Writer(ClientContext context, String table) {
     this(context, new Text(table));
   }
@@ -64,7 +64,7 @@ public class Writer {
     checkArgument(extent != null, "extent is null");
     checkArgument(server != null, "server is null");
     checkArgument(context != null, "context is null");
-    
+
     TabletClientService.Iface client = null;
     try {
       client = ThriftUtil.getTServerClient(server, context);
@@ -76,16 +76,16 @@ public class Writer {
       ThriftUtil.returnClient((TServiceClient) client);
     }
   }
-  
+
   public void update(Mutation m) throws AccumuloException, AccumuloSecurityException, ConstraintViolationException, TableNotFoundException {
     checkArgument(m != null, "m is null");
-    
+
     if (m.size() == 0)
       throw new IllegalArgumentException("Can not add empty mutations");
-    
+
     while (true) {
       TabletLocation tabLoc = TabletLocator.getLocator(context, table).locateTablet(context, new Text(m.getRow()), false, true);
-      
+
       if (tabLoc == null) {
         log.trace("No tablet location found for row " + new String(m.getRow(), UTF_8));
         UtilWaitThread.sleep(500);
@@ -108,9 +108,9 @@ public class Writer {
         log.error("error sending update to " + parsedLocation + ": " + e);
         TabletLocator.getLocator(context, table).invalidateCache(tabLoc.tablet_extent);
       }
-      
+
       UtilWaitThread.sleep(500);
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/ZookeeperLockChecker.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/ZookeeperLockChecker.java b/core/src/main/java/org/apache/accumulo/core/client/impl/ZookeeperLockChecker.java
index be56ad4..a9c33b5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/ZookeeperLockChecker.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/ZookeeperLockChecker.java
@@ -26,21 +26,22 @@ import org.apache.accumulo.fate.zookeeper.ZooLock;
 import org.apache.zookeeper.KeeperException;
 
 /**
- * 
+ *
  */
 public class ZookeeperLockChecker implements TabletServerLockChecker {
-  
+
   private final ZooCache zc;
   private final String root;
 
   ZookeeperLockChecker(Instance instance) {
     this(instance, new ZooCacheFactory());
   }
+
   ZookeeperLockChecker(Instance instance, ZooCacheFactory zcf) {
     zc = zcf.getZooCache(instance.getZooKeepers(), instance.getZooKeepersSessionTimeOut());
     this.root = ZooUtil.getRoot(instance) + Constants.ZTSERVERS;
   }
-  
+
   @Override
   public boolean isLockHeld(String tserver, String session) {
     try {
@@ -51,10 +52,10 @@ public class ZookeeperLockChecker implements TabletServerLockChecker {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public void invalidateCache(String tserver) {
     zc.clear(root + "/" + tserver);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoder.java
index 12bcdd2..838e3cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BigIntegerLexicoder.java
@@ -27,58 +27,59 @@ import org.apache.accumulo.core.iterators.ValueFormatException;
 
 /**
  * A lexicoder to encode/decode a BigInteger to/from bytes that maintain its native Java sort order.
+ *
  * @since 1.6.0
  */
 public class BigIntegerLexicoder implements Lexicoder<BigInteger> {
-  
+
   @Override
   public byte[] encode(BigInteger v) {
-    
+
     try {
       byte[] bytes = v.toByteArray();
-      
+
       byte[] ret = new byte[4 + bytes.length];
-      
+
       DataOutputStream dos = new DataOutputStream(new FixedByteArrayOutputStream(ret));
-      
+
       // flip the sign bit
       bytes[0] = (byte) (0x80 ^ bytes[0]);
-      
+
       int len = bytes.length;
       if (v.signum() < 0)
         len = -len;
-      
+
       len = len ^ 0x80000000;
-      
+
       dos.writeInt(len);
       dos.write(bytes);
       dos.close();
-      
+
       return ret;
     } catch (IOException ioe) {
       throw new RuntimeException(ioe);
     }
-    
+
   }
-  
+
   @Override
   public BigInteger decode(byte[] b) throws ValueFormatException {
-    
+
     try {
       DataInputStream dis = new DataInputStream(new ByteArrayInputStream(b));
       int len = dis.readInt();
       len = len ^ 0x80000000;
       len = Math.abs(len);
-      
+
       byte[] bytes = new byte[len];
       dis.readFully(bytes);
-      
+
       bytes[0] = (byte) (0x80 ^ bytes[0]);
-      
+
       return new BigInteger(bytes);
     } catch (IOException ioe) {
       throw new RuntimeException(ioe);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BytesLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BytesLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BytesLexicoder.java
index b6fbab5..e018e0c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BytesLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/BytesLexicoder.java
@@ -19,19 +19,19 @@ package org.apache.accumulo.core.client.lexicoder;
 /**
  * For each of the methods, this lexicoder just passes the input through untouched. It is meant to be combined with other lexicoders like the
  * {@link ReverseLexicoder}.
- * 
+ *
  * @since 1.6.0
  */
 public class BytesLexicoder implements Lexicoder<byte[]> {
-  
+
   @Override
   public byte[] encode(byte[] data) {
     return data;
   }
-  
+
   @Override
   public byte[] decode(byte[] data) {
     return data;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoder.java
index 5a25ede..6310645 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/DoubleLexicoder.java
@@ -22,9 +22,9 @@ package org.apache.accumulo.core.client.lexicoder;
  * @since 1.6.0
  */
 public class DoubleLexicoder implements Lexicoder<Double> {
-  
+
   private ULongLexicoder longEncoder = new ULongLexicoder();
-  
+
   @Override
   public byte[] encode(Double d) {
     long l = Double.doubleToRawLongBits(d);
@@ -32,10 +32,10 @@ public class DoubleLexicoder implements Lexicoder<Double> {
       l = ~l;
     else
       l = l ^ 0x8000000000000000l;
-    
+
     return longEncoder.encode(l);
   }
-  
+
   @Override
   public Double decode(byte[] data) {
     long l = longEncoder.decode(data);
@@ -45,5 +45,5 @@ public class DoubleLexicoder implements Lexicoder<Double> {
       l = ~l;
     return Double.longBitsToDouble(l);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Encoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Encoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Encoder.java
index 313ac6d..f1249ca 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Encoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Encoder.java
@@ -18,8 +18,9 @@ package org.apache.accumulo.core.client.lexicoder;
 
 /**
  * An encoder represents a typed object that can be encoded/decoded to/from a byte array.
+ *
  * @since 1.6.0
  */
 public interface Encoder<T> extends org.apache.accumulo.core.iterators.TypedValueCombiner.Encoder<T> {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/IntegerLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/IntegerLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/IntegerLexicoder.java
index 8226421..12b515a 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/IntegerLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/IntegerLexicoder.java
@@ -17,22 +17,23 @@
 package org.apache.accumulo.core.client.lexicoder;
 
 /**
- * A lexicoder for signed integers. The encoding sorts Integer.MIN_VALUE first and Integer.MAX_VALUE last. The encoding sorts -2 before -1. It
- * corresponds to the sort order of Integer.
+ * A lexicoder for signed integers. The encoding sorts Integer.MIN_VALUE first and Integer.MAX_VALUE last. The encoding sorts -2 before -1. It corresponds to
+ * the sort order of Integer.
+ *
  * @since 1.6.0
  */
 public class IntegerLexicoder implements Lexicoder<Integer> {
-  
+
   private UIntegerLexicoder uil = new UIntegerLexicoder();
-  
+
   @Override
   public byte[] encode(Integer i) {
     return uil.encode(i ^ 0x80000000);
   }
-  
+
   @Override
   public Integer decode(byte[] data) {
     return uil.decode(data) ^ 0x80000000;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Lexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Lexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Lexicoder.java
index 2c7d4ae..b2ef8d2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Lexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/Lexicoder.java
@@ -19,9 +19,9 @@ package org.apache.accumulo.core.client.lexicoder;
 /**
  * A lexicoder provides methods to convert to/from byte arrays. The byte arrays are constructed so that their sort order corresponds their parameterized class's
  * native Java sort order.
- * 
+ *
  * @since 1.6.0
  */
 public interface Lexicoder<T> extends Encoder<T> {
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ListLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ListLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ListLexicoder.java
index 06eb962..5ecee88 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ListLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ListLexicoder.java
@@ -26,42 +26,42 @@ import java.util.List;
 
 /**
  * A lexicoder to encode/decode a Java List to/from a byte array where the concatenation of each encoded element sorts lexicographically.
- * 
+ *
  * @since 1.6.0
  */
 public class ListLexicoder<LT> implements Lexicoder<List<LT>> {
-  
+
   private Lexicoder<LT> lexicoder;
-  
+
   public ListLexicoder(Lexicoder<LT> lexicoder) {
     this.lexicoder = lexicoder;
   }
-  
+
   /**
    * @return a byte array containing the concatenation of each element in the list encoded.
    */
   @Override
   public byte[] encode(List<LT> v) {
     byte[][] encElements = new byte[v.size()][];
-    
+
     int index = 0;
     for (LT element : v) {
       encElements[index++] = escape(lexicoder.encode(element));
     }
-    
+
     return concat(encElements);
   }
-  
+
   @Override
   public List<LT> decode(byte[] b) {
-    
+
     byte[][] escapedElements = split(b);
     ArrayList<LT> ret = new ArrayList<LT>(escapedElements.length);
-    
+
     for (byte[] escapedElement : escapedElements) {
       ret.add(lexicoder.decode(unescape(escapedElement)));
     }
-    
+
     return ret;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/LongLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/LongLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/LongLexicoder.java
index 9106e26..f70a83c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/LongLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/LongLexicoder.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.core.client.lexicoder;
 /**
  * Signed long lexicoder. The encoding sorts Long.MIN_VALUE first and Long.MAX_VALUE last. The encoding sorts -2l before -1l. It corresponds to the native Java
  * sort order of Long.
- * 
+ *
  * @since 1.6.0
  */
 public class LongLexicoder extends ULongLexicoder {
@@ -27,7 +27,7 @@ public class LongLexicoder extends ULongLexicoder {
   public byte[] encode(Long l) {
     return super.encode(l ^ 0x8000000000000000l);
   }
-  
+
   @Override
   public Long decode(byte[] data) {
     return super.decode(data) ^ 0x8000000000000000l;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/PairLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/PairLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/PairLexicoder.java
index e185630..22af289 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/PairLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/PairLexicoder.java
@@ -24,53 +24,54 @@ import static org.apache.accumulo.core.client.lexicoder.impl.ByteUtils.unescape;
 import org.apache.accumulo.core.util.ComparablePair;
 
 /**
- * This class is a lexicoder that sorts a ComparablePair. Each item in the pair is encoded with the given lexicoder and concatenated together.
- * This makes it easy to construct a sortable key based on two components. There are many examples of this- but a key/value relationship is a great one.
+ * This class is a lexicoder that sorts a ComparablePair. Each item in the pair is encoded with the given lexicoder and concatenated together. This makes it
+ * easy to construct a sortable key based on two components. There are many examples of this- but a key/value relationship is a great one.
  *
  * If we decided we wanted a two-component key where the first component is a string and the second component a date which is reverse sorted, we can do so with
  * the following example:
  *
  * <pre>
- * {@code
- * StringLexicoder stringEncoder = new StringLexicoder();
- * ReverseLexicoder<Date> dateEncoder = new ReverseLexicoder<Date>(new DateLexicoder());
- * PairLexicoder<String,Date> pairLexicoder = new PairLexicoder<String,Date>(stringEncoder, dateEncoder);
- * byte[] pair1 = pairLexicoder.encode(new ComparablePair<String,Date>("com.google", new Date()));
- * byte[] pair2 = pairLexicoder.encode(new ComparablePair<String,Date>("com.google", new Date(System.currentTimeMillis() + 500)));
- * byte[] pair3 = pairLexicoder.encode(new ComparablePair<String,Date>("org.apache", new Date(System.currentTimeMillis() + 1000)));
+ * {
+ *   &#064;code
+ *   StringLexicoder stringEncoder = new StringLexicoder();
+ *   ReverseLexicoder&lt;Date&gt; dateEncoder = new ReverseLexicoder&lt;Date&gt;(new DateLexicoder());
+ *   PairLexicoder&lt;String,Date&gt; pairLexicoder = new PairLexicoder&lt;String,Date&gt;(stringEncoder, dateEncoder);
+ *   byte[] pair1 = pairLexicoder.encode(new ComparablePair&lt;String,Date&gt;(&quot;com.google&quot;, new Date()));
+ *   byte[] pair2 = pairLexicoder.encode(new ComparablePair&lt;String,Date&gt;(&quot;com.google&quot;, new Date(System.currentTimeMillis() + 500)));
+ *   byte[] pair3 = pairLexicoder.encode(new ComparablePair&lt;String,Date&gt;(&quot;org.apache&quot;, new Date(System.currentTimeMillis() + 1000)));
  * }
  * </pre>
- * 
+ *
  * In the example, pair2 will be sorted before pair1. pair3 will occur last since 'org' is sorted after 'com'. If we just used a {@link DateLexicoder} instead
  * of a {@link ReverseLexicoder}, pair1 would have been sorted before pair2.
- * 
+ *
  * @since 1.6.0
  */
 
 public class PairLexicoder<A extends Comparable<A>,B extends Comparable<B>> implements Lexicoder<ComparablePair<A,B>> {
-  
+
   private Lexicoder<A> firstLexicoder;
   private Lexicoder<B> secondLexicoder;
-  
+
   public PairLexicoder(Lexicoder<A> firstLexicoder, Lexicoder<B> secondLexicoder) {
     this.firstLexicoder = firstLexicoder;
     this.secondLexicoder = secondLexicoder;
   }
-  
+
   @Override
   public byte[] encode(ComparablePair<A,B> data) {
     return concat(escape(firstLexicoder.encode(data.getFirst())), escape(secondLexicoder.encode(data.getSecond())));
   }
-  
+
   @Override
   public ComparablePair<A,B> decode(byte[] data) {
-    
+
     byte[][] fields = split(data);
     if (fields.length != 2) {
       throw new RuntimeException("Data does not have 2 fields, it has " + fields.length);
     }
-    
+
     return new ComparablePair<A,B>(firstLexicoder.decode(unescape(fields[0])), secondLexicoder.decode(unescape(fields[1])));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ReverseLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ReverseLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ReverseLexicoder.java
index 1ced085..1a6999b 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ReverseLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ReverseLexicoder.java
@@ -23,14 +23,14 @@ import static org.apache.accumulo.core.client.lexicoder.impl.ByteUtils.unescape;
  * A lexicoder that flips the sort order from another lexicoder. If this is applied to {@link DateLexicoder}, the most recent date will be sorted first and the
  * oldest date will be sorted last. If it's applied to {@link LongLexicoder}, the Long.MAX_VALUE will be sorted first and Long.MIN_VALUE will be sorted last,
  * etc...
- * 
+ *
  * @since 1.6.0
  */
 
 public class ReverseLexicoder<T> implements Lexicoder<T> {
-  
+
   private Lexicoder<T> lexicoder;
-  
+
   /**
    * @param lexicoder
    *          The lexicoder who's sort order will be flipped.
@@ -38,27 +38,27 @@ public class ReverseLexicoder<T> implements Lexicoder<T> {
   public ReverseLexicoder(Lexicoder<T> lexicoder) {
     this.lexicoder = lexicoder;
   }
-  
+
   @Override
   public byte[] encode(T data) {
     byte[] bytes = escape(lexicoder.encode(data));
     byte[] ret = new byte[bytes.length + 1];
-    
+
     for (int i = 0; i < bytes.length; i++)
       ret[i] = (byte) (0xff - (0xff & bytes[i]));
-    
+
     ret[bytes.length] = (byte) 0xff;
-    
+
     return ret;
   }
-  
+
   @Override
   public T decode(byte[] data) {
     byte ret[] = new byte[data.length - 1];
-    
+
     for (int i = 0; i < ret.length; i++)
       ret[i] = (byte) (0xff - (0xff & data[i]));
-    
+
     return lexicoder.decode(unescape(ret));
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/StringLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/StringLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/StringLexicoder.java
index 165248b..9d5e07e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/StringLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/StringLexicoder.java
@@ -21,20 +21,20 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 /**
  * This lexicoder encodes/decodes a given String to/from bytes without further processing. It can be combined with other encoders like the
  * {@link ReverseLexicoder} to flip the default sort order.
- * 
+ *
  * @since 1.6.0
  */
 
 public class StringLexicoder implements Lexicoder<String> {
-  
+
   @Override
   public byte[] encode(String data) {
     return data.getBytes(UTF_8);
   }
-  
+
   @Override
   public String decode(byte[] data) {
     return new String(data, UTF_8);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/TextLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/TextLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/TextLexicoder.java
index eb9d554..08837a5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/TextLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/TextLexicoder.java
@@ -22,20 +22,20 @@ import org.apache.hadoop.io.Text;
 /**
  * A lexicoder that preserves a Text's native sort order. It can be combined with other encoders like the {@link ReverseLexicoder} to flip the default sort
  * order.
- * 
+ *
  * @since 1.6.0
  */
 
 public class TextLexicoder implements Lexicoder<Text> {
-  
+
   @Override
   public byte[] encode(Text data) {
     return TextUtil.getBytes(data);
   }
-  
+
   @Override
   public Text decode(byte[] data) {
     return new Text(data);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoder.java
index 81398ae..f8842a9 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UIntegerLexicoder.java
@@ -19,57 +19,57 @@ package org.apache.accumulo.core.client.lexicoder;
 /**
  * A lexicoder for an unsigned integer. It sorts 0 before -1 and does not preserve the native sort order of a Java integer because Java does not contain an
  * unsigned integer. If Java had an unsigned integer type, this would correspond to its sort order.
- * 
+ *
  * @since 1.6.0
  */
 public class UIntegerLexicoder implements Lexicoder<Integer> {
-  
+
   @Override
   public byte[] encode(Integer i) {
     int shift = 56;
     int index;
     int prefix = i < 0 ? 0xff : 0x00;
-    
+
     for (index = 0; index < 4; index++) {
       if (((i >>> shift) & 0xff) != prefix)
         break;
-      
+
       shift -= 8;
     }
-    
+
     byte ret[] = new byte[5 - index];
     ret[0] = (byte) (4 - index);
     for (index = 1; index < ret.length; index++) {
       ret[index] = (byte) (i >>> shift);
       shift -= 8;
     }
-    
+
     if (i < 0)
       ret[0] = (byte) (8 - ret[0]);
-    
+
     return ret;
-    
+
   }
-  
+
   @Override
   public Integer decode(byte[] data) {
-    
+
     if (data[0] < 0 || data[0] > 8)
       throw new IllegalArgumentException("Unexpected length " + (0xff & data[0]));
-    
+
     int i = 0;
     int shift = 0;
-    
+
     for (int idx = data.length - 1; idx >= 1; idx--) {
       i += (data[idx] & 0xffl) << shift;
       shift += 8;
     }
-    
+
     // fill in 0xff prefix
     if (data[0] > 4)
       i |= -1 << ((8 - data[0]) << 3);
-    
+
     return i;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ULongLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ULongLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ULongLexicoder.java
index 1539fe7..a3dcab5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ULongLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/ULongLexicoder.java
@@ -19,56 +19,56 @@ package org.apache.accumulo.core.client.lexicoder;
 /**
  * Unsigned long lexicoder. The lexicographic encoding sorts first 0l and -1l last. This encoding does not correspond to the sort of Long because it does not
  * consider the sign bit. If Java had an unsigned long type, this encoder would correspond to its sort order.
- * 
+ *
  * @since 1.6.0
  */
 public class ULongLexicoder implements Lexicoder<Long> {
-  
+
   @Override
   public byte[] encode(Long l) {
     int shift = 56;
     int index;
     int prefix = l < 0 ? 0xff : 0x00;
-    
+
     for (index = 0; index < 8; index++) {
       if (((l >>> shift) & 0xff) != prefix)
         break;
-      
+
       shift -= 8;
     }
-    
+
     byte ret[] = new byte[9 - index];
     ret[0] = (byte) (8 - index);
     for (index = 1; index < ret.length; index++) {
       ret[index] = (byte) (l >>> shift);
       shift -= 8;
     }
-    
+
     if (l < 0)
       ret[0] = (byte) (16 - ret[0]);
-    
+
     return ret;
-    
+
   }
-  
+
   @Override
   public Long decode(byte[] data) {
-    
+
     long l = 0;
     int shift = 0;
-    
+
     if (data[0] < 0 || data[0] > 16)
       throw new IllegalArgumentException("Unexpected length " + (0xff & data[0]));
-    
+
     for (int i = data.length - 1; i >= 1; i--) {
       l += (data[i] & 0xffl) << shift;
       shift += 8;
     }
-    
+
     // fill in 0xff prefix
     if (data[0] > 8)
       l |= -1l << ((16 - data[0]) << 3);
-    
+
     return l;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UUIDLexicoder.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UUIDLexicoder.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UUIDLexicoder.java
index 4611b25..43fa1a4 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UUIDLexicoder.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/UUIDLexicoder.java
@@ -27,11 +27,11 @@ import org.apache.accumulo.core.iterators.ValueFormatException;
 
 /**
  * A lexicoder for a UUID that maintains its lexicographic sorting order.
- * 
+ *
  * @since 1.6.0
  */
 public class UUIDLexicoder implements Lexicoder<UUID> {
-  
+
   /**
    * @see <a href="http://www.ietf.org/rfc/rfc4122.txt"> RFC 4122: A Universally Unique IDentifier (UUID) URN Namespace, "Rules for Lexical Equivalence" in
    *      Section 3.</a>
@@ -41,18 +41,18 @@ public class UUIDLexicoder implements Lexicoder<UUID> {
     try {
       byte ret[] = new byte[16];
       DataOutputStream out = new DataOutputStream(new FixedByteArrayOutputStream(ret));
-      
+
       out.writeLong(uuid.getMostSignificantBits() ^ 0x8000000000000000l);
       out.writeLong(uuid.getLeastSignificantBits() ^ 0x8000000000000000l);
-      
+
       out.close();
-      
+
       return ret;
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public UUID decode(byte[] b) throws ValueFormatException {
     try {
@@ -62,5 +62,5 @@ public class UUIDLexicoder implements Lexicoder<UUID> {
       throw new RuntimeException(e);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/ByteUtils.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/ByteUtils.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/ByteUtils.java
index d418767..4973de8 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/ByteUtils.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/ByteUtils.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.core.client.lexicoder.impl;
 import java.util.ArrayList;
 
 public class ByteUtils {
-  
+
   /**
    * Escapes 0x00 with 0x01 0x01 and 0x01 with 0x01 0x02
    */
@@ -30,13 +30,13 @@ public class ByteUtils {
         escapeCount++;
       }
     }
-    
+
     if (escapeCount == 0)
       return in;
-    
+
     byte ret[] = new byte[escapeCount + in.length];
     int index = 0;
-    
+
     for (int i = 0; i < in.length; i++) {
       switch (in[i]) {
         case 0x00:
@@ -51,7 +51,7 @@ public class ByteUtils {
           ret[index++] = in[i];
       }
     }
-    
+
     return ret;
   }
 
@@ -66,12 +66,12 @@ public class ByteUtils {
         i++;
       }
     }
-    
+
     if (escapeCount == 0)
       return in;
-    
+
     byte ret[] = new byte[in.length - escapeCount];
-    
+
     int index = 0;
     for (int i = 0; i < in.length; i++) {
       if (in[i] == 0x01) {
@@ -80,9 +80,9 @@ public class ByteUtils {
       } else {
         ret[index++] = in[i];
       }
-      
+
     }
-    
+
     return ret;
   }
 
@@ -91,24 +91,24 @@ public class ByteUtils {
    */
   public static byte[][] split(byte[] data) {
     ArrayList<Integer> offsets = new ArrayList<Integer>();
-    
+
     for (int i = 0; i < data.length; i++) {
       if (data[i] == 0x00) {
         offsets.add(i);
       }
     }
-    
+
     offsets.add(data.length);
-    
+
     byte[][] ret = new byte[offsets.size()][];
-    
+
     int index = 0;
     for (int i = 0; i < offsets.size(); i++) {
       ret[i] = new byte[offsets.get(i) - index];
       System.arraycopy(data, index, ret[i], 0, ret[i].length);
       index = offsets.get(i) + 1;
     }
-    
+
     return ret;
   }
 
@@ -120,18 +120,18 @@ public class ByteUtils {
     for (byte[] field : fields) {
       len += field.length;
     }
-    
+
     byte ret[] = new byte[len + fields.length - 1];
     int index = 0;
-    
+
     for (byte[] field : fields) {
       System.arraycopy(field, 0, ret, index, field.length);
       index += field.length;
       if (index < ret.length)
         ret[index++] = 0x00;
     }
-    
+
     return ret;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/FixedByteArrayOutputStream.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/FixedByteArrayOutputStream.java b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/FixedByteArrayOutputStream.java
index dc10d7b..e3a87da 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/FixedByteArrayOutputStream.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/lexicoder/impl/FixedByteArrayOutputStream.java
@@ -23,23 +23,23 @@ import java.io.OutputStream;
  * Uses a fixed length array and will not grow in size dynamically like the {@link java.io.ByteArrayOutputStream}.
  */
 public class FixedByteArrayOutputStream extends OutputStream {
-  
+
   private int i;
   byte out[];
-  
+
   public FixedByteArrayOutputStream(byte out[]) {
     this.out = out;
   }
-  
+
   @Override
   public void write(int b) throws IOException {
     out[i++] = (byte) b;
   }
-  
+
   @Override
   public void write(byte b[], int off, int len) throws IOException {
     System.arraycopy(b, off, out, i, len);
     i += len;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloFileOutputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloFileOutputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloFileOutputFormat.java
index cfaaa58..2f2b4b2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloFileOutputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloFileOutputFormat.java
@@ -43,7 +43,7 @@ import org.apache.log4j.Logger;
 /**
  * This class allows MapReduce jobs to write output in the Accumulo data file format.<br />
  * Care should be taken to write only sorted data (sorted by {@link Key}), as this is an important requirement of Accumulo data files.
- * 
+ *
  * <p>
  * The output path to be created must be specified via {@link AccumuloFileOutputFormat#setOutputPath(JobConf, Path)}. This is inherited from
  * {@link FileOutputFormat#setOutputPath(JobConf, Path)}. Other methods from {@link FileOutputFormat} are not supported and may be ignored or cause failures.
@@ -51,14 +51,14 @@ import org.apache.log4j.Logger;
  * supported at this time.
  */
 public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
-  
+
   private static final Class<?> CLASS = AccumuloFileOutputFormat.class;
   protected static final Logger log = Logger.getLogger(CLASS);
-  
+
   /**
    * This helper method provides an AccumuloConfiguration object constructed from the Accumulo defaults, and overridden with Accumulo properties that have been
    * stored in the Job's configuration.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @since 1.5.0
@@ -66,10 +66,10 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   protected static AccumuloConfiguration getAccumuloConfiguration(JobConf job) {
     return FileOutputConfigurator.getAccumuloConfiguration(CLASS, job);
   }
-  
+
   /**
    * Sets the compression type to use for data blocks. Specifying a compression may require additional libraries to be available to your Job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param compressionType
@@ -79,14 +79,14 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setCompressionType(JobConf job, String compressionType) {
     FileOutputConfigurator.setCompressionType(CLASS, job, compressionType);
   }
-  
+
   /**
    * Sets the size for data blocks within each file.<br />
    * Data blocks are a span of key/value pairs stored in the file that are compressed and indexed as a group.
-   * 
+   *
    * <p>
    * Making this value smaller may increase seek performance, but at the cost of increasing the size of the indexes (which can also affect seek performance).
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param dataBlockSize
@@ -96,10 +96,10 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setDataBlockSize(JobConf job, long dataBlockSize) {
     FileOutputConfigurator.setDataBlockSize(CLASS, job, dataBlockSize);
   }
-  
+
   /**
    * Sets the size for file blocks in the file system; file blocks are managed, and replicated, by the underlying file system.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param fileBlockSize
@@ -109,11 +109,11 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setFileBlockSize(JobConf job, long fileBlockSize) {
     FileOutputConfigurator.setFileBlockSize(CLASS, job, fileBlockSize);
   }
-  
+
   /**
    * Sets the size for index blocks within each file; smaller blocks means a deeper index hierarchy within the file, while larger blocks mean a more shallow
    * index hierarchy within the file. This can affect the performance of queries.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param indexBlockSize
@@ -123,10 +123,10 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setIndexBlockSize(JobConf job, long indexBlockSize) {
     FileOutputConfigurator.setIndexBlockSize(CLASS, job, indexBlockSize);
   }
-  
+
   /**
    * Sets the file system replication factor for the resulting file, overriding the file system default.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param replication
@@ -136,37 +136,37 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setReplication(JobConf job, int replication) {
     FileOutputConfigurator.setReplication(CLASS, job, replication);
   }
-  
+
   @Override
   public RecordWriter<Key,Value> getRecordWriter(FileSystem ignored, JobConf job, String name, Progressable progress) throws IOException {
     // get the path of the temporary output file
     final Configuration conf = job;
     final AccumuloConfiguration acuConf = getAccumuloConfiguration(job);
-    
+
     final String extension = acuConf.get(Property.TABLE_FILE_TYPE);
     final Path file = new Path(getWorkOutputPath(job), getUniqueName(job, "part") + "." + extension);
-    
+
     final LRUMap validVisibilities = new LRUMap(ConfiguratorBase.getVisibilityCacheSize(conf));
-    
+
     return new RecordWriter<Key,Value>() {
       FileSKVWriter out = null;
-      
+
       @Override
       public void close(Reporter reporter) throws IOException {
         if (out != null)
           out.close();
       }
-      
+
       @Override
       public void write(Key key, Value value) throws IOException {
-        
+
         Boolean wasChecked = (Boolean) validVisibilities.get(key.getColumnVisibilityData());
         if (wasChecked == null) {
           byte[] cv = key.getColumnVisibilityData().toArray();
           new ColumnVisibility(cv);
           validVisibilities.put(new ArrayByteSequence(Arrays.copyOf(cv, cv.length)), Boolean.TRUE);
         }
-        
+
         if (out == null) {
           out = FileOperations.getInstance().openWriter(file.toString(), file.getFileSystem(conf), conf, acuConf);
           out.startDefaultLocalityGroup();
@@ -175,5 +175,5 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
       }
     };
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormat.java
index 18e286a..2cdc236 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloInputFormat.java
@@ -35,17 +35,16 @@ import org.apache.log4j.Level;
 /**
  * This class allows MapReduce jobs to use Accumulo as the source of data. This {@link InputFormat} provides keys and values of type {@link Key} and
  * {@link Value} to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloInputFormat#setConnectorInfo(JobConf, String, AuthenticationToken)}
  * <li>{@link AccumuloInputFormat#setConnectorInfo(JobConf, String, String)}
  * <li>{@link AccumuloInputFormat#setScanAuthorizations(JobConf, Authorizations)}
- * <li>{@link AccumuloInputFormat#setZooKeeperInstance(JobConf, ClientConfiguration)} OR
- * {@link AccumuloInputFormat#setMockInstance(JobConf, String)}
+ * <li>{@link AccumuloInputFormat#setZooKeeperInstance(JobConf, ClientConfiguration)} OR {@link AccumuloInputFormat#setMockInstance(JobConf, String)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 public class AccumuloInputFormat extends InputFormatBase<Key,Value> {
@@ -53,7 +52,7 @@ public class AccumuloInputFormat extends InputFormatBase<Key,Value> {
   @Override
   public RecordReader<Key,Value> getRecordReader(InputSplit split, JobConf job, Reporter reporter) throws IOException {
     log.setLevel(getLogLevel(job));
-    
+
     // Override the log level from the configuration as if the RangeInputSplit has one it's the more correct one to use.
     if (split instanceof org.apache.accumulo.core.client.mapreduce.RangeInputSplit) {
       org.apache.accumulo.core.client.mapreduce.RangeInputSplit risplit = (org.apache.accumulo.core.client.mapreduce.RangeInputSplit) split;
@@ -62,7 +61,7 @@ public class AccumuloInputFormat extends InputFormatBase<Key,Value> {
         log.setLevel(level);
       }
     }
-    
+
     RecordReaderBase<Key,Value> recordReader = new RecordReaderBase<Key,Value>() {
 
       @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloMultiTableInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloMultiTableInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloMultiTableInputFormat.java
index bbafef5..00a79f2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloMultiTableInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloMultiTableInputFormat.java
@@ -33,9 +33,9 @@ import org.apache.hadoop.mapred.Reporter;
 /**
  * This class allows MapReduce jobs to use multiple Accumulo tables as the source of data. This {@link org.apache.hadoop.mapred.InputFormat} provides keys and
  * values of type {@link Key} and {@link Value} to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloInputFormat#setConnectorInfo(JobConf, String, org.apache.accumulo.core.client.security.tokens.AuthenticationToken)}
  * <li>{@link AccumuloInputFormat#setConnectorInfo(JobConf, String, String)}
@@ -43,7 +43,7 @@ import org.apache.hadoop.mapred.Reporter;
  * <li>{@link AccumuloInputFormat#setZooKeeperInstance(JobConf, ClientConfiguration)} OR {@link AccumuloInputFormat#setMockInstance(JobConf, String)}
  * <li>{@link AccumuloMultiTableInputFormat#setInputTableConfigs(org.apache.hadoop.mapred.JobConf, java.util.Map)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 
@@ -51,7 +51,7 @@ public class AccumuloMultiTableInputFormat extends AbstractInputFormat<Key,Value
 
   /**
    * Sets the {@link InputTableConfig} objects on the given Hadoop configuration
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param configs

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormat.java
index a32a8b8..f877ec6 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloOutputFormat.java
@@ -428,7 +428,7 @@ public class AccumuloOutputFormat implements OutputFormat<Text,Mutation> {
         try {
           addTable(table);
         } catch (final Exception e) {
-          log.error("Could not add table '"+table.toString()+"'", e);
+          log.error("Could not add table '" + table.toString() + "'", e);
           throw new IOException(e);
         }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormat.java
index 673c5b8..6f257ff 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/AccumuloRowInputFormat.java
@@ -36,16 +36,16 @@ import org.apache.hadoop.mapred.Reporter;
 /**
  * This class allows MapReduce jobs to use Accumulo as the source of data. This {@link InputFormat} provides row names as {@link Text} as keys, and a
  * corresponding {@link PeekingIterator} as a value, which in turn makes the {@link Key}/{@link Value} pairs for that row available to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloRowInputFormat#setConnectorInfo(JobConf, String, AuthenticationToken)}
  * <li>{@link AccumuloRowInputFormat#setInputTableName(JobConf, String)}
  * <li>{@link AccumuloRowInputFormat#setScanAuthorizations(JobConf, Authorizations)}
  * <li>{@link AccumuloRowInputFormat#setZooKeeperInstance(JobConf, ClientConfiguration)} OR {@link AccumuloRowInputFormat#setMockInstance(JobConf, String)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 public class AccumuloRowInputFormat extends InputFormatBase<Text,PeekingIterator<Entry<Key,Value>>> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/InputFormatBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/InputFormatBase.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/InputFormatBase.java
index 0cee355..5a8f592 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/InputFormatBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/InputFormatBase.java
@@ -53,7 +53,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param tableName
@@ -66,7 +66,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the table name from the configuration.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return the table name
@@ -79,7 +79,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Sets the input ranges to scan for this job. If not set, the entire table will be scanned.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param ranges
@@ -92,7 +92,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the ranges to scan over from a job.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return the ranges
@@ -107,7 +107,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Restricts the columns that will be mapped over for this job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param columnFamilyColumnQualifierPairs
@@ -121,7 +121,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the columns to be mapped over from this job.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return a set of columns
@@ -134,7 +134,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Encode an iterator on the input for this job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param cfg
@@ -147,7 +147,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets a list of the iterator settings (for iterators to apply to a scanner) from this configuration.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return a list of iterators
@@ -161,10 +161,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -178,7 +178,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return false if the feature is disabled, true otherwise
@@ -191,10 +191,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Controls the use of the {@link IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -207,7 +207,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -221,10 +221,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
   /**
    * Controls the use of the {@link ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack to be constructed within the Map
    * task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -237,7 +237,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -253,26 +253,26 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -285,7 +285,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @param job
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -298,7 +298,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Initializes an Accumulo {@link org.apache.accumulo.core.client.impl.TabletLocator} based on the configuration.
-   * 
+   *
    * @param job
    *          the Hadoop job for the configured job
    * @return an Accumulo tablet locator
@@ -332,7 +332,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
     /**
      * Apply the configured iterators to the scanner.
-     * 
+     *
      * @param iterators
      *          the iterators to set
      * @param scanner
@@ -346,7 +346,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
     /**
      * Apply the configured iterators from the configuration to the scanner.
-     * 
+     *
      * @param job
      *          the job configuration
      * @param scanner

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapred/RangeInputSplit.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapred/RangeInputSplit.java b/core/src/main/java/org/apache/accumulo/core/client/mapred/RangeInputSplit.java
index 3fd2ab0..29274aa 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapred/RangeInputSplit.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapred/RangeInputSplit.java
@@ -33,7 +33,7 @@ public class RangeInputSplit extends org.apache.accumulo.core.client.mapreduce.R
   public RangeInputSplit(RangeInputSplit split) throws IOException {
     super(split);
   }
-  
+
   protected RangeInputSplit(String table, String tableId, Range range, String[] locations) {
     super(table, tableId, range, locations);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloFileOutputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloFileOutputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloFileOutputFormat.java
index 196fb04..c68dd56 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloFileOutputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloFileOutputFormat.java
@@ -41,7 +41,7 @@ import org.apache.log4j.Logger;
 /**
  * This class allows MapReduce jobs to write output in the Accumulo data file format.<br />
  * Care should be taken to write only sorted data (sorted by {@link Key}), as this is an important requirement of Accumulo data files.
- * 
+ *
  * <p>
  * The output path to be created must be specified via {@link AccumuloFileOutputFormat#setOutputPath(Job, Path)}. This is inherited from
  * {@link FileOutputFormat#setOutputPath(Job, Path)}. Other methods from {@link FileOutputFormat} are not supported and may be ignored or cause failures. Using
@@ -49,14 +49,14 @@ import org.apache.log4j.Logger;
  * supported at this time.
  */
 public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
-  
+
   private static final Class<?> CLASS = AccumuloFileOutputFormat.class;
   protected static final Logger log = Logger.getLogger(CLASS);
-  
+
   /**
    * This helper method provides an AccumuloConfiguration object constructed from the Accumulo defaults, and overridden with Accumulo properties that have been
    * stored in the Job's configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @since 1.5.0
@@ -64,10 +64,10 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   protected static AccumuloConfiguration getAccumuloConfiguration(JobContext context) {
     return FileOutputConfigurator.getAccumuloConfiguration(CLASS, InputFormatBase.getConfiguration(context));
   }
-  
+
   /**
    * Sets the compression type to use for data blocks. Specifying a compression may require additional libraries to be available to your Job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param compressionType
@@ -77,14 +77,14 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setCompressionType(Job job, String compressionType) {
     FileOutputConfigurator.setCompressionType(CLASS, job.getConfiguration(), compressionType);
   }
-  
+
   /**
    * Sets the size for data blocks within each file.<br />
    * Data blocks are a span of key/value pairs stored in the file that are compressed and indexed as a group.
-   * 
+   *
    * <p>
    * Making this value smaller may increase seek performance, but at the cost of increasing the size of the indexes (which can also affect seek performance).
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param dataBlockSize
@@ -94,10 +94,10 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setDataBlockSize(Job job, long dataBlockSize) {
     FileOutputConfigurator.setDataBlockSize(CLASS, job.getConfiguration(), dataBlockSize);
   }
-  
+
   /**
    * Sets the size for file blocks in the file system; file blocks are managed, and replicated, by the underlying file system.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param fileBlockSize
@@ -107,11 +107,11 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setFileBlockSize(Job job, long fileBlockSize) {
     FileOutputConfigurator.setFileBlockSize(CLASS, job.getConfiguration(), fileBlockSize);
   }
-  
+
   /**
    * Sets the size for index blocks within each file; smaller blocks means a deeper index hierarchy within the file, while larger blocks mean a more shallow
    * index hierarchy within the file. This can affect the performance of queries.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param indexBlockSize
@@ -121,10 +121,10 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setIndexBlockSize(Job job, long indexBlockSize) {
     FileOutputConfigurator.setIndexBlockSize(CLASS, job.getConfiguration(), indexBlockSize);
   }
-  
+
   /**
    * Sets the file system replication factor for the resulting file, overriding the file system default.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param replication
@@ -134,37 +134,37 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
   public static void setReplication(Job job, int replication) {
     FileOutputConfigurator.setReplication(CLASS, job.getConfiguration(), replication);
   }
-  
+
   @Override
   public RecordWriter<Key,Value> getRecordWriter(TaskAttemptContext context) throws IOException {
     // get the path of the temporary output file
     final Configuration conf = InputFormatBase.getConfiguration(context);
     final AccumuloConfiguration acuConf = getAccumuloConfiguration(context);
-    
+
     final String extension = acuConf.get(Property.TABLE_FILE_TYPE);
     final Path file = this.getDefaultWorkFile(context, "." + extension);
-    
+
     final LRUMap validVisibilities = new LRUMap(1000);
-    
+
     return new RecordWriter<Key,Value>() {
       FileSKVWriter out = null;
-      
+
       @Override
       public void close(TaskAttemptContext context) throws IOException {
         if (out != null)
           out.close();
       }
-      
+
       @Override
       public void write(Key key, Value value) throws IOException {
-        
+
         Boolean wasChecked = (Boolean) validVisibilities.get(key.getColumnVisibilityData());
         if (wasChecked == null) {
           byte[] cv = key.getColumnVisibilityData().toArray();
           new ColumnVisibility(cv);
           validVisibilities.put(new ArrayByteSequence(Arrays.copyOf(cv, cv.length)), Boolean.TRUE);
         }
-        
+
         if (out == null) {
           out = FileOperations.getInstance().openWriter(file.toString(), file.getFileSystem(conf), conf, acuConf);
           out.startDefaultLocalityGroup();
@@ -173,5 +173,5 @@ public class AccumuloFileOutputFormat extends FileOutputFormat<Key,Value> {
       }
     };
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormat.java
index 21a0280..b98cb77 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloInputFormat.java
@@ -35,15 +35,15 @@ import org.apache.log4j.Level;
 /**
  * This class allows MapReduce jobs to use Accumulo as the source of data. This {@link InputFormat} provides keys and values of type {@link Key} and
  * {@link Value} to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloInputFormat#setConnectorInfo(Job, String, AuthenticationToken)}
  * <li>{@link AccumuloInputFormat#setScanAuthorizations(Job, Authorizations)}
  * <li>{@link AccumuloInputFormat#setZooKeeperInstance(Job, ClientConfiguration)} OR {@link AccumuloInputFormat#setMockInstance(Job, String)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 public class AccumuloInputFormat extends InputFormatBase<Key,Value> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloMultiTableInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloMultiTableInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloMultiTableInputFormat.java
index af1001f..010a94f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloMultiTableInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloMultiTableInputFormat.java
@@ -39,23 +39,23 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
 /**
  * This class allows MapReduce jobs to use multiple Accumulo tables as the source of data. This {@link org.apache.hadoop.mapreduce.InputFormat} provides keys
  * and values of type {@link Key} and {@link Value} to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloMultiTableInputFormat#setConnectorInfo(Job, String, AuthenticationToken)}
  * <li>{@link AccumuloMultiTableInputFormat#setScanAuthorizations(Job, Authorizations)}
  * <li>{@link AccumuloMultiTableInputFormat#setZooKeeperInstance(Job, ClientConfiguration)} OR {@link AccumuloInputFormat#setMockInstance(Job, String)}
  * <li>{@link AccumuloMultiTableInputFormat#setInputTableConfigs(Job, Map)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 public class AccumuloMultiTableInputFormat extends AbstractInputFormat<Key,Value> {
 
   /**
    * Sets the {@link InputTableConfig} objects on the given Hadoop configuration
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param configs
@@ -87,11 +87,11 @@ public class AccumuloMultiTableInputFormat extends AbstractInputFormat<Key,Value
 
       @Override
       protected void setupIterators(TaskAttemptContext context, Scanner scanner, String tableName, RangeInputSplit split) {
-        List<IteratorSetting> iterators = split.getIterators(); 
+        List<IteratorSetting> iterators = split.getIterators();
         if (null == iterators) {
           iterators = getInputTableConfig(context, tableName).getIterators();
         }
-        
+
         for (IteratorSetting setting : iterators) {
           scanner.addScanIterator(setting);
         }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormat.java
index e220c00..9a8ab58 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloOutputFormat.java
@@ -428,7 +428,7 @@ public class AccumuloOutputFormat extends OutputFormat<Text,Mutation> {
         try {
           addTable(table);
         } catch (Exception e) {
-          log.error("Could not add table '"+table+"'", e);
+          log.error("Could not add table '" + table + "'", e);
           throw new IOException(e);
         }
 


[63/66] [abbrv] accumulo git commit: Merge branch '1.5' into 1.6

Posted by ct...@apache.org.
Merge branch '1.5' into 1.6

Conflicts:
	core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
	core/src/main/java/org/apache/accumulo/core/file/rfile/bcfile/TFile.java
	core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
	core/src/main/java/org/apache/accumulo/core/security/crypto/DefaultSecretKeyEncryptionStrategy.java
	core/src/main/java/org/apache/accumulo/core/util/format/BinaryFormatter.java
	core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
	core/src/main/java/org/apache/accumulo/core/util/shell/commands/HistoryCommand.java
	server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
	start/src/main/java/org/apache/accumulo/start/Main.java


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/1368d09c
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/1368d09c
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/1368d09c

Branch: refs/heads/master
Commit: 1368d09c198c898a856e9fe3e25150ade33fdb97
Parents: 627e525 901d60e
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 21:29:28 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 21:29:28 2015 -0500

----------------------------------------------------------------------
 .../org/apache/accumulo/core/cli/ClientOpts.java     |  6 +++---
 .../core/client/admin/InstanceOperations.java        |  6 +++---
 .../apache/accumulo/core/file/rfile/CreateEmpty.java |  4 ++--
 .../accumulo/core/file/rfile/bcfile/Utils.java       |  1 +
 .../org/apache/accumulo/core/iterators/Combiner.java |  2 +-
 .../apache/accumulo/core/iterators/LongCombiner.java |  4 ++--
 .../accumulo/core/iterators/OptionDescriber.java     |  4 ++--
 .../accumulo/core/iterators/TypedValueCombiner.java  | 10 +++++-----
 .../core/iterators/user/SummingArrayCombiner.java    |  5 +++--
 .../accumulo/core/util/shell/commands/DUCommand.java |  3 ++-
 .../core/util/shell/commands/EGrepCommand.java       |  3 ++-
 .../util/shell/commands/QuotedStringTokenizer.java   | 15 ++++++---------
 .../examples/simple/mapreduce/TableToFile.java       |  4 ++--
 .../java/org/apache/accumulo/server/Accumulo.java    |  4 ++--
 .../java/org/apache/accumulo/tserver/Compactor.java  |  5 +++--
 .../main/java/org/apache/accumulo/start/Main.java    |  4 ++--
 .../start/classloader/AccumuloClassLoader.java       |  4 ++--
 .../test/randomwalk/security/CreateTable.java        |  5 ++---
 .../test/randomwalk/security/CreateUser.java         |  5 ++---
 19 files changed, 47 insertions(+), 47 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
----------------------------------------------------------------------
diff --cc core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
index f1465bc,82ac73d..8bb8b3f
--- a/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
@@@ -166,14 -163,6 +166,14 @@@ public class ClientOpts extends Help 
    @Parameter(names = "--site-file", description = "Read the given accumulo site file to find the accumulo instance")
    public String siteFile = null;
  
 +  @Parameter(names = "--ssl", description = "Connect to accumulo over SSL")
 +  public boolean sslEnabled = false;
 +
-   @Parameter(
-       names = "--config-file",
-       description = "Read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
++  @Parameter(names = "--config-file", description = "Read the given client config file. "
++      + "If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, "
++      + "which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
 +  public String clientConfigFile = null;
 +
    public void startDebugLogging() {
      if (debug)
        Logger.getLogger(Constants.CORE_PACKAGE_NAME).setLevel(Level.TRACE);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
----------------------------------------------------------------------
diff --cc core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
index 49dd1d5,8cb656c..cc9d282
--- a/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/admin/InstanceOperations.java
@@@ -103,10 -103,10 +103,10 @@@ public interface InstanceOperations 
     * Throws an exception if a tablet server can not be contacted.
     *
     * @param tserver
-    *          The tablet server address should be of the form <ip address>:<port>
+    *          The tablet server address should be of the form {@code <ip address>:<port>}
     * @since 1.5.0
     */
 -  public void ping(String tserver) throws AccumuloException;
 +  void ping(String tserver) throws AccumuloException;
  
    /**
     * Test to see if the instance can load the given class as the given type. This check does not consider per table classpaths, see

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
----------------------------------------------------------------------
diff --cc core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
index cd6bff8,8d54088..dcbaf9e
--- a/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/rfile/CreateEmpty.java
@@@ -60,9 -58,9 +60,9 @@@ public class CreateEmpty 
  
    static class Opts extends Help {
      @Parameter(names = {"-c", "--codec"}, description = "the compression codec to use.", validateWith = IsSupportedCompressionAlgorithm.class)
 -    String codec = TFile.COMPRESSION_NONE;
 +    String codec = Compression.COMPRESSION_NONE;
-     @Parameter(
-         description = " <path> { <path> ... } Each path given is a URL. Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
+     @Parameter(description = " <path> { <path> ... } Each path given is a URL. "
+         + "Relative paths are resolved according to the default filesystem defined in your Hadoop configuration, which is usually an HDFS instance.",
          required = true, validateWith = NamedLikeRFile.class)
      List<String> files = new ArrayList<String>();
    }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/iterators/Combiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/iterators/OptionDescriber.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/iterators/TypedValueCombiner.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/core/src/main/java/org/apache/accumulo/core/util/shell/commands/DUCommand.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
----------------------------------------------------------------------
diff --cc server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
index 1b93c05,0000000..147a7c2
mode 100644,000000..100644
--- a/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/Accumulo.java
@@@ -1,336 -1,0 +1,336 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.server;
 +
 +import static com.google.common.base.Charsets.UTF_8;
 +
 +import java.io.File;
 +import java.io.FileInputStream;
 +import java.io.IOException;
 +import java.io.InputStream;
 +import java.net.InetAddress;
 +import java.net.UnknownHostException;
 +import java.util.Arrays;
 +import java.util.Map.Entry;
 +import java.util.TreeMap;
 +
 +import org.apache.accumulo.core.Constants;
 +import org.apache.accumulo.core.client.AccumuloException;
 +import org.apache.accumulo.core.client.Instance;
 +import org.apache.accumulo.core.conf.AccumuloConfiguration;
 +import org.apache.accumulo.core.conf.Property;
 +import org.apache.accumulo.core.trace.DistributedTrace;
 +import org.apache.accumulo.core.util.AddressUtil;
 +import org.apache.accumulo.core.util.UtilWaitThread;
 +import org.apache.accumulo.core.util.Version;
 +import org.apache.accumulo.core.volume.Volume;
 +import org.apache.accumulo.core.zookeeper.ZooUtil;
 +import org.apache.accumulo.fate.ReadOnlyStore;
 +import org.apache.accumulo.fate.ReadOnlyTStore;
 +import org.apache.accumulo.fate.ZooStore;
 +import org.apache.accumulo.server.client.HdfsZooInstance;
 +import org.apache.accumulo.server.conf.ServerConfiguration;
 +import org.apache.accumulo.server.fs.VolumeManager;
 +import org.apache.accumulo.server.util.time.SimpleTimer;
 +import org.apache.accumulo.server.watcher.Log4jConfiguration;
 +import org.apache.accumulo.server.watcher.MonitorLog4jWatcher;
 +import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 +import org.apache.hadoop.fs.FileStatus;
 +import org.apache.hadoop.fs.FileSystem;
 +import org.apache.hadoop.fs.Path;
 +import org.apache.log4j.Logger;
 +import org.apache.log4j.helpers.LogLog;
 +import org.apache.zookeeper.KeeperException;
 +
 +public class Accumulo {
 +
 +  private static final Logger log = Logger.getLogger(Accumulo.class);
 +
 +  public static synchronized void updateAccumuloVersion(VolumeManager fs, int oldVersion) {
 +    for (Volume volume : fs.getVolumes()) {
 +      try {
 +        if (getAccumuloPersistentVersion(fs) == oldVersion) {
 +          log.debug("Attempting to upgrade " + volume);
 +          Path dataVersionLocation = ServerConstants.getDataVersionLocation(volume);
 +          fs.create(new Path(dataVersionLocation, Integer.toString(ServerConstants.DATA_VERSION))).close();
 +          // TODO document failure mode & recovery if FS permissions cause above to work and below to fail ACCUMULO-2596
 +          Path prevDataVersionLoc = new Path(dataVersionLocation, Integer.toString(oldVersion));
 +          if (!fs.delete(prevDataVersionLoc)) {
 +            throw new RuntimeException("Could not delete previous data version location (" + prevDataVersionLoc + ") for " + volume);
 +          }
 +        }
 +      } catch (IOException e) {
 +        throw new RuntimeException("Unable to set accumulo version: an error occurred.", e);
 +      }
 +    }
 +  }
 +
 +  public static synchronized int getAccumuloPersistentVersion(FileSystem fs, Path path) {
 +    int dataVersion;
 +    try {
 +      FileStatus[] files = fs.listStatus(path);
 +      if (files == null || files.length == 0) {
 +        dataVersion = -1; // assume it is 0.5 or earlier
 +      } else {
 +        dataVersion = Integer.parseInt(files[0].getPath().getName());
 +      }
 +      return dataVersion;
 +    } catch (IOException e) {
 +      throw new RuntimeException("Unable to read accumulo version: an error occurred.", e);
 +    }
 +  }
 +
 +  public static synchronized int getAccumuloPersistentVersion(VolumeManager fs) {
 +    // It doesn't matter which Volume is used as they should all have the data version stored
 +    Volume v = fs.getVolumes().iterator().next();
 +    Path path = ServerConstants.getDataVersionLocation(v);
 +    return getAccumuloPersistentVersion(v.getFileSystem(), path);
 +  }
 +
 +  public static synchronized Path getAccumuloInstanceIdPath(VolumeManager fs) {
 +    // It doesn't matter which Volume is used as they should all have the instance ID stored
 +    Volume v = fs.getVolumes().iterator().next();
 +    return ServerConstants.getInstanceIdLocation(v);
 +  }
 +
 +  public static void enableTracing(String address, String application) {
 +    try {
 +      DistributedTrace.enable(HdfsZooInstance.getInstance(), ZooReaderWriter.getInstance(), application, address);
 +    } catch (Exception ex) {
 +      log.error("creating remote sink for trace spans", ex);
 +    }
 +  }
 +
 +  /**
 +   * Finds the best log4j configuration file. A generic file is used only if an application-specific file is not available. An XML file is preferred over a
 +   * properties file, if possible.
 +   *
 +   * @param confDir
 +   *          directory where configuration files should reside
 +   * @param application
 +   *          application name for configuration file name
 +   * @return configuration file name
 +   */
 +  static String locateLogConfig(String confDir, String application) {
 +    String explicitConfigFile = System.getProperty("log4j.configuration");
 +    if (explicitConfigFile != null) {
 +      return explicitConfigFile;
 +    }
 +    String[] configFiles = {String.format("%s/%s_logger.xml", confDir, application), String.format("%s/%s_logger.properties", confDir, application),
 +        String.format("%s/generic_logger.xml", confDir), String.format("%s/generic_logger.properties", confDir)};
 +    String defaultConfigFile = configFiles[2]; // generic_logger.xml
 +    for (String f : configFiles) {
 +      if (new File(f).exists()) {
 +        return f;
 +      }
 +    }
 +    return defaultConfigFile;
 +  }
 +
 +  public static void setupLogging(String application) throws UnknownHostException {
 +    System.setProperty("org.apache.accumulo.core.application", application);
 +
 +    if (System.getenv("ACCUMULO_LOG_DIR") != null)
 +      System.setProperty("org.apache.accumulo.core.dir.log", System.getenv("ACCUMULO_LOG_DIR"));
 +    else
 +      System.setProperty("org.apache.accumulo.core.dir.log", System.getenv("ACCUMULO_HOME") + "/logs/");
 +
 +    String localhost = InetAddress.getLocalHost().getHostName();
 +    System.setProperty("org.apache.accumulo.core.ip.localhost.hostname", localhost);
 +
 +    // Use a specific log config, if it exists
 +    String logConfigFile = locateLogConfig(System.getenv("ACCUMULO_CONF_DIR"), application);
 +    // Turn off messages about not being able to reach the remote logger... we protect against that.
 +    LogLog.setQuietMode(true);
 +
 +    // Set up local file-based logging right away
 +    Log4jConfiguration logConf = new Log4jConfiguration(logConfigFile);
 +    logConf.resetLogger();
 +  }
 +
 +  public static void init(VolumeManager fs, ServerConfiguration serverConfig, String application) throws IOException {
 +    final AccumuloConfiguration conf = serverConfig.getConfiguration();
 +    final Instance instance = serverConfig.getInstance();
 +
 +    // Use a specific log config, if it exists
 +    final String logConfigFile = locateLogConfig(System.getenv("ACCUMULO_CONF_DIR"), application);
 +
 +    // Set up polling log4j updates and log-forwarding using information advertised in zookeeper by the monitor
 +    MonitorLog4jWatcher logConfigWatcher = new MonitorLog4jWatcher(instance.getInstanceID(), logConfigFile);
 +    logConfigWatcher.setDelay(5000L);
 +    logConfigWatcher.start();
 +
 +    // Makes sure the log-forwarding to the monitor is configured
 +    int logPort = conf.getPort(Property.MONITOR_LOG4J_PORT);
 +    System.setProperty("org.apache.accumulo.core.host.log.port", Integer.toString(logPort));
 +
 +    log.info(application + " starting");
 +    log.info("Instance " + serverConfig.getInstance().getInstanceID());
 +    int dataVersion = Accumulo.getAccumuloPersistentVersion(fs);
 +    log.info("Data Version " + dataVersion);
 +    Accumulo.waitForZookeeperAndHdfs(fs);
 +
 +    Version codeVersion = new Version(Constants.VERSION);
 +    if (!(canUpgradeFromDataVersion(dataVersion))) {
 +      throw new RuntimeException("This version of accumulo (" + codeVersion + ") is not compatible with files stored using data version " + dataVersion);
 +    }
 +
 +    TreeMap<String,String> sortedProps = new TreeMap<String,String>();
 +    for (Entry<String,String> entry : conf)
 +      sortedProps.put(entry.getKey(), entry.getValue());
 +
 +    for (Entry<String,String> entry : sortedProps.entrySet()) {
 +      String key = entry.getKey();
 +      log.info(key + " = " + (Property.isSensitive(key) ? "<hidden>" : entry.getValue()));
 +    }
 +
 +    monitorSwappiness();
 +
 +    // Encourage users to configure TLS
 +    final String SSL = "SSL";
 +    for (Property sslProtocolProperty : Arrays.asList(Property.RPC_SSL_CLIENT_PROTOCOL, Property.RPC_SSL_ENABLED_PROTOCOLS,
 +        Property.MONITOR_SSL_INCLUDE_PROTOCOLS)) {
 +      String value = conf.get(sslProtocolProperty);
 +      if (value.contains(SSL)) {
 +        log.warn("It is recommended that " + sslProtocolProperty + " only allow TLS");
 +      }
 +    }
 +
 +  }
 +
 +  /**
 +   * Sanity check that the current persistent version is allowed to upgrade to the version of Accumulo running.
 +   *
 +   * @param dataVersion
 +   *          the version that is persisted in the backing Volumes
 +   */
 +  public static boolean canUpgradeFromDataVersion(final int dataVersion) {
 +    return dataVersion == ServerConstants.DATA_VERSION || dataVersion == ServerConstants.PREV_DATA_VERSION
 +        || dataVersion == ServerConstants.TWO_DATA_VERSIONS_AGO;
 +  }
 +
 +  /**
 +   * Does the data version number stored in the backing Volumes indicate we need to upgrade something?
 +   */
 +  public static boolean persistentVersionNeedsUpgrade(final int accumuloPersistentVersion) {
 +    return accumuloPersistentVersion == ServerConstants.TWO_DATA_VERSIONS_AGO || accumuloPersistentVersion == ServerConstants.PREV_DATA_VERSION;
 +  }
 +
 +  /**
 +   *
 +   */
 +  public static void monitorSwappiness() {
 +    SimpleTimer.getInstance().schedule(new Runnable() {
 +      @Override
 +      public void run() {
 +        try {
 +          String procFile = "/proc/sys/vm/swappiness";
 +          File swappiness = new File(procFile);
 +          if (swappiness.exists() && swappiness.canRead()) {
 +            InputStream is = new FileInputStream(procFile);
 +            try {
 +              byte[] buffer = new byte[10];
 +              int bytes = is.read(buffer);
 +              String setting = new String(buffer, 0, bytes, UTF_8);
 +              setting = setting.trim();
 +              if (bytes > 0 && Integer.parseInt(setting) > 10) {
 +                log.warn("System swappiness setting is greater than ten (" + setting + ") which can cause time-sensitive operations to be delayed. "
 +                    + " Accumulo is time sensitive because it needs to maintain distributed lock agreement.");
 +              }
 +            } finally {
 +              is.close();
 +            }
 +          }
 +        } catch (Throwable t) {
 +          log.error(t, t);
 +        }
 +      }
 +    }, 1000, 10 * 60 * 1000);
 +  }
 +
 +  public static void waitForZookeeperAndHdfs(VolumeManager fs) {
 +    log.info("Attempting to talk to zookeeper");
 +    while (true) {
 +      try {
 +        ZooReaderWriter.getInstance().getChildren(Constants.ZROOT);
 +        break;
 +      } catch (InterruptedException e) {
 +        // ignored
 +      } catch (KeeperException ex) {
 +        log.info("Waiting for accumulo to be initialized");
 +        UtilWaitThread.sleep(1000);
 +      }
 +    }
 +    log.info("Zookeeper connected and initialized, attemping to talk to HDFS");
 +    long sleep = 1000;
 +    int unknownHostTries = 3;
 +    while (true) {
 +      try {
 +        if (fs.isReady())
 +          break;
 +        log.warn("Waiting for the NameNode to leave safemode");
 +      } catch (IOException ex) {
 +        log.warn("Unable to connect to HDFS", ex);
 +      } catch (IllegalArgumentException exception) {
 +        /* Unwrap the UnknownHostException so we can deal with it directly */
 +        if (exception.getCause() instanceof UnknownHostException) {
 +          if (unknownHostTries > 0) {
 +            log.warn("Unable to connect to HDFS, will retry. cause: " + exception.getCause());
 +            /* We need to make sure our sleep period is long enough to avoid getting a cached failure of the host lookup. */
 +            sleep = Math.max(sleep, (AddressUtil.getAddressCacheNegativeTtl((UnknownHostException) (exception.getCause())) + 1) * 1000);
 +          } else {
 +            log.error("Unable to connect to HDFS and have exceeded max number of retries.", exception);
 +            throw exception;
 +          }
 +          unknownHostTries--;
 +        } else {
 +          throw exception;
 +        }
 +      }
 +      log.info("Backing off due to failure; current sleep period is " + sleep / 1000. + " seconds");
 +      UtilWaitThread.sleep(sleep);
 +      /* Back off to give transient failures more time to clear. */
 +      sleep = Math.min(60 * 1000, sleep * 2);
 +    }
 +    log.info("Connected to HDFS");
 +  }
 +
 +  /**
 +   * Exit loudly if there are outstanding Fate operations. Since Fate serializes class names, we need to make sure there are no queued transactions from a
 +   * previous version before continuing an upgrade. The status of the operations is irrelevant; those in SUCCESSFUL status cause the same problem as those just
 +   * queued.
 +   *
 +   * Note that the Master should not allow write access to Fate until after all upgrade steps are complete.
 +   *
 +   * Should be called as a guard before performing any upgrade steps, after determining that an upgrade is needed.
 +   *
 +   * see ACCUMULO-2519
 +   */
 +  public static void abortIfFateTransactions() {
 +    try {
 +      final ReadOnlyTStore<Accumulo> fate = new ReadOnlyStore<Accumulo>(new ZooStore<Accumulo>(
 +          ZooUtil.getRoot(HdfsZooInstance.getInstance()) + Constants.ZFATE, ZooReaderWriter.getInstance()));
 +      if (!(fate.list().isEmpty())) {
-         throw new AccumuloException(
-             "Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. Please see the README document for instructions on what to do under your previous version.");
++        throw new AccumuloException("Aborting upgrade because there are outstanding FATE transactions from a previous Accumulo version. "
++            + "Please see the README document for instructions on what to do under your previous version.");
 +      }
 +    } catch (Exception exception) {
 +      log.fatal("Problem verifying Fate readiness", exception);
 +      System.exit(1);
 +    }
 +  }
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
----------------------------------------------------------------------
diff --cc server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
index 822171c,0000000..381f75c
mode 100644,000000..100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/Compactor.java
@@@ -1,548 -1,0 +1,549 @@@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one or more
 + * contributor license agreements.  See the NOTICE file distributed with
 + * this work for additional information regarding copyright ownership.
 + * The ASF licenses this file to You under the Apache License, Version 2.0
 + * (the "License"); you may not use this file except in compliance with
 + * the License.  You may obtain a copy of the License at
 + *
 + *     http://www.apache.org/licenses/LICENSE-2.0
 + *
 + * Unless required by applicable law or agreed to in writing, software
 + * distributed under the License is distributed on an "AS IS" BASIS,
 + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 + * See the License for the specific language governing permissions and
 + * limitations under the License.
 + */
 +package org.apache.accumulo.tserver;
 +
 +import java.io.IOException;
 +import java.text.DateFormat;
 +import java.text.SimpleDateFormat;
 +import java.util.ArrayList;
 +import java.util.Collections;
 +import java.util.Date;
 +import java.util.HashMap;
 +import java.util.HashSet;
 +import java.util.List;
 +import java.util.Map;
 +import java.util.Map.Entry;
 +import java.util.Set;
 +import java.util.concurrent.Callable;
 +import java.util.concurrent.atomic.AtomicLong;
 +
 +import org.apache.accumulo.core.client.IteratorSetting;
 +import org.apache.accumulo.core.conf.AccumuloConfiguration;
 +import org.apache.accumulo.core.data.ByteSequence;
 +import org.apache.accumulo.core.data.Key;
 +import org.apache.accumulo.core.data.KeyExtent;
 +import org.apache.accumulo.core.data.Value;
 +import org.apache.accumulo.core.data.thrift.IterInfo;
 +import org.apache.accumulo.core.file.FileOperations;
 +import org.apache.accumulo.core.file.FileSKVIterator;
 +import org.apache.accumulo.core.file.FileSKVWriter;
 +import org.apache.accumulo.core.iterators.IteratorEnvironment;
 +import org.apache.accumulo.core.iterators.IteratorUtil;
 +import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
 +import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 +import org.apache.accumulo.core.iterators.WrappingIterator;
 +import org.apache.accumulo.core.iterators.system.ColumnFamilySkippingIterator;
 +import org.apache.accumulo.core.iterators.system.DeletingIterator;
 +import org.apache.accumulo.core.iterators.system.MultiIterator;
 +import org.apache.accumulo.core.iterators.system.TimeSettingIterator;
 +import org.apache.accumulo.core.metadata.schema.DataFileValue;
 +import org.apache.accumulo.core.tabletserver.thrift.ActiveCompaction;
 +import org.apache.accumulo.core.tabletserver.thrift.CompactionReason;
 +import org.apache.accumulo.core.tabletserver.thrift.CompactionType;
 +import org.apache.accumulo.core.util.LocalityGroupUtil;
 +import org.apache.accumulo.core.util.LocalityGroupUtil.LocalityGroupConfigurationError;
 +import org.apache.accumulo.server.fs.FileRef;
 +import org.apache.accumulo.server.fs.VolumeManager;
 +import org.apache.accumulo.server.problems.ProblemReport;
 +import org.apache.accumulo.server.problems.ProblemReportingIterator;
 +import org.apache.accumulo.server.problems.ProblemReports;
 +import org.apache.accumulo.server.problems.ProblemType;
 +import org.apache.accumulo.trace.instrument.Span;
 +import org.apache.accumulo.trace.instrument.Trace;
 +import org.apache.accumulo.tserver.Tablet.MinorCompactionReason;
 +import org.apache.accumulo.tserver.compaction.MajorCompactionReason;
 +import org.apache.hadoop.conf.Configuration;
 +import org.apache.hadoop.fs.FileSystem;
 +import org.apache.log4j.Logger;
 +
 +public class Compactor implements Callable<CompactionStats> {
 +
 +  public static class CountingIterator extends WrappingIterator {
 +
 +    private long count;
 +    private ArrayList<CountingIterator> deepCopies;
 +    private AtomicLong entriesRead;
 +
 +    @Override
 +    public CountingIterator deepCopy(IteratorEnvironment env) {
 +      return new CountingIterator(this, env);
 +    }
 +
 +    private CountingIterator(CountingIterator other, IteratorEnvironment env) {
 +      setSource(other.getSource().deepCopy(env));
 +      count = 0;
 +      this.deepCopies = other.deepCopies;
 +      this.entriesRead = other.entriesRead;
 +      deepCopies.add(this);
 +    }
 +
 +    public CountingIterator(SortedKeyValueIterator<Key,Value> source, AtomicLong entriesRead) {
 +      deepCopies = new ArrayList<Compactor.CountingIterator>();
 +      this.setSource(source);
 +      count = 0;
 +      this.entriesRead = entriesRead;
 +    }
 +
 +    @Override
 +    public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) {
 +      throw new UnsupportedOperationException();
 +    }
 +
 +    @Override
 +    public void next() throws IOException {
 +      super.next();
 +      count++;
 +      if (count % 1024 == 0) {
 +        entriesRead.addAndGet(1024);
 +      }
 +    }
 +
 +    public long getCount() {
 +      long sum = 0;
 +      for (CountingIterator dc : deepCopies) {
 +        sum += dc.count;
 +      }
 +
 +      return count + sum;
 +    }
 +  }
 +
 +  private static final Logger log = Logger.getLogger(Compactor.class);
 +
 +  static class CompactionCanceledException extends Exception {
 +    private static final long serialVersionUID = 1L;
 +  }
 +
 +  interface CompactionEnv {
 +    boolean isCompactionEnabled();
 +
 +    IteratorScope getIteratorScope();
 +  }
 +
 +  private Map<FileRef,DataFileValue> filesToCompact;
 +  private InMemoryMap imm;
 +  private FileRef outputFile;
 +  private boolean propogateDeletes;
 +  private AccumuloConfiguration acuTableConf;
 +  private CompactionEnv env;
 +  private Configuration conf;
 +  private VolumeManager fs;
 +  protected KeyExtent extent;
 +  private List<IteratorSetting> iterators;
 +
 +  // things to report
 +  private String currentLocalityGroup = "";
 +  private long startTime;
 +
 +  private MajorCompactionReason reason;
 +  protected MinorCompactionReason mincReason;
 +
 +  private AtomicLong entriesRead = new AtomicLong(0);
 +  private AtomicLong entriesWritten = new AtomicLong(0);
 +  private DateFormat dateFormatter = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss.SSS");
 +
 +  private static AtomicLong nextCompactorID = new AtomicLong(0);
 +
 +  // a unique id to identify a compactor
 +  private long compactorID = nextCompactorID.getAndIncrement();
 +
 +  protected volatile Thread thread;
 +
 +  private synchronized void setLocalityGroup(String name) {
 +    this.currentLocalityGroup = name;
 +  }
 +
 +  private void clearStats() {
 +    entriesRead.set(0);
 +    entriesWritten.set(0);
 +  }
 +
 +  protected static final Set<Compactor> runningCompactions = Collections.synchronizedSet(new HashSet<Compactor>());
 +
 +  public static class CompactionInfo {
 +
 +    private Compactor compactor;
 +    private String localityGroup;
 +    private long entriesRead;
 +    private long entriesWritten;
 +
 +    CompactionInfo(Compactor compactor) {
 +      this.localityGroup = compactor.currentLocalityGroup;
 +      this.entriesRead = compactor.entriesRead.get();
 +      this.entriesWritten = compactor.entriesWritten.get();
 +      this.compactor = compactor;
 +    }
 +
 +    public long getID() {
 +      return compactor.compactorID;
 +    }
 +
 +    public KeyExtent getExtent() {
 +      return compactor.getExtent();
 +    }
 +
 +    public long getEntriesRead() {
 +      return entriesRead;
 +    }
 +
 +    public long getEntriesWritten() {
 +      return entriesWritten;
 +    }
 +
 +    public Thread getThread() {
 +      return compactor.thread;
 +    }
 +
 +    public String getOutputFile() {
 +      return compactor.getOutputFile();
 +    }
 +
 +    public ActiveCompaction toThrift() {
 +
 +      CompactionType type;
 +
 +      if (compactor.imm != null)
 +        if (compactor.filesToCompact.size() > 0)
 +          type = CompactionType.MERGE;
 +        else
 +          type = CompactionType.MINOR;
 +      else if (!compactor.propogateDeletes)
 +        type = CompactionType.FULL;
 +      else
 +        type = CompactionType.MAJOR;
 +
 +      CompactionReason reason;
 +
-       if (compactor.imm != null)
++      if (compactor.imm != null) {
 +        switch (compactor.mincReason) {
 +          case USER:
 +            reason = CompactionReason.USER;
 +            break;
 +          case CLOSE:
 +            reason = CompactionReason.CLOSE;
 +            break;
 +          case SYSTEM:
 +          default:
 +            reason = CompactionReason.SYSTEM;
 +            break;
 +        }
-       else
++      } else {
 +        switch (compactor.reason) {
 +          case USER:
 +            reason = CompactionReason.USER;
 +            break;
 +          case CHOP:
 +            reason = CompactionReason.CHOP;
 +            break;
 +          case IDLE:
 +            reason = CompactionReason.IDLE;
 +            break;
 +          case NORMAL:
 +          default:
 +            reason = CompactionReason.SYSTEM;
 +            break;
 +        }
++      }
 +
 +      List<IterInfo> iiList = new ArrayList<IterInfo>();
 +      Map<String,Map<String,String>> iterOptions = new HashMap<String,Map<String,String>>();
 +
 +      for (IteratorSetting iterSetting : compactor.iterators) {
 +        iiList.add(new IterInfo(iterSetting.getPriority(), iterSetting.getIteratorClass(), iterSetting.getName()));
 +        iterOptions.put(iterSetting.getName(), iterSetting.getOptions());
 +      }
 +      List<String> filesToCompact = new ArrayList<String>();
 +      for (FileRef ref : compactor.filesToCompact.keySet())
 +        filesToCompact.add(ref.toString());
 +      return new ActiveCompaction(compactor.extent.toThrift(), System.currentTimeMillis() - compactor.startTime, filesToCompact,
 +          compactor.outputFile.toString(), type, reason, localityGroup, entriesRead, entriesWritten, iiList, iterOptions);
 +    }
 +  }
 +
 +  public static List<CompactionInfo> getRunningCompactions() {
 +    ArrayList<CompactionInfo> compactions = new ArrayList<Compactor.CompactionInfo>();
 +
 +    synchronized (runningCompactions) {
 +      for (Compactor compactor : runningCompactions) {
 +        compactions.add(new CompactionInfo(compactor));
 +      }
 +    }
 +
 +    return compactions;
 +  }
 +
 +  Compactor(Configuration conf, VolumeManager fs, Map<FileRef,DataFileValue> files, InMemoryMap imm, FileRef outputFile, boolean propogateDeletes,
 +      AccumuloConfiguration acuTableConf, KeyExtent extent, CompactionEnv env, List<IteratorSetting> iterators, MajorCompactionReason reason) {
 +    this.extent = extent;
 +    this.conf = conf;
 +    this.fs = fs;
 +    this.filesToCompact = files;
 +    this.imm = imm;
 +    this.outputFile = outputFile;
 +    this.propogateDeletes = propogateDeletes;
 +    this.acuTableConf = acuTableConf;
 +    this.env = env;
 +    this.iterators = iterators;
 +    this.reason = reason;
 +
 +    startTime = System.currentTimeMillis();
 +  }
 +
 +  Compactor(Configuration conf, VolumeManager fs, Map<FileRef,DataFileValue> files, InMemoryMap imm, FileRef outputFile, boolean propogateDeletes,
 +      AccumuloConfiguration acuTableConf, KeyExtent extent, CompactionEnv env) {
 +    this(conf, fs, files, imm, outputFile, propogateDeletes, acuTableConf, extent, env, new ArrayList<IteratorSetting>(), null);
 +  }
 +
 +  public VolumeManager getFileSystem() {
 +    return fs;
 +  }
 +
 +  KeyExtent getExtent() {
 +    return extent;
 +  }
 +
 +  String getOutputFile() {
 +    return outputFile.toString();
 +  }
 +
 +  @Override
 +  public CompactionStats call() throws IOException, CompactionCanceledException {
 +
 +    FileSKVWriter mfw = null;
 +
 +    CompactionStats majCStats = new CompactionStats();
 +
 +    boolean remove = runningCompactions.add(this);
 +
 +    clearStats();
 +
 +    String oldThreadName = Thread.currentThread().getName();
 +    String newThreadName = "MajC compacting " + extent.toString() + " started " + dateFormatter.format(new Date()) + " file: " + outputFile;
 +    Thread.currentThread().setName(newThreadName);
 +    thread = Thread.currentThread();
 +    try {
 +      FileOperations fileFactory = FileOperations.getInstance();
 +      FileSystem ns = this.fs.getVolumeByPath(outputFile.path()).getFileSystem();
 +      mfw = fileFactory.openWriter(outputFile.path().toString(), ns, ns.getConf(), acuTableConf);
 +
 +      Map<String,Set<ByteSequence>> lGroups;
 +      try {
 +        lGroups = LocalityGroupUtil.getLocalityGroups(acuTableConf);
 +      } catch (LocalityGroupConfigurationError e) {
 +        throw new IOException(e);
 +      }
 +
 +      long t1 = System.currentTimeMillis();
 +
 +      HashSet<ByteSequence> allColumnFamilies = new HashSet<ByteSequence>();
 +
 +      if (mfw.supportsLocalityGroups()) {
 +        for (Entry<String,Set<ByteSequence>> entry : lGroups.entrySet()) {
 +          setLocalityGroup(entry.getKey());
 +          compactLocalityGroup(entry.getKey(), entry.getValue(), true, mfw, majCStats);
 +          allColumnFamilies.addAll(entry.getValue());
 +        }
 +      }
 +
 +      setLocalityGroup("");
 +      compactLocalityGroup(null, allColumnFamilies, false, mfw, majCStats);
 +
 +      long t2 = System.currentTimeMillis();
 +
 +      FileSKVWriter mfwTmp = mfw;
 +      mfw = null; // set this to null so we do not try to close it again in finally if the close fails
 +      mfwTmp.close(); // if the close fails it will cause the compaction to fail
 +
 +      // Verify the file, since hadoop 0.20.2 sometimes lies about the success of close()
 +      try {
 +        FileSKVIterator openReader = fileFactory.openReader(outputFile.path().toString(), false, ns, ns.getConf(), acuTableConf);
 +        openReader.close();
 +      } catch (IOException ex) {
 +        log.error("Verification of successful compaction fails!!! " + extent + " " + outputFile, ex);
 +        throw ex;
 +      }
 +
 +      log.debug(String.format("Compaction %s %,d read | %,d written | %,6d entries/sec | %6.3f secs", extent, majCStats.getEntriesRead(),
 +          majCStats.getEntriesWritten(), (int) (majCStats.getEntriesRead() / ((t2 - t1) / 1000.0)), (t2 - t1) / 1000.0));
 +
 +      majCStats.setFileSize(fileFactory.getFileSize(outputFile.path().toString(), ns, ns.getConf(), acuTableConf));
 +      return majCStats;
 +    } catch (IOException e) {
 +      log.error(e, e);
 +      throw e;
 +    } catch (RuntimeException e) {
 +      log.error(e, e);
 +      throw e;
 +    } finally {
 +      Thread.currentThread().setName(oldThreadName);
 +      if (remove) {
 +        thread = null;
 +        runningCompactions.remove(this);
 +      }
 +
 +      try {
 +        if (mfw != null) {
 +          // compaction must not have finished successfully, so close its output file
 +          try {
 +            mfw.close();
 +          } finally {
 +            if (!fs.deleteRecursively(outputFile.path()))
 +              if (fs.exists(outputFile.path()))
 +                log.error("Unable to delete " + outputFile);
 +          }
 +        }
 +      } catch (IOException e) {
 +        log.warn(e, e);
 +      } catch (RuntimeException exception) {
 +        log.warn(exception, exception);
 +      }
 +    }
 +  }
 +
 +  private List<SortedKeyValueIterator<Key,Value>> openMapDataFiles(String lgName, ArrayList<FileSKVIterator> readers) throws IOException {
 +
 +    List<SortedKeyValueIterator<Key,Value>> iters = new ArrayList<SortedKeyValueIterator<Key,Value>>(filesToCompact.size());
 +
 +    for (FileRef mapFile : filesToCompact.keySet()) {
 +      try {
 +
 +        FileOperations fileFactory = FileOperations.getInstance();
 +        FileSystem fs = this.fs.getVolumeByPath(mapFile.path()).getFileSystem();
 +        FileSKVIterator reader;
 +
 +        reader = fileFactory.openReader(mapFile.path().toString(), false, fs, conf, acuTableConf);
 +
 +        readers.add(reader);
 +
 +        SortedKeyValueIterator<Key,Value> iter = new ProblemReportingIterator(extent.getTableId().toString(), mapFile.path().toString(), false, reader);
 +
 +        if (filesToCompact.get(mapFile).isTimeSet()) {
 +          iter = new TimeSettingIterator(iter, filesToCompact.get(mapFile).getTime());
 +        }
 +
 +        iters.add(iter);
 +
 +      } catch (Throwable e) {
 +
 +        ProblemReports.getInstance().report(new ProblemReport(extent.getTableId().toString(), ProblemType.FILE_READ, mapFile.path().toString(), e));
 +
 +        log.warn("Some problem opening map file " + mapFile + " " + e.getMessage(), e);
 +        // failed to open some map file... close the ones that were opened
 +        for (FileSKVIterator reader : readers) {
 +          try {
 +            reader.close();
 +          } catch (Throwable e2) {
 +            log.warn("Failed to close map file", e2);
 +          }
 +        }
 +
 +        readers.clear();
 +
 +        if (e instanceof IOException)
 +          throw (IOException) e;
 +        throw new IOException("Failed to open map data files", e);
 +      }
 +    }
 +
 +    return iters;
 +  }
 +
 +  private void compactLocalityGroup(String lgName, Set<ByteSequence> columnFamilies, boolean inclusive, FileSKVWriter mfw, CompactionStats majCStats)
 +      throws IOException, CompactionCanceledException {
 +    ArrayList<FileSKVIterator> readers = new ArrayList<FileSKVIterator>(filesToCompact.size());
 +    Span span = Trace.start("compact");
 +    try {
 +      long entriesCompacted = 0;
 +      List<SortedKeyValueIterator<Key,Value>> iters = openMapDataFiles(lgName, readers);
 +
 +      if (imm != null) {
 +        iters.add(imm.compactionIterator());
 +      }
 +
 +      CountingIterator citr = new CountingIterator(new MultiIterator(iters, extent.toDataRange()), entriesRead);
 +      DeletingIterator delIter = new DeletingIterator(citr, propogateDeletes);
 +      ColumnFamilySkippingIterator cfsi = new ColumnFamilySkippingIterator(delIter);
 +
 +      // if(env.getIteratorScope() )
 +
 +      TabletIteratorEnvironment iterEnv;
 +      if (env.getIteratorScope() == IteratorScope.majc)
 +        iterEnv = new TabletIteratorEnvironment(IteratorScope.majc, !propogateDeletes, acuTableConf);
 +      else if (env.getIteratorScope() == IteratorScope.minc)
 +        iterEnv = new TabletIteratorEnvironment(IteratorScope.minc, acuTableConf);
 +      else
 +        throw new IllegalArgumentException();
 +
 +      SortedKeyValueIterator<Key,Value> itr = iterEnv.getTopLevelIterator(IteratorUtil.loadIterators(env.getIteratorScope(), cfsi, extent, acuTableConf,
 +          iterators, iterEnv));
 +
 +      itr.seek(extent.toDataRange(), columnFamilies, inclusive);
 +
 +      if (!inclusive) {
 +        mfw.startDefaultLocalityGroup();
 +      } else {
 +        mfw.startNewLocalityGroup(lgName, columnFamilies);
 +      }
 +
 +      Span write = Trace.start("write");
 +      try {
 +        while (itr.hasTop() && env.isCompactionEnabled()) {
 +          mfw.append(itr.getTopKey(), itr.getTopValue());
 +          itr.next();
 +          entriesCompacted++;
 +
 +          if (entriesCompacted % 1024 == 0) {
 +            // Periodically update stats, do not want to do this too often since its volatile
 +            entriesWritten.addAndGet(1024);
 +          }
 +        }
 +
 +        if (itr.hasTop() && !env.isCompactionEnabled()) {
 +          // cancel major compaction operation
 +          try {
 +            try {
 +              mfw.close();
 +            } catch (IOException e) {
 +              log.error(e, e);
 +            }
 +            fs.deleteRecursively(outputFile.path());
 +          } catch (Exception e) {
 +            log.warn("Failed to delete Canceled compaction output file " + outputFile, e);
 +          }
 +          throw new CompactionCanceledException();
 +        }
 +
 +      } finally {
 +        CompactionStats lgMajcStats = new CompactionStats(citr.getCount(), entriesCompacted);
 +        majCStats.add(lgMajcStats);
 +        write.stop();
 +      }
 +
 +    } finally {
 +      // close sequence files opened
 +      for (FileSKVIterator reader : readers) {
 +        try {
 +          reader.close();
 +        } catch (Throwable e) {
 +          log.warn("Failed to close map file", e);
 +        }
 +      }
 +      span.stop();
 +    }
 +  }
 +
 +}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/start/src/main/java/org/apache/accumulo/start/Main.java
----------------------------------------------------------------------
diff --cc start/src/main/java/org/apache/accumulo/start/Main.java
index c4aaea1,a80ebe6..b9753bf
--- a/start/src/main/java/org/apache/accumulo/start/Main.java
+++ b/start/src/main/java/org/apache/accumulo/start/Main.java
@@@ -175,28 -141,7 +175,28 @@@ public class Main 
    }
  
    private static void printUsage() {
-     System.out
-         .println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info | tracer | minicluster | proxy | zookeeper | create-token | info | version | help | jar <jar> [<main class>] args | <accumulo class> args");
+     System.out.println("accumulo init | master | tserver | monitor | shell | admin | gc | classpath | rfile-info | login-info "
 -        + "| tracer | proxy | zookeeper | info | version | help | <accumulo class> args");
++        + "| tracer | minicluster | proxy | zookeeper | create-token | info | version | help | jar <jar> [<main class>] args | <accumulo class> args");
 +  }
 +
 +  // feature: will work even if main class isn't in the JAR
 +  static Class<?> loadClassFromJar(String[] args, JarFile f, ClassLoader cl) throws IOException, ClassNotFoundException {
 +    ClassNotFoundException explicitNotFound = null;
 +    if (args.length >= 3) {
 +      try {
 +        return cl.loadClass(args[2]); // jar jar-file main-class
 +      } catch (ClassNotFoundException cnfe) {
 +        // assume this is the first argument, look for main class in JAR manifest
 +        explicitNotFound = cnfe;
 +      }
 +    }
 +    String mainClass = f.getManifest().getMainAttributes().getValue(Attributes.Name.MAIN_CLASS);
 +    if (mainClass == null) {
 +      if (explicitNotFound != null) {
 +        throw explicitNotFound;
 +      }
 +      throw new ClassNotFoundException("No main class was specified, and the JAR manifest does not specify one");
 +    }
 +    return cl.loadClass(mainClass);
    }
  }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/start/src/main/java/org/apache/accumulo/start/classloader/AccumuloClassLoader.java
----------------------------------------------------------------------

http://git-wip-us.apache.org/repos/asf/accumulo/blob/1368d09c/test/src/main/java/org/apache/accumulo/test/randomwalk/security/CreateTable.java
----------------------------------------------------------------------


[22/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/maven-plugin/src/it/plugin-test/src/main/java/org/apache/accumulo/plugin/CustomFilter.java
----------------------------------------------------------------------
diff --git a/maven-plugin/src/it/plugin-test/src/main/java/org/apache/accumulo/plugin/CustomFilter.java b/maven-plugin/src/it/plugin-test/src/main/java/org/apache/accumulo/plugin/CustomFilter.java
index 9a0497a..97be677 100644
--- a/maven-plugin/src/it/plugin-test/src/main/java/org/apache/accumulo/plugin/CustomFilter.java
+++ b/maven-plugin/src/it/plugin-test/src/main/java/org/apache/accumulo/plugin/CustomFilter.java
@@ -21,13 +21,13 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.Filter;
 
 /**
- * 
+ *
  */
 public class CustomFilter extends Filter {
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     return k.getColumnFamily().toString().equals("allowed");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/maven-plugin/src/it/plugin-test/src/test/java/org/apache/accumulo/plugin/PluginIT.java
----------------------------------------------------------------------
diff --git a/maven-plugin/src/it/plugin-test/src/test/java/org/apache/accumulo/plugin/PluginIT.java b/maven-plugin/src/it/plugin-test/src/test/java/org/apache/accumulo/plugin/PluginIT.java
index 1e3fe37..6d3267a 100644
--- a/maven-plugin/src/it/plugin-test/src/test/java/org/apache/accumulo/plugin/PluginIT.java
+++ b/maven-plugin/src/it/plugin-test/src/test/java/org/apache/accumulo/plugin/PluginIT.java
@@ -43,17 +43,17 @@ import org.junit.BeforeClass;
 import org.junit.Test;
 
 public class PluginIT {
-  
+
   private static Instance instance;
   private static Connector connector;
-  
+
   @BeforeClass
   public static void setUp() throws Exception {
     String instanceName = "plugin-it-instance";
     instance = new MiniAccumuloInstance(instanceName, new File("target/accumulo-maven-plugin/" + instanceName));
     connector = instance.getConnector("root", new PasswordToken("ITSecret"));
   }
-  
+
   @Test
   public void testInstanceConnection() {
     assertTrue(instance != null);
@@ -61,7 +61,7 @@ public class PluginIT {
     assertTrue(connector != null);
     assertTrue(connector instanceof Connector);
   }
-  
+
   @Test
   public void testCreateTable() throws AccumuloException, AccumuloSecurityException, TableExistsException, IOException {
     String tableName = "testCreateTable";
@@ -69,7 +69,7 @@ public class PluginIT {
     assertTrue(connector.tableOperations().exists(tableName));
     assertTrue(new File("target/accumulo-maven-plugin/" + instance.getInstanceName() + "/testCreateTablePassed").createNewFile());
   }
-  
+
   @Test
   public void writeToTable() throws AccumuloException, AccumuloSecurityException, TableExistsException, TableNotFoundException, IOException {
     String tableName = "writeToTable";
@@ -91,7 +91,7 @@ public class PluginIT {
     assertEquals(1, count);
     assertTrue(new File("target/accumulo-maven-plugin/" + instance.getInstanceName() + "/testWriteToTablePassed").createNewFile());
   }
-  
+
   @Test
   public void checkIterator() throws IOException, AccumuloException, AccumuloSecurityException, TableExistsException, TableNotFoundException {
     String tableName = "checkIterator";
@@ -108,7 +108,7 @@ public class PluginIT {
     m.put("allowed", "CQ3", "V3");
     bw.addMutation(m);
     bw.close();
-    
+
     // check filter
     Scanner scanner = connector.createScanner(tableName, Authorizations.EMPTY);
     IteratorSetting is = new IteratorSetting(5, CustomFilter.class);
@@ -119,7 +119,7 @@ public class PluginIT {
       assertEquals("allowed", entry.getKey().getColumnFamily().toString());
     }
     assertEquals(4, count);
-    
+
     // check filter negated
     scanner.clearScanIterators();
     CustomFilter.setNegate(is, true);
@@ -132,5 +132,5 @@ public class PluginIT {
     assertEquals(2, count);
     assertTrue(new File("target/accumulo-maven-plugin/" + instance.getInstanceName() + "/testCheckIteratorPassed").createNewFile());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/AbstractAccumuloMojo.java
----------------------------------------------------------------------
diff --git a/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/AbstractAccumuloMojo.java b/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/AbstractAccumuloMojo.java
index 798499d..5e6a905 100644
--- a/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/AbstractAccumuloMojo.java
+++ b/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/AbstractAccumuloMojo.java
@@ -43,16 +43,16 @@ public abstract class AbstractAccumuloMojo extends AbstractMojo {
     } else if (miniClasspath != null && !miniClasspath.isEmpty()) {
       classpathItems.addAll(Arrays.asList(miniClasspath.split(File.pathSeparator)));
     }
-    
+
     // Hack to prevent sisu-guava, a maven 3.0.4 dependency, from effecting normal accumulo behavior.
     String sisuGuava = null;
     for (String items : classpathItems)
       if (items.contains("sisu-guava"))
         sisuGuava = items;
-    
+
     if (sisuGuava != null)
       classpathItems.remove(sisuGuava);
-    
+
     macConfig.setClasspathItems(classpathItems.toArray(new String[classpathItems.size()]));
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/StopMojo.java
----------------------------------------------------------------------
diff --git a/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/StopMojo.java b/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/StopMojo.java
index 0bd36e4..5d51ed6 100644
--- a/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/StopMojo.java
+++ b/maven-plugin/src/main/java/org/apache/accumulo/maven/plugin/StopMojo.java
@@ -31,7 +31,7 @@ import org.apache.maven.plugins.annotations.ResolutionScope;
 @ThreadSafe
 @Mojo(name = "stop", defaultPhase = LifecyclePhase.POST_INTEGRATION_TEST, requiresDependencyResolution = ResolutionScope.TEST)
 public class StopMojo extends AbstractAccumuloMojo {
-  
+
   @Override
   public void execute() throws MojoExecutionException {
     for (MiniAccumuloClusterImpl mac : StartMojo.runningClusters) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/main/java/org/apache/accumulo/cluster/RemoteShell.java
----------------------------------------------------------------------
diff --git a/minicluster/src/main/java/org/apache/accumulo/cluster/RemoteShell.java b/minicluster/src/main/java/org/apache/accumulo/cluster/RemoteShell.java
index 5db279c..5a44acf 100644
--- a/minicluster/src/main/java/org/apache/accumulo/cluster/RemoteShell.java
+++ b/minicluster/src/main/java/org/apache/accumulo/cluster/RemoteShell.java
@@ -28,8 +28,7 @@ import org.slf4j.LoggerFactory;
 import com.google.common.base.Preconditions;
 
 /**
- * Execute a command, leveraging Hadoop's {@link ShellCommandExecutor}, on a remote host. SSH configuration provided
- * by {@link RemoteShellOptions}.
+ * Execute a command, leveraging Hadoop's {@link ShellCommandExecutor}, on a remote host. SSH configuration provided by {@link RemoteShellOptions}.
  */
 public class RemoteShell extends ShellCommandExecutor {
   private static final Logger log = LoggerFactory.getLogger(RemoteShell.class);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/main/java/org/apache/accumulo/cluster/standalone/StandaloneClusterControl.java
----------------------------------------------------------------------
diff --git a/minicluster/src/main/java/org/apache/accumulo/cluster/standalone/StandaloneClusterControl.java b/minicluster/src/main/java/org/apache/accumulo/cluster/standalone/StandaloneClusterControl.java
index 733f8d0..9b467bc 100644
--- a/minicluster/src/main/java/org/apache/accumulo/cluster/standalone/StandaloneClusterControl.java
+++ b/minicluster/src/main/java/org/apache/accumulo/cluster/standalone/StandaloneClusterControl.java
@@ -129,7 +129,7 @@ public class StandaloneClusterControl implements ClusterControl {
   public void adminStopAll() throws IOException {
     File confDir = getConfDir();
     String master = getHosts(new File(confDir, "masters")).get(0);
-    String[] cmd = new String[] { accumuloPath, Admin.class.getName(), "stopAll" };
+    String[] cmd = new String[] {accumuloPath, Admin.class.getName(), "stopAll"};
     Entry<Integer,String> pair = exec(master, cmd);
     if (0 != pair.getKey().intValue()) {
       throw new IOException("stopAll did not finish successfully, retcode=" + pair.getKey() + ", stdout=" + pair.getValue());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloConfig.java
----------------------------------------------------------------------
diff --git a/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloConfig.java b/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloConfig.java
index 68e30fa..c8c499d 100644
--- a/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloConfig.java
+++ b/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloConfig.java
@@ -24,17 +24,17 @@ import org.apache.accumulo.minicluster.impl.MiniAccumuloConfigImpl;
 
 /**
  * Holds configuration for {@link MiniAccumuloCluster}. Required configurations must be passed to constructor(s) and all other configurations are optional.
- * 
+ *
  * @since 1.5.0
  */
 public class MiniAccumuloConfig {
 
   private MiniAccumuloConfigImpl impl;
-  
+
   MiniAccumuloConfig(MiniAccumuloConfigImpl config) {
     this.impl = config;
   }
-  
+
   MiniAccumuloConfigImpl getImpl() {
     return impl;
   }
@@ -49,10 +49,10 @@ public class MiniAccumuloConfig {
   public MiniAccumuloConfig(File dir, String rootPassword) {
     this.impl = new MiniAccumuloConfigImpl(dir, rootPassword);
   }
-  
+
   /**
    * Calling this method is optional. If not set, it defaults to two.
-   * 
+   *
    * @param numTservers
    *          the number of tablet servers that mini accumulo cluster should start
    */
@@ -63,7 +63,7 @@ public class MiniAccumuloConfig {
 
   /**
    * Calling this method is optional. If not set, defaults to 'miniInstance'
-   * 
+   *
    * @since 1.6.0
    */
   public MiniAccumuloConfig setInstanceName(String instanceName) {
@@ -73,7 +73,7 @@ public class MiniAccumuloConfig {
 
   /**
    * Calling this method is optional. If not set, it defaults to an empty map.
-   * 
+   *
    * @param siteConfig
    *          key/values that you normally put in accumulo-site.xml can be put here.
    */
@@ -84,10 +84,10 @@ public class MiniAccumuloConfig {
 
   /**
    * Calling this method is optional. A random port is generated by default
-   * 
+   *
    * @param zooKeeperPort
    *          A valid (and unused) port to use for the zookeeper
-   * 
+   *
    * @since 1.6.0
    */
   public MiniAccumuloConfig setZooKeeperPort(int zooKeeperPort) {
@@ -97,10 +97,10 @@ public class MiniAccumuloConfig {
 
   /**
    * Configure the time to wait for ZooKeeper to startup. Calling this method is optional. The default is 20000 milliseconds
-   * 
+   *
    * @param zooKeeperStartupTime
    *          Time to wait for ZooKeeper to startup, in milliseconds
-   * 
+   *
    * @since 1.6.1
    */
   public MiniAccumuloConfig setZooKeeperStartupTime(long zooKeeperStartupTime) {
@@ -110,15 +110,15 @@ public class MiniAccumuloConfig {
 
   /**
    * Sets the amount of memory to use in the master process. Calling this method is optional. Default memory is 128M
-   * 
+   *
    * @param serverType
    *          the type of server to apply the memory settings
    * @param memory
    *          amount of memory to set
-   * 
+   *
    * @param memoryUnit
    *          the units for which to apply with the memory size
-   * 
+   *
    * @since 1.6.0
    */
   public MiniAccumuloConfig setMemory(ServerType serverType, long memory, MemoryUnit memoryUnit) {
@@ -129,13 +129,13 @@ public class MiniAccumuloConfig {
   /**
    * Sets the default memory size to use. This value is also used when a ServerType has not been configured explicitly. Calling this method is optional. Default
    * memory is 128M
-   * 
+   *
    * @param memory
    *          amount of memory to set
-   * 
+   *
    * @param memoryUnit
    *          the units for which to apply with the memory size
-   * 
+   *
    * @since 1.6.0
    */
   public MiniAccumuloConfig setDefaultMemory(long memory, MemoryUnit memoryUnit) {
@@ -152,7 +152,7 @@ public class MiniAccumuloConfig {
 
   /**
    * @return name of configured instance
-   * 
+   *
    * @since 1.6.0
    */
   public String getInstanceName() {
@@ -161,7 +161,7 @@ public class MiniAccumuloConfig {
 
   /**
    * @return The configured zookeeper port
-   * 
+   *
    * @since 1.6.0
    */
   public int getZooKeeperPort() {
@@ -171,9 +171,9 @@ public class MiniAccumuloConfig {
   /**
    * @param serverType
    *          get configuration for this server type
-   * 
+   *
    * @return memory configured in bytes, returns default if this server type is not configured
-   * 
+   *
    * @since 1.6.0
    */
   public long getMemory(ServerType serverType) {
@@ -182,7 +182,7 @@ public class MiniAccumuloConfig {
 
   /**
    * @return memory configured in bytes
-   * 
+   *
    * @since 1.6.0
    */
   public long getDefaultMemory() {
@@ -212,7 +212,7 @@ public class MiniAccumuloConfig {
 
   /**
    * @return is the current configuration in jdwpEnabled mode?
-   * 
+   *
    * @since 1.6.0
    */
   public boolean isJDWPEnabled() {
@@ -223,7 +223,7 @@ public class MiniAccumuloConfig {
    * @param jdwpEnabled
    *          should the processes run remote jdwpEnabled servers?
    * @return the current instance
-   * 
+   *
    * @since 1.6.0
    */
   public MiniAccumuloConfig setJDWPEnabled(boolean jdwpEnabled) {
@@ -233,7 +233,7 @@ public class MiniAccumuloConfig {
 
   /**
    * @return the paths to use for loading native libraries
-   * 
+   *
    * @since 1.6.0
    */
   public String[] getNativeLibPaths() {
@@ -242,10 +242,10 @@ public class MiniAccumuloConfig {
 
   /**
    * Sets the path for processes to use for loading native libraries
-   * 
+   *
    * @param nativePathItems
    *          the nativePathItems to set
-   * 
+   *
    * @since 1.6.0
    */
   public MiniAccumuloConfig setNativeLibPaths(String... nativePathItems) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloInstance.java
----------------------------------------------------------------------
diff --git a/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloInstance.java b/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloInstance.java
index e0c93a6..b94cb0b 100644
--- a/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloInstance.java
+++ b/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloInstance.java
@@ -29,7 +29,7 @@ import org.apache.commons.configuration.PropertiesConfiguration;
 import org.apache.hadoop.conf.Configuration;
 
 /**
- * 
+ *
  * @since 1.6.0
  */
 public class MiniAccumuloInstance extends ZooKeeperInstance {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloRunner.java
----------------------------------------------------------------------
diff --git a/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloRunner.java b/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloRunner.java
index cee0db7..c45abc0 100644
--- a/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloRunner.java
+++ b/minicluster/src/main/java/org/apache/accumulo/minicluster/MiniAccumuloRunner.java
@@ -30,16 +30,16 @@ import java.util.regex.Pattern;
 import org.apache.accumulo.core.cli.Help;
 import org.apache.accumulo.core.util.Pair;
 import org.apache.commons.io.FileUtils;
+import org.apache.log4j.Logger;
 
 import com.beust.jcommander.IStringConverter;
 import com.beust.jcommander.Parameter;
 import com.google.common.io.Files;
-import org.apache.log4j.Logger;
 
 /**
  * A runner for starting up a {@link MiniAccumuloCluster} from the command line using an optional configuration properties file. An example property file looks
  * like the following:
- * 
+ *
  * <pre>
  * rootPassword=secret
  * instanceName=testInstance
@@ -53,15 +53,15 @@ import org.apache.log4j.Logger;
  * shutdownPort=4446
  * site.instance.secret=HUSH
  * </pre>
- * 
+ *
  * All items in the properties file above are optional and a default value will be provided in their absence. Any site configuration properties (typically found
  * in the accumulo-site.xml file) should be prefixed with "site." in the properties file.
- * 
+ *
  * @since 1.6.0
  */
 public class MiniAccumuloRunner {
   private static final Logger log = Logger.getLogger(MiniAccumuloRunner.class);
-  
+
   private static final String ROOT_PASSWORD_PROP = "rootPassword";
   private static final String SHUTDOWN_PORT_PROP = "shutdownPort";
   private static final String DEFAULT_MEMORY_PROP = "defaultMemory";
@@ -132,7 +132,7 @@ public class MiniAccumuloRunner {
 
   /**
    * Runs the {@link MiniAccumuloCluster} given a -p argument with a property file. Establishes a shutdown port for asynchronous operation.
-   * 
+   *
    * @param args
    *          An optional -p argument can be specified with the path to a valid properties file.
    */
@@ -202,14 +202,14 @@ public class MiniAccumuloRunner {
           log.error("InterruptedException attempting to stop Accumulo.", e);
           return;
         }
-        
+
         try {
           FileUtils.deleteDirectory(miniDir);
         } catch (IOException e) {
           log.error("IOException attempting to clean up miniDir.", e);
           return;
         }
-        
+
         System.out.println("\nShut down gracefully on " + new Date());
       }
     });

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/main/java/org/apache/accumulo/minicluster/impl/ZooKeeperBindException.java
----------------------------------------------------------------------
diff --git a/minicluster/src/main/java/org/apache/accumulo/minicluster/impl/ZooKeeperBindException.java b/minicluster/src/main/java/org/apache/accumulo/minicluster/impl/ZooKeeperBindException.java
index 21bc972..50217ce 100644
--- a/minicluster/src/main/java/org/apache/accumulo/minicluster/impl/ZooKeeperBindException.java
+++ b/minicluster/src/main/java/org/apache/accumulo/minicluster/impl/ZooKeeperBindException.java
@@ -23,8 +23,7 @@ public class ZooKeeperBindException extends RuntimeException {
 
   private static final long serialVersionUID = 1L;
 
-  public ZooKeeperBindException() {
-  }
+  public ZooKeeperBindException() {}
 
   public ZooKeeperBindException(String message) {
     super(message);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/test/java/org/apache/accumulo/minicluster/MiniAccumuloClusterStartStopTest.java
----------------------------------------------------------------------
diff --git a/minicluster/src/test/java/org/apache/accumulo/minicluster/MiniAccumuloClusterStartStopTest.java b/minicluster/src/test/java/org/apache/accumulo/minicluster/MiniAccumuloClusterStartStopTest.java
index 6e8d936..10bb08a 100644
--- a/minicluster/src/test/java/org/apache/accumulo/minicluster/MiniAccumuloClusterStartStopTest.java
+++ b/minicluster/src/test/java/org/apache/accumulo/minicluster/MiniAccumuloClusterStartStopTest.java
@@ -29,14 +29,14 @@ import org.junit.Test;
 import org.junit.rules.TestName;
 
 public class MiniAccumuloClusterStartStopTest {
-  
+
   private static final Logger log = Logger.getLogger(MiniAccumuloClusterStartStopTest.class);
   private File baseDir = new File(System.getProperty("user.dir") + "/target/mini-tests/" + this.getClass().getName());
   private MiniAccumuloCluster accumulo;
 
   @Rule
   public TestName testName = new TestName();
-  
+
   @Before
   public void setupTestCluster() throws IOException {
     baseDir.mkdirs();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/CleanShutdownMacTest.java
----------------------------------------------------------------------
diff --git a/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/CleanShutdownMacTest.java b/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/CleanShutdownMacTest.java
index 800e91b..81f9d3d 100644
--- a/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/CleanShutdownMacTest.java
+++ b/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/CleanShutdownMacTest.java
@@ -47,7 +47,7 @@ public class CleanShutdownMacTest {
     cluster.setShutdownExecutor(mockService);
 
     EasyMock.expect(future.get()).andReturn(0).anyTimes();
-    EasyMock.expect(mockService.<Integer>submit(EasyMock.anyObject(Callable.class))).andReturn(future).anyTimes();
+    EasyMock.expect(mockService.<Integer> submit(EasyMock.anyObject(Callable.class))).andReturn(future).anyTimes();
     EasyMock.expect(mockService.shutdownNow()).andReturn(Collections.<Runnable> emptyList()).once();
 
     EasyMock.replay(mockService, future);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/MiniAccumuloClusterImplTest.java
----------------------------------------------------------------------
diff --git a/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/MiniAccumuloClusterImplTest.java b/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/MiniAccumuloClusterImplTest.java
index 41ad88c..3339fb0 100644
--- a/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/MiniAccumuloClusterImplTest.java
+++ b/minicluster/src/test/java/org/apache/accumulo/minicluster/impl/MiniAccumuloClusterImplTest.java
@@ -76,7 +76,8 @@ public class MiniAccumuloClusterImplTest {
     testTableID = tableops.tableIdMap().get(TEST_TABLE);
 
     Scanner s = conn.createScanner(TEST_TABLE, Authorizations.EMPTY);
-    for (@SuppressWarnings("unused") Entry<Key,Value> e : s) {}
+    for (@SuppressWarnings("unused")
+    Entry<Key,Value> e : s) {}
   }
 
   @Test(timeout = 10000)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/proxy/src/main/java/org/apache/accumulo/proxy/Proxy.java
----------------------------------------------------------------------
diff --git a/proxy/src/main/java/org/apache/accumulo/proxy/Proxy.java b/proxy/src/main/java/org/apache/accumulo/proxy/Proxy.java
index 4b048eb..7eb4fbf 100644
--- a/proxy/src/main/java/org/apache/accumulo/proxy/Proxy.java
+++ b/proxy/src/main/java/org/apache/accumulo/proxy/Proxy.java
@@ -42,9 +42,9 @@ import com.beust.jcommander.Parameter;
 import com.google.common.io.Files;
 
 public class Proxy {
-  
+
   private static final Logger log = Logger.getLogger(Proxy.class);
-  
+
   public static class PropertiesConverter implements IStringConverter<Properties> {
     @Override
     public Properties convert(String fileName) {
@@ -63,36 +63,36 @@ public class Proxy {
       return prop;
     }
   }
-  
+
   public static class Opts extends Help {
     @Parameter(names = "-p", required = true, description = "properties file name", converter = PropertiesConverter.class)
     Properties prop;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(Proxy.class.getName(), args);
-    
+
     boolean useMini = Boolean.parseBoolean(opts.prop.getProperty("useMiniAccumulo", "false"));
     boolean useMock = Boolean.parseBoolean(opts.prop.getProperty("useMockInstance", "false"));
     String instance = opts.prop.getProperty("instance");
     String zookeepers = opts.prop.getProperty("zookeepers");
-    
+
     if (!useMini && !useMock && instance == null) {
       System.err.println("Properties file must contain one of : useMiniAccumulo=true, useMockInstance=true, or instance=<instance name>");
       System.exit(1);
     }
-    
+
     if (instance != null && zookeepers == null) {
       System.err.println("When instance is set in properties file, zookeepers must also be set.");
       System.exit(1);
     }
-    
+
     if (!opts.prop.containsKey("port")) {
       System.err.println("No port property");
       System.exit(1);
     }
-    
+
     if (useMini) {
       log.info("Creating mini cluster");
       final File folder = Files.createTempDir();
@@ -114,35 +114,35 @@ public class Proxy {
         }
       });
     }
-    
+
     Class<? extends TProtocolFactory> protoFactoryClass = Class.forName(opts.prop.getProperty("protocolFactory", TCompactProtocol.Factory.class.getName()))
         .asSubclass(TProtocolFactory.class);
     int port = Integer.parseInt(opts.prop.getProperty("port"));
     TServer server = createProxyServer(AccumuloProxy.class, ProxyServer.class, port, protoFactoryClass, opts.prop);
     server.serve();
   }
-  
+
   public static TServer createProxyServer(Class<?> api, Class<?> implementor, final int port, Class<? extends TProtocolFactory> protoClass,
       Properties properties) throws Exception {
     final TNonblockingServerSocket socket = new TNonblockingServerSocket(port);
-    
+
     // create the implementor
     Object impl = implementor.getConstructor(Properties.class).newInstance(properties);
-    
+
     Class<?> proxyProcClass = Class.forName(api.getName() + "$Processor");
     Class<?> proxyIfaceClass = Class.forName(api.getName() + "$Iface");
 
     @SuppressWarnings("unchecked")
     Constructor<? extends TProcessor> proxyProcConstructor = (Constructor<? extends TProcessor>) proxyProcClass.getConstructor(proxyIfaceClass);
-    
+
     final TProcessor processor = proxyProcConstructor.newInstance(RpcWrapper.service(impl));
-    
+
     THsHaServer.Args args = new THsHaServer.Args(socket);
     args.processor(processor);
     final long maxFrameSize = AccumuloConfiguration.getMemoryInBytes(properties.getProperty("maxFrameSize", "16M"));
     if (maxFrameSize > Integer.MAX_VALUE)
       throw new RuntimeException(maxFrameSize + " is larger than MAX_INT");
-    args.transportFactory(new TFramedTransport.Factory((int)maxFrameSize));
+    args.transportFactory(new TFramedTransport.Factory((int) maxFrameSize));
     args.protocolFactory(protoClass.newInstance());
     args.maxReadBufferBytes = maxFrameSize;
     return new THsHaServer(args);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/proxy/src/main/java/org/apache/accumulo/proxy/ProxyServer.java
----------------------------------------------------------------------
diff --git a/proxy/src/main/java/org/apache/accumulo/proxy/ProxyServer.java b/proxy/src/main/java/org/apache/accumulo/proxy/ProxyServer.java
index b1f039c..af12815 100644
--- a/proxy/src/main/java/org/apache/accumulo/proxy/ProxyServer.java
+++ b/proxy/src/main/java/org/apache/accumulo/proxy/ProxyServer.java
@@ -41,9 +41,9 @@ import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.BatchScanner;
 import org.apache.accumulo.core.client.BatchWriter;
 import org.apache.accumulo.core.client.BatchWriterConfig;
+import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.ConditionalWriter;
 import org.apache.accumulo.core.client.ConditionalWriter.Result;
-import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.ConditionalWriterConfig;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.Instance;
@@ -115,26 +115,26 @@ import com.google.common.cache.RemovalNotification;
 
 /**
  * Proxy Server exposing the Accumulo API via Thrift..
- * 
+ *
  * @since 1.5
  */
 public class ProxyServer implements AccumuloProxy.Iface {
-  
+
   public static final Logger logger = Logger.getLogger(ProxyServer.class);
   protected Instance instance;
-  
+
   protected Class<? extends AuthenticationToken> tokenClass;
-  
+
   static protected class ScannerPlusIterator {
     public ScannerBase scanner;
     public Iterator<Map.Entry<Key,Value>> iterator;
   }
-  
+
   static protected class BatchWriterPlusException {
     public BatchWriter writer;
     public MutationsRejectedException exception = null;
   }
-  
+
   static class CloseWriter implements RemovalListener<UUID,BatchWriterPlusException> {
     @Override
     public void onRemoval(RemovalNotification<UUID,BatchWriterPlusException> notification) {
@@ -147,10 +147,10 @@ public class ProxyServer implements AccumuloProxy.Iface {
         logger.warn(e, e);
       }
     }
-    
+
     public CloseWriter() {}
   }
-  
+
   static class CloseScanner implements RemovalListener<UUID,ScannerPlusIterator> {
     @Override
     public void onRemoval(RemovalNotification<UUID,ScannerPlusIterator> notification) {
@@ -160,10 +160,10 @@ public class ProxyServer implements AccumuloProxy.Iface {
         scanner.close();
       }
     }
-    
+
     public CloseScanner() {}
   }
-  
+
   public static class CloseConditionalWriter implements RemovalListener<UUID,ConditionalWriter> {
     @Override
     public void onRemoval(RemovalNotification<UUID,ConditionalWriter> notification) {
@@ -174,40 +174,42 @@ public class ProxyServer implements AccumuloProxy.Iface {
   protected Cache<UUID,ScannerPlusIterator> scannerCache;
   protected Cache<UUID,BatchWriterPlusException> writerCache;
   protected Cache<UUID,ConditionalWriter> conditionalWriterCache;
-  
+
   public ProxyServer(Properties props) {
-    
+
     String useMock = props.getProperty("useMockInstance");
     if (useMock != null && Boolean.parseBoolean(useMock))
       instance = new MockInstance();
     else
-      instance = new ZooKeeperInstance(ClientConfiguration.loadDefault().withInstance(props.getProperty("instance")).withZkHosts(props.getProperty("zookeepers")));
-    
+      instance = new ZooKeeperInstance(ClientConfiguration.loadDefault().withInstance(props.getProperty("instance"))
+          .withZkHosts(props.getProperty("zookeepers")));
+
     try {
       String tokenProp = props.getProperty("tokenClass", PasswordToken.class.getName());
       tokenClass = Class.forName(tokenProp).asSubclass(AuthenticationToken.class);
     } catch (ClassNotFoundException e) {
       throw new RuntimeException(e);
     }
-    
+
     scannerCache = CacheBuilder.newBuilder().expireAfterAccess(10, TimeUnit.MINUTES).maximumSize(1000).removalListener(new CloseScanner()).build();
-    
+
     writerCache = CacheBuilder.newBuilder().expireAfterAccess(10, TimeUnit.MINUTES).maximumSize(1000).removalListener(new CloseWriter()).build();
-    
+
     conditionalWriterCache = CacheBuilder.newBuilder().expireAfterAccess(10, TimeUnit.MINUTES).maximumSize(1000).removalListener(new CloseConditionalWriter())
         .build();
   }
-  
+
   protected Connector getConnector(ByteBuffer login) throws Exception {
     String[] pair = new String(login.array(), login.position(), login.remaining(), UTF_8).split(",", 2);
     if (instance.getInstanceID().equals(pair[0])) {
       Credentials creds = Credentials.deserialize(pair[1]);
       return instance.getConnector(creds.getPrincipal(), creds.getToken());
     } else {
-      throw new org.apache.accumulo.core.client.AccumuloSecurityException(pair[0], org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode.INVALID_INSTANCEID);
+      throw new org.apache.accumulo.core.client.AccumuloSecurityException(pair[0],
+          org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode.INVALID_INSTANCEID);
     }
   }
-  
+
   private void handleAccumuloException(AccumuloException e) throws org.apache.accumulo.proxy.thrift.TableNotFoundException,
       org.apache.accumulo.proxy.thrift.AccumuloException {
     if (e.getCause() instanceof ThriftTableOperationException) {
@@ -218,14 +220,14 @@ public class ProxyServer implements AccumuloProxy.Iface {
     }
     throw new org.apache.accumulo.proxy.thrift.AccumuloException(e.toString());
   }
-  
+
   private void handleAccumuloSecurityException(AccumuloSecurityException e) throws org.apache.accumulo.proxy.thrift.TableNotFoundException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException {
     if (e.getSecurityErrorCode().equals(SecurityErrorCode.TABLE_DOESNT_EXIST))
       throw new org.apache.accumulo.proxy.thrift.TableNotFoundException(e.toString());
     throw new org.apache.accumulo.proxy.thrift.AccumuloSecurityException(e.toString());
   }
-  
+
   private void handleExceptionTNF(Exception ex) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
     try {
@@ -244,7 +246,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new org.apache.accumulo.proxy.thrift.AccumuloException(e.toString());
     }
   }
-  
+
   private void handleExceptionTEE(Exception ex) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException,
       org.apache.accumulo.proxy.thrift.TableExistsException, TException {
@@ -262,7 +264,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new org.apache.accumulo.proxy.thrift.AccumuloException(e.toString());
     }
   }
-  
+
   private void handleExceptionMRE(Exception ex) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException,
       org.apache.accumulo.proxy.thrift.MutationsRejectedException, TException {
@@ -280,7 +282,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new org.apache.accumulo.proxy.thrift.AccumuloException(e.toString());
     }
   }
-  
+
   private void handleException(Exception ex) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
     try {
@@ -293,11 +295,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new org.apache.accumulo.proxy.thrift.AccumuloException(e.toString());
     }
   }
-  
+
   @Override
   public int addConstraint(ByteBuffer login, String tableName, String constraintClassName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
-    
+
     try {
       return getConnector(login).tableOperations().addConstraint(tableName, constraintClassName);
     } catch (Exception e) {
@@ -305,11 +307,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return -1;
     }
   }
-  
+
   @Override
   public void addSplits(ByteBuffer login, String tableName, Set<ByteBuffer> splits) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
-    
+
     try {
       SortedSet<Text> sorted = new TreeSet<Text>();
       for (ByteBuffer split : splits) {
@@ -320,7 +322,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void clearLocatorCache(ByteBuffer login, String tableName) throws org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
     try {
@@ -331,7 +333,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e.toString());
     }
   }
-  
+
   @Override
   public void compactTable(ByteBuffer login, String tableName, ByteBuffer startRow, ByteBuffer endRow,
       List<org.apache.accumulo.proxy.thrift.IteratorSetting> iterators, boolean flush, boolean wait, CompactionStrategyConfig compactionStrategy)
@@ -340,7 +342,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     try {
       CompactionConfig compactionConfig = new CompactionConfig().setStartRow(ByteBufferUtil.toText(startRow)).setEndRow(ByteBufferUtil.toText(endRow))
           .setIterators(getIteratorSettings(iterators)).setFlush(flush).setWait(wait);
-      
+
       if (compactionStrategy != null) {
         org.apache.accumulo.core.client.admin.CompactionStrategyConfig ccc = new org.apache.accumulo.core.client.admin.CompactionStrategyConfig(
             compactionStrategy.getClassName());
@@ -354,18 +356,18 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void cancelCompaction(ByteBuffer login, String tableName) throws org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
       org.apache.accumulo.proxy.thrift.TableNotFoundException, org.apache.accumulo.proxy.thrift.AccumuloException, TException {
-    
+
     try {
       getConnector(login).tableOperations().cancelCompaction(tableName);
     } catch (Exception e) {
       handleExceptionTNF(e);
     }
   }
-  
+
   private List<IteratorSetting> getIteratorSettings(List<org.apache.accumulo.proxy.thrift.IteratorSetting> iterators) {
     List<IteratorSetting> result = new ArrayList<IteratorSetting>();
     if (iterators != null) {
@@ -375,7 +377,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     }
     return result;
   }
-  
+
   @Override
   public void createTable(ByteBuffer login, String tableName, boolean versioningIter, org.apache.accumulo.proxy.thrift.TimeType type)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -383,7 +385,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     try {
       if (type == null)
         type = org.apache.accumulo.proxy.thrift.TimeType.MILLIS;
-      
+
       NewTableConfiguration tConfig = new NewTableConfiguration().setTimeType(TimeType.valueOf(type.toString()));
       if (!versioningIter)
         tConfig = tConfig.withoutDefaultIterators();
@@ -394,7 +396,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void deleteTable(ByteBuffer login, String tableName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -404,7 +406,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void deleteRows(ByteBuffer login, String tableName, ByteBuffer startRow, ByteBuffer endRow) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -414,7 +416,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public boolean tableExists(ByteBuffer login, String tableName) throws TException {
     try {
@@ -423,7 +425,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   @Override
   public void flushTable(ByteBuffer login, String tableName, ByteBuffer startRow, ByteBuffer endRow, boolean wait)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -434,7 +436,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public Map<String,Set<String>> getLocalityGroups(ByteBuffer login, String tableName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -454,7 +456,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public ByteBuffer getMaxRow(ByteBuffer login, String tableName, Set<ByteBuffer> auths, ByteBuffer startRow, boolean startInclusive, ByteBuffer endRow,
       boolean endInclusive) throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -476,13 +478,13 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public Map<String,String> getTableProperties(ByteBuffer login, String tableName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
     try {
       Map<String,String> ret = new HashMap<String,String>();
-      
+
       for (Map.Entry<String,String> entry : getConnector(login).tableOperations().getProperties(tableName)) {
         ret.put(entry.getKey(), entry.getValue());
       }
@@ -492,7 +494,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public List<ByteBuffer> listSplits(ByteBuffer login, String tableName, int maxSplits) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -508,7 +510,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public Set<String> listTables(ByteBuffer login) throws TException {
     try {
@@ -517,11 +519,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   @Override
   public Map<String,Integer> listConstraints(ByteBuffer login, String tableName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
-    
+
     try {
       return getConnector(login).tableOperations().listConstraints(tableName);
     } catch (Exception e) {
@@ -529,7 +531,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public void mergeTablets(ByteBuffer login, String tableName, ByteBuffer startRow, ByteBuffer endRow)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -540,7 +542,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void offlineTable(ByteBuffer login, String tableName, boolean wait) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -550,7 +552,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void onlineTable(ByteBuffer login, String tableName, boolean wait) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -560,18 +562,18 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void removeConstraint(ByteBuffer login, String tableName, int constraint) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
-    
+
     try {
       getConnector(login).tableOperations().removeConstraint(tableName, constraint);
     } catch (Exception e) {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void removeTableProperty(ByteBuffer login, String tableName, String property) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -581,7 +583,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void renameTable(ByteBuffer login, String oldTableName, String newTableName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException,
@@ -592,7 +594,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTEE(e);
     }
   }
-  
+
   @Override
   public void setLocalityGroups(ByteBuffer login, String tableName, Map<String,Set<String>> groupStrings)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -610,7 +612,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void setTableProperty(ByteBuffer login, String tableName, String property, String value) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -620,7 +622,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public Map<String,String> tableIdMap(ByteBuffer login) throws TException {
     try {
@@ -629,7 +631,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   @Override
   public List<DiskUsage> getDiskUsage(ByteBuffer login, Set<String> tables) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -648,7 +650,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public Map<String,String> getSiteConfiguration(ByteBuffer login) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -659,7 +661,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public Map<String,String> getSystemConfiguration(ByteBuffer login) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -670,7 +672,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public List<String> getTabletServers(ByteBuffer login) throws TException {
     try {
@@ -679,7 +681,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   @Override
   public List<org.apache.accumulo.proxy.thrift.ActiveScan> getActiveScans(ByteBuffer login, String tserver)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -734,11 +736,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public List<org.apache.accumulo.proxy.thrift.ActiveCompaction> getActiveCompactions(ByteBuffer login, String tserver)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
-    
+
     try {
       List<org.apache.accumulo.proxy.thrift.ActiveCompaction> result = new ArrayList<org.apache.accumulo.proxy.thrift.ActiveCompaction>();
       List<ActiveCompaction> active = getConnector(login).instanceOperations().getActiveCompactions(tserver);
@@ -758,7 +760,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
         pcomp.outputFile = comp.getOutputFile();
         pcomp.reason = CompactionReason.valueOf(comp.getReason().toString());
         pcomp.type = CompactionType.valueOf(comp.getType().toString());
-        
+
         pcomp.iterators = new ArrayList<org.apache.accumulo.proxy.thrift.IteratorSetting>();
         if (comp.getIterators() != null) {
           for (IteratorSetting setting : comp.getIterators()) {
@@ -775,7 +777,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public void removeProperty(ByteBuffer login, String property) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -785,7 +787,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void setProperty(ByteBuffer login, String property, String value) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -795,7 +797,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public boolean testClassLoad(ByteBuffer login, String className, String asTypeName) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -806,7 +808,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return false;
     }
   }
-  
+
   @Override
   public boolean authenticateUser(ByteBuffer login, String user, Map<String,String> properties) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -817,7 +819,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return false;
     }
   }
-  
+
   @Override
   public void changeUserAuthorizations(ByteBuffer login, String user, Set<ByteBuffer> authorizations)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -831,7 +833,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void changeLocalUserPassword(ByteBuffer login, String user, ByteBuffer password) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -841,7 +843,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void createLocalUser(ByteBuffer login, String user, ByteBuffer password) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -851,7 +853,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void dropLocalUser(ByteBuffer login, String user) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -861,7 +863,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public List<ByteBuffer> getUserAuthorizations(ByteBuffer login, String user) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -872,7 +874,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public void grantSystemPermission(ByteBuffer login, String user, org.apache.accumulo.proxy.thrift.SystemPermission perm)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -882,7 +884,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void grantTablePermission(ByteBuffer login, String user, String table, org.apache.accumulo.proxy.thrift.TablePermission perm)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -893,7 +895,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public boolean hasSystemPermission(ByteBuffer login, String user, org.apache.accumulo.proxy.thrift.SystemPermission perm)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -904,7 +906,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return false;
     }
   }
-  
+
   @Override
   public boolean hasTablePermission(ByteBuffer login, String user, String table, org.apache.accumulo.proxy.thrift.TablePermission perm)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -916,7 +918,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return false;
     }
   }
-  
+
   @Override
   public Set<String> listLocalUsers(ByteBuffer login) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -927,7 +929,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public void revokeSystemPermission(ByteBuffer login, String user, org.apache.accumulo.proxy.thrift.SystemPermission perm)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -937,7 +939,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public void revokeTablePermission(ByteBuffer login, String user, String table, org.apache.accumulo.proxy.thrift.TablePermission perm)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -948,7 +950,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   private Authorizations getAuthorizations(Set<ByteBuffer> authorizations) {
     List<String> auths = new ArrayList<String>();
     for (ByteBuffer bbauth : authorizations) {
@@ -956,13 +958,13 @@ public class ProxyServer implements AccumuloProxy.Iface {
     }
     return new Authorizations(auths.toArray(new String[0]));
   }
-  
+
   @Override
   public String createScanner(ByteBuffer login, String tableName, ScanOptions opts) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
     try {
       Connector connector = getConnector(login);
-      
+
       Authorizations auth;
       if (opts != null && opts.isSetAuthorizations()) {
         auth = getAuthorizations(opts.authorizations);
@@ -970,7 +972,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
         auth = connector.securityOperations().getUserAuthorizations(connector.whoami());
       }
       Scanner scanner = connector.createScanner(tableName, auth);
-      
+
       if (opts != null) {
         if (opts.iterators != null) {
           for (org.apache.accumulo.proxy.thrift.IteratorSetting iter : opts.iterators) {
@@ -992,9 +994,9 @@ public class ProxyServer implements AccumuloProxy.Iface {
           }
         }
       }
-      
+
       UUID uuid = UUID.randomUUID();
-      
+
       ScannerPlusIterator spi = new ScannerPlusIterator();
       spi.scanner = scanner;
       spi.iterator = scanner.iterator();
@@ -1005,13 +1007,13 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public String createBatchScanner(ByteBuffer login, String tableName, BatchScanOptions opts) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
     try {
       Connector connector = getConnector(login);
-      
+
       int threads = 10;
       Authorizations auth;
       if (opts != null && opts.isSetAuthorizations()) {
@@ -1021,9 +1023,9 @@ public class ProxyServer implements AccumuloProxy.Iface {
       }
       if (opts != null && opts.threads > 0)
         threads = opts.threads;
-      
+
       BatchScanner scanner = connector.createBatchScanner(tableName, auth, threads);
-      
+
       if (opts != null) {
         if (opts.iterators != null) {
           for (org.apache.accumulo.proxy.thrift.IteratorSetting iter : opts.iterators) {
@@ -1031,9 +1033,9 @@ public class ProxyServer implements AccumuloProxy.Iface {
             scanner.addScanIterator(is);
           }
         }
-        
+
         ArrayList<Range> ranges = new ArrayList<Range>();
-        
+
         if (opts.ranges == null) {
           ranges.add(new Range());
         } else {
@@ -1044,7 +1046,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
           }
         }
         scanner.setRanges(ranges);
-        
+
         if (opts.columns != null) {
           for (ScanColumn col : opts.columns) {
             if (col.isSetColQualifier())
@@ -1054,9 +1056,9 @@ public class ProxyServer implements AccumuloProxy.Iface {
           }
         }
       }
-      
+
       UUID uuid = UUID.randomUUID();
-      
+
       ScannerPlusIterator spi = new ScannerPlusIterator();
       spi.scanner = scanner;
       spi.iterator = scanner.iterator();
@@ -1067,34 +1069,34 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   private ScannerPlusIterator getScanner(String scanner) throws UnknownScanner {
-    
+
     UUID uuid = null;
     try {
       uuid = UUID.fromString(scanner);
     } catch (IllegalArgumentException e) {
       throw new UnknownScanner(e.getMessage());
     }
-    
+
     ScannerPlusIterator spi = scannerCache.getIfPresent(uuid);
     if (spi == null) {
       throw new UnknownScanner("Scanner never existed or no longer exists");
     }
     return spi;
   }
-  
+
   @Override
   public boolean hasNext(String scanner) throws UnknownScanner, TException {
     ScannerPlusIterator spi = getScanner(scanner);
-    
+
     return (spi.iterator.hasNext());
   }
-  
+
   @Override
   public KeyValueAndPeek nextEntry(String scanner) throws NoMoreEntriesException, UnknownScanner, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
       TException {
-    
+
     ScanResult scanResult = nextK(scanner, 1);
     if (scanResult.results.size() > 0) {
       return new KeyValueAndPeek(scanResult.results.get(0), scanResult.isMore());
@@ -1102,11 +1104,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new NoMoreEntriesException();
     }
   }
-  
+
   @Override
   public ScanResult nextK(String scanner, int k) throws NoMoreEntriesException, UnknownScanner, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
       TException {
-    
+
     // fetch the scanner
     ScannerPlusIterator spi = getScanner(scanner);
     Iterator<Map.Entry<Key,Value>> batchScanner = spi.iterator;
@@ -1129,7 +1131,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return ret;
     }
   }
-  
+
   @Override
   public void closeScanner(String scanner) throws UnknownScanner, TException {
     UUID uuid = null;
@@ -1138,7 +1140,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     } catch (IllegalArgumentException e) {
       throw new UnknownScanner(e.getMessage());
     }
-    
+
     try {
       if (scannerCache.asMap().remove(uuid) == null) {
         throw new UnknownScanner("Scanner never existed or no longer exists");
@@ -1149,7 +1151,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e.toString());
     }
   }
-  
+
   @Override
   public void updateAndFlush(ByteBuffer login, String tableName, Map<ByteBuffer,List<ColumnUpdate>> cells)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1165,15 +1167,15 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionMRE(e);
     }
   }
-  
+
   private static final ColumnVisibility EMPTY_VIS = new ColumnVisibility();
-  
+
   private void addCellsToWriter(Map<ByteBuffer,List<ColumnUpdate>> cells, BatchWriterPlusException bwpe) {
     if (bwpe.exception != null)
       return;
-    
+
     HashMap<Text,ColumnVisibility> vizMap = new HashMap<Text,ColumnVisibility>();
-    
+
     for (Map.Entry<ByteBuffer,List<ColumnUpdate>> entry : cells.entrySet()) {
       Mutation m = new Mutation(ByteBufferUtil.toBytes(entry.getKey()));
       addUpdatesToMutation(vizMap, m, entry.getValue());
@@ -1184,7 +1186,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       }
     }
   }
-  
+
   private void addUpdatesToMutation(HashMap<Text,ColumnVisibility> vizMap, Mutation m, List<ColumnUpdate> cu) {
     for (ColumnUpdate update : cu) {
       ColumnVisibility viz = EMPTY_VIS;
@@ -1209,7 +1211,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       }
     }
   }
-  
+
   private static ColumnVisibility getCahcedCV(HashMap<Text,ColumnVisibility> vizMap, byte[] cv) {
     ColumnVisibility viz;
     Text vizText = new Text(cv);
@@ -1219,7 +1221,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     }
     return viz;
   }
-  
+
   @Override
   public String createWriter(ByteBuffer login, String tableName, WriterOptions opts) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
@@ -1233,7 +1235,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public void update(String writer, Map<ByteBuffer,List<ColumnUpdate>> cells) throws TException {
     try {
@@ -1243,7 +1245,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       // just drop it, this is a oneway thrift call and throwing a TException seems to make all subsequent thrift calls fail
     }
   }
-  
+
   @Override
   public void flush(String writer) throws UnknownWriter, org.apache.accumulo.proxy.thrift.MutationsRejectedException, TException {
     try {
@@ -1259,7 +1261,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   @Override
   public void closeWriter(String writer) throws UnknownWriter, org.apache.accumulo.proxy.thrift.MutationsRejectedException, TException {
     try {
@@ -1276,7 +1278,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   private BatchWriterPlusException getWriter(String writer) throws UnknownWriter {
     UUID uuid = null;
     try {
@@ -1284,14 +1286,14 @@ public class ProxyServer implements AccumuloProxy.Iface {
     } catch (IllegalArgumentException iae) {
       throw new UnknownWriter(iae.getMessage());
     }
-    
+
     BatchWriterPlusException bwpe = writerCache.getIfPresent(uuid);
     if (bwpe == null) {
       throw new UnknownWriter("Writer never existed or no longer exists");
     }
     return bwpe;
   }
-  
+
   private BatchWriterPlusException getWriter(ByteBuffer login, String tableName, WriterOptions opts) throws Exception {
     BatchWriterConfig cfg = new BatchWriterConfig();
     if (opts != null) {
@@ -1311,7 +1313,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     result.writer = getConnector(login).createBatchWriter(tableName, cfg);
     return result;
   }
-  
+
   private org.apache.accumulo.core.client.Durability getDurability(Durability durability) {
     switch (durability) {
       case DEFAULT:
@@ -1331,11 +1333,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
   private IteratorSetting getIteratorSetting(org.apache.accumulo.proxy.thrift.IteratorSetting setting) {
     return new IteratorSetting(setting.priority, setting.name, setting.iteratorClass, setting.getProperties());
   }
-  
+
   private IteratorScope getIteratorScope(org.apache.accumulo.proxy.thrift.IteratorScope scope) {
     return IteratorScope.valueOf(scope.toString().toLowerCase());
   }
-  
+
   private EnumSet<IteratorScope> getIteratorScopes(Set<org.apache.accumulo.proxy.thrift.IteratorScope> scopes) {
     EnumSet<IteratorScope> scopes_ = EnumSet.noneOf(IteratorScope.class);
     for (org.apache.accumulo.proxy.thrift.IteratorScope scope : scopes) {
@@ -1343,7 +1345,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     }
     return scopes_;
   }
-  
+
   private EnumSet<org.apache.accumulo.proxy.thrift.IteratorScope> getProxyIteratorScopes(Set<IteratorScope> scopes) {
     EnumSet<org.apache.accumulo.proxy.thrift.IteratorScope> scopes_ = EnumSet.noneOf(org.apache.accumulo.proxy.thrift.IteratorScope.class);
     for (IteratorScope scope : scopes) {
@@ -1351,7 +1353,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     }
     return scopes_;
   }
-  
+
   @Override
   public void attachIterator(ByteBuffer login, String tableName, org.apache.accumulo.proxy.thrift.IteratorSetting setting,
       Set<org.apache.accumulo.proxy.thrift.IteratorScope> scopes) throws org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1362,7 +1364,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void checkIteratorConflicts(ByteBuffer login, String tableName, org.apache.accumulo.proxy.thrift.IteratorSetting setting,
       Set<org.apache.accumulo.proxy.thrift.IteratorScope> scopes) throws org.apache.accumulo.proxy.thrift.AccumuloException,
@@ -1373,7 +1375,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void cloneTable(ByteBuffer login, String tableName, String newTableName, boolean flush, Map<String,String> propertiesToSet,
       Set<String> propertiesToExclude) throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1381,28 +1383,28 @@ public class ProxyServer implements AccumuloProxy.Iface {
     try {
       propertiesToExclude = propertiesToExclude == null ? new HashSet<String>() : propertiesToExclude;
       propertiesToSet = propertiesToSet == null ? new HashMap<String,String>() : propertiesToSet;
-      
+
       getConnector(login).tableOperations().clone(tableName, newTableName, flush, propertiesToSet, propertiesToExclude);
     } catch (Exception e) {
       handleExceptionTEE(e);
     }
   }
-  
+
   @Override
   public void exportTable(ByteBuffer login, String tableName, String exportDir) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
-    
+
     try {
       getConnector(login).tableOperations().exportTable(tableName, exportDir);
     } catch (Exception e) {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public void importTable(ByteBuffer login, String tableName, String importDir) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, org.apache.accumulo.proxy.thrift.TableExistsException, TException {
-    
+
     try {
       getConnector(login).tableOperations().importTable(tableName, importDir);
     } catch (TableExistsException e) {
@@ -1411,7 +1413,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public org.apache.accumulo.proxy.thrift.IteratorSetting getIteratorSetting(ByteBuffer login, String tableName, String iteratorName,
       org.apache.accumulo.proxy.thrift.IteratorScope scope) throws org.apache.accumulo.proxy.thrift.AccumuloException,
@@ -1424,7 +1426,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public Map<String,Set<org.apache.accumulo.proxy.thrift.IteratorScope>> listIterators(ByteBuffer login, String tableName)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1441,7 +1443,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   @Override
   public void removeIterator(ByteBuffer login, String tableName, String iterName, Set<org.apache.accumulo.proxy.thrift.IteratorScope> scopes)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1452,7 +1454,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public Set<org.apache.accumulo.proxy.thrift.Range> splitRangeByTablets(ByteBuffer login, String tableName, org.apache.accumulo.proxy.thrift.Range range,
       int maxSplits) throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1469,11 +1471,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return null;
     }
   }
-  
+
   private org.apache.accumulo.proxy.thrift.Range getRange(Range r) {
     return new org.apache.accumulo.proxy.thrift.Range(getProxyKey(r.getStartKey()), r.isStartKeyInclusive(), getProxyKey(r.getEndKey()), r.isEndKeyInclusive());
   }
-  
+
   private org.apache.accumulo.proxy.thrift.Key getProxyKey(Key k) {
     if (k == null)
       return null;
@@ -1482,11 +1484,11 @@ public class ProxyServer implements AccumuloProxy.Iface {
     result.setTimestamp(k.getTimestamp());
     return result;
   }
-  
+
   private Range getRange(org.apache.accumulo.proxy.thrift.Range range) {
     return new Range(Util.fromThrift(range.start), Util.fromThrift(range.stop));
   }
-  
+
   @Override
   public void importDirectory(ByteBuffer login, String tableName, String importDir, String failureDir, boolean setTime)
       throws org.apache.accumulo.proxy.thrift.TableNotFoundException, org.apache.accumulo.proxy.thrift.AccumuloException,
@@ -1497,12 +1499,12 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleExceptionTNF(e);
     }
   }
-  
+
   @Override
   public org.apache.accumulo.proxy.thrift.Range getRowRange(ByteBuffer row) throws TException {
     return getRange(new Range(ByteBufferUtil.toText(row)));
   }
-  
+
   @Override
   public org.apache.accumulo.proxy.thrift.Key getFollowing(org.apache.accumulo.proxy.thrift.Key key, org.apache.accumulo.proxy.thrift.PartialKey part)
       throws TException {
@@ -1511,7 +1513,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     Key followingKey = key_.followingKey(part_);
     return getProxyKey(followingKey);
   }
-  
+
   @Override
   public void pingTabletServer(ByteBuffer login, String tserver) throws org.apache.accumulo.proxy.thrift.AccumuloException,
       org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
@@ -1521,7 +1523,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       handleException(e);
     }
   }
-  
+
   @Override
   public ByteBuffer login(String principal, Map<String,String> loginProperties) throws org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
     try {
@@ -1535,7 +1537,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       throw new TException(e);
     }
   }
-  
+
   private AuthenticationToken getToken(String principal, Map<String,String> properties) throws AccumuloSecurityException, AccumuloException {
     AuthenticationToken.Properties props = new AuthenticationToken.Properties();
     props.putAllStrings(properties);
@@ -1550,7 +1552,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
     token.init(props);
     return token;
   }
-  
+
   @Override
   public boolean testTableClassLoad(ByteBuffer login, String tableName, String className, String asTypeName)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1562,7 +1564,7 @@ public class ProxyServer implements AccumuloProxy.Iface {
       return false;
     }
   }
-  
+
   @Override
   public String createConditionalWriter(ByteBuffer login, String tableName, ConditionalWriterOptions options)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
@@ -1580,81 +1582,81 @@ public class ProxyServer implements AccumuloProxy.Iface {
         cwc.setAuthorizations(getAuthorizations(options.getAuthorizations()));
       if (options.isSetDurability() && options.getDurability() != null)
         cwc.setDurability(getDurability(options.getDurability()));
-      
+
       ConditionalWriter cw = getConnector(login).createConditionalWriter(tableName, cwc);
-      
+
       UUID id = UUID.randomUUID();
-      
+
       conditionalWriterCache.put(id, cw);
-      
+
       return id.toString();
     } catch (Exception e) {
       handleExceptionTNF(e);
       return null;
     }
   }
-  
+
   @Override
   public Map<ByteBuffer,ConditionalStatus> updateRowsConditionally(String conditionalWriter, Map<ByteBuffer,ConditionalUpdates> updates) throws UnknownWriter,
       org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException, TException {
-    
+
     ConditionalWriter cw = conditionalWriterCache.getIfPresent(UUID.fromString(conditionalWriter));
-    
+
     if (cw == null) {
       throw new UnknownWriter();
     }
-    
+
     try {
       HashMap<Text,ColumnVisibility> vizMap = new HashMap<Text,ColumnVisibility>();
-      
+
       ArrayList<ConditionalMutation> cmuts = new ArrayList<ConditionalMutation>(updates.size());
       for (Entry<ByteBuffer,ConditionalUpdates> cu : updates.entrySet()) {
         ConditionalMutation cmut = new ConditionalMutation(ByteBufferUtil.toBytes(cu.getKey()));
-        
+
         for (Condition tcond : cu.getValue().conditions) {
           org.apache.accumulo.core.data.Condition cond = new org.apache.accumulo.core.data.Condition(tcond.column.getColFamily(),
               tcond.column.getColQualifier());
-          
+
           if (tcond.getColumn().getColVisibility() != null && tcond.getColumn().getColVisibility().length > 0) {
             cond.setVisibility(getCahcedCV(vizMap, tcond.getColumn().getColVisibility()));
           }
-          
+
           if (tcond.isSetValue())
             cond.setValue(tcond.getValue());
-          
+
           if (tcond.isSetTimestamp())
             cond.setTimestamp(tcond.getTimestamp());
-          
+
           if (tcond.isSetIterators()) {
             cond.setIterators(getIteratorSettings(tcond.getIterators()).toArray(new IteratorSetting[tcond.getIterators().size()]));
           }
-          
+
           cmut.addCondition(cond);
         }
-        
+
         addUpdatesToMutation(vizMap, cmut, cu.getValue().updates);
-        
+
         cmuts.add(cmut);
       }
-      
+
       Iterator<Result> results = cw.write(cmuts.iterator());
-      
+
       HashMap<ByteBuffer,ConditionalStatus> resultMap = new HashMap<ByteBuffer,ConditionalStatus>();
-      
+
       while (results.hasNext()) {
         Result result = results.next();
         ByteBuffer row = ByteBuffer.wrap(result.getMutation().getRow());
         ConditionalStatus status = ConditionalStatus.valueOf(result.getStatus().name());
         resultMap.put(row, status);
       }
-      
+
       return resultMap;
     } catch (Exception e) {
       handleException(e);
       return null;
     }
   }
-  
+
   @Override
   public void closeConditionalWriter(String conditionalWriter) throws TException {
     ConditionalWriter cw = conditionalWriterCache.getIfPresent(UUID.fromString(conditionalWriter));
@@ -1663,12 +1665,12 @@ public class ProxyServer implements AccumuloProxy.Iface {
       conditionalWriterCache.invalidate(UUID.fromString(conditionalWriter));
     }
   }
-  
+
   @Override
   public ConditionalStatus updateRowConditionally(ByteBuffer login, String tableName, ByteBuffer row, ConditionalUpdates updates)
       throws org.apache.accumulo.proxy.thrift.AccumuloException, org.apache.accumulo.proxy.thrift.AccumuloSecurityException,
       org.apache.accumulo.proxy.thrift.TableNotFoundException, TException {
-    
+
     String cwid = createConditionalWriter(login, tableName, new ConditionalWriterOptions());
     try {
       return updateRowsConditionally(cwid, Collections.singletonMap(row, updates)).get(row);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/proxy/src/main/java/org/apache/accumulo/proxy/TestProxyClient.java
----------------------------------------------------------------------
diff --git a/proxy/src/main/java/org/apache/accumulo/proxy/TestProxyClient.java b/proxy/src/main/java/org/apache/accumulo/proxy/TestProxyClient.java
index 4850c19..99ebb38 100644
--- a/proxy/src/main/java/org/apache/accumulo/proxy/TestProxyClient.java
+++ b/proxy/src/main/java/org/apache/accumulo/proxy/TestProxyClient.java
@@ -42,14 +42,14 @@ import org.apache.thrift.transport.TTransport;
 import org.apache.thrift.transport.TTransportException;
 
 public class TestProxyClient {
-  
+
   protected AccumuloProxy.Client proxy;
   protected TTransport transport;
-  
+
   public TestProxyClient(String host, int port) throws TTransportException {
     this(host, port, new TCompactProtocol.Factory());
   }
-  
+
   public TestProxyClient(String host, int port, TProtocolFactory protoFactory) throws TTransportException {
     final TSocket socket = new TSocket(host, port);
     socket.setTimeout(600000);
@@ -58,41 +58,41 @@ public class TestProxyClient {
     proxy = new AccumuloProxy.Client(protocol);
     transport.open();
   }
-  
+
   public AccumuloProxy.Client proxy() {
     return proxy;
   }
-  
+
   public static void main(String[] args) throws Exception {
-    
+
     TestProxyClient tpc = new TestProxyClient("localhost", 42424);
     String principal = "root";
     Map<String,String> props = new TreeMap<String,String>();
     props.put("password", "secret");
-    
+
     System.out.println("Logging in");
     ByteBuffer login = tpc.proxy.login(principal, props);
-    
+
     System.out.println("Creating user: ");
     if (!tpc.proxy().listLocalUsers(login).contains("testuser")) {
       tpc.proxy().createLocalUser(login, "testuser", ByteBuffer.wrap("testpass".getBytes(UTF_8)));
     }
     System.out.println("UserList: " + tpc.proxy().listLocalUsers(login));
-    
+
     System.out.println("Listing: " + tpc.proxy().listTables(login));
-    
+
     System.out.println("Deleting: ");
     String testTable = "testtableOMGOMGOMG";
-    
+
     System.out.println("Creating: ");
-    
+
     if (tpc.proxy().tableExists(login, testTable))
       tpc.proxy().deleteTable(login, testTable);
-    
+
     tpc.proxy().createTable(login, testTable, true, TimeType.MILLIS);
-    
+
     System.out.println("Listing: " + tpc.proxy().listTables(login));
-    
+
     System.out.println("Writing: ");
     Date start = new Date();
     Date then = new Date();
@@ -104,7 +104,7 @@ public class TestProxyClient {
       ColumnUpdate update = new ColumnUpdate(ByteBuffer.wrap(("cf" + i).getBytes(UTF_8)), ByteBuffer.wrap(("cq" + i).getBytes(UTF_8)));
       update.setValue(Util.randStringBuffer(10));
       mutations.put(ByteBuffer.wrap(result.getBytes(UTF_8)), Collections.singletonList(update));
-      
+
       if (i % 1000 == 0) {
         tpc.proxy().updateAndFlush(login, testTable, mutations);
         mutations.clear();
@@ -113,12 +113,12 @@ public class TestProxyClient {
     tpc.proxy().updateAndFlush(login, testTable, mutations);
     Date end = new Date();
     System.out.println(" End of writing: " + (end.getTime() - start.getTime()));
-    
+
     tpc.proxy().deleteTable(login, testTable);
     tpc.proxy().createTable(login, testTable, true, TimeType.MILLIS);
-    
+
     // Thread.sleep(1000);
-    
+
     System.out.println("Writing async: ");
     start = new Date();
     then = new Date();
@@ -134,7 +134,7 @@ public class TestProxyClient {
       tpc.proxy().update(writer, mutations);
       mutations.clear();
     }
-    
+
     end = new Date();
     System.out.println(" End of writing: " + (end.getTime() - start.getTime()));
     start = end;
@@ -142,29 +142,29 @@ public class TestProxyClient {
     tpc.proxy().closeWriter(writer);
     end = new Date();
     System.out.println(" End of closing: " + (end.getTime() - start.getTime()));
-    
+
     System.out.println("Reading: ");
-    
+
     String regex = "cf1.*";
-    
+
     IteratorSetting is = new IteratorSetting(50, regex, RegExFilter.class);
     RegExFilter.setRegexs(is, null, regex, null, null, false);
-    
+
     String cookie = tpc.proxy().createScanner(login, testTable, null);
-    
+
     int i = 0;
     start = new Date();
     then = new Date();
     boolean hasNext = true;
-    
+
     int k = 1000;
     while (hasNext) {
       ScanResult kvList = tpc.proxy().nextK(cookie, k);
-      
+
       Date now = new Date();
       System.out.println(i + " " + (now.getTime() - then.getTime()));
       then = now;
-      
+
       i += kvList.getResultsSize();
       // for (TKeyValue kv:kvList.getResults()) System.out.println(new Key(kv.getKey()));
       hasNext = kvList.isMore();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/proxy/src/main/java/org/apache/accumulo/proxy/Util.java
----------------------------------------------------------------------
diff --git a/proxy/src/main/java/org/apache/accumulo/proxy/Util.java b/proxy/src/main/java/org/apache/accumulo/proxy/Util.java
index bc0db72..29e5a40 100644
--- a/proxy/src/main/java/org/apache/accumulo/proxy/Util.java
+++ b/proxy/src/main/java/org/apache/accumulo/proxy/Util.java
@@ -26,35 +26,35 @@ import org.apache.accumulo.proxy.thrift.IteratorSetting;
 import org.apache.accumulo.proxy.thrift.Key;
 
 public class Util {
-  
+
   private static Random random = new Random(0);
-  
+
   public static String randString(int numbytes) {
     return new BigInteger(numbytes * 5, random).toString(32);
   }
-  
+
   public static ByteBuffer randStringBuffer(int numbytes) {
     return ByteBuffer.wrap(new BigInteger(numbytes * 5, random).toString(32).getBytes(UTF_8));
   }
-  
+
   public static IteratorSetting iteratorSetting2ProxyIteratorSetting(org.apache.accumulo.core.client.IteratorSetting is) {
     return new IteratorSetting(is.getPriority(), is.getName(), is.getIteratorClass(), is.getOptions());
   }
-  
+
   public static Key toThrift(org.apache.accumulo.core.data.Key key) {
     Key pkey = new Key(ByteBuffer.wrap(key.getRow().getBytes()), ByteBuffer.wrap(key.getColumnFamily().getBytes()), ByteBuffer.wrap(key.getColumnQualifier()
         .getBytes()), ByteBuffer.wrap(key.getColumnVisibility().getBytes()));
     pkey.setTimestamp(key.getTimestamp());
     return pkey;
   }
-  
+
   public static org.apache.accumulo.core.data.Key fromThrift(Key pkey) {
     if (pkey == null)
       return null;
-    return new org.apache.accumulo.core.data.Key(deNullify(pkey.getRow()), deNullify(pkey.getColFamily()), deNullify(pkey.getColQualifier()), deNullify(pkey.getColVisibility()),
-        pkey.getTimestamp());
+    return new org.apache.accumulo.core.data.Key(deNullify(pkey.getRow()), deNullify(pkey.getColFamily()), deNullify(pkey.getColQualifier()),
+        deNullify(pkey.getColVisibility()), pkey.getTimestamp());
   }
-  
+
   protected static byte[] deNullify(byte[] in) {
     if (in == null)
       return new byte[0];


[24/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataIngest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataIngest.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataIngest.java
index 52ea0bd..e899ff5 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataIngest.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataIngest.java
@@ -49,13 +49,13 @@ public class FileDataIngest {
   public static final String REFS_FILE_EXT = "filext";
   public static final ByteSequence CHUNK_CF_BS = new ArrayByteSequence(CHUNK_CF.getBytes(), 0, CHUNK_CF.getLength());
   public static final ByteSequence REFS_CF_BS = new ArrayByteSequence(REFS_CF.getBytes(), 0, REFS_CF.getLength());
-  
+
   int chunkSize;
   byte[] chunkSizeBytes;
   byte[] buf;
   MessageDigest md5digest;
   ColumnVisibility cv;
-  
+
   public FileDataIngest(int chunkSize, ColumnVisibility colvis) {
     this.chunkSize = chunkSize;
     chunkSizeBytes = intToBytes(chunkSize);
@@ -67,35 +67,35 @@ public class FileDataIngest {
     }
     cv = colvis;
   }
-  
+
   public String insertFileData(String filename, BatchWriter bw) throws MutationsRejectedException, IOException {
     if (chunkSize == 0)
       return "";
     md5digest.reset();
     String uid = hexString(md5digest.digest(filename.getBytes()));
-    
+
     // read through file once, calculating hashes
     md5digest.reset();
     InputStream fis = null;
     int numRead = 0;
     try {
-	    fis = new FileInputStream(filename);
-	    numRead = fis.read(buf);
-	    while (numRead >= 0) {
-	      if (numRead > 0) {
-	        md5digest.update(buf, 0, numRead);
-	      }
-	      numRead = fis.read(buf);
-	    }
+      fis = new FileInputStream(filename);
+      numRead = fis.read(buf);
+      while (numRead >= 0) {
+        if (numRead > 0) {
+          md5digest.update(buf, 0, numRead);
+        }
+        numRead = fis.read(buf);
+      }
     } finally {
       if (fis != null) {
-    	  fis.close();
+        fis.close();
       }
     }
-    
+
     String hash = hexString(md5digest.digest());
     Text row = new Text(hash);
-    
+
     // write info to accumulo
     Mutation m = new Mutation(row);
     m.put(REFS_CF, KeyUtil.buildNullSepText(uid, REFS_ORIG_FILE), cv, new Value(filename.getBytes()));
@@ -103,34 +103,34 @@ public class FileDataIngest {
     if (fext != null)
       m.put(REFS_CF, KeyUtil.buildNullSepText(uid, REFS_FILE_EXT), cv, new Value(fext.getBytes()));
     bw.addMutation(m);
-    
+
     // read through file again, writing chunks to accumulo
     int chunkCount = 0;
     try {
-	    fis = new FileInputStream(filename);
-	    numRead = fis.read(buf);
-	    while (numRead >= 0) {
-	      while (numRead < buf.length) {
-	        int moreRead = fis.read(buf, numRead, buf.length - numRead);
-	        if (moreRead > 0)
-	          numRead += moreRead;
-	        else if (moreRead < 0)
-	          break;
-	      }
-	      m = new Mutation(row);
-	      Text chunkCQ = new Text(chunkSizeBytes);
-	      chunkCQ.append(intToBytes(chunkCount), 0, 4);
-	      m.put(CHUNK_CF, chunkCQ, cv, new Value(buf, 0, numRead));
-	      bw.addMutation(m);
-	      if (chunkCount == Integer.MAX_VALUE)
-	        throw new RuntimeException("too many chunks for file " + filename + ", try raising chunk size");
-	      chunkCount++;
-	      numRead = fis.read(buf);
-	    }
+      fis = new FileInputStream(filename);
+      numRead = fis.read(buf);
+      while (numRead >= 0) {
+        while (numRead < buf.length) {
+          int moreRead = fis.read(buf, numRead, buf.length - numRead);
+          if (moreRead > 0)
+            numRead += moreRead;
+          else if (moreRead < 0)
+            break;
+        }
+        m = new Mutation(row);
+        Text chunkCQ = new Text(chunkSizeBytes);
+        chunkCQ.append(intToBytes(chunkCount), 0, 4);
+        m.put(CHUNK_CF, chunkCQ, cv, new Value(buf, 0, numRead));
+        bw.addMutation(m);
+        if (chunkCount == Integer.MAX_VALUE)
+          throw new RuntimeException("too many chunks for file " + filename + ", try raising chunk size");
+        chunkCount++;
+        numRead = fis.read(buf);
+      }
     } finally {
-    	if (fis != null) {
-    		fis.close();
-    	}
+      if (fis != null) {
+        fis.close();
+      }
     }
     m = new Mutation(row);
     Text chunkCQ = new Text(chunkSizeBytes);
@@ -139,14 +139,14 @@ public class FileDataIngest {
     bw.addMutation(m);
     return hash;
   }
-  
+
   public static int bytesToInt(byte[] b, int offset) {
     if (b.length <= offset + 3)
       throw new NumberFormatException("couldn't pull integer from bytes at offset " + offset);
     int i = (((b[offset] & 255) << 24) + ((b[offset + 1] & 255) << 16) + ((b[offset + 2] & 255) << 8) + ((b[offset + 3] & 255) << 0));
     return i;
   }
-  
+
   public static byte[] intToBytes(int l) {
     byte[] b = new byte[4];
     b[0] = (byte) (l >>> 24);
@@ -155,13 +155,13 @@ public class FileDataIngest {
     b[3] = (byte) (l >>> 0);
     return b;
   }
-  
+
   private static String getExt(String filename) {
     if (filename.indexOf(".") == -1)
       return null;
     return filename.substring(filename.lastIndexOf(".") + 1);
   }
-  
+
   public String hexString(byte[] bytes) {
     StringBuilder sb = new StringBuilder();
     for (byte b : bytes) {
@@ -169,24 +169,23 @@ public class FileDataIngest {
     }
     return sb.toString();
   }
-  
+
   public static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--vis", description="use a given visibility for the new counts", converter=VisibilityConverter.class)
+    @Parameter(names = "--vis", description = "use a given visibility for the new counts", converter = VisibilityConverter.class)
     ColumnVisibility visibility = new ColumnVisibility();
-    
-    @Parameter(names="--chunk", description="size of the chunks used to store partial files")
-    int chunkSize = 64*1024;
-    
-    @Parameter(description="<file> { <file> ... }")
+
+    @Parameter(names = "--chunk", description = "size of the chunks used to store partial files")
+    int chunkSize = 64 * 1024;
+
+    @Parameter(description = "<file> { <file> ... }")
     List<String> files = new ArrayList<String>();
   }
-  
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(FileDataIngest.class.getName(), args, bwOpts);
-    
+
     Connector conn = opts.getConnector();
     if (!conn.tableOperations().exists(opts.getTableName())) {
       conn.tableOperations().create(opts.getTableName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataQuery.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataQuery.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataQuery.java
index ee55fcc..75e32ae 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataQuery.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/FileDataQuery.java
@@ -44,20 +44,20 @@ public class FileDataQuery {
   List<Entry<Key,Value>> lastRefs;
   private ChunkInputStream cis;
   Scanner scanner;
-  
-  public FileDataQuery(String instanceName, String zooKeepers, String user, AuthenticationToken token, String tableName, Authorizations auths) throws AccumuloException,
-      AccumuloSecurityException, TableNotFoundException {
+
+  public FileDataQuery(String instanceName, String zooKeepers, String user, AuthenticationToken token, String tableName, Authorizations auths)
+      throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     ZooKeeperInstance instance = new ZooKeeperInstance(ClientConfiguration.loadDefault().withInstance(instanceName).withZkHosts(zooKeepers));
     conn = instance.getConnector(user, token);
     lastRefs = new ArrayList<Entry<Key,Value>>();
     cis = new ChunkInputStream();
     scanner = conn.createScanner(tableName, auths);
   }
-  
+
   public List<Entry<Key,Value>> getLastRefs() {
     return lastRefs;
   }
-  
+
   public ChunkInputStream getData(String hash) throws IOException {
     scanner.setRange(new Range(hash));
     scanner.setBatchSize(1);
@@ -73,7 +73,7 @@ public class FileDataQuery {
     cis.setSource(pi);
     return cis;
   }
-  
+
   public String getSomeData(String hash, int numBytes) throws IOException {
     ChunkInputStream is = getData(hash);
     byte[] buf = new byte[numBytes];

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/KeyUtil.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/KeyUtil.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/KeyUtil.java
index d0ebcb0..2f09785 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/KeyUtil.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/KeyUtil.java
@@ -25,10 +25,10 @@ import org.apache.hadoop.io.Text;
  */
 public class KeyUtil {
   public static final byte[] nullbyte = new byte[] {0};
-  
+
   /**
    * Join some number of strings using a null byte separator into a text object.
-   * 
+   *
    * @param s
    *          strings
    * @return a text object containing the strings separated by null bytes
@@ -41,10 +41,10 @@ public class KeyUtil {
     }
     return t;
   }
-  
+
   /**
    * Split a text object using a null byte separator into an array of strings.
-   * 
+   *
    * @param t
    *          null-byte separated text object
    * @return an array of strings

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/VisibilityCombiner.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/VisibilityCombiner.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/VisibilityCombiner.java
index a3e2bcc..ab2e7fc 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/VisibilityCombiner.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/filedata/VisibilityCombiner.java
@@ -24,16 +24,16 @@ import org.apache.accumulo.core.data.ByteSequence;
  * A utility for merging visibilities into the form {@code (VIS1)|(VIS2)|...|(VISN)}. Used by the {@link ChunkCombiner}.
  */
 public class VisibilityCombiner {
-  
+
   private TreeSet<String> visibilities = new TreeSet<String>();
-  
+
   void add(ByteSequence cv) {
     if (cv.length() == 0)
       return;
-    
+
     int depth = 0;
     int offset = 0;
-    
+
     for (int i = 0; i < cv.length(); i++) {
       switch (cv.byteAt(i)) {
         case '(':
@@ -49,25 +49,25 @@ public class VisibilityCombiner {
             insert(cv.subSequence(offset, i));
             offset = i + 1;
           }
-          
+
           break;
       }
     }
-    
+
     insert(cv.subSequence(offset, cv.length()));
-    
+
     if (depth != 0)
       throw new IllegalArgumentException("Invalid vis " + cv);
-    
+
   }
-  
+
   private void insert(ByteSequence cv) {
     for (int i = 0; i < cv.length(); i++) {
-      
+
     }
-    
+
     String cvs = cv.toString();
-    
+
     if (cvs.charAt(0) != '(')
       cvs = "(" + cvs + ")";
     else {
@@ -85,14 +85,14 @@ public class VisibilityCombiner {
             break;
         }
       }
-      
+
       if (depthZeroCloses > 1)
         cvs = "(" + cvs + ")";
     }
-    
+
     visibilities.add(cvs);
   }
-  
+
   byte[] get() {
     StringBuilder sb = new StringBuilder();
     String sep = "";
@@ -101,7 +101,7 @@ public class VisibilityCombiner {
       sep = "|";
       sb.append(cvs);
     }
-    
+
     return sb.toString().getBytes();
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/InsertWithBatchWriter.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/InsertWithBatchWriter.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/InsertWithBatchWriter.java
index 74d8548..0e60086 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/InsertWithBatchWriter.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/InsertWithBatchWriter.java
@@ -34,20 +34,20 @@ import org.apache.hadoop.io.Text;
  * Inserts 10K rows (50K entries) into accumulo with each row having 5 entries.
  */
 public class InsertWithBatchWriter {
-  
+
   public static void main(String[] args) throws AccumuloException, AccumuloSecurityException, MutationsRejectedException, TableExistsException,
       TableNotFoundException {
     ClientOnRequiredTable opts = new ClientOnRequiredTable();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(InsertWithBatchWriter.class.getName(), args, bwOpts);
-    
+
     Connector connector = opts.getConnector();
     MultiTableBatchWriter mtbw = connector.createMultiTableBatchWriter(bwOpts.getBatchWriterConfig());
-    
+
     if (!connector.tableOperations().exists(opts.getTableName()))
       connector.tableOperations().create(opts.getTableName());
     BatchWriter bw = mtbw.getBatchWriter(opts.getTableName());
-    
+
     Text colf = new Text("colfam");
     System.out.println("writing ...");
     for (int i = 0; i < 10000; i++) {
@@ -61,5 +61,5 @@ public class InsertWithBatchWriter {
     }
     mtbw.close();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/ReadData.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/ReadData.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/ReadData.java
index 4eaa31f..041c15f 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/ReadData.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/helloworld/ReadData.java
@@ -38,13 +38,13 @@ import com.beust.jcommander.Parameter;
  * Reads all data between two rows; all data after a given row; or all data in a table, depending on the number of arguments given.
  */
 public class ReadData {
-  
+
   private static final Logger log = Logger.getLogger(ReadData.class);
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--startKey")
+    @Parameter(names = "--startKey")
     String startKey;
-    @Parameter(names="--endKey")
+    @Parameter(names = "--endKey")
     String endKey;
   }
 
@@ -52,9 +52,9 @@ public class ReadData {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(ReadData.class.getName(), args, scanOpts);
-    
+
     Connector connector = opts.getConnector();
-    
+
     Scanner scan = connector.createScanner(opts.getTableName(), opts.auths);
     scan.setBatchSize(scanOpts.scanBatchSize);
     Key start = null;
@@ -65,7 +65,7 @@ public class ReadData {
       end = new Key(new Text(opts.endKey));
     scan.setRange(new Range(start, end));
     Iterator<Entry<Key,Value>> iter = scan.iterator();
-    
+
     while (iter.hasNext()) {
       Entry<Key,Value> e = iter.next();
       Text colf = e.getKey().getColumnFamily();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/isolation/InterferenceTest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/isolation/InterferenceTest.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/isolation/InterferenceTest.java
index fd6d159..bbfca48 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/isolation/InterferenceTest.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/isolation/InterferenceTest.java
@@ -38,43 +38,43 @@ import com.beust.jcommander.Parameter;
 /**
  * This example shows how a concurrent reader and writer can interfere with each other. It creates two threads that run forever reading and writing to the same
  * table.
- * 
+ *
  * When the example is run with isolation enabled, no interference will be observed.
- * 
+ *
  * When the example is run with out isolation, the reader will see partial mutations of a row.
- * 
+ *
  */
 
 public class InterferenceTest {
-  
+
   private static final int NUM_ROWS = 500;
   private static final int NUM_COLUMNS = 113; // scanner batches 1000 by default, so make num columns not a multiple of 10
   private static final Logger log = Logger.getLogger(InterferenceTest.class);
-  
+
   static class Writer implements Runnable {
-    
+
     private final BatchWriter bw;
     private final long iterations;
-    
+
     Writer(BatchWriter bw, long iterations) {
       this.bw = bw;
-      this.iterations = iterations; 
+      this.iterations = iterations;
     }
-    
+
     @Override
     public void run() {
       int row = 0;
       int value = 0;
-      
+
       for (long i = 0; i < iterations; i++) {
         Mutation m = new Mutation(new Text(String.format("%03d", row)));
         row = (row + 1) % NUM_ROWS;
-        
+
         for (int cq = 0; cq < NUM_COLUMNS; cq++)
           m.put(new Text("000"), new Text(String.format("%04d", cq)), new Value(("" + value).getBytes()));
-        
+
         value++;
-        
+
         try {
           bw.addMutation(m);
         } catch (MutationsRejectedException e) {
@@ -89,80 +89,79 @@ public class InterferenceTest {
       }
     }
   }
-  
+
   static class Reader implements Runnable {
-    
+
     private Scanner scanner;
     volatile boolean stop = false;
-    
+
     Reader(Scanner scanner) {
       this.scanner = scanner;
     }
-    
+
     @Override
     public void run() {
       while (!stop) {
         ByteSequence row = null;
         int count = 0;
-        
+
         // all columns in a row should have the same value,
         // use this hash set to track that
         HashSet<String> values = new HashSet<String>();
-        
+
         for (Entry<Key,Value> entry : scanner) {
           if (row == null)
             row = entry.getKey().getRowData();
-          
+
           if (!row.equals(entry.getKey().getRowData())) {
             if (count != NUM_COLUMNS)
               System.err.println("ERROR Did not see " + NUM_COLUMNS + " columns in row " + row);
-            
+
             if (values.size() > 1)
               System.err.println("ERROR Columns in row " + row + " had multiple values " + values);
-            
+
             row = entry.getKey().getRowData();
             count = 0;
             values.clear();
           }
-          
+
           count++;
-          
+
           values.add(entry.getValue().toString());
         }
-        
+
         if (count > 0 && count != NUM_COLUMNS)
           System.err.println("ERROR Did not see " + NUM_COLUMNS + " columns in row " + row);
-        
+
         if (values.size() > 1)
           System.err.println("ERROR Columns in row " + row + " had multiple values " + values);
       }
     }
-    
+
     public void stopNow() {
       stop = true;
     }
   }
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--iterations", description="number of times to run", required=true)
+    @Parameter(names = "--iterations", description = "number of times to run", required = true)
     long iterations = 0;
-    @Parameter(names="--isolated", description="use isolated scans")
+    @Parameter(names = "--isolated", description = "use isolated scans")
     boolean isolated = false;
   }
-  
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(InterferenceTest.class.getName(), args, bwOpts);
-    
+
     if (opts.iterations < 1)
       opts.iterations = Long.MAX_VALUE;
-    
+
     Connector conn = opts.getConnector();
     if (!conn.tableOperations().exists(opts.getTableName()))
       conn.tableOperations().create(opts.getTableName());
-    
+
     Thread writer = new Thread(new Writer(conn.createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig()), opts.iterations));
     writer.start();
     Reader r;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
index 351a51c..01fbb8f 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TableToFile.java
@@ -102,7 +102,7 @@ public class TableToFile extends Configured implements Tool {
   }
 
   /**
-   * 
+   *
    * @param args
    *          instanceName zookeepers username password table columns outputpath
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TeraSortIngest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TeraSortIngest.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TeraSortIngest.java
index ade6ce1..8c48877 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TeraSortIngest.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TeraSortIngest.java
@@ -6,9 +6,9 @@
  * to you under the Apache License, Version 2.0 (the
  * "License"); you may not use this file except in compliance
  * with the License. You may obtain a copy of the License at
- * 
+ *
  * http://www.apache.org/licenses/LICENSE-2.0
- * 
+ *
  * Unless required by applicable law or agreed to in writing, software
  * distributed under the License is distributed on an "AS IS" BASIS,
  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
@@ -59,12 +59,12 @@ import com.beust.jcommander.Parameter;
  * <li>The rowid is the right justified row id as a int.
  * <li>The filler consists of 7 runs of 10 characters from 'A' to 'Z'.
  * </ul>
- * 
+ *
  * This TeraSort is slightly modified to allow for variable length key sizes and value sizes. The row length isn't variable. To generate a terabyte of data in
  * the same way TeraSort does use 10000000000 rows and 10/10 byte key length and 78/78 byte value length. Along with the 10 byte row id and \r\n this gives you
  * 100 byte row * 10000000000 rows = 1tb. Min/Max ranges for key and value parameters are inclusive/inclusive respectively.
- * 
- * 
+ *
+ *
  */
 public class TeraSortIngest extends Configured implements Tool {
   /**
@@ -202,7 +202,7 @@ public class TeraSortIngest extends Configured implements Tool {
 
     /**
      * Start the random number generator on the given iteration.
-     * 
+     *
      * @param initalIteration
      *          the iteration number to start on
      */
@@ -290,7 +290,7 @@ public class TeraSortIngest extends Configured implements Tool {
 
     /**
      * Add the required filler bytes. Each row consists of 7 blocks of 10 characters and 1 block of 8 characters.
-     * 
+     *
      * @param rowId
      *          the current row number
      */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TokenFileWordCount.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TokenFileWordCount.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TokenFileWordCount.java
index 7bb7e69..7822910 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TokenFileWordCount.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/TokenFileWordCount.java
@@ -36,10 +36,10 @@ import org.apache.log4j.Logger;
 /**
  * A simple map reduce job that inserts word counts into accumulo. See the README for instructions on how to run this. This version does not use the ClientOpts
  * class to parse arguments as an example of using AccumuloInputFormat and AccumuloOutputFormat directly. See README.mapred for more details.
- * 
+ *
  */
 public class TokenFileWordCount extends Configured implements Tool {
-  
+
   private static final Logger log = Logger.getLogger(TokenFileWordCount.class);
 
   public static class MapClass extends Mapper<LongWritable,Text,Text,Mutation> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/WordCount.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/WordCount.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/WordCount.java
index 4f1f861..8ead101 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/WordCount.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/WordCount.java
@@ -32,18 +32,18 @@ import org.apache.hadoop.mapreduce.Mapper;
 import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
+import org.apache.log4j.Logger;
 
 import com.beust.jcommander.Parameter;
-import org.apache.log4j.Logger;
 
 /**
  * A simple map reduce job that inserts word counts into accumulo. See the README for instructions on how to run this.
- * 
+ *
  */
 public class WordCount extends Configured implements Tool {
 
   private static final Logger log = Logger.getLogger(WordCount.class);
-  
+
   static class Opts extends MapReduceClientOnRequiredTable {
     @Parameter(names = "--input", description = "input directory")
     String inputDirectory;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/SetupTable.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/SetupTable.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/SetupTable.java
index ac96e9d..8651c39 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/SetupTable.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/SetupTable.java
@@ -27,12 +27,12 @@ import org.apache.hadoop.io.Text;
 import com.beust.jcommander.Parameter;
 
 public class SetupTable {
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(description="<split> { <split> ... } ")
+    @Parameter(description = "<split> { <split> ... } ")
     List<String> splits = new ArrayList<String>();
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(SetupTable.class.getName(), args);
@@ -45,6 +45,6 @@ public class SetupTable {
         intialPartitions.add(new Text(split));
       }
       conn.tableOperations().addSplits(opts.getTableName(), intialPartitions);
-    } 
+    }
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/VerifyIngest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/VerifyIngest.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/VerifyIngest.java
index 61d3f7e..fd54058 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/VerifyIngest.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/mapreduce/bulk/VerifyIngest.java
@@ -35,52 +35,52 @@ import com.beust.jcommander.Parameter;
 
 public class VerifyIngest {
   private static final Logger log = Logger.getLogger(VerifyIngest.class);
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--start-row")
+    @Parameter(names = "--start-row")
     int startRow = 0;
-    @Parameter(names="--count", required=true, description="number of rows to verify")
+    @Parameter(names = "--count", required = true, description = "number of rows to verify")
     int numRows = 0;
   }
-  
+
   public static void main(String[] args) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     Opts opts = new Opts();
     opts.parseArgs(VerifyIngest.class.getName(), args);
-    
+
     Connector connector = opts.getConnector();
     Scanner scanner = connector.createScanner(opts.getTableName(), opts.auths);
-    
+
     scanner.setRange(new Range(new Text(String.format("row_%08d", opts.startRow)), null));
-    
+
     Iterator<Entry<Key,Value>> si = scanner.iterator();
-    
+
     boolean ok = true;
-    
+
     for (int i = opts.startRow; i < opts.numRows; i++) {
-      
+
       if (si.hasNext()) {
         Entry<Key,Value> entry = si.next();
-        
+
         if (!entry.getKey().getRow().toString().equals(String.format("row_%08d", i))) {
           log.error("unexpected row key " + entry.getKey().getRow().toString() + " expected " + String.format("row_%08d", i));
           ok = false;
         }
-        
+
         if (!entry.getValue().toString().equals(String.format("value_%08d", i))) {
           log.error("unexpected value " + entry.getValue().toString() + " expected " + String.format("value_%08d", i));
           ok = false;
         }
-        
+
       } else {
         log.error("no more rows, expected " + String.format("row_%08d", i));
         ok = false;
         break;
       }
-      
+
     }
-    
+
     if (ok)
       System.out.println("OK");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/reservations/ARS.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/reservations/ARS.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/reservations/ARS.java
index e9dc2aa..6f47abd 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/reservations/ARS.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/reservations/ARS.java
@@ -22,9 +22,9 @@ import java.util.Map.Entry;
 
 import jline.console.ConsoleReader;
 
+import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.ConditionalWriter;
 import org.apache.accumulo.core.client.ConditionalWriter.Status;
-import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.ConditionalWriterConfig;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.client.IsolatedScanner;
@@ -52,43 +52,43 @@ import org.apache.log4j.Logger;
 // 10 threads per node.
 
 public class ARS {
-  
+
   private static final Logger log = Logger.getLogger(ARS.class);
-  
+
   private Connector conn;
   private String rTable;
-  
+
   public enum ReservationResult {
     RESERVED, WAIT_LISTED
   }
-  
+
   public ARS(Connector conn, String rTable) {
     this.conn = conn;
     this.rTable = rTable;
   }
-  
+
   public List<String> setCapacity(String what, String when, int count) {
     // EXCERCISE implement this method which atomically sets a capacity and returns anyone who was moved to the wait list if the capacity was decreased
-    
+
     throw new UnsupportedOperationException();
   }
-  
+
   public ReservationResult reserve(String what, String when, String who) throws Exception {
-    
+
     String row = what + ":" + when;
-    
+
     // EXCERCISE This code assumes there is no reservation and tries to create one. If a reservation exist then the update will fail. This is a good strategy
     // when it is expected there are usually no reservations. Could modify the code to scan first.
-    
+
     // The following mutation requires that the column tx:seq does not exist and will fail if it does.
     ConditionalMutation update = new ConditionalMutation(row, new Condition("tx", "seq"));
     update.put("tx", "seq", "0");
     update.put("res", String.format("%04d", 0), who);
-    
+
     ReservationResult result = ReservationResult.RESERVED;
-    
+
     ConditionalWriter cwriter = conn.createConditionalWriter(rTable, new ConditionalWriterConfig());
-    
+
     try {
       while (true) {
         Status status = cwriter.write(update).getStatus();
@@ -102,24 +102,24 @@ public class ARS {
           default:
             throw new RuntimeException("Unexpected status " + status);
         }
-        
+
         // EXCERCISE in the case of many threads trying to reserve a slot, this approach of immediately retrying is inefficient. Exponential back-off is good
         // general solution to solve contention problems like this. However in this particular case, exponential back-off could penalize the earliest threads
         // that attempted to make a reservation by putting them later in the list. A more complex solution could involve having independent sub-queues within
         // the row that approximately maintain arrival order and use exponential back off to fairly merge the sub-queues into the main queue.
-        
+
         // it is important to use an isolated scanner so that only whole mutations are seen
         Scanner scanner = new IsolatedScanner(conn.createScanner(rTable, Authorizations.EMPTY));
         scanner.setRange(new Range(row));
-        
+
         int seq = -1;
         int maxReservation = -1;
-        
+
         for (Entry<Key,Value> entry : scanner) {
           String cf = entry.getKey().getColumnFamilyData().toString();
           String cq = entry.getKey().getColumnQualifierData().toString();
           String val = entry.getValue().toString();
-          
+
           if (cf.equals("tx") && cq.equals("seq")) {
             seq = Integer.parseInt(val);
           } else if (cf.equals("res")) {
@@ -130,21 +130,21 @@ public class ARS {
                 return ReservationResult.RESERVED; // already have the first reservation
               else
                 return ReservationResult.WAIT_LISTED; // already on wait list
-                
+
             // EXCERCISE the way this code finds the max reservation is very inefficient.... it would be better if it did not have to scan the entire row.
             // One possibility is to just use the sequence number. Could also consider sorting the data in another way and/or using an iterator.
             maxReservation = Integer.parseInt(cq);
           }
         }
-        
+
         Condition condition = new Condition("tx", "seq");
         if (seq >= 0)
           condition.setValue(seq + ""); // only expect a seq # if one was seen
-          
+
         update = new ConditionalMutation(row, condition);
         update.put("tx", "seq", (seq + 1) + "");
         update.put("res", String.format("%04d", maxReservation + 1), who);
-        
+
         // EXCERCISE if set capacity is implemented, then result should take capacity into account
         if (maxReservation == -1)
           result = ReservationResult.RESERVED; // if successful, will be first reservation
@@ -154,48 +154,48 @@ public class ARS {
     } finally {
       cwriter.close();
     }
-    
+
   }
-  
+
   public void cancel(String what, String when, String who) throws Exception {
-    
+
     String row = what + ":" + when;
-    
+
     // Even though this method is only deleting a column, its important to use a conditional writer. By updating the seq # when deleting a reservation, it
     // will cause any concurrent reservations to retry. If this delete were done using a batch writer, then a concurrent reservation could report WAIT_LISTED
     // when it actually got the reservation.
-    
+
     ConditionalWriter cwriter = conn.createConditionalWriter(rTable, new ConditionalWriterConfig());
-    
+
     try {
       while (true) {
-        
+
         // its important to use an isolated scanner so that only whole mutations are seen
         Scanner scanner = new IsolatedScanner(conn.createScanner(rTable, Authorizations.EMPTY));
         scanner.setRange(new Range(row));
-        
+
         int seq = -1;
         String reservation = null;
-        
+
         for (Entry<Key,Value> entry : scanner) {
           String cf = entry.getKey().getColumnFamilyData().toString();
           String cq = entry.getKey().getColumnQualifierData().toString();
           String val = entry.getValue().toString();
-          
+
           // EXCERCISE avoid linear scan
-          
+
           if (cf.equals("tx") && cq.equals("seq")) {
             seq = Integer.parseInt(val);
           } else if (cf.equals("res") && val.equals(who)) {
             reservation = cq;
           }
         }
-        
+
         if (reservation != null) {
           ConditionalMutation update = new ConditionalMutation(row, new Condition("tx", "seq").setValue(seq + ""));
           update.putDelete("res", reservation);
           update.put("tx", "seq", (seq + 1) + "");
-          
+
           Status status = cwriter.write(update).getStatus();
           switch (status) {
             case ACCEPTED:
@@ -209,50 +209,50 @@ public class ARS {
             default:
               throw new RuntimeException("Unexpected status " + status);
           }
-          
+
         } else {
           // not reserved, nothing to do
           break;
         }
-        
+
       }
     } finally {
       cwriter.close();
     }
   }
-  
+
   public List<String> list(String what, String when) throws Exception {
     String row = what + ":" + when;
-    
+
     // its important to use an isolated scanner so that only whole mutations are seen
     Scanner scanner = new IsolatedScanner(conn.createScanner(rTable, Authorizations.EMPTY));
     scanner.setRange(new Range(row));
     scanner.fetchColumnFamily(new Text("res"));
-    
+
     List<String> reservations = new ArrayList<String>();
-    
+
     for (Entry<Key,Value> entry : scanner) {
       String val = entry.getValue().toString();
       reservations.add(val);
     }
-    
+
     return reservations;
   }
-  
+
   public static void main(String[] args) throws Exception {
     final ConsoleReader reader = new ConsoleReader();
     ARS ars = null;
-    
+
     while (true) {
       String line = reader.readLine(">");
       if (line == null)
         break;
-      
+
       final String[] tokens = line.split("\\s+");
-      
+
       if (tokens[0].equals("reserve") && tokens.length >= 4 && ars != null) {
         // start up multiple threads all trying to reserve the same resource, no more than one should succeed
-        
+
         final ARS fars = ars;
         ArrayList<Thread> threads = new ArrayList<Thread>();
         for (int i = 3; i < tokens.length; i++) {
@@ -267,16 +267,16 @@ public class ARS {
               }
             }
           };
-          
+
           threads.add(new Thread(reservationTask));
         }
-        
+
         for (Thread thread : threads)
           thread.start();
-        
+
         for (Thread thread : threads)
           thread.join();
-        
+
       } else if (tokens[0].equals("cancel") && tokens.length == 4 && ars != null) {
         ars.cancel(tokens[1], tokens[2], tokens[3]);
       } else if (tokens[0].equals("list") && tokens.length == 3 && ars != null) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/ContinuousQuery.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/ContinuousQuery.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/ContinuousQuery.java
index 5367a44..8d05922 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/ContinuousQuery.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/ContinuousQuery.java
@@ -20,8 +20,8 @@ import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.Collections;
 import java.util.Map.Entry;
-import java.util.concurrent.TimeUnit;
 import java.util.Random;
+import java.util.concurrent.TimeUnit;
 
 import org.apache.accumulo.core.cli.BatchScannerOpts;
 import org.apache.accumulo.core.cli.ClientOpts;
@@ -40,48 +40,48 @@ import com.beust.jcommander.Parameter;
 /**
  * Using the doc2word table created by Reverse.java, this program randomly selects N words per document. Then it continually queries a random set of words in
  * the shard table (created by {@link Index}) using the {@link IntersectingIterator}.
- * 
+ *
  * See docs/examples/README.shard for instructions.
  */
 
 public class ContinuousQuery {
-  
+
   static class Opts extends ClientOpts {
-    @Parameter(names="--shardTable", required=true, description="name of the shard table")
+    @Parameter(names = "--shardTable", required = true, description = "name of the shard table")
     String table = null;
-    @Parameter(names="--doc2Term", required=true, description="name of the doc2Term table")
+    @Parameter(names = "--doc2Term", required = true, description = "name of the doc2Term table")
     String doc2Term;
-    @Parameter(names="--terms", required=true, description="the number of terms in the query")
+    @Parameter(names = "--terms", required = true, description = "the number of terms in the query")
     int numTerms;
-    @Parameter(names="--count", description="the number of queries to run")
+    @Parameter(names = "--count", description = "the number of queries to run")
     long iterations = Long.MAX_VALUE;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     BatchScannerOpts bsOpts = new BatchScannerOpts();
     opts.parseArgs(ContinuousQuery.class.getName(), args, bsOpts);
-    
+
     Connector conn = opts.getConnector();
-    
+
     ArrayList<Text[]> randTerms = findRandomTerms(conn.createScanner(opts.doc2Term, opts.auths), opts.numTerms);
-    
+
     Random rand = new Random();
-    
+
     BatchScanner bs = conn.createBatchScanner(opts.table, opts.auths, bsOpts.scanThreads);
     bs.setTimeout(bsOpts.scanTimeout, TimeUnit.MILLISECONDS);
-    
+
     for (long i = 0; i < opts.iterations; i += 1) {
       Text[] columns = randTerms.get(rand.nextInt(randTerms.size()));
-      
+
       bs.clearScanIterators();
       bs.clearColumns();
-      
+
       IteratorSetting ii = new IteratorSetting(20, "ii", IntersectingIterator.class);
       IntersectingIterator.setColumnFamilies(ii, columns);
       bs.addScanIterator(ii);
       bs.setRanges(Collections.singleton(new Range()));
-      
+
       long t1 = System.currentTimeMillis();
       int count = 0;
       for (@SuppressWarnings("unused")
@@ -89,44 +89,44 @@ public class ContinuousQuery {
         count++;
       }
       long t2 = System.currentTimeMillis();
-      
+
       System.out.printf("  %s %,d %6.3f%n", Arrays.asList(columns), count, (t2 - t1) / 1000.0);
     }
-    
+
     bs.close();
-    
+
   }
-  
+
   private static ArrayList<Text[]> findRandomTerms(Scanner scanner, int numTerms) {
-    
+
     Text currentRow = null;
-    
+
     ArrayList<Text> words = new ArrayList<Text>();
     ArrayList<Text[]> ret = new ArrayList<Text[]>();
-    
+
     Random rand = new Random();
-    
+
     for (Entry<Key,Value> entry : scanner) {
       Key key = entry.getKey();
-      
+
       if (currentRow == null)
         currentRow = key.getRow();
-      
+
       if (!currentRow.equals(key.getRow())) {
         selectRandomWords(words, ret, rand, numTerms);
         words.clear();
         currentRow = key.getRow();
       }
-      
+
       words.add(key.getColumnFamily());
-      
+
     }
-    
+
     selectRandomWords(words, ret, rand, numTerms);
-    
+
     return ret;
   }
-  
+
   private static void selectRandomWords(ArrayList<Text> words, ArrayList<Text[]> ret, Random rand, int numTerms) {
     if (words.size() >= numTerms) {
       Collections.shuffle(words, rand);
@@ -134,7 +134,7 @@ public class ContinuousQuery {
       for (int i = 0; i < docWords.length; i++) {
         docWords[i] = words.get(i);
       }
-      
+
       ret.add(docWords);
     }
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Index.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Index.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Index.java
index accb3a0..3564be4 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Index.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Index.java
@@ -33,41 +33,41 @@ import com.beust.jcommander.Parameter;
 
 /**
  * This program indexes a set of documents given on the command line into a shard table.
- * 
+ *
  * What it writes to the table is row = partition id, column family = term, column qualifier = document id.
- * 
+ *
  * See docs/examples/README.shard for instructions.
  */
 
 public class Index {
-  
+
   static Text genPartition(int partition) {
     return new Text(String.format("%08x", Math.abs(partition)));
   }
-  
+
   public static void index(int numPartitions, Text docId, String doc, String splitRegex, BatchWriter bw) throws Exception {
-    
+
     String[] tokens = doc.split(splitRegex);
-    
+
     Text partition = genPartition(doc.hashCode() % numPartitions);
-    
+
     Mutation m = new Mutation(partition);
-    
+
     HashSet<String> tokensSeen = new HashSet<String>();
-    
+
     for (String token : tokens) {
       token = token.toLowerCase();
-      
+
       if (!tokensSeen.contains(token)) {
         tokensSeen.add(token);
         m.put(new Text(token), docId, new Value(new byte[0]));
       }
     }
-    
+
     if (m.size() > 0)
       bw.addMutation(m);
   }
-  
+
   public static void index(int numPartitions, File src, String splitRegex, BatchWriter bw) throws Exception {
     if (src.isDirectory()) {
       for (File child : src.listFiles()) {
@@ -75,41 +75,41 @@ public class Index {
       }
     } else {
       FileReader fr = new FileReader(src);
-      
+
       StringBuilder sb = new StringBuilder();
-      
+
       char data[] = new char[4096];
       int len;
       while ((len = fr.read(data)) != -1) {
         sb.append(data, 0, len);
       }
-      
+
       fr.close();
-      
+
       index(numPartitions, new Text(src.getAbsolutePath()), sb.toString(), splitRegex, bw);
     }
-    
+
   }
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(names="--partitions", required=true, description="the number of shards to create")
+    @Parameter(names = "--partitions", required = true, description = "the number of shards to create")
     int partitions;
-    @Parameter(required=true, description="<file> { <file> ... }")
+    @Parameter(required = true, description = "<file> { <file> ... }")
     List<String> files = new ArrayList<String>();
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(Index.class.getName(), args, bwOpts);
-    
+
     String splitRegex = "\\W+";
-    
+
     BatchWriter bw = opts.getConnector().createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig());
     for (String filename : opts.files) {
       index(opts.partitions, new File(filename), splitRegex, bw);
     }
     bw.close();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Query.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Query.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Query.java
index b0502a7..41d5dc7 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Query.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Query.java
@@ -37,19 +37,19 @@ import com.beust.jcommander.Parameter;
 
 /**
  * This program queries a set of terms in the shard table (populated by {@link Index}) using the {@link IntersectingIterator}.
- * 
+ *
  * See docs/examples/README.shard for instructions.
  */
 
 public class Query {
-  
+
   static class Opts extends ClientOnRequiredTable {
-    @Parameter(description=" term { <term> ... }")
+    @Parameter(description = " term { <term> ... }")
     List<String> terms = new ArrayList<String>();
   }
-  
+
   public static List<String> query(BatchScanner bs, List<String> terms) {
-    
+
     Text columns[] = new Text[terms.size()];
     int i = 0;
     for (String term : terms) {
@@ -65,7 +65,7 @@ public class Query {
     }
     return result;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     BatchScannerOpts bsOpts = new BatchScannerOpts();
@@ -77,5 +77,5 @@ public class Query {
     for (String entry : query(bs, opts.terms))
       System.out.println("  " + entry);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Reverse.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Reverse.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Reverse.java
index 4d03425..dbcbe5f 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Reverse.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shard/Reverse.java
@@ -34,39 +34,39 @@ import com.beust.jcommander.Parameter;
 /**
  * The program reads an accumulo table written by {@link Index} and writes out to another table. It writes out a mapping of documents to terms. The document to
  * term mapping is used by {@link ContinuousQuery}.
- * 
+ *
  * See docs/examples/README.shard for instructions.
  */
 
 public class Reverse {
-  
+
   static class Opts extends ClientOpts {
-    @Parameter(names="--shardTable")
+    @Parameter(names = "--shardTable")
     String shardTable = "shard";
-    @Parameter(names="--doc2Term")
+    @Parameter(names = "--doc2Term")
     String doc2TermTable = "doc2Term";
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(Reverse.class.getName(), args, scanOpts, bwOpts);
-    
+
     Connector conn = opts.getConnector();
-    
+
     Scanner scanner = conn.createScanner(opts.shardTable, opts.auths);
     scanner.setBatchSize(scanOpts.scanBatchSize);
     BatchWriter bw = conn.createBatchWriter(opts.doc2TermTable, bwOpts.getBatchWriterConfig());
-    
+
     for (Entry<Key,Value> entry : scanner) {
       Key key = entry.getKey();
       Mutation m = new Mutation(key.getColumnQualifier());
       m.put(key.getColumnFamily(), new Text(), new Value(new byte[0]));
       bw.addMutation(m);
     }
-    
+
     bw.close();
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/DebugCommand.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/DebugCommand.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/DebugCommand.java
index 6960898..728aa08 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/DebugCommand.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/DebugCommand.java
@@ -39,5 +39,5 @@ public class DebugCommand extends Command {
   public int numArgs() {
     return 0;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/MyAppShellExtension.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/MyAppShellExtension.java b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/MyAppShellExtension.java
index 1ba5ad3..e37acbd 100644
--- a/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/MyAppShellExtension.java
+++ b/examples/simple/src/main/java/org/apache/accumulo/examples/simple/shell/MyAppShellExtension.java
@@ -16,18 +16,18 @@
  */
 package org.apache.accumulo.examples.simple.shell;
 
-import org.apache.accumulo.shell.ShellExtension;
 import org.apache.accumulo.shell.Shell.Command;
+import org.apache.accumulo.shell.ShellExtension;
 
 public class MyAppShellExtension extends ShellExtension {
-  
+
   public String getExtensionName() {
     return "MyApp";
   }
-  
+
   @Override
   public Command[] getCommands() {
-    return new Command[] { new DebugCommand() };
+    return new Command[] {new DebugCommand()};
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/test/java/org/apache/accumulo/examples/simple/dirlist/CountTest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/dirlist/CountTest.java b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/dirlist/CountTest.java
index 0addc77..72ff442 100644
--- a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/dirlist/CountTest.java
+++ b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/dirlist/CountTest.java
@@ -40,9 +40,9 @@ import org.apache.hadoop.io.Text;
 import org.apache.log4j.Logger;
 
 public class CountTest extends TestCase {
-  
+
   private static final Logger log = Logger.getLogger(CountTest.class);
-  
+
   {
     try {
       Connector conn = new MockInstance("counttest").getConnector("root", new PasswordToken(""));
@@ -64,12 +64,12 @@ public class CountTest extends TestCase {
       log.error("Could not add mutations in initializer.", e);
     }
   }
-  
+
   public void test() throws Exception {
     Scanner scanner = new MockInstance("counttest").getConnector("root", new PasswordToken("")).createScanner("dirlisttable", new Authorizations());
     scanner.fetchColumn(new Text("dir"), new Text("counts"));
     assertFalse(scanner.iterator().hasNext());
-    
+
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
@@ -80,13 +80,13 @@ public class CountTest extends TestCase {
     opts.password = new Opts.Password("");
     FileCount fc = new FileCount(opts, scanOpts, bwOpts);
     fc.run();
-    
+
     ArrayList<Pair<String,String>> expected = new ArrayList<Pair<String,String>>();
     expected.add(new Pair<String,String>(QueryUtil.getRow("").toString(), "1,0,3,3"));
     expected.add(new Pair<String,String>(QueryUtil.getRow("/local").toString(), "2,1,2,3"));
     expected.add(new Pair<String,String>(QueryUtil.getRow("/local/user1").toString(), "0,2,0,2"));
     expected.add(new Pair<String,String>(QueryUtil.getRow("/local/user2").toString(), "0,0,0,0"));
-    
+
     int i = 0;
     for (Entry<Key,Value> e : scanner) {
       assertEquals(e.getKey().getRow().toString(), expected.get(i).getFirst());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkCombinerTest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkCombinerTest.java b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkCombinerTest.java
index b380b2b..6d1467a 100644
--- a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkCombinerTest.java
+++ b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkCombinerTest.java
@@ -22,9 +22,11 @@ import java.util.Collections;
 import java.util.HashSet;
 import java.util.Iterator;
 import java.util.Map;
+import java.util.Map.Entry;
 import java.util.SortedMap;
 import java.util.TreeMap;
-import java.util.Map.Entry;
+
+import junit.framework.TestCase;
 
 import org.apache.accumulo.core.data.ByteSequence;
 import org.apache.accumulo.core.data.Key;
@@ -33,24 +35,20 @@ import org.apache.accumulo.core.data.Range;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.IteratorEnvironment;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
-import org.apache.accumulo.examples.simple.filedata.ChunkCombiner;
-import org.apache.accumulo.examples.simple.filedata.FileDataIngest;
-
-import junit.framework.TestCase;
 
 public class ChunkCombinerTest extends TestCase {
-  
+
   public static class MapIterator implements SortedKeyValueIterator<Key,Value> {
     private Iterator<Entry<Key,Value>> iter;
     private Entry<Key,Value> entry;
     Collection<ByteSequence> columnFamilies;
     private SortedMap<Key,Value> map;
     private Range range;
-    
+
     public MapIterator deepCopy(IteratorEnvironment env) {
       return new MapIterator(map);
     }
-    
+
     private MapIterator(SortedMap<Key,Value> map) {
       this.map = map;
       iter = map.entrySet().iterator();
@@ -60,22 +58,22 @@ public class ChunkCombinerTest extends TestCase {
       else
         entry = null;
     }
-    
+
     @Override
     public Key getTopKey() {
       return entry.getKey();
     }
-    
+
     @Override
     public Value getTopValue() {
       return entry.getValue();
     }
-    
+
     @Override
     public boolean hasTop() {
       return entry != null;
     }
-    
+
     @Override
     public void next() throws IOException {
       entry = null;
@@ -90,7 +88,7 @@ public class ChunkCombinerTest extends TestCase {
         break;
       }
     }
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
       if (!inclusive) {
@@ -98,93 +96,93 @@ public class ChunkCombinerTest extends TestCase {
       }
       this.columnFamilies = columnFamilies;
       this.range = range;
-      
+
       Key key = range.getStartKey();
       if (key == null) {
         key = new Key();
       }
-      
+
       iter = map.tailMap(key).entrySet().iterator();
       next();
       while (hasTop() && range.beforeStartKey(getTopKey())) {
         next();
       }
     }
-    
+
     public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
       throw new UnsupportedOperationException();
     }
   }
-  
+
   private TreeMap<Key,Value> row1;
   private TreeMap<Key,Value> row2;
   private TreeMap<Key,Value> row3;
   private TreeMap<Key,Value> allRows;
-  
+
   private TreeMap<Key,Value> cRow1;
   private TreeMap<Key,Value> cRow2;
   private TreeMap<Key,Value> cRow3;
   private TreeMap<Key,Value> allCRows;
-  
+
   private TreeMap<Key,Value> cOnlyRow1;
   private TreeMap<Key,Value> cOnlyRow2;
   private TreeMap<Key,Value> cOnlyRow3;
   private TreeMap<Key,Value> allCOnlyRows;
-  
+
   private TreeMap<Key,Value> badrow;
-  
+
   @Override
   protected void setUp() {
     row1 = new TreeMap<Key,Value>();
     row2 = new TreeMap<Key,Value>();
     row3 = new TreeMap<Key,Value>();
     allRows = new TreeMap<Key,Value>();
-    
+
     cRow1 = new TreeMap<Key,Value>();
     cRow2 = new TreeMap<Key,Value>();
     cRow3 = new TreeMap<Key,Value>();
     allCRows = new TreeMap<Key,Value>();
-    
+
     cOnlyRow1 = new TreeMap<Key,Value>();
     cOnlyRow2 = new TreeMap<Key,Value>();
     cOnlyRow3 = new TreeMap<Key,Value>();
     allCOnlyRows = new TreeMap<Key,Value>();
-    
+
     badrow = new TreeMap<Key,Value>();
-    
+
     String refs = FileDataIngest.REFS_CF.toString();
     String fileext = FileDataIngest.REFS_FILE_EXT;
     String filename = FileDataIngest.REFS_ORIG_FILE;
     String chunk_cf = FileDataIngest.CHUNK_CF.toString();
-    
+
     row1.put(new Key("row1", refs, "hash1\0" + fileext, "C"), new Value("jpg".getBytes()));
     row1.put(new Key("row1", refs, "hash1\0" + filename, "D"), new Value("foo1.jpg".getBytes()));
     row1.put(new Key("row1", chunk_cf, "0000", "A"), new Value("V1".getBytes()));
     row1.put(new Key("row1", chunk_cf, "0000", "B"), new Value("V1".getBytes()));
     row1.put(new Key("row1", chunk_cf, "0001", "A"), new Value("V2".getBytes()));
     row1.put(new Key("row1", chunk_cf, "0001", "B"), new Value("V2".getBytes()));
-    
+
     cRow1.put(new Key("row1", refs, "hash1\0" + fileext, "C"), new Value("jpg".getBytes()));
     cRow1.put(new Key("row1", refs, "hash1\0" + filename, "D"), new Value("foo1.jpg".getBytes()));
     cRow1.put(new Key("row1", chunk_cf, "0000", "(C)|(D)"), new Value("V1".getBytes()));
     cRow1.put(new Key("row1", chunk_cf, "0001", "(C)|(D)"), new Value("V2".getBytes()));
-    
+
     cOnlyRow1.put(new Key("row1", chunk_cf, "0000", "(C)|(D)"), new Value("V1".getBytes()));
     cOnlyRow1.put(new Key("row1", chunk_cf, "0001", "(C)|(D)"), new Value("V2".getBytes()));
-    
+
     row2.put(new Key("row2", refs, "hash1\0" + fileext, "A"), new Value("jpg".getBytes()));
     row2.put(new Key("row2", refs, "hash1\0" + filename, "B"), new Value("foo1.jpg".getBytes()));
     row2.put(new Key("row2", chunk_cf, "0000", "A|B"), new Value("V1".getBytes()));
     row2.put(new Key("row2", chunk_cf, "0000", "A"), new Value("V1".getBytes()));
     row2.put(new Key("row2", chunk_cf, "0000", "(A)|(B)"), new Value("V1".getBytes()));
     row2.put(new Key("row2a", chunk_cf, "0000", "C"), new Value("V1".getBytes()));
-    
+
     cRow2.put(new Key("row2", refs, "hash1\0" + fileext, "A"), new Value("jpg".getBytes()));
     cRow2.put(new Key("row2", refs, "hash1\0" + filename, "B"), new Value("foo1.jpg".getBytes()));
     cRow2.put(new Key("row2", chunk_cf, "0000", "(A)|(B)"), new Value("V1".getBytes()));
-    
+
     cOnlyRow2.put(new Key("row2", chunk_cf, "0000", "(A)|(B)"), new Value("V1".getBytes()));
-    
+
     row3.put(new Key("row3", refs, "hash1\0w", "(A&B)|(C&(D|E))"), new Value("".getBytes()));
     row3.put(new Key("row3", refs, "hash1\0x", "A&B"), new Value("".getBytes()));
     row3.put(new Key("row3", refs, "hash1\0y", "(A&B)"), new Value("".getBytes()));
@@ -193,39 +191,39 @@ public class ChunkCombinerTest extends TestCase {
     row3.put(new Key("row3", chunk_cf, "0000", "A&B", 20), new Value("V1".getBytes()));
     row3.put(new Key("row3", chunk_cf, "0000", "(A&B)", 10), new Value("V1".getBytes()));
     row3.put(new Key("row3", chunk_cf, "0000", "(F|G)&(D|E)", 10), new Value("V1".getBytes()));
-    
+
     cRow3.put(new Key("row3", refs, "hash1\0w", "(A&B)|(C&(D|E))"), new Value("".getBytes()));
     cRow3.put(new Key("row3", refs, "hash1\0x", "A&B"), new Value("".getBytes()));
     cRow3.put(new Key("row3", refs, "hash1\0y", "(A&B)"), new Value("".getBytes()));
     cRow3.put(new Key("row3", refs, "hash1\0z", "(F|G)&(D|E)"), new Value("".getBytes()));
     cRow3.put(new Key("row3", chunk_cf, "0000", "((F|G)&(D|E))|(A&B)|(C&(D|E))", 20), new Value("V1".getBytes()));
-    
+
     cOnlyRow3.put(new Key("row3", chunk_cf, "0000", "((F|G)&(D|E))|(A&B)|(C&(D|E))", 20), new Value("V1".getBytes()));
-    
+
     badrow.put(new Key("row1", chunk_cf, "0000", "A"), new Value("V1".getBytes()));
     badrow.put(new Key("row1", chunk_cf, "0000", "B"), new Value("V2".getBytes()));
-    
+
     allRows.putAll(row1);
     allRows.putAll(row2);
     allRows.putAll(row3);
-    
+
     allCRows.putAll(cRow1);
     allCRows.putAll(cRow2);
     allCRows.putAll(cRow3);
-    
+
     allCOnlyRows.putAll(cOnlyRow1);
     allCOnlyRows.putAll(cOnlyRow2);
     allCOnlyRows.putAll(cOnlyRow3);
   }
-  
+
   private static final Collection<ByteSequence> emptyColfs = new HashSet<ByteSequence>();
-  
+
   public void test1() throws IOException {
     runTest(false, allRows, allCRows, emptyColfs);
     runTest(true, allRows, allCRows, emptyColfs);
     runTest(false, allRows, allCOnlyRows, Collections.singleton(FileDataIngest.CHUNK_CF_BS));
     runTest(true, allRows, allCOnlyRows, Collections.singleton(FileDataIngest.CHUNK_CF_BS));
-    
+
     try {
       runTest(true, badrow, null, emptyColfs);
       assertNotNull(null);
@@ -233,26 +231,26 @@ public class ChunkCombinerTest extends TestCase {
       assertNull(null);
     }
   }
-  
+
   private void runTest(boolean reseek, TreeMap<Key,Value> source, TreeMap<Key,Value> result, Collection<ByteSequence> cols) throws IOException {
     MapIterator src = new MapIterator(source);
     SortedKeyValueIterator<Key,Value> iter = new ChunkCombiner();
     iter.init(src, null, null);
     iter = iter.deepCopy(null);
     iter.seek(new Range(), cols, true);
-    
+
     TreeMap<Key,Value> seen = new TreeMap<Key,Value>();
-    
+
     while (iter.hasTop()) {
       assertFalse("already contains " + iter.getTopKey(), seen.containsKey(iter.getTopKey()));
       seen.put(new Key(iter.getTopKey()), new Value(iter.getTopValue()));
-      
+
       if (reseek)
         iter.seek(new Range(iter.getTopKey().followingKey(PartialKey.ROW_COLFAM_COLQUAL), true, null, true), cols, true);
       else
         iter.next();
     }
-    
+
     assertEquals(result, seen);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStreamTest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStreamTest.java b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStreamTest.java
index 54279e4..68af6d8 100644
--- a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStreamTest.java
+++ b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/ChunkInputStreamTest.java
@@ -48,7 +48,7 @@ public class ChunkInputStreamTest extends TestCase {
   List<Entry<Key,Value>> data;
   List<Entry<Key,Value>> baddata;
   List<Entry<Key,Value>> multidata;
-  
+
   {
     data = new ArrayList<Entry<Key,Value>>();
     addData(data, "a", "refs", "id\0ext", "A&B", "ext");
@@ -92,17 +92,17 @@ public class ChunkInputStreamTest extends TestCase {
     addData(multidata, "c", "~chunk", 100, 0, "A&B", "asdfjkl;");
     addData(multidata, "c", "~chunk", 100, 1, "B&C", "");
   }
-  
+
   public static void addData(List<Entry<Key,Value>> data, String row, String cf, String cq, String vis, String value) {
     data.add(new KeyValue(new Key(new Text(row), new Text(cf), new Text(cq), new Text(vis)), value.getBytes()));
   }
-  
+
   public static void addData(List<Entry<Key,Value>> data, String row, String cf, int chunkSize, int chunkCount, String vis, String value) {
     Text chunkCQ = new Text(FileDataIngest.intToBytes(chunkSize));
     chunkCQ.append(FileDataIngest.intToBytes(chunkCount), 0, 4);
     data.add(new KeyValue(new Key(new Text(row), new Text(cf), chunkCQ, new Text(vis)), value.getBytes()));
   }
-  
+
   public void testExceptionOnMultipleSetSourceWithoutClose() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(data.iterator());
@@ -116,11 +116,11 @@ public class ChunkInputStreamTest extends TestCase {
     }
     cis.close();
   }
-  
+
   public void testExceptionOnGetVisBeforeClose() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(data.iterator());
-    
+
     cis.setSource(pi);
     try {
       cis.getVisibilities();
@@ -131,13 +131,13 @@ public class ChunkInputStreamTest extends TestCase {
     cis.close();
     cis.getVisibilities();
   }
-  
+
   public void testReadIntoBufferSmallerThanChunks() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     byte[] b = new byte[5];
-    
+
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(data.iterator());
-    
+
     cis.setSource(pi);
     int read;
     assertEquals(read = cis.read(b), 5);
@@ -145,7 +145,7 @@ public class ChunkInputStreamTest extends TestCase {
     assertEquals(read = cis.read(b), 3);
     assertEquals(new String(b, 0, read), "kl;");
     assertEquals(read = cis.read(b), -1);
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 5);
     assertEquals(new String(b, 0, read), "qwert");
@@ -154,7 +154,7 @@ public class ChunkInputStreamTest extends TestCase {
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[A&B, B&C, D]");
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 5);
     assertEquals(new String(b, 0, read), "asdfj");
@@ -167,11 +167,11 @@ public class ChunkInputStreamTest extends TestCase {
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[A&B]");
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), -1);
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 5);
     assertEquals(new String(b, 0, read), "asdfj");
@@ -179,53 +179,53 @@ public class ChunkInputStreamTest extends TestCase {
     assertEquals(new String(b, 0, read), "kl;");
     assertEquals(read = cis.read(b), -1);
     cis.close();
-    
+
     assertFalse(pi.hasNext());
   }
-  
+
   public void testReadIntoBufferLargerThanChunks() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     byte[] b = new byte[20];
     int read;
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(data.iterator());
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 10);
     assertEquals(new String(b, 0, read), "qwertyuiop");
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[A&B, B&C, D]");
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 16);
     assertEquals(new String(b, 0, read), "asdfjkl;asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[A&B]");
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), -1);
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     cis.close();
-    
+
     assertFalse(pi.hasNext());
   }
-  
+
   public void testWithAccumulo() throws AccumuloException, AccumuloSecurityException, TableExistsException, TableNotFoundException, IOException {
     Connector conn = new MockInstance().getConnector("root", new PasswordToken(""));
     conn.tableOperations().create("test");
     BatchWriter bw = conn.createBatchWriter("test", new BatchWriterConfig());
-    
+
     for (Entry<Key,Value> e : data) {
       Key k = e.getKey();
       Mutation m = new Mutation(k.getRow());
@@ -233,46 +233,46 @@ public class ChunkInputStreamTest extends TestCase {
       bw.addMutation(m);
     }
     bw.close();
-    
+
     Scanner scan = conn.createScanner("test", new Authorizations("A", "B", "C", "D"));
-    
+
     ChunkInputStream cis = new ChunkInputStream();
     byte[] b = new byte[20];
     int read;
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(scan.iterator());
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 10);
     assertEquals(new String(b, 0, read), "qwertyuiop");
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[A&B, B&C, D]");
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 16);
     assertEquals(new String(b, 0, read), "asdfjkl;asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[A&B]");
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), -1);
     cis.close();
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     cis.close();
-    
+
     assertFalse(pi.hasNext());
   }
-  
+
   private static void assumeExceptionOnRead(ChunkInputStream cis, byte[] b) {
     try {
       cis.read(b);
@@ -282,7 +282,7 @@ public class ChunkInputStreamTest extends TestCase {
       assertNull(null);
     }
   }
-  
+
   private static void assumeExceptionOnClose(ChunkInputStream cis) {
     try {
       cis.close();
@@ -292,41 +292,41 @@ public class ChunkInputStreamTest extends TestCase {
       assertNull(null);
     }
   }
-  
+
   public void testBadData() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     byte[] b = new byte[20];
     int read;
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(baddata.iterator());
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assumeExceptionOnClose(cis);
     // can still get visibilities after exception -- bad?
     assertEquals(cis.getVisibilities().toString(), "[A]");
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assumeExceptionOnClose(cis);
     assertEquals(cis.getVisibilities().toString(), "[B, C]");
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assumeExceptionOnClose(cis);
     assertEquals(cis.getVisibilities().toString(), "[D, E]");
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[F, G]");
     cis.close();
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     cis.close();
     assertEquals(cis.getVisibilities().toString(), "[I, J]");
-    
+
     try {
       cis.setSource(pi);
       assertNotNull(null);
@@ -335,49 +335,49 @@ public class ChunkInputStreamTest extends TestCase {
     }
     assumeExceptionOnClose(cis);
     assertEquals(cis.getVisibilities().toString(), "[K]");
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[L]");
     cis.close();
-    
+
     assertFalse(pi.hasNext());
-    
+
     pi = new PeekingIterator<Entry<Key,Value>>(baddata.iterator());
     cis.setSource(pi);
     assumeExceptionOnClose(cis);
   }
-  
+
   public void testBadDataWithoutClosing() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     byte[] b = new byte[20];
     int read;
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(baddata.iterator());
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     // can still get visibilities after exception -- bad?
     assertEquals(cis.getVisibilities().toString(), "[A]");
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assertEquals(cis.getVisibilities().toString(), "[B, C]");
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assertEquals(cis.getVisibilities().toString(), "[D, E]");
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[F, G]");
     cis.close();
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assertEquals(cis.getVisibilities().toString(), "[I, J]");
-    
+
     try {
       cis.setSource(pi);
       assertNotNull(null);
@@ -385,51 +385,51 @@ public class ChunkInputStreamTest extends TestCase {
       assertNull(null);
     }
     assertEquals(cis.getVisibilities().toString(), "[K]");
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), -1);
     assertEquals(cis.getVisibilities().toString(), "[L]");
     cis.close();
-    
+
     assertFalse(pi.hasNext());
-    
+
     pi = new PeekingIterator<Entry<Key,Value>>(baddata.iterator());
     cis.setSource(pi);
     assumeExceptionOnClose(cis);
   }
-  
+
   public void testMultipleChunkSizes() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     byte[] b = new byte[20];
     int read;
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(multidata.iterator());
-    
+
     b = new byte[20];
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(read = cis.read(b), -1);
     cis.close();
     assertEquals(cis.getVisibilities().toString(), "[A&B]");
-    
+
     cis.setSource(pi);
     assumeExceptionOnRead(cis, b);
     assertEquals(cis.getVisibilities().toString(), "[A&B]");
-    
+
     cis.setSource(pi);
     assertEquals(read = cis.read(b), 8);
     assertEquals(new String(b, 0, read), "asdfjkl;");
     assertEquals(read = cis.read(b), -1);
     cis.close();
     assertEquals(cis.getVisibilities().toString(), "[A&B, B&C]");
-    
+
     assertFalse(pi.hasNext());
   }
-  
+
   public void testSingleByteRead() throws IOException {
     ChunkInputStream cis = new ChunkInputStream();
     PeekingIterator<Entry<Key,Value>> pi = new PeekingIterator<Entry<Key,Value>>(data.iterator());
-    
+
     cis.setSource(pi);
     assertEquals((byte) 'a', (byte) cis.read());
     assertEquals((byte) 's', (byte) cis.read());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/KeyUtilTest.java
----------------------------------------------------------------------
diff --git a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/KeyUtilTest.java b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/KeyUtilTest.java
index 8307069..e93331a 100644
--- a/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/KeyUtilTest.java
+++ b/examples/simple/src/test/java/org/apache/accumulo/examples/simple/filedata/KeyUtilTest.java
@@ -18,14 +18,13 @@ package org.apache.accumulo.examples.simple.filedata;
 
 import junit.framework.TestCase;
 
-import org.apache.accumulo.examples.simple.filedata.KeyUtil;
 import org.apache.hadoop.io.Text;
 
 public class KeyUtilTest extends TestCase {
   public static void checkSeps(String... s) {
     Text t = KeyUtil.buildNullSepText(s);
     String[] rets = KeyUtil.splitNullSepText(t);
-    
+
     int length = 0;
     for (String str : s)
       length += str.length();
@@ -34,7 +33,7 @@ public class KeyUtilTest extends TestCase {
     for (int i = 0; i < s.length; i++)
       assertEquals(s[i], rets[i]);
   }
-  
+
   public void testNullSep() {
     checkSeps("abc", "d", "", "efgh");
     checkSeps("ab", "");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/fate/src/main/java/org/apache/accumulo/fate/AdminUtil.java
----------------------------------------------------------------------
diff --git a/fate/src/main/java/org/apache/accumulo/fate/AdminUtil.java b/fate/src/main/java/org/apache/accumulo/fate/AdminUtil.java
index 60ff4b4..c12efb5 100644
--- a/fate/src/main/java/org/apache/accumulo/fate/AdminUtil.java
+++ b/fate/src/main/java/org/apache/accumulo/fate/AdminUtil.java
@@ -40,19 +40,19 @@ import org.apache.zookeeper.KeeperException;
  */
 public class AdminUtil<T> {
   private static final Logger log = Logger.getLogger(AdminUtil.class);
-  
+
   private boolean exitOnError = false;
-  
+
   /**
    * Default constructor
    */
   public AdminUtil() {
     this(true);
   }
-  
+
   /**
    * Constructor
-   * 
+   *
    * @param exitOnError
    *          <code>System.exit(1)</code> on error if true
    */
@@ -60,37 +60,37 @@ public class AdminUtil<T> {
     super();
     this.exitOnError = exitOnError;
   }
-  
+
   public void print(ReadOnlyTStore<T> zs, IZooReaderWriter zk, String lockPath) throws KeeperException, InterruptedException {
     print(zs, zk, lockPath, new Formatter(System.out), null, null);
   }
-  
+
   public void print(ReadOnlyTStore<T> zs, IZooReaderWriter zk, String lockPath, Formatter fmt, Set<Long> filterTxid, EnumSet<TStatus> filterStatus)
       throws KeeperException, InterruptedException {
     Map<Long,List<String>> heldLocks = new HashMap<Long,List<String>>();
     Map<Long,List<String>> waitingLocks = new HashMap<Long,List<String>>();
-    
+
     List<String> lockedIds = zk.getChildren(lockPath);
-    
+
     for (String id : lockedIds) {
       try {
         List<String> lockNodes = zk.getChildren(lockPath + "/" + id);
         lockNodes = new ArrayList<String>(lockNodes);
         Collections.sort(lockNodes);
-        
+
         int pos = 0;
         boolean sawWriteLock = false;
-        
+
         for (String node : lockNodes) {
           try {
             byte[] data = zk.getData(lockPath + "/" + id + "/" + node, null);
             String lda[] = new String(data, UTF_8).split(":");
-            
+
             if (lda[0].charAt(0) == 'W')
               sawWriteLock = true;
-            
+
             Map<Long,List<String>> locks;
-            
+
             if (pos == 0) {
               locks = heldLocks;
             } else {
@@ -100,77 +100,77 @@ public class AdminUtil<T> {
                 locks = waitingLocks;
               }
             }
-            
+
             List<String> tables = locks.get(Long.parseLong(lda[1], 16));
             if (tables == null) {
               tables = new ArrayList<String>();
               locks.put(Long.parseLong(lda[1], 16), tables);
             }
-            
+
             tables.add(lda[0].charAt(0) + ":" + id);
-            
+
           } catch (Exception e) {
             log.error(e);
           }
           pos++;
         }
-        
+
       } catch (Exception e) {
-        log.error("Failed to read locks for "+id+" continuing.", e);
+        log.error("Failed to read locks for " + id + " continuing.", e);
         fmt.format("Failed to read locks for %s continuing", id);
       }
     }
-    
+
     List<Long> transactions = zs.list();
-    
+
     long txCount = 0;
     for (Long tid : transactions) {
-      
+
       zs.reserve(tid);
-      
+
       String debug = (String) zs.getProperty(tid, "debug");
-      
+
       List<String> hlocks = heldLocks.remove(tid);
       if (hlocks == null)
         hlocks = Collections.emptyList();
-      
+
       List<String> wlocks = waitingLocks.remove(tid);
       if (wlocks == null)
         wlocks = Collections.emptyList();
-      
+
       String top = null;
       ReadOnlyRepo<T> repo = zs.top(tid);
       if (repo != null)
         top = repo.getDescription();
-      
+
       TStatus status = null;
       status = zs.getStatus(tid);
-      
+
       zs.unreserve(tid, 0);
-      
+
       if ((filterTxid != null && !filterTxid.contains(tid)) || (filterStatus != null && !filterStatus.contains(status)))
         continue;
-      
+
       ++txCount;
       fmt.format("txid: %016x  status: %-18s  op: %-15s  locked: %-15s locking: %-15s top: %s%n", tid, status, debug, hlocks, wlocks, top);
     }
     fmt.format(" %s transactions", txCount);
-    
+
     if (heldLocks.size() != 0 || waitingLocks.size() != 0) {
       fmt.format("%nThe following locks did not have an associated FATE operation%n");
       for (Entry<Long,List<String>> entry : heldLocks.entrySet())
         fmt.format("txid: %016x  locked: %s%n", entry.getKey(), entry.getValue());
-      
+
       for (Entry<Long,List<String>> entry : waitingLocks.entrySet())
         fmt.format("txid: %016x  locking: %s%n", entry.getKey(), entry.getValue());
     }
   }
-  
+
   public boolean prepDelete(TStore<T> zs, IZooReaderWriter zk, String path, String txidStr) {
     if (!checkGlobalLock(zk, path)) {
       return false;
     }
-    
+
     long txid;
     try {
       txid = Long.parseLong(txidStr, 16);
@@ -185,7 +185,7 @@ public class AdminUtil<T> {
       case UNKNOWN:
         System.out.printf("Invalid transaction ID: %016x%n", txid);
         break;
-      
+
       case IN_PROGRESS:
       case NEW:
       case FAILED:
@@ -196,16 +196,16 @@ public class AdminUtil<T> {
         state = true;
         break;
     }
-    
+
     zs.unreserve(txid, 0);
     return state;
   }
-  
+
   public boolean prepFail(TStore<T> zs, IZooReaderWriter zk, String path, String txidStr) {
     if (!checkGlobalLock(zk, path)) {
       return false;
     }
-    
+
     long txid;
     try {
       txid = Long.parseLong(txidStr, 16);
@@ -220,33 +220,33 @@ public class AdminUtil<T> {
       case UNKNOWN:
         System.out.printf("Invalid transaction ID: %016x%n", txid);
         break;
-      
+
       case IN_PROGRESS:
       case NEW:
         System.out.printf("Failing transaction: %016x (%s)%n", txid, ts);
         zs.setStatus(txid, TStatus.FAILED_IN_PROGRESS);
         state = true;
         break;
-      
+
       case SUCCESSFUL:
         System.out.printf("Transaction already completed: %016x (%s)%n", txid, ts);
         break;
-      
+
       case FAILED:
       case FAILED_IN_PROGRESS:
         System.out.printf("Transaction already failed: %016x (%s)%n", txid, ts);
         state = true;
         break;
     }
-    
+
     zs.unreserve(txid, 0);
     return state;
   }
-  
+
   public void deleteLocks(TStore<T> zs, IZooReaderWriter zk, String path, String txidStr) throws KeeperException, InterruptedException {
     // delete any locks assoc w/ fate operation
     List<String> lockedIds = zk.getChildren(path);
-    
+
     for (String id : lockedIds) {
       List<String> lockNodes = zk.getChildren(path + "/" + id);
       for (String node : lockNodes) {
@@ -258,7 +258,7 @@ public class AdminUtil<T> {
       }
     }
   }
-  
+
   public boolean checkGlobalLock(IZooReaderWriter zk, String path) {
     try {
       if (ZooLock.getLockData(zk.getZooKeeper(), path) != null) {


[45/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocatorImpl.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocatorImpl.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocatorImpl.java
index 1d49647..659d877 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocatorImpl.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletLocatorImpl.java
@@ -52,23 +52,23 @@ import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 
 public class TabletLocatorImpl extends TabletLocator {
-  
+
   private static final Logger log = Logger.getLogger(TabletLocatorImpl.class);
-  
+
   // there seems to be a bug in TreeMap.tailMap related to
   // putting null in the treemap.. therefore instead of
   // putting null, put MAX_TEXT
   static final Text MAX_TEXT = new Text();
-  
+
   private static class EndRowComparator implements Comparator<Text>, Serializable {
-    
+
     private static final long serialVersionUID = 1L;
 
     @Override
     public int compare(Text o1, Text o2) {
-      
+
       int ret;
-      
+
       if (o1 == MAX_TEXT)
         if (o2 == MAX_TEXT)
           ret = 0;
@@ -78,21 +78,21 @@ public class TabletLocatorImpl extends TabletLocator {
         ret = -1;
       else
         ret = o1.compareTo(o2);
-      
+
       return ret;
     }
-    
+
   }
-  
+
   static final EndRowComparator endRowComparator = new EndRowComparator();
-  
+
   protected Text tableId;
   protected TabletLocator parent;
   protected TreeMap<Text,TabletLocation> metaCache = new TreeMap<Text,TabletLocation>(endRowComparator);
   protected TabletLocationObtainer locationObtainer;
   private TabletServerLockChecker lockChecker;
   protected Text lastTabletRow;
-  
+
   private TreeSet<KeyExtent> badExtents = new TreeSet<KeyExtent>();
   private ReentrantReadWriteLock rwLock = new ReentrantReadWriteLock();
   private final Lock rLock = rwLock.readLock();
@@ -104,11 +104,11 @@ public class TabletLocatorImpl extends TabletLocator {
      */
     TabletLocations lookupTablet(ClientContext context, TabletLocation src, Text row, Text stopRow, TabletLocator parent) throws AccumuloSecurityException,
         AccumuloException;
-    
+
     List<TabletLocation> lookupTablets(ClientContext context, String tserver, Map<KeyExtent,List<Range>> map, TabletLocator parent)
         throws AccumuloSecurityException, AccumuloException;
   }
-  
+
   public static interface TabletServerLockChecker {
     boolean isLockHeld(String tserver, String session);
 
@@ -116,36 +116,36 @@ public class TabletLocatorImpl extends TabletLocator {
   }
 
   private class LockCheckerSession {
-    
+
     private HashSet<Pair<String,String>> okLocks = new HashSet<Pair<String,String>>();
     private HashSet<Pair<String,String>> invalidLocks = new HashSet<Pair<String,String>>();
-    
+
     private TabletLocation checkLock(TabletLocation tl) {
       // the goal of this class is to minimize calls out to lockChecker under that assumption that its a resource synchronized among many threads... want to
       // avoid fine grained synchronization when binning lots of mutations or ranges... remember decisions from the lockChecker in thread local unsynchronized
       // memory
-      
+
       if (tl == null)
         return null;
 
       Pair<String,String> lock = new Pair<String,String>(tl.tablet_location, tl.tablet_session);
-      
+
       if (okLocks.contains(lock))
         return tl;
-      
+
       if (invalidLocks.contains(lock))
         return null;
-      
+
       if (lockChecker.isLockHeld(tl.tablet_location, tl.tablet_session)) {
         okLocks.add(lock);
         return tl;
       }
-      
+
       if (log.isTraceEnabled())
         log.trace("Tablet server " + tl.tablet_location + " " + tl.tablet_session + " no longer holds its lock");
-      
+
       invalidLocks.add(lock);
-      
+
       return null;
     }
   }
@@ -155,34 +155,34 @@ public class TabletLocatorImpl extends TabletLocator {
     this.parent = parent;
     this.locationObtainer = tlo;
     this.lockChecker = tslc;
-    
+
     this.lastTabletRow = new Text(tableId);
     lastTabletRow.append(new byte[] {'<'}, 0, 1);
   }
-  
+
   @Override
   public <T extends Mutation> void binMutations(ClientContext context, List<T> mutations, Map<String,TabletServerMutations<T>> binnedMutations, List<T> failures)
       throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
-    
+
     OpTimer opTimer = null;
     if (log.isTraceEnabled())
       opTimer = new OpTimer(log, Level.TRACE).start("Binning " + mutations.size() + " mutations for table " + tableId);
-    
+
     ArrayList<T> notInCache = new ArrayList<T>();
     Text row = new Text();
-    
+
     LockCheckerSession lcSession = new LockCheckerSession();
 
     rLock.lock();
     try {
       processInvalidated(context, lcSession);
-      
+
       // for this to be efficient rows need to be in sorted order, but always sorting is slow... therefore only sort the
       // stuff not in the cache.... it is most efficient to pass _locateTablet rows in sorted order
-      
+
       // For this to be efficient, need to avoid fine grained synchronization and fine grained logging.
       // Therefore methods called by this are not synchronized and should not log.
-      
+
       for (T mutation : mutations) {
         row.set(mutation.getRow());
         TabletLocation tl = locateTabletInCache(row);
@@ -192,7 +192,7 @@ public class TabletLocatorImpl extends TabletLocator {
     } finally {
       rLock.unlock();
     }
-    
+
     if (notInCache.size() > 0) {
       Collections.sort(notInCache, new Comparator<Mutation>() {
         @Override
@@ -200,7 +200,7 @@ public class TabletLocatorImpl extends TabletLocator {
           return WritableComparator.compareBytes(o1.getRow(), 0, o1.getRow().length, o2.getRow(), 0, o2.getRow().length);
         }
       });
-      
+
       wLock.lock();
       try {
         boolean failed = false;
@@ -211,11 +211,11 @@ public class TabletLocatorImpl extends TabletLocator {
             failures.add(mutation);
             continue;
           }
-          
+
           row.set(mutation.getRow());
-          
+
           TabletLocation tl = _locateTablet(context, row, false, false, false, lcSession);
-          
+
           if (tl == null || !addMutation(binnedMutations, mutation, tl, lcSession)) {
             failures.add(mutation);
             failed = true;
@@ -233,7 +233,7 @@ public class TabletLocatorImpl extends TabletLocator {
   private <T extends Mutation> boolean addMutation(Map<String,TabletServerMutations<T>> binnedMutations, T mutation, TabletLocation tl,
       LockCheckerSession lcSession) {
     TabletServerMutations<T> tsm = binnedMutations.get(tl.tablet_location);
-    
+
     if (tsm == null) {
       // do lock check once per tserver here to make binning faster
       boolean lockHeld = lcSession.checkLock(tl) != null;
@@ -244,50 +244,50 @@ public class TabletLocatorImpl extends TabletLocator {
         return false;
       }
     }
-    
+
     // its possible the same tserver could be listed with different sessions
     if (tsm.getSession().equals(tl.tablet_session)) {
       tsm.addMutation(tl.tablet_extent, mutation);
       return true;
     }
-    
+
     return false;
   }
-  
+
   private List<Range> binRanges(ClientContext context, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges, boolean useCache,
       LockCheckerSession lcSession) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     List<Range> failures = new ArrayList<Range>();
     List<TabletLocation> tabletLocations = new ArrayList<TabletLocation>();
-    
+
     boolean lookupFailed = false;
-    
+
     l1: for (Range range : ranges) {
-      
+
       tabletLocations.clear();
-      
+
       Text startRow;
-      
+
       if (range.getStartKey() != null) {
         startRow = range.getStartKey().getRow();
       } else
         startRow = new Text();
-      
+
       TabletLocation tl = null;
-      
+
       if (useCache)
         tl = lcSession.checkLock(locateTabletInCache(startRow));
       else if (!lookupFailed)
         tl = _locateTablet(context, startRow, false, false, false, lcSession);
-      
+
       if (tl == null) {
         failures.add(range);
         if (!useCache)
           lookupFailed = true;
         continue;
       }
-      
+
       tabletLocations.add(tl);
-      
+
       while (tl.tablet_extent.getEndRow() != null && !range.afterEndKey(new Key(tl.tablet_extent.getEndRow()).followingKey(PartialKey.ROW))) {
         if (useCache) {
           Text row = new Text(tl.tablet_extent.getEndRow());
@@ -296,7 +296,7 @@ public class TabletLocatorImpl extends TabletLocator {
         } else {
           tl = _locateTablet(context, tl.tablet_extent.getEndRow(), true, false, false, lcSession);
         }
-        
+
         if (tl == null) {
           failures.add(range);
           if (!useCache)
@@ -309,46 +309,46 @@ public class TabletLocatorImpl extends TabletLocator {
       for (TabletLocation tl2 : tabletLocations) {
         TabletLocatorImpl.addRange(binnedRanges, tl2.tablet_location, tl2.tablet_extent, range);
       }
-      
+
     }
-    
+
     return failures;
   }
-  
+
   @Override
   public List<Range> binRanges(ClientContext context, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
-    
+
     /*
      * For this to be efficient, need to avoid fine grained synchronization and fine grained logging. Therefore methods called by this are not synchronized and
      * should not log.
      */
-    
+
     OpTimer opTimer = null;
     if (log.isTraceEnabled())
       opTimer = new OpTimer(log, Level.TRACE).start("Binning " + ranges.size() + " ranges for table " + tableId);
-    
+
     LockCheckerSession lcSession = new LockCheckerSession();
 
     List<Range> failures;
     rLock.lock();
     try {
       processInvalidated(context, lcSession);
-      
+
       // for this to be optimal, need to look ranges up in sorted order when
       // ranges are not present in cache... however do not want to always
       // sort ranges... therefore try binning ranges using only the cache
       // and sort whatever fails and retry
-      
+
       failures = binRanges(context, ranges, binnedRanges, true, lcSession);
     } finally {
       rLock.unlock();
     }
-    
+
     if (failures.size() > 0) {
       // sort failures by range start key
       Collections.sort(failures);
-      
+
       // try lookups again
       wLock.lock();
       try {
@@ -357,13 +357,13 @@ public class TabletLocatorImpl extends TabletLocator {
         wLock.unlock();
       }
     }
-    
+
     if (opTimer != null)
       opTimer.stop("Binned " + ranges.size() + " ranges for table " + tableId + " to " + binnedRanges.size() + " tservers in %DURATION%");
 
     return failures;
   }
-  
+
   @Override
   public void invalidateCache(KeyExtent failedExtent) {
     wLock.lock();
@@ -375,7 +375,7 @@ public class TabletLocatorImpl extends TabletLocator {
     if (log.isTraceEnabled())
       log.trace("Invalidated extent=" + failedExtent);
   }
-  
+
   @Override
   public void invalidateCache(Collection<KeyExtent> keySet) {
     wLock.lock();
@@ -387,11 +387,11 @@ public class TabletLocatorImpl extends TabletLocator {
     if (log.isTraceEnabled())
       log.trace("Invalidated " + keySet.size() + " cache entries for table " + tableId);
   }
-  
+
   @Override
   public void invalidateCache(Instance instance, String server) {
     int invalidatedCount = 0;
-    
+
     wLock.lock();
     try {
       for (TabletLocation cacheEntry : metaCache.values())
@@ -402,14 +402,14 @@ public class TabletLocatorImpl extends TabletLocator {
     } finally {
       wLock.unlock();
     }
-    
+
     lockChecker.invalidateCache(server);
 
     if (log.isTraceEnabled())
       log.trace("invalidated " + invalidatedCount + " cache entries  table=" + tableId + " server=" + server);
-    
+
   }
-  
+
   @Override
   public void invalidateCache() {
     int invalidatedCount;
@@ -423,18 +423,18 @@ public class TabletLocatorImpl extends TabletLocator {
     if (log.isTraceEnabled())
       log.trace("invalidated all " + invalidatedCount + " cache entries for table=" + tableId);
   }
-  
+
   @Override
   public TabletLocation locateTablet(ClientContext context, Text row, boolean skipRow, boolean retry) throws AccumuloException, AccumuloSecurityException,
       TableNotFoundException {
-    
+
     OpTimer opTimer = null;
     if (log.isTraceEnabled())
       opTimer = new OpTimer(log, Level.TRACE).start("Locating tablet  table=" + tableId + " row=" + TextUtil.truncate(row) + "  skipRow=" + skipRow + " retry="
           + retry);
-    
+
     while (true) {
-      
+
       LockCheckerSession lcSession = new LockCheckerSession();
       TabletLocation tl = _locateTablet(context, row, skipRow, retry, true, lcSession);
 
@@ -444,21 +444,21 @@ public class TabletLocatorImpl extends TabletLocator {
           log.trace("Failed to locate tablet containing row " + TextUtil.truncate(row) + " in table " + tableId + ", will retry...");
         continue;
       }
-      
+
       if (opTimer != null)
         opTimer.stop("Located tablet " + (tl == null ? null : tl.tablet_extent) + " at " + (tl == null ? null : tl.tablet_location) + " in %DURATION%");
-      
+
       return tl;
     }
   }
-  
+
   private void lookupTabletLocation(ClientContext context, Text row, boolean retry, LockCheckerSession lcSession) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
     Text metadataRow = new Text(tableId);
     metadataRow.append(new byte[] {';'}, 0, 1);
     metadataRow.append(row.getBytes(), 0, row.getLength());
     TabletLocation ptl = parent.locateTablet(context, metadataRow, false, retry);
-    
+
     if (ptl != null) {
       TabletLocations locations = locationObtainer.lookupTablet(context, ptl, metadataRow, lastTabletRow, parent);
       while (locations != null && locations.getLocations().isEmpty() && locations.getLocationless().isEmpty()) {
@@ -475,19 +475,19 @@ public class TabletLocatorImpl extends TabletLocator {
           break;
         }
       }
-      
+
       if (locations == null)
         return;
-      
+
       // cannot assume the list contains contiguous key extents... so it is probably
       // best to deal with each extent individually
-      
+
       Text lastEndRow = null;
       for (TabletLocation tabletLocation : locations.getLocations()) {
-        
+
         KeyExtent ke = tabletLocation.tablet_extent;
         TabletLocation locToCache;
-        
+
         // create new location if current prevEndRow == endRow
         if ((lastEndRow != null) && (ke.getPrevEndRow() != null) && ke.getPrevEndRow().equals(lastEndRow)) {
           locToCache = new TabletLocation(new KeyExtent(ke.getTableId(), ke.getEndRow(), lastEndRow), tabletLocation.tablet_location,
@@ -495,35 +495,35 @@ public class TabletLocatorImpl extends TabletLocator {
         } else {
           locToCache = tabletLocation;
         }
-        
+
         // save endRow for next iteration
         lastEndRow = locToCache.tablet_extent.getEndRow();
-        
+
         updateCache(locToCache, lcSession);
       }
     }
-    
+
   }
-  
+
   private void updateCache(TabletLocation tabletLocation, LockCheckerSession lcSession) {
     if (!tabletLocation.tablet_extent.getTableId().equals(tableId)) {
       // sanity check
       throw new IllegalStateException("Unexpected extent returned " + tableId + "  " + tabletLocation.tablet_extent);
     }
-    
+
     if (tabletLocation.tablet_location == null) {
       // sanity check
       throw new IllegalStateException("Cannot add null locations to cache " + tableId + "  " + tabletLocation.tablet_extent);
     }
-    
+
     if (!tabletLocation.tablet_extent.getTableId().equals(tableId)) {
       // sanity check
       throw new IllegalStateException("Cannot add other table ids to locations cache " + tableId + "  " + tabletLocation.tablet_extent);
     }
-    
+
     // clear out any overlapping extents in cache
     removeOverlapping(metaCache, tabletLocation.tablet_extent);
-    
+
     // do not add to cache unless lock is held
     if (lcSession.checkLock(tabletLocation) == null)
       return;
@@ -533,14 +533,14 @@ public class TabletLocatorImpl extends TabletLocator {
     if (er == null)
       er = MAX_TEXT;
     metaCache.put(er, tabletLocation);
-    
+
     if (badExtents.size() > 0)
       removeOverlapping(badExtents, tabletLocation.tablet_extent);
   }
-  
+
   static void removeOverlapping(TreeMap<Text,TabletLocation> metaCache, KeyExtent nke) {
     Iterator<Entry<Text,TabletLocation>> iter = null;
-    
+
     if (nke.getPrevEndRow() == null) {
       iter = metaCache.entrySet().iterator();
     } else {
@@ -548,30 +548,30 @@ public class TabletLocatorImpl extends TabletLocator {
       SortedMap<Text,TabletLocation> tailMap = metaCache.tailMap(row);
       iter = tailMap.entrySet().iterator();
     }
-    
+
     while (iter.hasNext()) {
       Entry<Text,TabletLocation> entry = iter.next();
-      
+
       KeyExtent ke = entry.getValue().tablet_extent;
-      
+
       if (stopRemoving(nke, ke)) {
         break;
       }
-      
+
       iter.remove();
     }
   }
-  
+
   private static boolean stopRemoving(KeyExtent nke, KeyExtent ke) {
     return ke.getPrevEndRow() != null && nke.getEndRow() != null && ke.getPrevEndRow().compareTo(nke.getEndRow()) >= 0;
   }
-  
+
   private static Text rowAfterPrevRow(KeyExtent nke) {
     Text row = new Text(nke.getPrevEndRow());
     row.append(new byte[] {0}, 0, 1);
     return row;
   }
-  
+
   static void removeOverlapping(TreeSet<KeyExtent> extents, KeyExtent nke) {
     for (KeyExtent overlapping : KeyExtent.findOverlapping(nke, extents)) {
       extents.remove(overlapping);
@@ -579,9 +579,9 @@ public class TabletLocatorImpl extends TabletLocator {
   }
 
   private TabletLocation locateTabletInCache(Text row) {
-    
+
     Entry<Text,TabletLocation> entry = metaCache.ceilingEntry(row);
-    
+
     if (entry != null) {
       KeyExtent ke = entry.getValue().tablet_extent;
       if (ke.getPrevEndRow() == null || ke.getPrevEndRow().compareTo(row) < 0) {
@@ -590,17 +590,17 @@ public class TabletLocatorImpl extends TabletLocator {
     }
     return null;
   }
-  
+
   protected TabletLocation _locateTablet(ClientContext context, Text row, boolean skipRow, boolean retry, boolean lock, LockCheckerSession lcSession)
       throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
-    
+
     if (skipRow) {
       row = new Text(row);
       row.append(new byte[] {0}, 0, 1);
     }
-    
+
     TabletLocation tl;
-    
+
     if (lock)
       rLock.lock();
     try {
@@ -610,30 +610,30 @@ public class TabletLocatorImpl extends TabletLocator {
       if (lock)
         rLock.unlock();
     }
-    
+
     if (tl == null) {
       if (lock)
         wLock.lock();
       try {
         // not in cache, so obtain info
         lookupTabletLocation(context, row, retry, lcSession);
-        
+
         tl = lcSession.checkLock(locateTabletInCache(row));
       } finally {
         if (lock)
           wLock.unlock();
       }
     }
-    
+
     return tl;
   }
-  
+
   private void processInvalidated(ClientContext context, LockCheckerSession lcSession) throws AccumuloSecurityException, AccumuloException,
       TableNotFoundException {
-    
+
     if (badExtents.size() == 0)
       return;
-    
+
     final boolean writeLockHeld = rwLock.isWriteLockedByCurrentThread();
     try {
       if (!writeLockHeld) {
@@ -642,27 +642,27 @@ public class TabletLocatorImpl extends TabletLocator {
         if (badExtents.size() == 0)
           return;
       }
-      
+
       List<Range> lookups = new ArrayList<Range>(badExtents.size());
-      
+
       for (KeyExtent be : badExtents) {
         lookups.add(be.toMetadataRange());
         removeOverlapping(metaCache, be);
       }
-      
+
       lookups = Range.mergeOverlapping(lookups);
-      
+
       Map<String,Map<KeyExtent,List<Range>>> binnedRanges = new HashMap<String,Map<KeyExtent,List<Range>>>();
-      
+
       parent.binRanges(context, lookups, binnedRanges);
-      
+
       // randomize server order
       ArrayList<String> tabletServers = new ArrayList<String>(binnedRanges.keySet());
       Collections.shuffle(tabletServers);
-      
+
       for (String tserver : tabletServers) {
         List<TabletLocation> locations = locationObtainer.lookupTablets(context, tserver, binnedRanges.get(tserver), parent);
-        
+
         for (TabletLocation tabletLocation : locations) {
           updateCache(tabletLocation, lcSession);
         }
@@ -674,21 +674,21 @@ public class TabletLocatorImpl extends TabletLocator {
       }
     }
   }
-  
+
   protected static void addRange(Map<String,Map<KeyExtent,List<Range>>> binnedRanges, String location, KeyExtent ke, Range range) {
     Map<KeyExtent,List<Range>> tablets = binnedRanges.get(location);
     if (tablets == null) {
       tablets = new HashMap<KeyExtent,List<Range>>();
       binnedRanges.put(location, tablets);
     }
-    
+
     List<Range> tabletsRanges = tablets.get(ke);
     if (tabletsRanges == null) {
       tabletsRanges = new ArrayList<Range>();
       tablets.put(ke, tabletsRanges);
     }
-    
+
     tabletsRanges.add(range);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchDeleter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchDeleter.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchDeleter.java
index b88571a..d3b26dc 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchDeleter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchDeleter.java
@@ -33,11 +33,11 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.ColumnVisibility;
 
 public class TabletServerBatchDeleter extends TabletServerBatchReader implements BatchDeleter {
-  
+
   private final ClientContext context;
   private String tableId;
   private BatchWriterConfig bwConfig;
-  
+
   public TabletServerBatchDeleter(ClientContext context, String tableId, Authorizations authorizations, int numQueryThreads, BatchWriterConfig bwConfig)
       throws TableNotFoundException {
     super(context, tableId, authorizations, numQueryThreads);
@@ -46,7 +46,7 @@ public class TabletServerBatchDeleter extends TabletServerBatchReader implements
     this.bwConfig = bwConfig;
     super.addScanIterator(new IteratorSetting(Integer.MAX_VALUE, BatchDeleter.class.getName() + ".NOVALUE", SortedKeyIterator.class));
   }
-  
+
   @Override
   public void delete() throws MutationsRejectedException, TableNotFoundException {
     BatchWriter bw = null;
@@ -65,5 +65,5 @@ public class TabletServerBatchDeleter extends TabletServerBatchReader implements
         bw.close();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReader.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReader.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReader.java
index 2a79f05..a7422c2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReader.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReader.java
@@ -34,25 +34,25 @@ import org.apache.log4j.Logger;
 
 public class TabletServerBatchReader extends ScannerOptions implements BatchScanner {
   private static final Logger log = Logger.getLogger(TabletServerBatchReader.class);
-  
+
   private String table;
   private int numThreads;
   private ExecutorService queryThreadPool;
-  
+
   private final ClientContext context;
   private ArrayList<Range> ranges;
-  
+
   private Authorizations authorizations = Authorizations.EMPTY;
   private Throwable ex = null;
-  
+
   private static int nextBatchReaderInstance = 1;
-  
+
   private static synchronized int getNextBatchReaderInstance() {
     return nextBatchReaderInstance++;
   }
-  
+
   private final int batchReaderInstance = getNextBatchReaderInstance();
-  
+
   public TabletServerBatchReader(ClientContext context, String table, Authorizations authorizations, int numQueryThreads) {
     checkArgument(context != null, "context is null");
     checkArgument(table != null, "table is null");
@@ -61,18 +61,18 @@ public class TabletServerBatchReader extends ScannerOptions implements BatchScan
     this.authorizations = authorizations;
     this.table = table;
     this.numThreads = numQueryThreads;
-    
+
     queryThreadPool = new SimpleThreadPool(numQueryThreads, "batch scanner " + batchReaderInstance + "-");
-    
+
     ranges = null;
     ex = new Throwable();
   }
-  
+
   @Override
   public void close() {
     queryThreadPool.shutdownNow();
   }
-  
+
   /**
    * Warning: do not rely upon finalize to close this class. Finalize is not guaranteed to be called.
    */
@@ -83,31 +83,31 @@ public class TabletServerBatchReader extends ScannerOptions implements BatchScan
       close();
     }
   }
-  
+
   @Override
   public void setRanges(Collection<Range> ranges) {
     if (ranges == null || ranges.size() == 0) {
       throw new IllegalArgumentException("ranges must be non null and contain at least 1 range");
     }
-    
+
     if (queryThreadPool.isShutdown()) {
       throw new IllegalStateException("batch reader closed");
     }
-    
+
     this.ranges = new ArrayList<Range>(ranges);
-    
+
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     if (ranges == null) {
       throw new IllegalStateException("ranges not set");
     }
-    
+
     if (queryThreadPool.isShutdown()) {
       throw new IllegalStateException("batch reader closed");
     }
-    
+
     return new TabletServerBatchReaderIterator(context, table, authorizations, ranges, numThreads, queryThreadPool, this, timeOut);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReaderIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReaderIterator.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReaderIterator.java
index eb80f8b..df330ed 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReaderIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchReaderIterator.java
@@ -73,9 +73,9 @@ import org.apache.thrift.transport.TTransportException;
 import com.google.common.net.HostAndPort;
 
 public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value>> {
-  
+
   private static final Logger log = Logger.getLogger(TabletServerBatchReaderIterator.class);
-  
+
   private final ClientContext context;
   private final Instance instance;
   private final String table;
@@ -83,30 +83,30 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
   private final int numThreads;
   private final ExecutorService queryThreadPool;
   private final ScannerOptions options;
-  
+
   private ArrayBlockingQueue<List<Entry<Key,Value>>> resultsQueue;
   private Iterator<Entry<Key,Value>> batchIterator;
   private List<Entry<Key,Value>> batch;
   private static final List<Entry<Key,Value>> LAST_BATCH = new ArrayList<Map.Entry<Key,Value>>();
   private final Object nextLock = new Object();
-  
+
   private long failSleepTime = 100;
-  
+
   private volatile Throwable fatalException = null;
-  
+
   private Map<String,TimeoutTracker> timeoutTrackers;
   private Set<String> timedoutServers;
   private long timeout;
-  
+
   private TabletLocator locator;
-  
+
   public interface ResultReceiver {
     void receive(List<Entry<Key,Value>> entries);
   }
-  
+
   public TabletServerBatchReaderIterator(ClientContext context, String table, Authorizations authorizations, ArrayList<Range> ranges, int numThreads,
       ExecutorService queryThreadPool, ScannerOptions scannerOptions, long timeout) {
-    
+
     this.context = context;
     this.instance = context.getInstance();
     this.table = table;
@@ -115,24 +115,24 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
     this.queryThreadPool = queryThreadPool;
     this.options = new ScannerOptions(scannerOptions);
     resultsQueue = new ArrayBlockingQueue<List<Entry<Key,Value>>>(numThreads);
-    
+
     this.locator = new TimeoutTabletLocator(TabletLocator.getLocator(context, new Text(table)), timeout);
-    
+
     timeoutTrackers = Collections.synchronizedMap(new HashMap<String,TabletServerBatchReaderIterator.TimeoutTracker>());
     timedoutServers = Collections.synchronizedSet(new HashSet<String>());
     this.timeout = timeout;
-    
+
     if (options.fetchedColumns.size() > 0) {
       ArrayList<Range> ranges2 = new ArrayList<Range>(ranges.size());
       for (Range range : ranges) {
         ranges2.add(range.bound(options.fetchedColumns.first(), options.fetchedColumns.last()));
       }
-      
+
       ranges = ranges2;
     }
-    
+
     ResultReceiver rr = new ResultReceiver() {
-      
+
       @Override
       public void receive(List<Entry<Key,Value>> entries) {
         try {
@@ -144,12 +144,12 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
             log.warn("Failed to add Batch Scan result", e);
           fatalException = e;
           throw new RuntimeException(e);
-          
+
         }
       }
-      
+
     };
-    
+
     try {
       lookup(ranges, rr);
     } catch (RuntimeException re) {
@@ -158,31 +158,31 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       throw new RuntimeException("Failed to create iterator", e);
     }
   }
-  
+
   @Override
   public boolean hasNext() {
     synchronized (nextLock) {
       if (batch == LAST_BATCH)
         return false;
-      
+
       if (batch != null && batchIterator.hasNext())
         return true;
-      
+
       // don't have one cached, try to cache one and return success
       try {
         batch = null;
         while (batch == null && fatalException == null && !queryThreadPool.isShutdown())
           batch = resultsQueue.poll(1, TimeUnit.SECONDS);
-        
+
         if (fatalException != null)
           if (fatalException instanceof RuntimeException)
             throw (RuntimeException) fatalException;
           else
             throw new RuntimeException(fatalException);
-        
+
         if (queryThreadPool.isShutdown())
           throw new RuntimeException("scanner closed");
-        
+
         batchIterator = batch.iterator();
         return batch != LAST_BATCH;
       } catch (InterruptedException e) {
@@ -190,7 +190,7 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       }
     }
   }
-  
+
   @Override
   public Entry<Key,Value> next() {
     // if there's one waiting, or hasNext() can get one, return it
@@ -201,33 +201,33 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
         throw new NoSuchElementException();
     }
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();
   }
-  
+
   private synchronized void lookup(List<Range> ranges, ResultReceiver receiver) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     List<Column> columns = new ArrayList<Column>(options.fetchedColumns);
     ranges = Range.mergeOverlapping(ranges);
-    
+
     Map<String,Map<KeyExtent,List<Range>>> binnedRanges = new HashMap<String,Map<KeyExtent,List<Range>>>();
-    
+
     binRanges(locator, ranges, binnedRanges);
-    
+
     doLookups(binnedRanges, receiver, columns);
   }
-  
+
   private void binRanges(TabletLocator tabletLocator, List<Range> ranges, Map<String,Map<KeyExtent,List<Range>>> binnedRanges) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
-    
+
     int lastFailureSize = Integer.MAX_VALUE;
-    
+
     while (true) {
-      
+
       binnedRanges.clear();
       List<Range> failures = tabletLocator.binRanges(context, ranges, binnedRanges);
-      
+
       if (failures.size() > 0) {
         // tried to only do table state checks when failures.size() == ranges.size(), however this did
         // not work because nothing ever invalidated entries in the tabletLocator cache... so even though
@@ -238,9 +238,9 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
             throw new TableDeletedException(table);
           else if (Tables.getTableState(instance, table) == TableState.OFFLINE)
             throw new TableOfflineException(instance, table);
-        
+
         lastFailureSize = failures.size();
-        
+
         if (log.isTraceEnabled())
           log.trace("Failed to bin " + failures.size() + " ranges, tablet locations were null, retrying in 100ms");
         try {
@@ -251,9 +251,9 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       } else {
         break;
       }
-      
+
     }
-    
+
     // truncate the ranges to within the tablets... this makes it easier to know what work
     // needs to be redone when failures occurs and tablets have merged or split
     Map<String,Map<KeyExtent,List<Range>>> binnedRanges2 = new HashMap<String,Map<KeyExtent,List<Range>>>();
@@ -268,47 +268,47 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
           clippedRanges.add(tabletRange.clip(range));
       }
     }
-    
+
     binnedRanges.clear();
     binnedRanges.putAll(binnedRanges2);
   }
-  
+
   private void processFailures(Map<KeyExtent,List<Range>> failures, ResultReceiver receiver, List<Column> columns) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException {
     if (log.isTraceEnabled())
       log.trace("Failed to execute multiscans against " + failures.size() + " tablets, retrying...");
-    
+
     try {
       Thread.sleep(failSleepTime);
     } catch (InterruptedException e) {
       Thread.currentThread().interrupt();
-      
+
       // We were interrupted (close called on batchscanner) just exit
       log.debug("Exiting failure processing on interrupt");
       return;
     }
 
     failSleepTime = Math.min(5000, failSleepTime * 2);
-    
+
     Map<String,Map<KeyExtent,List<Range>>> binnedRanges = new HashMap<String,Map<KeyExtent,List<Range>>>();
     List<Range> allRanges = new ArrayList<Range>();
-    
+
     for (List<Range> ranges : failures.values())
       allRanges.addAll(ranges);
-    
+
     // since the first call to binRanges clipped the ranges to within a tablet, we should not get only
     // bin to the set of failed tablets
     binRanges(locator, allRanges, binnedRanges);
-    
+
     doLookups(binnedRanges, receiver, columns);
   }
-  
+
   private String getTableInfo() {
     return Tables.getPrintableTableInfoFromId(instance, table);
   }
-  
+
   private class QueryTask implements Runnable {
-    
+
     private String tsLocation;
     private Map<KeyExtent,List<Range>> tabletsRanges;
     private ResultReceiver receiver;
@@ -316,7 +316,7 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
     private final Map<KeyExtent,List<Range>> failures;
     private List<Column> columns;
     private int semaphoreSize;
-    
+
     QueryTask(String tsLocation, Map<KeyExtent,List<Range>> tabletsRanges, Map<KeyExtent,List<Range>> failures, ResultReceiver receiver, List<Column> columns) {
       this.tsLocation = tsLocation;
       this.tabletsRanges = tabletsRanges;
@@ -324,12 +324,12 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       this.columns = columns;
       this.failures = failures;
     }
-    
+
     void setSemaphore(Semaphore semaphore, int semaphoreSize) {
       this.semaphore = semaphore;
       this.semaphoreSize = semaphoreSize;
     }
-    
+
     @Override
     public void run() {
       String threadName = Thread.currentThread().getName();
@@ -349,19 +349,19 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
             failures.putAll(tsFailures);
           }
         }
-        
+
       } catch (IOException e) {
         synchronized (failures) {
           failures.putAll(tsFailures);
           failures.putAll(unscanned);
         }
-        
+
         locator.invalidateCache(context.getInstance(), tsLocation);
         log.debug(e.getMessage(), e);
       } catch (AccumuloSecurityException e) {
         e.setTableInfo(getTableInfo());
         log.debug(e.getMessage(), e);
-        
+
         Tables.clearCache(instance);
         if (!Tables.exists(instance, table))
           fatalException = new TableDeletedException(table);
@@ -396,7 +396,7 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
               log.debug(t.getMessage(), t);
               fatalException = t;
             }
-            
+
             if (fatalException != null) {
               // we are finished with this batch query
               if (!resultsQueue.offer(LAST_BATCH)) {
@@ -423,16 +423,16 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
         }
       }
     }
-    
+
   }
-  
+
   private void doLookups(Map<String,Map<KeyExtent,List<Range>>> binnedRanges, final ResultReceiver receiver, List<Column> columns) {
-    
+
     if (timedoutServers.containsAll(binnedRanges.keySet())) {
       // all servers have timed out
       throw new TimedOutException(timedoutServers);
     }
-    
+
     // when there are lots of threads and a few tablet servers
     // it is good to break request to tablet servers up, the
     // following code determines if this is the case
@@ -442,16 +442,16 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       for (Entry<String,Map<KeyExtent,List<Range>>> entry : binnedRanges.entrySet()) {
         totalNumberOfTablets += entry.getValue().size();
       }
-      
+
       maxTabletsPerRequest = totalNumberOfTablets / numThreads;
       if (maxTabletsPerRequest == 0) {
         maxTabletsPerRequest = 1;
       }
-      
+
     }
-    
+
     Map<KeyExtent,List<Range>> failures = new HashMap<KeyExtent,List<Range>>();
-    
+
     if (timedoutServers.size() > 0) {
       // go ahead and fail any timed out servers
       for (Iterator<Entry<String,Map<KeyExtent,List<Range>>>> iterator = binnedRanges.entrySet().iterator(); iterator.hasNext();) {
@@ -462,16 +462,16 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
         }
       }
     }
-    
+
     // randomize tabletserver order... this will help when there are multiple
     // batch readers and writers running against accumulo
     List<String> locations = new ArrayList<String>(binnedRanges.keySet());
     Collections.shuffle(locations);
-    
+
     List<QueryTask> queryTasks = new ArrayList<QueryTask>();
-    
+
     for (final String tsLocation : locations) {
-      
+
       final Map<KeyExtent,List<Range>> tabletsRanges = binnedRanges.get(tsLocation);
       if (maxTabletsPerRequest == Integer.MAX_VALUE || tabletsRanges.size() == 1) {
         QueryTask queryTask = new QueryTask(tsLocation, tabletsRanges, failures, receiver, columns);
@@ -486,44 +486,44 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
             tabletSubset = new HashMap<KeyExtent,List<Range>>();
           }
         }
-        
+
         if (tabletSubset.size() > 0) {
           QueryTask queryTask = new QueryTask(tsLocation, tabletSubset, failures, receiver, columns);
           queryTasks.add(queryTask);
         }
       }
     }
-    
+
     final Semaphore semaphore = new Semaphore(queryTasks.size());
     semaphore.acquireUninterruptibly(queryTasks.size());
-    
+
     for (QueryTask queryTask : queryTasks) {
       queryTask.setSemaphore(semaphore, queryTasks.size());
       queryThreadPool.execute(new TraceRunnable(queryTask));
     }
   }
-  
+
   static void trackScanning(Map<KeyExtent,List<Range>> failures, Map<KeyExtent,List<Range>> unscanned, MultiScanResult scanResult) {
-    
+
     // translate returned failures, remove them from unscanned, and add them to failures
     Map<KeyExtent,List<Range>> retFailures = Translator.translate(scanResult.failures, Translators.TKET, new Translator.ListTranslator<TRange,Range>(
         Translators.TRT));
     unscanned.keySet().removeAll(retFailures.keySet());
     failures.putAll(retFailures);
-    
+
     // translate full scans and remove them from unscanned
     HashSet<KeyExtent> fullScans = new HashSet<KeyExtent>(Translator.translate(scanResult.fullScans, Translators.TKET));
     unscanned.keySet().removeAll(fullScans);
-    
+
     // remove partial scan from unscanned
     if (scanResult.partScan != null) {
       KeyExtent ke = new KeyExtent(scanResult.partScan);
       Key nextKey = new Key(scanResult.partNextKey);
-      
+
       ListIterator<Range> iterator = unscanned.get(ke).listIterator();
       while (iterator.hasNext()) {
         Range range = iterator.next();
-        
+
         if (range.afterEndKey(nextKey) || (nextKey.equals(range.getEndKey()) && scanResult.partNextKeyInclusive != range.isEndKeyInclusive())) {
           iterator.remove();
         } else if (range.contains(nextKey)) {
@@ -534,41 +534,41 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       }
     }
   }
-  
+
   private static class TimeoutTracker {
-    
+
     String server;
     Set<String> badServers;
     long timeOut;
     long activityTime;
     Long firstErrorTime = null;
-    
+
     TimeoutTracker(String server, Set<String> badServers, long timeOut) {
       this(timeOut);
       this.server = server;
       this.badServers = badServers;
     }
-    
+
     TimeoutTracker(long timeOut) {
       this.timeOut = timeOut;
     }
-    
+
     void startingScan() {
       activityTime = System.currentTimeMillis();
     }
-    
+
     void check() throws IOException {
       if (System.currentTimeMillis() - activityTime > timeOut) {
         badServers.add(server);
         throw new IOException("Time exceeded " + (System.currentTimeMillis() - activityTime) + " " + server);
       }
     }
-    
+
     void madeProgress() {
       activityTime = System.currentTimeMillis();
       firstErrorTime = null;
     }
-    
+
     void errorOccured(Exception e) {
       if (firstErrorTime == null) {
         firstErrorTime = activityTime;
@@ -576,28 +576,28 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
         badServers.add(server);
       }
     }
-    
+
     /**
      */
     public long getTimeOut() {
       return timeOut;
     }
   }
-  
+
   public static void doLookup(ClientContext context, String server, Map<KeyExtent,List<Range>> requested, Map<KeyExtent,List<Range>> failures,
       Map<KeyExtent,List<Range>> unscanned, ResultReceiver receiver, List<Column> columns, ScannerOptions options, Authorizations authorizations)
       throws IOException, AccumuloSecurityException, AccumuloServerException {
     doLookup(context, server, requested, failures, unscanned, receiver, columns, options, authorizations, new TimeoutTracker(Long.MAX_VALUE));
   }
-  
+
   static void doLookup(ClientContext context, String server, Map<KeyExtent,List<Range>> requested, Map<KeyExtent,List<Range>> failures,
       Map<KeyExtent,List<Range>> unscanned, ResultReceiver receiver, List<Column> columns, ScannerOptions options, Authorizations authorizations,
       TimeoutTracker timeoutTracker) throws IOException, AccumuloSecurityException, AccumuloServerException {
-    
+
     if (requested.size() == 0) {
       return;
     }
-    
+
     // copy requested to unscanned map. we will remove ranges as they are scanned in trackScanning()
     for (Entry<KeyExtent,List<Range>> entry : requested.entrySet()) {
       ArrayList<Range> ranges = new ArrayList<Range>();
@@ -606,7 +606,7 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       }
       unscanned.put(new KeyExtent(entry.getKey()), ranges);
     }
-    
+
     timeoutTracker.startingScan();
     TTransport transport = null;
     try {
@@ -618,13 +618,13 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
         client = ThriftUtil.getTServerClient(parsedServer, context);
 
       try {
-        
+
         OpTimer opTimer = new OpTimer(log, Level.TRACE).start("Starting multi scan, tserver=" + server + "  #tablets=" + requested.size() + "  #ranges="
             + sumSizes(requested.values()) + " ssil=" + options.serverSideIteratorList + " ssio=" + options.serverSideIteratorOptions);
-        
+
         TabletType ttype = TabletType.type(requested.keySet());
         boolean waitForWrites = !ThriftScanner.serversWaitedForWrites.get(ttype).contains(server);
-        
+
         Map<TKeyExtent,List<TRange>> thriftTabletRanges = Translator.translate(requested, Translators.KET, new Translator.ListTranslator<Range,TRange>(
             Translators.RT));
         InitialMultiScan imsr = client.startMultiScan(Tracer.traceInfo(), context.rpcCreds(), thriftTabletRanges,
@@ -634,48 +634,48 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
           ThriftScanner.serversWaitedForWrites.get(ttype).add(server.toString());
 
         MultiScanResult scanResult = imsr.result;
-        
+
         opTimer.stop("Got 1st multi scan results, #results=" + scanResult.results.size() + (scanResult.more ? "  scanID=" + imsr.scanID : "")
             + " in %DURATION%");
-        
+
         ArrayList<Entry<Key,Value>> entries = new ArrayList<Map.Entry<Key,Value>>(scanResult.results.size());
         for (TKeyValue kv : scanResult.results) {
           entries.add(new SimpleImmutableEntry<Key,Value>(new Key(kv.key), new Value(kv.value)));
         }
-        
+
         if (entries.size() > 0)
           receiver.receive(entries);
-        
+
         if (entries.size() > 0 || scanResult.fullScans.size() > 0)
           timeoutTracker.madeProgress();
-        
+
         trackScanning(failures, unscanned, scanResult);
-        
+
         while (scanResult.more) {
-          
+
           timeoutTracker.check();
-          
+
           opTimer.start("Continuing multi scan, scanid=" + imsr.scanID);
           scanResult = client.continueMultiScan(Tracer.traceInfo(), imsr.scanID);
           opTimer.stop("Got more multi scan results, #results=" + scanResult.results.size() + (scanResult.more ? "  scanID=" + imsr.scanID : "")
               + " in %DURATION%");
-          
+
           entries = new ArrayList<Map.Entry<Key,Value>>(scanResult.results.size());
           for (TKeyValue kv : scanResult.results) {
             entries.add(new SimpleImmutableEntry<Key,Value>(new Key(kv.key), new Value(kv.value)));
           }
-          
+
           if (entries.size() > 0)
             receiver.receive(entries);
-          
+
           if (entries.size() > 0 || scanResult.fullScans.size() > 0)
             timeoutTracker.madeProgress();
-          
+
           trackScanning(failures, unscanned, scanResult);
         }
-        
+
         client.closeMultiScan(Tracer.traceInfo(), imsr.scanID);
-        
+
       } finally {
         ThriftUtil.returnClient(client);
       }
@@ -700,14 +700,14 @@ public class TabletServerBatchReaderIterator implements Iterator<Entry<Key,Value
       ThriftTransportPool.getInstance().returnTransport(transport);
     }
   }
-  
+
   static int sumSizes(Collection<List<Range>> values) {
     int sum = 0;
-    
+
     for (List<Range> list : values) {
       sum += list.size();
     }
-    
+
     return sum;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchWriter.java
index c54c2f1..7abc826 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletServerBatchWriter.java
@@ -76,45 +76,45 @@ import com.google.common.net.HostAndPort;
 /*
  * Differences from previous TabletServerBatchWriter
  *   + As background threads finish sending mutations to tablet servers they decrement memory usage
- *   + Once the queue of unprocessed mutations reaches 50% it is always pushed to the background threads, 
- *      even if they are currently processing... new mutations are merged with mutations currently 
+ *   + Once the queue of unprocessed mutations reaches 50% it is always pushed to the background threads,
+ *      even if they are currently processing... new mutations are merged with mutations currently
  *      processing in the background
  *   + Failed mutations are held for 1000ms and then re-added to the unprocessed queue
  *   + Flush holds adding of new mutations so it does not wait indefinitely
- * 
+ *
  * Considerations
  *   + All background threads must catch and note Throwable
- *   + mutations for a single tablet server are only processed by one thread concurrently (if new mutations 
- *      come in for a tablet server while one thread is processing mutations for it, no other thread should 
+ *   + mutations for a single tablet server are only processed by one thread concurrently (if new mutations
+ *      come in for a tablet server while one thread is processing mutations for it, no other thread should
  *      start processing those mutations)
- *   
+ *
  * Memory accounting
  *   + when a mutation enters the system memory is incremented
  *   + when a mutation successfully leaves the system memory is decremented
- * 
- * 
- * 
+ *
+ *
+ *
  */
 
 public class TabletServerBatchWriter {
-  
+
   private static final Logger log = Logger.getLogger(TabletServerBatchWriter.class);
-  
+
   // basic configuration
   private final ClientContext context;
   private final long maxMem;
   private final long maxLatency;
   private final long timeout;
   private final Durability durability;
-  
+
   // state
   private boolean flushing;
   private boolean closed;
   private MutationSet mutations;
-  
+
   // background writer
   private final MutationWriter writer;
-  
+
   // latency timers
   private final Timer jtimer = new Timer("BatchWriterLatencyTimer", true);
   private final Map<String,TimeoutTracker> timeoutTrackers = Collections.synchronizedMap(new HashMap<String,TabletServerBatchWriter.TimeoutTracker>());
@@ -122,7 +122,7 @@ public class TabletServerBatchWriter {
   // stats
   private long totalMemUsed = 0;
   private long lastProcessingStartTime;
-  
+
   private long totalAdded = 0;
   private final AtomicLong totalSent = new AtomicLong(0);
   private final AtomicLong totalBinned = new AtomicLong(0);
@@ -132,7 +132,7 @@ public class TabletServerBatchWriter {
   private long initialGCTimes;
   private long initialCompileTimes;
   private double initialSystemLoad;
-  
+
   private int tabletServersBatchSum = 0;
   private int tabletBatchSum = 0;
   private int numBatches = 0;
@@ -140,7 +140,7 @@ public class TabletServerBatchWriter {
   private int minTabletBatch = Integer.MAX_VALUE;
   private int minTabletServersBatch = Integer.MAX_VALUE;
   private int maxTabletServersBatch = Integer.MIN_VALUE;
-  
+
   // error handling
   private final Violations violations = new Violations();
   private final Map<KeyExtent,Set<SecurityErrorCode>> authorizationFailures = new HashMap<KeyExtent,Set<SecurityErrorCode>>();
@@ -156,21 +156,21 @@ public class TabletServerBatchWriter {
     final long timeOut;
     long activityTime;
     Long firstErrorTime = null;
-    
+
     TimeoutTracker(String server, long timeOut) {
       this.timeOut = timeOut;
       this.server = server;
     }
-    
+
     void startingWrite() {
       activityTime = System.currentTimeMillis();
     }
-    
+
     void madeProgress() {
       activityTime = System.currentTimeMillis();
       firstErrorTime = null;
     }
-    
+
     void wroteNothing() {
       if (firstErrorTime == null) {
         firstErrorTime = activityTime;
@@ -178,16 +178,16 @@ public class TabletServerBatchWriter {
         throw new TimedOutException(Collections.singleton(server));
       }
     }
-    
+
     void errorOccured(Exception e) {
       wroteNothing();
     }
-    
+
     public long getTimeOut() {
       return timeOut;
     }
   }
-  
+
   public TabletServerBatchWriter(ClientContext context, BatchWriterConfig config) {
     this.context = context;
     this.maxMem = config.getMaxMemory();
@@ -196,9 +196,9 @@ public class TabletServerBatchWriter {
     this.mutations = new MutationSet();
     this.lastProcessingStartTime = System.currentTimeMillis();
     this.durability = config.getDurability();
-    
+
     this.writer = new MutationWriter(config.getMaxWriteThreads());
-    
+
     if (this.maxLatency != Long.MAX_VALUE) {
       jtimer.schedule(new TimerTask() {
         @Override
@@ -215,7 +215,7 @@ public class TabletServerBatchWriter {
       }, 0, this.maxLatency / 4);
     }
   }
-  
+
   private synchronized void startProcessing() {
     if (mutations.getMemoryUsed() == 0)
       return;
@@ -223,125 +223,125 @@ public class TabletServerBatchWriter {
     writer.addMutations(mutations);
     mutations = new MutationSet();
   }
-  
+
   private synchronized void decrementMemUsed(long amount) {
     totalMemUsed -= amount;
     this.notifyAll();
   }
-  
+
   public synchronized void addMutation(String table, Mutation m) throws MutationsRejectedException {
-    
+
     if (closed)
       throw new IllegalStateException("Closed");
     if (m.size() == 0)
       throw new IllegalArgumentException("Can not add empty mutations");
-    
+
     checkForFailures();
-    
+
     while ((totalMemUsed > maxMem || flushing) && !somethingFailed) {
       waitRTE();
     }
-    
+
     // do checks again since things could have changed while waiting and not holding lock
     if (closed)
       throw new IllegalStateException("Closed");
     checkForFailures();
-    
+
     if (startTime == 0) {
       startTime = System.currentTimeMillis();
-      
+
       List<GarbageCollectorMXBean> gcmBeans = ManagementFactory.getGarbageCollectorMXBeans();
       for (GarbageCollectorMXBean garbageCollectorMXBean : gcmBeans) {
         initialGCTimes += garbageCollectorMXBean.getCollectionTime();
       }
-      
+
       CompilationMXBean compMxBean = ManagementFactory.getCompilationMXBean();
       if (compMxBean.isCompilationTimeMonitoringSupported()) {
         initialCompileTimes = compMxBean.getTotalCompilationTime();
       }
-      
+
       initialSystemLoad = ManagementFactory.getOperatingSystemMXBean().getSystemLoadAverage();
     }
-    
+
     // create a copy of mutation so that after this method returns the user
     // is free to reuse the mutation object, like calling readFields... this
     // is important for the case where a mutation is passed from map to reduce
     // to batch writer... the map reduce code will keep passing the same mutation
     // object into the reduce method
     m = new Mutation(m);
-    
+
     totalMemUsed += m.estimatedMemoryUsed();
     mutations.addMutation(table, m);
     totalAdded++;
-    
+
     if (mutations.getMemoryUsed() >= maxMem / 2) {
       startProcessing();
       checkForFailures();
     }
   }
-  
+
   public void addMutation(String table, Iterator<Mutation> iterator) throws MutationsRejectedException {
     while (iterator.hasNext()) {
       addMutation(table, iterator.next());
     }
   }
-  
+
   public synchronized void flush() throws MutationsRejectedException {
-    
+
     if (closed)
       throw new IllegalStateException("Closed");
-    
+
     Span span = Trace.start("flush");
-    
+
     try {
       checkForFailures();
-      
+
       if (flushing) {
         // some other thread is currently flushing, so wait
         while (flushing && !somethingFailed)
           waitRTE();
-        
+
         checkForFailures();
-        
+
         return;
       }
-      
+
       flushing = true;
-      
+
       startProcessing();
       checkForFailures();
-      
+
       while (totalMemUsed > 0 && !somethingFailed) {
         waitRTE();
       }
-      
+
       flushing = false;
       this.notifyAll();
-      
+
       checkForFailures();
     } finally {
       span.stop();
       // somethingFailed = false;
     }
   }
-  
+
   public synchronized void close() throws MutationsRejectedException {
-    
+
     if (closed)
       return;
-    
+
     Span span = Trace.start("close");
     try {
       closed = true;
-      
+
       startProcessing();
-      
+
       while (totalMemUsed > 0 && !somethingFailed) {
         waitRTE();
       }
-      
+
       logStats();
-      
+
       checkForFailures();
     } finally {
       // make a best effort to release these resources
@@ -350,27 +350,27 @@ public class TabletServerBatchWriter {
       span.stop();
     }
   }
-  
+
   private void logStats() {
     long finishTime = System.currentTimeMillis();
-    
+
     long finalGCTimes = 0;
     List<GarbageCollectorMXBean> gcmBeans = ManagementFactory.getGarbageCollectorMXBeans();
     for (GarbageCollectorMXBean garbageCollectorMXBean : gcmBeans) {
       finalGCTimes += garbageCollectorMXBean.getCollectionTime();
     }
-    
+
     CompilationMXBean compMxBean = ManagementFactory.getCompilationMXBean();
     long finalCompileTimes = 0;
     if (compMxBean.isCompilationTimeMonitoringSupported()) {
       finalCompileTimes = compMxBean.getTotalCompilationTime();
     }
-    
+
     double averageRate = totalSent.get() / (totalSendTime.get() / 1000.0);
     double overallRate = totalAdded / ((finishTime - startTime) / 1000.0);
-    
+
     double finalSystemLoad = ManagementFactory.getOperatingSystemMXBean().getSystemLoadAverage();
-    
+
     if (log.isTraceEnabled()) {
       log.trace("");
       log.trace("TABLET SERVER BATCH WRITER STATISTICS");
@@ -400,39 +400,39 @@ public class TabletServerBatchWriter {
       log.trace(String.format("System load average : initial=%6.2f final=%6.2f", initialSystemLoad, finalSystemLoad));
     }
   }
-  
+
   private void updateSendStats(long count, long time) {
     totalSent.addAndGet(count);
     totalSendTime.addAndGet(time);
   }
-  
+
   public void updateBinningStats(int count, long time, Map<String,TabletServerMutations<Mutation>> binnedMutations) {
     totalBinTime.addAndGet(time);
     totalBinned.addAndGet(count);
     updateBatchStats(binnedMutations);
   }
-  
+
   private synchronized void updateBatchStats(Map<String,TabletServerMutations<Mutation>> binnedMutations) {
     tabletServersBatchSum += binnedMutations.size();
-    
+
     minTabletServersBatch = Math.min(minTabletServersBatch, binnedMutations.size());
     maxTabletServersBatch = Math.max(maxTabletServersBatch, binnedMutations.size());
-    
+
     int numTablets = 0;
-    
+
     for (Entry<String,TabletServerMutations<Mutation>> entry : binnedMutations.entrySet()) {
       TabletServerMutations<Mutation> tsm = entry.getValue();
       numTablets += tsm.getMutations().size();
     }
-    
+
     tabletBatchSum += numTablets;
-    
+
     minTabletBatch = Math.min(minTabletBatch, numTablets);
     maxTabletBatch = Math.max(maxTabletBatch, numTablets);
-    
+
     numBatches++;
   }
-  
+
   private void waitRTE() {
     try {
       wait();
@@ -440,9 +440,9 @@ public class TabletServerBatchWriter {
       throw new RuntimeException(e);
     }
   }
-  
+
   // BEGIN code for handling unrecoverable errors
-  
+
   private void updatedConstraintViolations(List<ConstraintViolationSummary> cvsList) {
     if (cvsList.size() > 0) {
       synchronized (this) {
@@ -452,28 +452,28 @@ public class TabletServerBatchWriter {
       }
     }
   }
-  
+
   private void updateAuthorizationFailures(Set<KeyExtent> keySet, SecurityErrorCode code) {
     HashMap<KeyExtent,SecurityErrorCode> map = new HashMap<KeyExtent,SecurityErrorCode>();
     for (KeyExtent ke : keySet)
       map.put(ke, code);
-    
+
     updateAuthorizationFailures(map);
   }
-  
+
   private void updateAuthorizationFailures(Map<KeyExtent,SecurityErrorCode> authorizationFailures) {
     if (authorizationFailures.size() > 0) {
-      
+
       // was a table deleted?
       HashSet<String> tableIds = new HashSet<String>();
       for (KeyExtent ke : authorizationFailures.keySet())
         tableIds.add(ke.getTableId().toString());
-      
+
       Tables.clearCache(context.getInstance());
       for (String tableId : tableIds)
         if (!Tables.exists(context.getInstance(), tableId))
           throw new TableDeletedException(tableId);
-      
+
       synchronized (this) {
         somethingFailed = true;
         mergeAuthorizationFailures(this.authorizationFailures, authorizationFailures);
@@ -481,7 +481,7 @@ public class TabletServerBatchWriter {
       }
     }
   }
-  
+
   private void mergeAuthorizationFailures(Map<KeyExtent,Set<SecurityErrorCode>> source, Map<KeyExtent,SecurityErrorCode> addition) {
     for (Entry<KeyExtent,SecurityErrorCode> entry : addition.entrySet()) {
       Set<SecurityErrorCode> secs = source.get(entry.getKey());
@@ -492,14 +492,14 @@ public class TabletServerBatchWriter {
       secs.add(entry.getValue());
     }
   }
-  
+
   private synchronized void updateServerErrors(String server, Exception e) {
     somethingFailed = true;
     this.serverSideErrors.add(server);
     this.notifyAll();
     log.error("Server side error on " + server + ": " + e);
   }
-  
+
   private synchronized void updateUnknownErrors(String msg, Throwable t) {
     somethingFailed = true;
     unknownErrors++;
@@ -510,29 +510,29 @@ public class TabletServerBatchWriter {
     else
       log.error(msg, t);
   }
-  
+
   private void checkForFailures() throws MutationsRejectedException {
     if (somethingFailed) {
       List<ConstraintViolationSummary> cvsList = violations.asList();
       HashMap<KeyExtent,Set<org.apache.accumulo.core.client.security.SecurityErrorCode>> af = new HashMap<KeyExtent,Set<org.apache.accumulo.core.client.security.SecurityErrorCode>>();
       for (Entry<KeyExtent,Set<SecurityErrorCode>> entry : authorizationFailures.entrySet()) {
         HashSet<org.apache.accumulo.core.client.security.SecurityErrorCode> codes = new HashSet<org.apache.accumulo.core.client.security.SecurityErrorCode>();
-        
+
         for (SecurityErrorCode sce : entry.getValue()) {
           codes.add(org.apache.accumulo.core.client.security.SecurityErrorCode.valueOf(sce.name()));
         }
-        
+
         af.put(entry.getKey(), codes);
       }
-      
+
       throw new MutationsRejectedException(context.getInstance(), cvsList, af, serverSideErrors, unknownErrors, lastUnknownError);
     }
   }
-  
+
   // END code for handling unrecoverable errors
-  
+
   // BEGIN code for handling failed mutations
-  
+
   /**
    * Add mutations that previously failed back into the mix
    */
@@ -542,16 +542,16 @@ public class TabletServerBatchWriter {
       startProcessing();
     }
   }
-  
+
   private class FailedMutations extends TimerTask {
-    
+
     private MutationSet recentFailures = null;
     private long initTime;
-    
+
     FailedMutations() {
       jtimer.schedule(this, 0, 500);
     }
-    
+
     private MutationSet init() {
       if (recentFailures == null) {
         recentFailures = new MutationSet();
@@ -559,35 +559,35 @@ public class TabletServerBatchWriter {
       }
       return recentFailures;
     }
-    
+
     synchronized void add(String table, ArrayList<Mutation> tableFailures) {
       init().addAll(table, tableFailures);
     }
-    
+
     synchronized void add(MutationSet failures) {
       init().addAll(failures);
     }
-    
+
     synchronized void add(String location, TabletServerMutations<Mutation> tsm) {
       init();
       for (Entry<KeyExtent,List<Mutation>> entry : tsm.getMutations().entrySet()) {
         recentFailures.addAll(entry.getKey().getTableId().toString(), entry.getValue());
       }
-      
+
     }
-    
+
     @Override
     public void run() {
       try {
         MutationSet rf = null;
-        
+
         synchronized (this) {
           if (recentFailures != null && System.currentTimeMillis() - initTime > 1000) {
             rf = recentFailures;
             recentFailures = null;
           }
         }
-        
+
         if (rf != null) {
           if (log.isTraceEnabled())
             log.trace("tid=" + Thread.currentThread().getId() + "  Requeuing " + rf.size() + " failed mutations");
@@ -599,26 +599,26 @@ public class TabletServerBatchWriter {
       }
     }
   }
-  
+
   // END code for handling failed mutations
-  
+
   // BEGIN code for sending mutations to tablet servers using background threads
-  
+
   private class MutationWriter {
-    
+
     private static final int MUTATION_BATCH_SIZE = 1 << 17;
     private final ExecutorService sendThreadPool;
     private final Map<String,TabletServerMutations<Mutation>> serversMutations;
     private final Set<String> queued;
     private final Map<String,TabletLocator> locators;
-    
+
     public MutationWriter(int numSendThreads) {
       serversMutations = new HashMap<String,TabletServerMutations<Mutation>>();
       queued = new HashSet<String>();
       sendThreadPool = new SimpleThreadPool(numSendThreads, this.getClass().getName());
       locators = new HashMap<String,TabletLocator>();
     }
-    
+
     private TabletLocator getLocator(String tableId) {
       TabletLocator ret = locators.get(tableId);
       if (ret == null) {
@@ -626,10 +626,10 @@ public class TabletServerBatchWriter {
         ret = new TimeoutTabletLocator(ret, timeout);
         locators.put(tableId, ret);
       }
-      
+
       return ret;
     }
-    
+
     private void binMutations(MutationSet mutationsToProcess, Map<String,TabletServerMutations<Mutation>> binnedMutations) {
       String tableId = null;
       try {
@@ -637,17 +637,17 @@ public class TabletServerBatchWriter {
         for (Entry<String,List<Mutation>> entry : es) {
           tableId = entry.getKey();
           TabletLocator locator = getLocator(tableId);
-          
+
           String table = entry.getKey();
           List<Mutation> tableMutations = entry.getValue();
-          
+
           if (tableMutations != null) {
             ArrayList<Mutation> tableFailures = new ArrayList<Mutation>();
             locator.binMutations(context, tableMutations, binnedMutations, tableFailures);
-            
+
             if (tableFailures.size() > 0) {
               failedMutations.add(table, tableFailures);
-              
+
               if (tableFailures.size() == tableMutations.size())
                 if (!Tables.exists(context.getInstance(), entry.getKey()))
                   throw new TableDeletedException(entry.getKey());
@@ -655,7 +655,7 @@ public class TabletServerBatchWriter {
                   throw new TableOfflineException(context.getInstance(), entry.getKey());
             }
           }
-          
+
         }
         return;
       } catch (AccumuloServerException ase) {
@@ -673,12 +673,12 @@ public class TabletServerBatchWriter {
       } catch (TableNotFoundException e) {
         updateUnknownErrors(e.getMessage(), e);
       }
-      
+
       // an error ocurred
       binnedMutations.clear();
-      
+
     }
-    
+
     void addMutations(MutationSet mutationsToSend) {
       Map<String,TabletServerMutations<Mutation>> binnedMutations = new HashMap<String,TabletServerMutations<Mutation>>();
       Span span = Trace.start("binMutations");
@@ -692,17 +692,17 @@ public class TabletServerBatchWriter {
       }
       addMutations(binnedMutations);
     }
-    
+
     private synchronized void addMutations(Map<String,TabletServerMutations<Mutation>> binnedMutations) {
-      
+
       int count = 0;
-      
+
       // merge mutations into existing mutations for a tablet server
       for (Entry<String,TabletServerMutations<Mutation>> entry : binnedMutations.entrySet()) {
         String server = entry.getKey();
-        
+
         TabletServerMutations<Mutation> currentMutations = serversMutations.get(server);
-        
+
         if (currentMutations == null) {
           serversMutations.put(server, entry.getValue());
         } else {
@@ -712,136 +712,136 @@ public class TabletServerBatchWriter {
             }
           }
         }
-        
+
         if (log.isTraceEnabled())
           for (Entry<KeyExtent,List<Mutation>> entry2 : entry.getValue().getMutations().entrySet())
             count += entry2.getValue().size();
-        
+
       }
-      
+
       if (count > 0 && log.isTraceEnabled())
         log.trace(String.format("Started sending %,d mutations to %,d tablet servers", count, binnedMutations.keySet().size()));
-      
+
       // randomize order of servers
       ArrayList<String> servers = new ArrayList<String>(binnedMutations.keySet());
       Collections.shuffle(servers);
-      
+
       for (String server : servers)
         if (!queued.contains(server)) {
           sendThreadPool.submit(Trace.wrap(new SendTask(server)));
           queued.add(server);
         }
     }
-    
+
     private synchronized TabletServerMutations<Mutation> getMutationsToSend(String server) {
       TabletServerMutations<Mutation> tsmuts = serversMutations.remove(server);
       if (tsmuts == null)
         queued.remove(server);
-      
+
       return tsmuts;
     }
-    
+
     class SendTask implements Runnable {
-      
+
       final private String location;
-      
+
       SendTask(String server) {
         this.location = server;
       }
-      
+
       @Override
       public void run() {
         try {
           TabletServerMutations<Mutation> tsmuts = getMutationsToSend(location);
-          
+
           while (tsmuts != null) {
             send(tsmuts);
             tsmuts = getMutationsToSend(location);
           }
-          
+
           return;
         } catch (Throwable t) {
           updateUnknownErrors("Failed to send tablet server " + location + " its batch : " + t.getMessage(), t);
         }
       }
-      
+
       public void send(TabletServerMutations<Mutation> tsm) throws AccumuloServerException, AccumuloSecurityException {
-        
+
         MutationSet failures = null;
-        
+
         String oldName = Thread.currentThread().getName();
-        
+
         Map<KeyExtent,List<Mutation>> mutationBatch = tsm.getMutations();
         try {
-          
+
           long count = 0;
           for (List<Mutation> list : mutationBatch.values()) {
             count += list.size();
           }
           String msg = "sending " + String.format("%,d", count) + " mutations to " + String.format("%,d", mutationBatch.size()) + " tablets at " + location;
           Thread.currentThread().setName(msg);
-          
+
           Span span = Trace.start("sendMutations");
           try {
-            
+
             TimeoutTracker timeoutTracker = timeoutTrackers.get(location);
             if (timeoutTracker == null) {
               timeoutTracker = new TimeoutTracker(location, timeout);
               timeoutTrackers.put(location, timeoutTracker);
             }
-            
+
             long st1 = System.currentTimeMillis();
             failures = sendMutationsToTabletServer(location, mutationBatch, timeoutTracker);
             long st2 = System.currentTimeMillis();
             if (log.isTraceEnabled())
               log.trace("sent " + String.format("%,d", count) + " mutations to " + location + " in "
                   + String.format("%.2f secs (%,.2f mutations/sec) with %,d failures", (st2 - st1) / 1000.0, count / ((st2 - st1) / 1000.0), failures.size()));
-            
+
             long successBytes = 0;
             for (Entry<KeyExtent,List<Mutation>> entry : mutationBatch.entrySet()) {
               for (Mutation mutation : entry.getValue()) {
                 successBytes += mutation.estimatedMemoryUsed();
               }
             }
-            
+
             if (failures.size() > 0) {
               failedMutations.add(failures);
               successBytes -= failures.getMemoryUsed();
             }
-            
+
             updateSendStats(count, st2 - st1);
             decrementMemUsed(successBytes);
-            
+
           } finally {
             span.stop();
           }
         } catch (IOException e) {
           if (log.isTraceEnabled())
             log.trace("failed to send mutations to " + location + " : " + e.getMessage());
-          
+
           HashSet<String> tables = new HashSet<String>();
           for (KeyExtent ke : mutationBatch.keySet())
             tables.add(ke.getTableId().toString());
-          
+
           for (String table : tables)
             TabletLocator.getLocator(context, new Text(table)).invalidateCache(context.getInstance(), location);
-          
+
           failedMutations.add(location, tsm);
         } finally {
           Thread.currentThread().setName(oldName);
         }
       }
     }
-    
+
     private MutationSet sendMutationsToTabletServer(String location, Map<KeyExtent,List<Mutation>> tabMuts, TimeoutTracker timeoutTracker) throws IOException,
         AccumuloSecurityException, AccumuloServerException {
       if (tabMuts.size() == 0) {
         return new MutationSet();
       }
       TInfo tinfo = Tracer.traceInfo();
-      
+
       timeoutTracker.startingWrite();
-      
+
       try {
         final HostAndPort parsedServer = HostAndPort.fromString(location);
         final TabletClientService.Iface client;
@@ -853,10 +853,10 @@ public class TabletServerBatchWriter {
 
         try {
           MutationSet allFailures = new MutationSet();
-          
+
           if (tabMuts.size() == 1 && tabMuts.values().iterator().next().size() == 1) {
             Entry<KeyExtent,List<Mutation>> entry = tabMuts.entrySet().iterator().next();
-            
+
             try {
               client.update(tinfo, context.rpcCreds(), entry.getKey().toThrift(), entry.getValue().get(0).toThrift(), DurabilityImpl.toThrift(durability));
             } catch (NotServingTabletException e) {
@@ -867,9 +867,9 @@ public class TabletServerBatchWriter {
             }
             timeoutTracker.madeProgress();
           } else {
-            
+
             long usid = client.startUpdate(tinfo, context.rpcCreds(), DurabilityImpl.toThrift(durability));
-            
+
             List<TMutation> updates = new ArrayList<TMutation>();
             for (Entry<KeyExtent,List<Mutation>> entry : tabMuts.entrySet()) {
               long size = 0;
@@ -880,34 +880,34 @@ public class TabletServerBatchWriter {
                   updates.add(mutation.toThrift());
                   size += mutation.numBytes();
                 }
-                
+
                 client.applyUpdates(tinfo, usid, entry.getKey().toThrift(), updates);
                 updates.clear();
                 size = 0;
               }
             }
-            
+
             UpdateErrors updateErrors = client.closeUpdate(tinfo, usid);
-            
+
             Map<KeyExtent,Long> failures = Translator.translate(updateErrors.failedExtents, Translators.TKET);
             updatedConstraintViolations(Translator.translate(updateErrors.violationSummaries, Translators.TCVST));
             updateAuthorizationFailures(Translator.translate(updateErrors.authorizationFailures, Translators.TKET));
-            
+
             long totalCommitted = 0;
-            
+
             for (Entry<KeyExtent,Long> entry : failures.entrySet()) {
               KeyExtent failedExtent = entry.getKey();
               int numCommitted = (int) (long) entry.getValue();
               totalCommitted += numCommitted;
-              
+
               String table = failedExtent.getTableId().toString();
-              
+
               TabletLocator.getLocator(context, new Text(table)).invalidateCache(failedExtent);
-              
+
               ArrayList<Mutation> mutations = (ArrayList<Mutation>) tabMuts.get(failedExtent);
               allFailures.addAll(table, mutations.subList(numCommitted, mutations.size()));
             }
-            
+
             if (failures.keySet().containsAll(tabMuts.keySet()) && totalCommitted == 0) {
               // nothing was successfully written
               timeoutTracker.wroteNothing();
@@ -936,34 +936,34 @@ public class TabletServerBatchWriter {
       }
     }
   }
-  
+
   // END code for sending mutations to tablet servers using background threads
-  
+
   private static class MutationSet {
-    
+
     private final HashMap<String,List<Mutation>> mutations;
     private int memoryUsed = 0;
-    
+
     MutationSet() {
       mutations = new HashMap<String,List<Mutation>>();
     }
-    
+
     void addMutation(String table, Mutation mutation) {
       List<Mutation> tabMutList = mutations.get(table);
       if (tabMutList == null) {
         tabMutList = new ArrayList<Mutation>();
         mutations.put(table, tabMutList);
       }
-      
+
       tabMutList.add(mutation);
-      
+
       memoryUsed += mutation.estimatedMemoryUsed();
     }
-    
+
     Map<String,List<Mutation>> getMutations() {
       return mutations;
     }
-    
+
     int size() {
       int result = 0;
       for (List<Mutation> perTable : mutations.values()) {
@@ -971,28 +971,28 @@ public class TabletServerBatchWriter {
       }
       return result;
     }
-    
+
     public void addAll(MutationSet failures) {
       Set<Entry<String,List<Mutation>>> es = failures.getMutations().entrySet();
-      
+
       for (Entry<String,List<Mutation>> entry : es) {
         String table = entry.getKey();
-        
+
         for (Mutation mutation : entry.getValue()) {
           addMutation(table, mutation);
         }
       }
     }
-    
+
     public void addAll(String table, List<Mutation> mutations) {
       for (Mutation mutation : mutations) {
         addMutation(table, mutation);
       }
     }
-    
+
     public int getMemoryUsed() {
       return memoryUsed;
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/impl/TabletType.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletType.java b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletType.java
index 3adcca9..d57bf94 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/impl/TabletType.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/impl/TabletType.java
@@ -22,7 +22,7 @@ import org.apache.accumulo.core.data.KeyExtent;
 
 public enum TabletType {
   ROOT, METADATA, USER;
-  
+
   public static TabletType type(KeyExtent ke) {
     if (ke.isRootTablet())
       return ROOT;
@@ -30,20 +30,20 @@ public enum TabletType {
       return METADATA;
     return USER;
   }
-  
+
   public static TabletType type(Collection<KeyExtent> extents) {
     if (extents.size() == 0)
       throw new IllegalArgumentException();
-    
+
     TabletType ttype = null;
-    
+
     for (KeyExtent extent : extents) {
       if (ttype == null)
         ttype = type(extent);
       else if (ttype != type(extent))
         throw new IllegalArgumentException("multiple extent types not allowed " + ttype + " " + type(extent));
     }
-    
+
     return ttype;
   }
 }


[53/66] [abbrv] accumulo git commit: ACCUMULO-3451 Add minimal checkstyle enforcement

Posted by ct...@apache.org.
ACCUMULO-3451 Add minimal checkstyle enforcement


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/d2c116ff
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/d2c116ff
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/d2c116ff

Branch: refs/heads/1.5
Commit: d2c116ffae59672ff995033e81be73260c4d5c25
Parents: c2155f4
Author: Christopher Tubbs <ct...@apache.org>
Authored: Tue Dec 23 18:51:04 2014 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:23:36 2015 -0500

----------------------------------------------------------------------
 pom.xml | 108 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 108 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/d2c116ff/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 8316a33..711a21d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -367,6 +367,11 @@
     <pluginManagement>
       <plugins>
         <plugin>
+          <groupId>org.apache.maven.plugins</groupId>
+          <artifactId>maven-checkstyle-plugin</artifactId>
+          <version>2.13</version>
+        </plugin>
+        <plugin>
           <groupId>com.google.code.sortpom</groupId>
           <artifactId>maven-sortpom-plugin</artifactId>
           <version>2.1.0</version>
@@ -566,6 +571,19 @@
                 <pluginExecution>
                   <pluginExecutionFilter>
                     <groupId>org.apache.maven.plugins</groupId>
+                    <artifactId>maven-checkstyle-plugin</artifactId>
+                    <versionRange>[2.13,)</versionRange>
+                    <goals>
+                      <goal>check</goal>
+                    </goals>
+                  </pluginExecutionFilter>
+                  <action>
+                    <ignore />
+                  </action>
+                </pluginExecution>
+                <pluginExecution>
+                  <pluginExecutionFilter>
+                    <groupId>org.apache.maven.plugins</groupId>
                     <artifactId>maven-dependency-plugin</artifactId>
                     <versionRange>[2.0,)</versionRange>
                     <goals>
@@ -709,6 +727,96 @@
         </executions>
       </plugin>
       <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-checkstyle-plugin</artifactId>
+        <configuration>
+          <checkstyleRules>
+            <module name="Checker">
+              <property name="charset" value="UTF-8" />
+              <property name="severity" value="warning" />
+              <!-- Checks for whitespace                               -->
+              <!-- See http://checkstyle.sf.net/config_whitespace.html -->
+              <module name="FileTabCharacter">
+                <property name="eachLine" value="true" />
+              </module>
+              <module name="TreeWalker">
+                <module name="OuterTypeFilename" />
+                <module name="LineLength">
+                  <!-- needs extra, because Eclipse formatter ignores the ending left brace -->
+                  <property name="max" value="200"/>
+                  <property name="ignorePattern" value="^package.*|^import.*|a href|href|http://|https://|ftp://"/>
+                </module>
+                <module name="AvoidStarImport" />
+                <module name="NoLineWrap" />
+                <module name="LeftCurly">
+                  <property name="maxLineLength" value="160" />
+                </module>
+                <module name="RightCurly" />
+                <module name="RightCurly">
+                  <property name="option" value="alone" />
+                  <property name="tokens" value="CLASS_DEF, METHOD_DEF, CTOR_DEF, LITERAL_FOR, LITERAL_WHILE, LITERAL_DO, STATIC_INIT, INSTANCE_INIT" />
+                </module>
+                <module name="SeparatorWrap">
+                  <property name="tokens" value="DOT" />
+                  <property name="option" value="nl" />
+                </module>
+                <module name="SeparatorWrap">
+                  <property name="tokens" value="COMMA" />
+                  <property name="option" value="EOL" />
+                </module>
+                <module name="PackageName">
+                  <property name="format" value="^[a-z]+(\.[a-z][a-zA-Z0-9]*)*$" />
+                </module>
+                <module name="MethodTypeParameterName">
+                  <property name="format" value="(^[A-Z][0-9]?)$|([A-Z][a-zA-Z0-9]*[T]$)" />
+                </module>
+                <module name="MethodParamPad" />
+                <module name="OperatorWrap">
+                  <property name="option" value="NL" />
+                  <property name="tokens" value="BAND, BOR, BSR, BXOR, DIV, EQUAL, GE, GT, LAND, LE, LITERAL_INSTANCEOF, LOR, LT, MINUS, MOD, NOT_EQUAL, QUESTION, SL, SR, STAR " />
+                </module>
+                <module name="AnnotationLocation">
+                  <property name="tokens" value="CLASS_DEF, INTERFACE_DEF, ENUM_DEF, METHOD_DEF, CTOR_DEF" />
+                </module>
+                <module name="AnnotationLocation">
+                  <property name="tokens" value="VARIABLE_DEF" />
+                  <property name="allowSamelineMultipleAnnotations" value="true" />
+                </module>
+                <module name="NonEmptyAtclauseDescription" />
+                <module name="JavadocTagContinuationIndentation" />
+                <module name="JavadocMethod">
+                  <property name="allowMissingJavadoc" value="true" />
+                  <property name="allowMissingParamTags" value="true" />
+                  <property name="allowMissingThrowsTags" value="true" />
+                  <property name="allowMissingReturnTag" value="true" />
+                  <property name="allowedAnnotations" value="Override,Test,BeforeClass,AfterClass,Before,After" />
+                  <property name="allowThrowsTagsForSubclasses" value="true" />
+                </module>
+                <module name="SingleLineJavadoc" />
+              </module>
+            </module>
+          </checkstyleRules>
+          <violationSeverity>warning</violationSeverity>
+          <includeTestSourceDirectory>true</includeTestSourceDirectory>
+          <excludes>**/thrift/*.java</excludes>
+        </configuration>
+        <dependencies>
+          <dependency>
+            <groupId>com.puppycrawl.tools</groupId>
+            <artifactId>checkstyle</artifactId>
+            <version>6.1.1</version>
+          </dependency>
+        </dependencies>
+        <executions>
+          <execution>
+            <id>check-style</id>
+            <goals>
+              <goal>check</goal>
+            </goals>
+          </execution>
+        </executions>
+      </plugin>
+      <plugin>
         <groupId>com.github.koraktor</groupId>
         <artifactId>mavanagaiata</artifactId>
         <executions>


[35/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/conf/PerColumnIteratorConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/conf/PerColumnIteratorConfig.java b/core/src/main/java/org/apache/accumulo/core/iterators/conf/PerColumnIteratorConfig.java
index 1d8a1b3..310776aa 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/conf/PerColumnIteratorConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/conf/PerColumnIteratorConfig.java
@@ -20,56 +20,56 @@ import org.apache.hadoop.io.Text;
 
 /**
  * @deprecated since 1.4
- * 
+ *
  * @see org.apache.accumulo.core.client.IteratorSetting.Column
  * @see org.apache.accumulo.core.iterators.Combiner#setColumns(org.apache.accumulo.core.client.IteratorSetting, java.util.List)
  */
 @Deprecated
 public class PerColumnIteratorConfig {
-  
+
   private String parameter;
   private Text colq;
   private Text colf;
-  
+
   public PerColumnIteratorConfig(Text columnFamily, String parameter) {
     this.colf = columnFamily;
     this.colq = null;
     this.parameter = parameter;
   }
-  
+
   public PerColumnIteratorConfig(Text columnFamily, Text columnQualifier, String parameter) {
     this.colf = columnFamily;
     this.colq = columnQualifier;
     this.parameter = parameter;
   }
-  
+
   public Text getColumnFamily() {
     return colf;
   }
-  
+
   public Text getColumnQualifier() {
     return colq;
   }
-  
+
   public String encodeColumns() {
     return encodeColumns(this);
   }
-  
+
   public String getClassName() {
     return parameter;
   }
-  
+
   private static String encodeColumns(PerColumnIteratorConfig pcic) {
     return ColumnSet.encodeColumns(pcic.colf, pcic.colq);
   }
-  
+
   public static String encodeColumns(Text columnFamily, Text columnQualifier) {
     return ColumnSet.encodeColumns(columnFamily, columnQualifier);
   }
 
   public static PerColumnIteratorConfig decodeColumns(String columns, String className) {
     String[] cols = columns.split(":");
-    
+
     if (cols.length == 1) {
       return new PerColumnIteratorConfig(ColumnSet.decode(cols[0]), className);
     } else if (cols.length == 2) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIterator.java
index 7df57c5..350c4cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnFamilySkippingIterator.java
@@ -33,27 +33,27 @@ import org.apache.accumulo.core.iterators.SkippingIterator;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 public class ColumnFamilySkippingIterator extends SkippingIterator implements InterruptibleIterator {
-  
+
   protected Set<ByteSequence> colFamSet = null;
   protected TreeSet<ByteSequence> sortedColFams = null;
-  
+
   protected boolean inclusive = false;
   protected Range range;
-  
+
   public ColumnFamilySkippingIterator(SortedKeyValueIterator<Key,Value> source) {
     this.setSource(source);
   }
-  
+
   protected ColumnFamilySkippingIterator(SortedKeyValueIterator<Key,Value> source, Set<ByteSequence> colFamSet, boolean inclusive) {
     this(source);
     this.colFamSet = colFamSet;
     this.inclusive = inclusive;
   }
-  
+
   @Override
   protected void consume() throws IOException {
     int count = 0;
-    
+
     if (inclusive)
       while (getSource().hasTop() && !colFamSet.contains(getSource().getTopKey().getColumnFamilyData())) {
         if (count < 10) {
@@ -70,7 +70,7 @@ public class ColumnFamilySkippingIterator extends SkippingIterator implements In
             // seek to the next column family in the sorted list of column families
             reseek(new Key(getSource().getTopKey().getRowData().toArray(), higherCF.toArray(), new byte[0], new byte[0], Long.MAX_VALUE));
           }
-          
+
           count = 0;
         }
       }
@@ -86,7 +86,7 @@ public class ColumnFamilySkippingIterator extends SkippingIterator implements In
         }
       }
   }
-  
+
   private void reseek(Key key) throws IOException {
     if (range.afterEndKey(key)) {
       range = new Range(range.getEndKey(), true, range.getEndKey(), range.isEndKeyInclusive());
@@ -96,36 +96,36 @@ public class ColumnFamilySkippingIterator extends SkippingIterator implements In
       getSource().seek(range, colFamSet, inclusive);
     }
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new ColumnFamilySkippingIterator(getSource().deepCopy(env), colFamSet, inclusive);
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-    
+
     if (columnFamilies instanceof Set<?>) {
       colFamSet = (Set<ByteSequence>) columnFamilies;
     } else {
       colFamSet = new HashSet<ByteSequence>();
       colFamSet.addAll(columnFamilies);
     }
-    
+
     if (inclusive) {
       sortedColFams = new TreeSet<ByteSequence>(colFamSet);
     } else {
       sortedColFams = null;
     }
-    
+
     this.range = range;
     this.inclusive = inclusive;
     super.seek(range, colFamSet, inclusive);
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     ((InterruptibleIterator) getSource()).setInterruptFlag(flag);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnQualifierFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnQualifierFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnQualifierFilter.java
index d5ca3b4..6e0f6e1 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnQualifierFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/ColumnQualifierFilter.java
@@ -34,14 +34,14 @@ public class ColumnQualifierFilter extends Filter {
   private boolean scanColumns;
   private HashSet<ByteSequence> columnFamilies;
   private HashMap<ByteSequence,HashSet<ByteSequence>> columnsQualifiers;
-  
+
   public ColumnQualifierFilter() {}
-  
+
   public ColumnQualifierFilter(SortedKeyValueIterator<Key,Value> iterator, Set<Column> columns) {
     setSource(iterator);
     init(columns);
   }
-  
+
   public ColumnQualifierFilter(SortedKeyValueIterator<Key,Value> iterator, HashSet<ByteSequence> columnFamilies,
       HashMap<ByteSequence,HashSet<ByteSequence>> columnsQualifiers, boolean scanColumns) {
     setSource(iterator);
@@ -49,25 +49,25 @@ public class ColumnQualifierFilter extends Filter {
     this.columnsQualifiers = columnsQualifiers;
     this.scanColumns = scanColumns;
   }
-  
+
   public boolean accept(Key key, Value v) {
     if (!scanColumns)
       return true;
-    
+
     if (columnFamilies.contains(key.getColumnFamilyData()))
       return true;
-    
+
     HashSet<ByteSequence> cfset = columnsQualifiers.get(key.getColumnQualifierData());
     // ensure the columm qualifier goes with a paired column family,
     // it is possible that a column qualifier could occur with a
     // column family it was not paired with
     return cfset != null && cfset.contains(key.getColumnFamilyData());
   }
-  
+
   public void init(Set<Column> columns) {
     this.columnFamilies = new HashSet<ByteSequence>();
     this.columnsQualifiers = new HashMap<ByteSequence,HashSet<ByteSequence>>();
-    
+
     for (Iterator<Column> iter = columns.iterator(); iter.hasNext();) {
       Column col = iter.next();
       if (col.columnQualifier != null) {
@@ -77,18 +77,18 @@ public class ColumnQualifierFilter extends Filter {
           cfset = new HashSet<ByteSequence>();
           this.columnsQualifiers.put(cq, cfset);
         }
-        
+
         cfset.add(new ArrayByteSequence(col.columnFamily));
       } else {
         // this whole column family should pass
         columnFamilies.add(new ArrayByteSequence(col.columnFamily));
       }
     }
-    
+
     // only take action when column qualifies are present
     scanColumns = this.columnsQualifiers.size() > 0;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new ColumnQualifierFilter(getSource().deepCopy(env), columnFamilies, columnsQualifiers, scanColumns);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/CountingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/CountingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/CountingIterator.java
index 010136a..b75ce67 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/CountingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/CountingIterator.java
@@ -26,34 +26,34 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.accumulo.core.iterators.WrappingIterator;
 
 public class CountingIterator extends WrappingIterator {
-  
+
   private long count;
-  
+
   public CountingIterator deepCopy(IteratorEnvironment env) {
     return new CountingIterator(this, env);
   }
-  
+
   private CountingIterator(CountingIterator other, IteratorEnvironment env) {
     setSource(other.getSource().deepCopy(env));
     count = 0;
   }
-  
+
   public CountingIterator(SortedKeyValueIterator<Key,Value> source) {
     this.setSource(source);
     count = 0;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void next() throws IOException {
     super.next();
     count++;
   }
-  
+
   public long getCount() {
     return count;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/DeletingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/DeletingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/DeletingIterator.java
index e770351..1e7bd0d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/DeletingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/DeletingIterator.java
@@ -33,23 +33,23 @@ import org.apache.accumulo.core.iterators.WrappingIterator;
 public class DeletingIterator extends WrappingIterator {
   private boolean propogateDeletes;
   private Key workKey = new Key();
-  
+
   public DeletingIterator deepCopy(IteratorEnvironment env) {
     return new DeletingIterator(this, env);
   }
-  
+
   public DeletingIterator(DeletingIterator other, IteratorEnvironment env) {
     setSource(other.getSource().deepCopy(env));
     propogateDeletes = other.propogateDeletes;
   }
-  
+
   public DeletingIterator() {}
-  
+
   public DeletingIterator(SortedKeyValueIterator<Key,Value> iterator, boolean propogateDeletes) throws IOException {
     this.setSource(iterator);
     this.propogateDeletes = propogateDeletes;
   }
-  
+
   @Override
   public void next() throws IOException {
     if (super.getTopKey().isDeleted())
@@ -58,26 +58,26 @@ public class DeletingIterator extends WrappingIterator {
       getSource().next();
     findTop();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     // do not want to seek to the middle of a row
     Range seekRange = IteratorUtil.maximizeStartKeyTimeStamp(range);
-    
+
     super.seek(seekRange, columnFamilies, inclusive);
     findTop();
-    
+
     if (range.getStartKey() != null) {
       while (getSource().hasTop() && getSource().getTopKey().compareTo(range.getStartKey(), PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME) < 0) {
         next();
       }
-      
+
       while (hasTop() && range.beforeStartKey(getTopKey())) {
         next();
       }
     }
   }
-  
+
   private void findTop() throws IOException {
     if (!propogateDeletes) {
       while (getSource().hasTop() && getSource().getTopKey().isDeleted()) {
@@ -85,18 +85,18 @@ public class DeletingIterator extends WrappingIterator {
       }
     }
   }
-  
+
   private void skipRowColumn() throws IOException {
     workKey.set(getSource().getTopKey());
-    
+
     Key keyToSkip = workKey;
     getSource().next();
-    
+
     while (getSource().hasTop() && getSource().getTopKey().equals(keyToSkip, PartialKey.ROW_COLFAM_COLQUAL_COLVIS)) {
       getSource().next();
     }
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/LocalityGroupIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/LocalityGroupIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/LocalityGroupIterator.java
index 71d2d8b..b2fae6d 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/LocalityGroupIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/LocalityGroupIterator.java
@@ -34,7 +34,7 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.commons.lang.mutable.MutableLong;
 
 /**
- * 
+ *
  */
 public class LocalityGroupIterator extends HeapIterator implements InterruptibleIterator {
 
@@ -45,12 +45,12 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
       this(localityGroup.columnFamilies, localityGroup.isDefaultLocalityGroup);
       this.iterator = (InterruptibleIterator) localityGroup.iterator.deepCopy(env);
     }
-    
+
     public LocalityGroup(InterruptibleIterator iterator, Map<ByteSequence,MutableLong> columnFamilies, boolean isDefaultLocalityGroup) {
       this(columnFamilies, isDefaultLocalityGroup);
       this.iterator = iterator;
     }
-    
+
     public LocalityGroup(Map<ByteSequence,MutableLong> columnFamilies, boolean isDefaultLocalityGroup) {
       this.isDefaultLocalityGroup = isDefaultLocalityGroup;
       this.columnFamilies = columnFamilies;
@@ -64,7 +64,7 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
     protected Map<ByteSequence,MutableLong> columnFamilies;
     private InterruptibleIterator iterator;
   }
-  
+
   private LocalityGroup groups[];
   private Set<ByteSequence> nonDefaultColumnFamilies;
   private AtomicBoolean interruptFlag;
@@ -79,13 +79,13 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   public static final int seek(HeapIterator hiter, LocalityGroup[] groups, Set<ByteSequence> nonDefaultColumnFamilies, Range range,
       Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     hiter.clear();
-    
+
     int numLGSeeked = 0;
-    
+
     Set<ByteSequence> cfSet;
     if (columnFamilies.size() > 0)
       if (columnFamilies instanceof Set<?>) {
@@ -101,13 +101,13 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
       // when include is set to true it means this locality groups contains
       // wanted column families
       boolean include = false;
-      
+
       if (cfSet.size() == 0) {
         include = !inclusive;
       } else if (lgr.isDefaultLocalityGroup && lgr.columnFamilies == null) {
         // do not know what column families are in the default locality group,
         // only know what column families are not in it
-        
+
         if (inclusive) {
           if (!nonDefaultColumnFamilies.containsAll(cfSet)) {
             // default LG may contain wanted and unwanted column families
@@ -123,7 +123,7 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
          * Need to consider the following cases for inclusive and exclusive (lgcf:locality group column family set, cf:column family set) lgcf and cf are
          * disjoint lgcf and cf are the same cf contains lgcf lgcf contains cf lgccf and cf intersect but neither is a subset of the other
          */
-        
+
         for (Entry<ByteSequence,MutableLong> entry : lgr.columnFamilies.entrySet())
           if (entry.getValue().longValue() > 0)
             if (cfSet.contains(entry.getKey())) {
@@ -140,7 +140,7 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
         numLGSeeked++;
       }// every column family is excluded, zero count, or not present
     }
-    
+
     return numLGSeeked;
   }
 
@@ -152,16 +152,16 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     LocalityGroup[] groupsCopy = new LocalityGroup[groups.length];
-    
+
     for (int i = 0; i < groups.length; i++) {
       groupsCopy[i] = new LocalityGroup(groups[i], env);
       if (interruptFlag != null)
         groupsCopy[i].getIterator().setInterruptFlag(interruptFlag);
     }
-    
+
     return new LocalityGroupIterator(groupsCopy, nonDefaultColumnFamilies);
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     this.interruptFlag = flag;
@@ -169,5 +169,5 @@ public class LocalityGroupIterator extends HeapIterator implements Interruptible
       lgr.getIterator().setInterruptFlag(flag);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/MapFileIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/MapFileIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/MapFileIterator.java
index 37a234c..b9a4e31 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/MapFileIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/MapFileIterator.java
@@ -49,7 +49,7 @@ public class MapFileIterator implements FileSKVIterator {
   private int interruptCheckCount = 0;
   private FileSystem fs;
   private String dirName;
-  
+
   public MapFileIterator(AccumuloConfiguration acuconf, FileSystem fs, String dir, Configuration conf) throws IOException {
     this.reader = MapFileUtil.openMapFile(acuconf, fs, dir, conf);
     this.fs = fs;
@@ -60,59 +60,59 @@ public class MapFileIterator implements FileSKVIterator {
   public void setInterruptFlag(AtomicBoolean flag) {
     this.interruptFlag = flag;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public boolean hasTop() {
     return topKey != null;
   }
-  
+
   @Override
   public void next() throws IOException {
     if (interruptFlag != null && interruptCheckCount++ % 100 == 0 && interruptFlag.get())
       throw new IterationInterruptedException();
-    
+
     reader.next(topKey, topValue);
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     if (columnFamilies.size() != 0 || inclusive) {
       throw new IllegalArgumentException("I do not know how to filter column families");
     }
-    
+
     if (range == null)
       throw new IllegalArgumentException("Cannot seek to null range");
-    
+
     if (interruptFlag != null && interruptFlag.get())
       throw new IterationInterruptedException();
-    
+
     Key key = range.getStartKey();
     if (key == null) {
       key = new Key();
     }
-    
+
     reader.seek(key);
-    
+
     while (hasTop() && range.beforeStartKey(getTopKey())) {
       next();
     }
   }
-  
+
   @Override
   public Key getTopKey() {
     return topKey;
   }
-  
+
   @Override
   public Value getTopValue() {
     return topValue;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     try {
@@ -125,17 +125,17 @@ public class MapFileIterator implements FileSKVIterator {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public Key getFirstKey() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public Key getLastKey() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public DataInputStream getMetaStore(String name) throws IOException {
     Path path = new Path(this.dirName, name);
@@ -143,12 +143,12 @@ public class MapFileIterator implements FileSKVIterator {
       throw new NoSuchMetaStoreException("name = " + name);
     return fs.open(path);
   }
-  
+
   @Override
   public void closeDeepCopies() throws IOException {
     // nothing to do, deep copies are externally managed/closed
   }
-  
+
   @Override
   public void close() throws IOException {
     reader.close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/MultiIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/MultiIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/MultiIterator.java
index f406fee..26ad8e9 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/MultiIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/MultiIterator.java
@@ -32,21 +32,21 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 /**
  * An iterator capable of iterating over other iterators in sorted order.
- * 
- * 
- * 
+ *
+ *
+ *
  */
 
 public class MultiIterator extends HeapIterator {
-  
+
   private List<SortedKeyValueIterator<Key,Value>> iters;
   private Range fence;
-  
+
   // deep copy with no seek/scan state
   public MultiIterator deepCopy(IteratorEnvironment env) {
     return new MultiIterator(this, env);
   }
-  
+
   private MultiIterator(MultiIterator other, IteratorEnvironment env) {
     super(other.iters.size());
     this.iters = new ArrayList<SortedKeyValueIterator<Key,Value>>();
@@ -55,58 +55,58 @@ public class MultiIterator extends HeapIterator {
       iters.add(iter.deepCopy(env));
     }
   }
-  
+
   private void init() {
     for (SortedKeyValueIterator<Key,Value> skvi : iters)
       addSource(skvi);
   }
-  
+
   private MultiIterator(List<SortedKeyValueIterator<Key,Value>> iters, Range seekFence, boolean init) {
     super(iters.size());
-    
+
     if (seekFence != null && init) {
       // throw this exception because multi-iterator does not seek on init, therefore the
       // fence would not be enforced in anyway, so do not want to give the impression it
       // will enforce this
       throw new IllegalArgumentException("Initializing not supported when seek fence set");
     }
-    
+
     this.fence = seekFence;
     this.iters = iters;
-    
+
     if (init) {
       init();
     }
   }
-  
+
   public MultiIterator(List<SortedKeyValueIterator<Key,Value>> iters, Range seekFence) {
     this(iters, seekFence, false);
   }
-  
+
   public MultiIterator(List<SortedKeyValueIterator<Key,Value>> iters2, KeyExtent extent) {
     this(iters2, new Range(extent.getPrevEndRow(), false, extent.getEndRow(), true), false);
   }
-  
+
   public MultiIterator(List<SortedKeyValueIterator<Key,Value>> readers, boolean init) {
     this(readers, (Range) null, init);
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     clear();
-    
+
     if (fence != null) {
       range = fence.clip(range, true);
       if (range == null)
         return;
     }
-    
+
     for (SortedKeyValueIterator<Key,Value> skvi : iters) {
       skvi.seek(range, columnFamilies, inclusive);
       addSource(skvi);
     }
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/SequenceFileIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/SequenceFileIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/SequenceFileIterator.java
index f593ee2..266b638 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/SequenceFileIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/SequenceFileIterator.java
@@ -33,89 +33,89 @@ import org.apache.hadoop.io.SequenceFile;
 import org.apache.hadoop.io.SequenceFile.Reader;
 
 public class SequenceFileIterator implements FileSKVIterator {
-  
+
   private Reader reader;
   private Value top_value;
   private Key top_key;
   private boolean readValue;
-  
+
   public SequenceFileIterator deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException("SequenceFileIterator does not yet support cloning");
   }
-  
+
   @Override
   public void closeDeepCopies() throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   public SequenceFileIterator(SequenceFile.Reader reader, boolean readValue) throws IOException {
     this.reader = reader;
     this.readValue = readValue;
-    
+
     top_key = new Key();
-    
+
     if (readValue)
       top_value = new Value();
-    
+
     next();
   }
-  
+
   public Key getTopKey() {
     return top_key;
   }
-  
+
   public Value getTopValue() {
     return top_value;
   }
-  
+
   public boolean hasTop() {
     return top_key != null;
   }
-  
+
   public void next() throws IOException {
     boolean valid;
     if (readValue)
       valid = reader.next(top_key, top_value);
     else
       valid = reader.next(top_key);
-    
+
     if (!valid) {
       top_key = null;
       top_value = null;
     }
-    
+
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     throw new UnsupportedOperationException("seek() not supported");
   }
-  
+
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
-    
+
   }
-  
+
   @Override
   public void close() throws IOException {
     reader.close();
   }
-  
+
   @Override
   public Key getFirstKey() throws IOException {
     throw new UnsupportedOperationException("getFirstKey() not supported");
   }
-  
+
   @Override
   public Key getLastKey() throws IOException {
     throw new UnsupportedOperationException("getLastKey() not supported");
   }
-  
+
   @Override
   public DataInputStream getMetaStore(String name) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/StatsIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/StatsIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/StatsIterator.java
index 1f23577..f92d1ec 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/StatsIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/StatsIterator.java
@@ -29,36 +29,36 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.accumulo.core.iterators.WrappingIterator;
 
 /**
- * 
+ *
  */
 public class StatsIterator extends WrappingIterator {
-  
+
   private int numRead = 0;
   private AtomicLong seekCounter;
   private AtomicLong readCounter;
-  
+
   public StatsIterator(SortedKeyValueIterator<Key,Value> source, AtomicLong seekCounter, AtomicLong readCounter) {
     super.setSource(source);
     this.seekCounter = seekCounter;
     this.readCounter = readCounter;
   }
-  
+
   @Override
   public void next() throws IOException {
     super.next();
     numRead++;
-    
+
     if (numRead % 23 == 0) {
       readCounter.addAndGet(numRead);
       numRead = 0;
     }
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new StatsIterator(getSource().deepCopy(env), seekCounter, readCounter);
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     super.seek(range, columnFamilies, inclusive);
@@ -66,7 +66,7 @@ public class StatsIterator extends WrappingIterator {
     readCounter.addAndGet(numRead);
     numRead = 0;
   }
-  
+
   public void report() {
     readCounter.addAndGet(numRead);
     numRead = 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/TimeSettingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/TimeSettingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/TimeSettingIterator.java
index 4eef14d..3e1b7a9 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/TimeSettingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/TimeSettingIterator.java
@@ -30,48 +30,48 @@ import org.apache.accumulo.core.iterators.IteratorUtil;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 public class TimeSettingIterator implements InterruptibleIterator {
-  
+
   private SortedKeyValueIterator<Key,Value> source;
   private long time;
   private Range range;
-  
+
   public TimeSettingIterator(SortedKeyValueIterator<Key,Value> source, long time) {
     this.source = source;
     this.time = time;
   }
-  
+
   @Override
   public Key getTopKey() {
     Key key = new Key(source.getTopKey());
     key.setTimestamp(time);
     return key;
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     ((InterruptibleIterator) source).setInterruptFlag(flag);
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new TimeSettingIterator(source.deepCopy(env), time);
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
-    
+
   }
-  
+
   @Override
   public boolean hasTop() {
     return source.hasTop() && !range.afterEndKey(getTopKey());
   }
-  
+
   @Override
   public void next() throws IOException {
     source.next();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     Range seekRange = IteratorUtil.maximizeStartKeyTimeStamp(range);
@@ -81,12 +81,12 @@ public class TimeSettingIterator implements InterruptibleIterator {
     while (hasTop() && range.beforeStartKey(getTopKey())) {
       next();
     }
-    
+
   }
-  
+
   @Override
   public Value getTopValue() {
     return source.getTopValue();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/system/VisibilityFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/system/VisibilityFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/system/VisibilityFilter.java
index 15c33fa..4bbd819 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/system/VisibilityFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/system/VisibilityFilter.java
@@ -37,11 +37,11 @@ public class VisibilityFilter extends Filter {
   protected LRUMap cache;
   protected Text tmpVis;
   protected Authorizations authorizations;
-  
+
   private static final Logger log = Logger.getLogger(VisibilityFilter.class);
-  
+
   public VisibilityFilter() {}
-  
+
   public VisibilityFilter(SortedKeyValueIterator<Key,Value> iterator, Authorizations authorizations, byte[] defaultVisibility) {
     setSource(iterator);
     this.ve = new VisibilityEvaluator(authorizations);
@@ -50,25 +50,25 @@ public class VisibilityFilter extends Filter {
     this.cache = new LRUMap(1000);
     this.tmpVis = new Text();
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new VisibilityFilter(getSource().deepCopy(env), authorizations, TextUtil.getBytes(defaultVisibility));
   }
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     Text testVis = k.getColumnVisibility(tmpVis);
-    
+
     if (testVis.getLength() == 0 && defaultVisibility.getLength() == 0)
       return true;
     else if (testVis.getLength() == 0)
       testVis = defaultVisibility;
-    
+
     Boolean b = (Boolean) cache.get(testVis);
     if (b != null)
       return b;
-    
+
     try {
       Boolean bb = ve.evaluate(new ColumnVisibility(testVis));
       cache.put(new Text(testVis), bb);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/AgeOffFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/AgeOffFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/AgeOffFilter.java
index 6e9a571..705f990 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/AgeOffFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/AgeOffFilter.java
@@ -28,7 +28,7 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 /**
  * A filter that ages off key/value pairs based on the Key's timestamp. It removes an entry if its timestamp is less than currentTime - threshold.
- * 
+ *
  * This filter requires a "ttl" option, in milliseconds, to determine the age off threshold.
  */
 public class AgeOffFilter extends Filter {
@@ -36,10 +36,10 @@ public class AgeOffFilter extends Filter {
   private static final String CURRENT_TIME = "currentTime";
   private long threshold;
   private long currentTime;
-  
+
   /**
    * Accepts entries whose timestamps are less than currentTime - threshold.
-   * 
+   *
    * @see org.apache.accumulo.core.iterators.Filter#accept(org.apache.accumulo.core.data.Key, org.apache.accumulo.core.data.Value)
    */
   @Override
@@ -48,29 +48,29 @@ public class AgeOffFilter extends Filter {
       return false;
     return true;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     if (options == null)
       throw new IllegalArgumentException(TTL + " must be set for AgeOffFilter");
-    
+
     String ttl = options.get(TTL);
     if (ttl == null)
       throw new IllegalArgumentException(TTL + " must be set for AgeOffFilter");
-    
+
     super.init(source, options, env);
     threshold = -1;
     threshold = Long.parseLong(ttl);
-    
+
     String time = options.get(CURRENT_TIME);
     if (time != null)
       currentTime = Long.parseLong(time);
     else
       currentTime = System.currentTimeMillis();
-    
+
     // add sanity checks for threshold and currentTime?
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     AgeOffFilter copy = (AgeOffFilter) super.deepCopy(env);
@@ -78,7 +78,7 @@ public class AgeOffFilter extends Filter {
     copy.threshold = threshold;
     return copy;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -88,7 +88,7 @@ public class AgeOffFilter extends Filter {
     io.setDescription("AgeOffFilter removes entries with timestamps more than <ttl> milliseconds old");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
@@ -100,10 +100,10 @@ public class AgeOffFilter extends Filter {
     }
     return true;
   }
-  
+
   /**
    * A convenience method for setting the age off threshold.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param ttl
@@ -112,10 +112,10 @@ public class AgeOffFilter extends Filter {
   public static void setTTL(IteratorSetting is, Long ttl) {
     is.addOption(TTL, Long.toString(ttl));
   }
-  
+
   /**
    * A convenience method for setting the current time (from which to measure the age off threshold).
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param currentTime

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/BigDecimalCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/BigDecimalCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/BigDecimalCombiner.java
index a9cd774..86c9e73 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/BigDecimalCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/BigDecimalCombiner.java
@@ -31,18 +31,17 @@ import org.apache.accumulo.core.iterators.TypedValueCombiner;
 import org.apache.accumulo.core.iterators.ValueFormatException;
 
 /**
- * A family of combiners that treat values as BigDecimals, encoding and 
- * decoding using the built-in BigDecimal String input/output functions.
+ * A family of combiners that treat values as BigDecimals, encoding and decoding using the built-in BigDecimal String input/output functions.
  */
 public abstract class BigDecimalCombiner extends TypedValueCombiner<BigDecimal> {
   private final static BigDecimalEncoder BDE = new BigDecimalEncoder();
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     setEncoder(BDE);
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -50,14 +49,14 @@ public abstract class BigDecimalCombiner extends TypedValueCombiner<BigDecimal>
     io.setDescription("bigdecimalcombiner interprets Values as BigDecimals before combining");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
       return false;
     return true;
   }
-  
+
   public static class BigDecimalSummingCombiner extends BigDecimalCombiner {
     @Override
     public BigDecimal typedReduce(Key key, Iterator<BigDecimal> iter) {
@@ -70,7 +69,7 @@ public abstract class BigDecimalCombiner extends TypedValueCombiner<BigDecimal>
       return sum;
     }
   }
-  
+
   public static class BigDecimalMaxCombiner extends BigDecimalCombiner {
     @Override
     public BigDecimal typedReduce(Key key, Iterator<BigDecimal> iter) {
@@ -83,7 +82,7 @@ public abstract class BigDecimalCombiner extends TypedValueCombiner<BigDecimal>
       return max;
     }
   }
-  
+
   public static class BigDecimalMinCombiner extends BigDecimalCombiner {
     @Override
     public BigDecimal typedReduce(Key key, Iterator<BigDecimal> iter) {
@@ -96,17 +95,17 @@ public abstract class BigDecimalCombiner extends TypedValueCombiner<BigDecimal>
       return min;
     }
   }
-  
+
   /**
    * Provides the ability to encode scientific notation.
-   * 
+   *
    */
   public static class BigDecimalEncoder implements org.apache.accumulo.core.iterators.TypedValueCombiner.Encoder<BigDecimal> {
     @Override
     public byte[] encode(BigDecimal v) {
       return v.toString().getBytes(UTF_8);
     }
-    
+
     @Override
     public BigDecimal decode(byte[] b) throws ValueFormatException {
       try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnAgeOffFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnAgeOffFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnAgeOffFilter.java
index 51e9ed3..c3da5c1 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnAgeOffFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnAgeOffFilter.java
@@ -39,14 +39,14 @@ public class ColumnAgeOffFilter extends Filter {
   public static class TTLSet extends ColumnToClassMapping<Long> {
     public TTLSet(Map<String,String> objectStrings) {
       super();
-      
+
       for (Entry<String,String> entry : objectStrings.entrySet()) {
         String column = entry.getKey();
         String ttl = entry.getValue();
         Long l = Long.parseLong(ttl);
-        
+
         Pair<Text,Text> colPair = ColumnSet.decodeColumns(column);
-        
+
         if (colPair.getSecond() == null) {
           addObject(colPair.getFirst(), l);
         } else {
@@ -55,10 +55,10 @@ public class ColumnAgeOffFilter extends Filter {
       }
     }
   }
-  
+
   TTLSet ttls;
   long currentTime = 0;
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     Long threshold = ttls.getObject(k);
@@ -68,14 +68,14 @@ public class ColumnAgeOffFilter extends Filter {
       return false;
     return true;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     this.ttls = new TTLSet(options);
     currentTime = System.currentTimeMillis();
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     ColumnAgeOffFilter copy = (ColumnAgeOffFilter) super.deepCopy(env);
@@ -83,11 +83,11 @@ public class ColumnAgeOffFilter extends Filter {
     copy.ttls = ttls;
     return copy;
   }
-  
+
   public void overrideCurrentTime(long ts) {
     this.currentTime = ts;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();
@@ -96,7 +96,7 @@ public class ColumnAgeOffFilter extends Filter {
     io.addUnnamedOption("<col fam>[:<col qual>] <Long> (escape non-alphanum chars using %<hex>)");
     return io;
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (super.validateOptions(options) == false)
@@ -108,10 +108,10 @@ public class ColumnAgeOffFilter extends Filter {
     }
     return true;
   }
-  
+
   /**
    * A convenience method for adding or changing an age off threshold for a column.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param column
@@ -122,10 +122,10 @@ public class ColumnAgeOffFilter extends Filter {
   public static void addTTL(IteratorSetting is, IteratorSetting.Column column, Long ttl) {
     is.addOption(ColumnSet.encodeColumns(column.getFirst(), column.getSecond()), Long.toString(ttl));
   }
-  
+
   /**
    * A convenience method for removing an age off threshold for a column.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param column

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilter.java
index 5dfcd17..5de72d5 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/ColumnSliceFilter.java
@@ -16,6 +16,9 @@
  */
 package org.apache.accumulo.core.iterators.user;
 
+import java.io.IOException;
+import java.util.Map;
+
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
@@ -23,93 +26,90 @@ import org.apache.accumulo.core.iterators.Filter;
 import org.apache.accumulo.core.iterators.IteratorEnvironment;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
-import java.io.IOException;
-import java.util.Map;
-
 public class ColumnSliceFilter extends Filter {
-    public static final String START_BOUND = "startBound";
-    public static final String START_INCLUSIVE = "startInclusive";
-    public static final String END_BOUND = "endBound";
-    public static final String END_INCLUSIVE = "endInclusive";
+  public static final String START_BOUND = "startBound";
+  public static final String START_INCLUSIVE = "startInclusive";
+  public static final String END_BOUND = "endBound";
+  public static final String END_INCLUSIVE = "endInclusive";
 
-    private String startBound;
-    private String endBound;
-    private boolean startInclusive;
-    private boolean endInclusive;
+  private String startBound;
+  private String endBound;
+  private boolean startInclusive;
+  private boolean endInclusive;
 
-    @Override
-    public boolean accept(Key key, Value value) {
-        String colQ = key.getColumnQualifier().toString();
-        return (startBound == null || (startInclusive ? (colQ.compareTo(startBound) >= 0) : (colQ.compareTo(startBound) > 0)))
-                && (endBound == null || (endInclusive ? (colQ.compareTo(endBound) <= 0) : (colQ.compareTo(endBound) < 0)));
-    }
+  @Override
+  public boolean accept(Key key, Value value) {
+    String colQ = key.getColumnQualifier().toString();
+    return (startBound == null || (startInclusive ? (colQ.compareTo(startBound) >= 0) : (colQ.compareTo(startBound) > 0)))
+        && (endBound == null || (endInclusive ? (colQ.compareTo(endBound) <= 0) : (colQ.compareTo(endBound) < 0)));
+  }
 
-    @Override
-    public void init(SortedKeyValueIterator<Key, Value> source, Map<String, String> options, IteratorEnvironment env) throws IOException {
-        super.init(source, options, env);
-        if (options.containsKey(START_BOUND)) {
-            startBound = options.get(START_BOUND);
-        } else {
-            startBound = null;
-        }
-
-        if (options.containsKey(START_INCLUSIVE)) {
-            startInclusive = Boolean.parseBoolean(options.get(START_INCLUSIVE));
-        } else {
-            startInclusive = true;
-        }
-
-        if (options.containsKey(END_BOUND)) {
-            endBound = options.get(END_BOUND);
-        } else {
-            endBound = null;
-        }
+  @Override
+  public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
+    super.init(source, options, env);
+    if (options.containsKey(START_BOUND)) {
+      startBound = options.get(START_BOUND);
+    } else {
+      startBound = null;
+    }
 
-        if (options.containsKey(END_INCLUSIVE)) {
-            endInclusive = Boolean.parseBoolean(options.get(END_INCLUSIVE));
-        } else {
-            endInclusive = false;
-        }
+    if (options.containsKey(START_INCLUSIVE)) {
+      startInclusive = Boolean.parseBoolean(options.get(START_INCLUSIVE));
+    } else {
+      startInclusive = true;
     }
 
-    @Override
-    public IteratorOptions describeOptions() {
-        IteratorOptions io = super.describeOptions();
-        io.setName("columnSlice");
-        io.setDescription("The ColumnSliceFilter/Iterator allows you to filter for key/value pairs based on a lexicographic range of column qualifier names");
-        io.addNamedOption(START_BOUND, "start string in slice");
-        io.addNamedOption(END_BOUND, "end string in slice");
-        io.addNamedOption(START_INCLUSIVE, "include the start bound in the result set");
-        io.addNamedOption(END_INCLUSIVE, "include the end bound in the result set");
-        return io;
+    if (options.containsKey(END_BOUND)) {
+      endBound = options.get(END_BOUND);
+    } else {
+      endBound = null;
     }
 
-    public static void setSlice(IteratorSetting si, String start, String end) {
-        setSlice(si, start, true, end, false);
+    if (options.containsKey(END_INCLUSIVE)) {
+      endInclusive = Boolean.parseBoolean(options.get(END_INCLUSIVE));
+    } else {
+      endInclusive = false;
     }
+  }
 
-    public static void setSlice(IteratorSetting si, String start, boolean startInclusive, String end, boolean endInclusive) {
-        if (start != null && end != null && (start.compareTo(end) > 0 || (start.compareTo(end) == 0 && (!startInclusive || !endInclusive)))) {
-            throw new IllegalArgumentException("Start key must be less than end key or equal with both sides inclusive in range (" + start + ", " + end + ")");
-        }
+  @Override
+  public IteratorOptions describeOptions() {
+    IteratorOptions io = super.describeOptions();
+    io.setName("columnSlice");
+    io.setDescription("The ColumnSliceFilter/Iterator allows you to filter for key/value pairs based on a lexicographic range of column qualifier names");
+    io.addNamedOption(START_BOUND, "start string in slice");
+    io.addNamedOption(END_BOUND, "end string in slice");
+    io.addNamedOption(START_INCLUSIVE, "include the start bound in the result set");
+    io.addNamedOption(END_INCLUSIVE, "include the end bound in the result set");
+    return io;
+  }
 
-        if (start != null) {
-            si.addOption(START_BOUND, start);
-        }
-        if (end != null) {
-            si.addOption(END_BOUND, end);
-        }
-        si.addOption(START_INCLUSIVE, String.valueOf(startInclusive));
-        si.addOption(END_INCLUSIVE, String.valueOf(endInclusive));
+  public static void setSlice(IteratorSetting si, String start, String end) {
+    setSlice(si, start, true, end, false);
+  }
+
+  public static void setSlice(IteratorSetting si, String start, boolean startInclusive, String end, boolean endInclusive) {
+    if (start != null && end != null && (start.compareTo(end) > 0 || (start.compareTo(end) == 0 && (!startInclusive || !endInclusive)))) {
+      throw new IllegalArgumentException("Start key must be less than end key or equal with both sides inclusive in range (" + start + ", " + end + ")");
     }
 
-    @Override
-    public SortedKeyValueIterator<Key, Value> deepCopy(IteratorEnvironment env) {
-        ColumnSliceFilter result = (ColumnSliceFilter) super.deepCopy(env);
-        result.startBound = startBound;
-        result.startInclusive = startInclusive;
-        result.endBound = endBound;
-        result.endInclusive = endInclusive;
-        return result;
+    if (start != null) {
+      si.addOption(START_BOUND, start);
+    }
+    if (end != null) {
+      si.addOption(END_BOUND, end);
     }
+    si.addOption(START_INCLUSIVE, String.valueOf(startInclusive));
+    si.addOption(END_INCLUSIVE, String.valueOf(endInclusive));
+  }
+
+  @Override
+  public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
+    ColumnSliceFilter result = (ColumnSliceFilter) super.deepCopy(env);
+    result.startBound = startBound;
+    result.startInclusive = startInclusive;
+    result.endBound = endBound;
+    result.endInclusive = endInclusive;
+    return result;
+  }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/GrepIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/GrepIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/GrepIterator.java
index e859363..043a729 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/GrepIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/GrepIterator.java
@@ -34,43 +34,43 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
  * This iterator provides exact string matching. It searches both the Key and Value for the string. The string to match is specified by the "term" option.
  */
 public class GrepIterator extends Filter {
-  
+
   private byte term[];
-  
+
   @Override
   public boolean accept(Key k, Value v) {
     return match(v.get()) || match(k.getRowData()) || match(k.getColumnFamilyData()) || match(k.getColumnQualifierData());
   }
-  
+
   private boolean match(ByteSequence bs) {
     return indexOf(bs.getBackingArray(), bs.offset(), bs.length(), term) >= 0;
   }
-  
+
   private boolean match(byte[] ba) {
     return indexOf(ba, 0, ba.length, term) >= 0;
   }
-  
+
   // copied code below from java string and modified
-  
+
   private static int indexOf(byte[] source, int sourceOffset, int sourceCount, byte[] target) {
     byte first = target[0];
     int targetCount = target.length;
     int max = sourceOffset + (sourceCount - targetCount);
-    
+
     for (int i = sourceOffset; i <= max; i++) {
       /* Look for first character. */
       if (source[i] != first) {
         while (++i <= max && source[i] != first)
           continue;
       }
-      
+
       /* Found first character, now look at the rest of v2 */
       if (i <= max) {
         int j = i + 1;
         int end = j + targetCount - 1;
         for (int k = 1; j < end && source[j] == target[k]; j++, k++)
           continue;
-        
+
         if (j == end) {
           /* Found whole string. */
           return i - sourceOffset;
@@ -79,20 +79,20 @@ public class GrepIterator extends Filter {
     }
     return -1;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     GrepIterator copy = (GrepIterator) super.deepCopy(env);
     copy.term = Arrays.copyOf(term, term.length);
     return copy;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     term = options.get("term").getBytes(UTF_8);
   }
-  
+
   /**
    * Encode the grep term as an option for a ScanIterator
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/IndexedDocIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/IndexedDocIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/IndexedDocIterator.java
index 2e9f049..9ef2bf6 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/IndexedDocIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/IndexedDocIterator.java
@@ -40,37 +40,37 @@ import org.slf4j.LoggerFactory;
  * docID. As with the IntersectingIterator, documents are grouped together and indexed into a single row of an Accumulo table. This allows a tablet server to
  * perform boolean AND operations on terms in the index. This iterator also stores the document contents in a separate column family in the same row so that the
  * full document can be returned with each query.
- * 
+ *
  * The table structure should have the following form:
- * 
+ *
  * row: shardID, colfam: docColf\0doctype, colqual: docID, value: doc
- * 
+ *
  * row: shardID, colfam: indexColf, colqual: term\0doctype\0docID\0info, value: (empty)
- * 
+ *
  * When you configure this iterator with a set of terms, it will return only the docIDs and docs that appear with all of the specified terms. The result will
  * have the following form:
- * 
+ *
  * row: shardID, colfam: indexColf, colqual: doctype\0docID\0info, value: doc
- * 
+ *
  * This iterator is commonly used with BatchScanner or AccumuloInputFormat, to parallelize the search over all shardIDs.
  */
 public class IndexedDocIterator extends IntersectingIterator {
   private static final Logger log = LoggerFactory.getLogger(IndexedDocIterator.class);
   public static final Text DEFAULT_INDEX_COLF = new Text("i");
   public static final Text DEFAULT_DOC_COLF = new Text("e");
-  
+
   private static final String indexFamilyOptionName = "indexFamily";
   private static final String docFamilyOptionName = "docFamily";
-  
+
   private Text indexColf = DEFAULT_INDEX_COLF;
   private Text docColf = DEFAULT_DOC_COLF;
   private Set<ByteSequence> indexColfSet;
   private Set<ByteSequence> docColfSet;
-  
+
   private static final byte[] nullByte = {0};
-  
+
   public SortedKeyValueIterator<Key,Value> docSource;
-  
+
   @Override
   protected Key buildKey(Text partition, Text term, Text docID) {
     Text colq = new Text(term);
@@ -79,18 +79,18 @@ public class IndexedDocIterator extends IntersectingIterator {
     colq.append(nullByte, 0, 1);
     return new Key(partition, indexColf, colq);
   }
-  
+
   @Override
   protected Key buildKey(Text partition, Text term) {
     Text colq = new Text(term);
     return new Key(partition, indexColf, colq);
   }
-  
+
   @Override
   protected Text getDocID(Key key) {
     return parseDocID(key);
   }
-  
+
   public static Text parseDocID(Key key) {
     Text colq = key.getColumnQualifier();
     int firstZeroIndex = colq.find("\0");
@@ -113,7 +113,7 @@ public class IndexedDocIterator extends IntersectingIterator {
     }
     return docID;
   }
-  
+
   @Override
   protected Text getTerm(Key key) {
     if (indexColf.compareTo(key.getColumnFamily().getBytes(), 0, indexColf.getLength()) < 0) {
@@ -127,7 +127,7 @@ public class IndexedDocIterator extends IntersectingIterator {
     term.set(colq.getBytes(), 0, zeroIndex);
     return term;
   }
-  
+
   @Override
   synchronized public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
@@ -137,23 +137,23 @@ public class IndexedDocIterator extends IntersectingIterator {
       docColf = new Text(options.get(docFamilyOptionName));
     docSource = source.deepCopy(env);
     indexColfSet = Collections.singleton((ByteSequence) new ArrayByteSequence(indexColf.getBytes(), 0, indexColf.getLength()));
-    
+
     for (TermSource ts : this.sources) {
       ts.seekColfams = indexColfSet;
     }
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> seekColumnFamilies, boolean inclusive) throws IOException {
     super.seek(range, null, true);
-    
+
   }
-  
+
   @Override
   protected void advanceToIntersection() throws IOException {
     super.advanceToIntersection();
@@ -169,7 +169,7 @@ public class IndexedDocIterator extends IntersectingIterator {
     }
     log.debug("got doc value: " + value.toString());
   }
-  
+
   protected Key buildDocKey() {
     if (log.isTraceEnabled())
       log.trace("building doc key for " + currentPartition + " " + currentDocID);
@@ -189,10 +189,10 @@ public class IndexedDocIterator extends IntersectingIterator {
       log.trace("built doc key for seek: " + k.toString());
     return k;
   }
-  
+
   /**
    * A convenience method for setting the index column family.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param indexColf
@@ -201,10 +201,10 @@ public class IndexedDocIterator extends IntersectingIterator {
   public static void setIndexColf(IteratorSetting is, String indexColf) {
     is.addOption(indexFamilyOptionName, indexColf);
   }
-  
+
   /**
    * A convenience method for setting the document column family prefix.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param docColfPrefix
@@ -213,10 +213,10 @@ public class IndexedDocIterator extends IntersectingIterator {
   public static void setDocColfPrefix(IteratorSetting is, String docColfPrefix) {
     is.addOption(docFamilyOptionName, docColfPrefix);
   }
-  
+
   /**
    * A convenience method for setting the index column family and document column family prefix.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param indexColf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/IntersectingIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/IntersectingIterator.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/IntersectingIterator.java
index 732a76c..63d6a34 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/IntersectingIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/IntersectingIterator.java
@@ -39,68 +39,68 @@ import org.apache.hadoop.io.Text;
 /**
  * This iterator facilitates document-partitioned indexing. It involves grouping a set of documents together and indexing those documents into a single row of
  * an Accumulo table. This allows a tablet server to perform boolean AND operations on terms in the index.
- * 
+ *
  * The table structure should have the following form:
- * 
+ *
  * row: shardID, colfam: term, colqual: docID
- * 
+ *
  * When you configure this iterator with a set of terms (column families), it will return only the docIDs that appear with all of the specified terms. The
  * result will have an empty column family, as follows:
- * 
+ *
  * row: shardID, colfam: (empty), colqual: docID
- * 
+ *
  * This iterator is commonly used with BatchScanner or AccumuloInputFormat, to parallelize the search over all shardIDs.
- * 
+ *
  * This iterator will *ignore* any columnFamilies passed to {@link #seek(Range, Collection, boolean)} as it performs intersections over terms. Extending classes
  * should override the {@link TermSource#seekColfams} in their implementation's {@link #init(SortedKeyValueIterator, Map, IteratorEnvironment)} method.
- * 
+ *
  * README.shard in docs/examples shows an example of using the IntersectingIterator.
  */
 public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
-  
+
   protected Text nullText = new Text();
-  
+
   protected Text getPartition(Key key) {
     return key.getRow();
   }
-  
+
   protected Text getTerm(Key key) {
     return key.getColumnFamily();
   }
-  
+
   protected Text getDocID(Key key) {
     return key.getColumnQualifier();
   }
-  
+
   protected Key buildKey(Text partition, Text term) {
     return new Key(partition, (term == null) ? nullText : term);
   }
-  
+
   protected Key buildKey(Text partition, Text term, Text docID) {
     return new Key(partition, (term == null) ? nullText : term, docID);
   }
-  
+
   protected Key buildFollowingPartitionKey(Key key) {
     return key.followingKey(PartialKey.ROW);
   }
-  
+
   public static class TermSource {
     public SortedKeyValueIterator<Key,Value> iter;
     public Text term;
     public Collection<ByteSequence> seekColfams;
     public boolean notFlag;
-    
+
     public TermSource(TermSource other) {
       this.iter = other.iter;
       this.term = other.term;
       this.notFlag = other.notFlag;
       this.seekColfams = other.seekColfams;
     }
-    
+
     public TermSource(SortedKeyValueIterator<Key,Value> iter, Text term) {
       this(iter, term, false);
     }
-    
+
     public TermSource(SortedKeyValueIterator<Key,Value> iter, Text term, boolean notFlag) {
       this.iter = iter;
       this.term = term;
@@ -108,32 +108,32 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
       // The desired column families for this source is the term itself
       this.seekColfams = Collections.<ByteSequence> singletonList(new ArrayByteSequence(term.getBytes(), 0, term.getLength()));
     }
-    
+
     public String getTermString() {
       return (this.term == null) ? "Iterator" : this.term.toString();
     }
   }
-  
+
   protected TermSource[] sources;
   int sourcesCount = 0;
-  
+
   Range overallRange;
-  
+
   // query-time settings
   protected Text currentPartition = null;
   protected Text currentDocID = new Text(emptyByteArray);
   static final byte[] emptyByteArray = new byte[0];
-  
+
   protected Key topKey = null;
   protected Value value = new Value(emptyByteArray);
-  
+
   public IntersectingIterator() {}
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new IntersectingIterator(this, env);
   }
-  
+
   private IntersectingIterator(IntersectingIterator other, IteratorEnvironment env) {
     if (other.sources != null) {
       sourcesCount = other.sourcesCount;
@@ -143,29 +143,29 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
       }
     }
   }
-  
+
   @Override
   public Key getTopKey() {
     return topKey;
   }
-  
+
   @Override
   public Value getTopValue() {
     // we don't really care about values
     return value;
   }
-  
+
   @Override
   public boolean hasTop() {
     return currentPartition != null;
   }
-  
+
   // precondition: currentRow is not null
   private boolean seekOneSource(int sourceID) throws IOException {
     // find the next key in the appropriate column family that is at or beyond the cursor (currentRow, currentCQ)
     // advance the cursor if this source goes beyond it
     // return whether we advanced the cursor
-    
+
     // within this loop progress must be made in one of the following forms:
     // - currentRow or currentCQ must be increased
     // - the given source must advance its iterator
@@ -174,7 +174,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     // - the given source is out of data and currentRow is set to null
     // - the given source has advanced beyond the endRow and currentRow is set to null
     boolean advancedCursor = false;
-    
+
     if (sources[sourceID].notFlag) {
       while (true) {
         if (sources[sourceID].iter.hasTop() == false) {
@@ -194,7 +194,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
         int partitionCompare = currentPartition.compareTo(getPartition(sources[sourceID].iter.getTopKey()));
         // check if this source is already at or beyond currentRow
         // if not, then seek to at least the current row
-        
+
         if (partitionCompare > 0) {
           // seek to at least the currentRow
           Key seekKey = buildKey(currentPartition, sources[sourceID].term);
@@ -224,7 +224,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
             break;
           }
         }
-        
+
         // we have verified that we are in currentRow and the correct column family
         // make sure we are at or beyond columnQualifier
         Text docID = getDocID(sources[sourceID].iter.getTopKey());
@@ -260,7 +260,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
         // check if we're past the end key
         int endCompare = -1;
         // we should compare the row to the end of the range
-        
+
         if (overallRange.getEndKey() != null) {
           endCompare = overallRange.getEndKey().getRow().compareTo(sources[sourceID].iter.getTopKey().getRow());
           if ((!overallRange.isEndKeyInclusive() && endCompare <= 0) || endCompare < 0) {
@@ -289,7 +289,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
         // we have verified that the current source is positioned in currentRow
         // now we must make sure we're in the right columnFamily in the current row
         // Note: Iterators are auto-magically set to the correct columnFamily
-        
+
         if (sources[sourceID].term != null) {
           int termCompare = sources[sourceID].term.compareTo(getTerm(sources[sourceID].iter.getTopKey()));
           // check if this source is already on the right columnFamily
@@ -343,7 +343,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     return advancedCursor;
   }
-  
+
   @Override
   public void next() throws IOException {
     if (currentPartition == null) {
@@ -354,7 +354,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     sources[0].iter.next();
     advanceToIntersection();
   }
-  
+
   protected void advanceToIntersection() throws IOException {
     boolean cursorChanged = true;
     while (cursorChanged) {
@@ -373,16 +373,16 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     topKey = buildKey(currentPartition, nullText, currentDocID);
   }
-  
+
   public static String stringTopKey(SortedKeyValueIterator<Key,Value> iter) {
     if (iter.hasTop())
       return iter.getTopKey().toString();
     return "";
   }
-  
+
   private static final String columnFamiliesOptionName = "columnFamilies";
   private static final String notFlagOptionName = "notFlag";
-  
+
   /**
    * @return encoded columns
    */
@@ -394,7 +394,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     return sb.toString();
   }
-  
+
   /**
    * @return encoded flags
    */
@@ -408,7 +408,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     return Base64.encodeBase64String(bytes);
   }
-  
+
   protected static Text[] decodeColumns(String columns) {
     String[] columnStrings = columns.split("\n");
     Text[] columnTexts = new Text[columnStrings.length];
@@ -417,7 +417,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     return columnTexts;
   }
-  
+
   /**
    * @return decoded flags
    */
@@ -425,7 +425,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     // return null of there were no flags
     if (flags == null)
       return null;
-    
+
     byte[] bytes = Base64.decodeBase64(flags.getBytes(UTF_8));
     boolean[] bFlags = new boolean[bytes.length];
     for (int i = 0; i < bytes.length; i++) {
@@ -436,16 +436,16 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     return bFlags;
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     Text[] terms = decodeColumns(options.get(columnFamiliesOptionName));
     boolean[] notFlag = decodeBooleans(options.get(notFlagOptionName));
-    
+
     if (terms.length < 1) {
       throw new IllegalArgumentException("IntersectionIterator requires one or more columns families");
     }
-    
+
     // Scan the not flags.
     // There must be at least one term that isn't negated
     // And we are going to re-order such that the first term is not a ! term
@@ -469,7 +469,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
         throw new IllegalArgumentException("IntersectionIterator requires at lest one column family without not");
       }
     }
-    
+
     sources = new TermSource[terms.length];
     sources[0] = new TermSource(source, terms[0]);
     for (int i = 1; i < terms.length; i++) {
@@ -477,13 +477,13 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     sourcesCount = terms.length;
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> seekColumnFamilies, boolean inclusive) throws IOException {
     overallRange = new Range(range);
     currentPartition = new Text();
     currentDocID.set(emptyByteArray);
-    
+
     // seek each of the sources to the right column family within the row given by key
     for (int i = 0; i < sourcesCount; i++) {
       Key sourceKey;
@@ -502,7 +502,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     }
     advanceToIntersection();
   }
-  
+
   /**
    * @deprecated since 1.6.0
    */
@@ -526,7 +526,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
     sources[sourcesCount] = new TermSource(source.deepCopy(env), term, notFlag);
     sourcesCount++;
   }
-  
+
   /**
    * Encode the columns to be used when iterating.
    */
@@ -535,7 +535,7 @@ public class IntersectingIterator implements SortedKeyValueIterator<Key,Value> {
       throw new IllegalArgumentException("Must supply at least one term to intersect");
     cfg.addOption(IntersectingIterator.columnFamiliesOptionName, IntersectingIterator.encodeColumns(columns));
   }
-  
+
   /**
    * Encode columns and NOT flags indicating which columns should be negated (docIDs will be excluded if matching negated columns, instead of included).
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/LargeRowFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/LargeRowFilter.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/LargeRowFilter.java
index 4db379b..59a5dec 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/LargeRowFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/LargeRowFilter.java
@@ -41,40 +41,40 @@ import org.apache.hadoop.io.Text;
 /**
  * This iterator suppresses rows that exceed a specified number of columns. Once a row exceeds the threshold, a marker is emitted and the row is always
  * suppressed by this iterator after that point in time.
- * 
+ *
  * This iterator works in a similar way to the RowDeletingIterator. See its javadoc about locality groups.
  */
 public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, OptionDescriber {
-  
+
   public static final Value SUPPRESS_ROW_VALUE = new Value("SUPPRESS_ROW".getBytes(UTF_8));
-  
+
   private static final ByteSequence EMPTY = new ArrayByteSequence(new byte[] {});
-  
+
   /* key into hash map, value refers to the row supression limit (maxColumns) */
   private static final String MAX_COLUMNS = "max_columns";
-  
+
   private SortedKeyValueIterator<Key,Value> source;
-  
+
   // a cache of keys
   private ArrayList<Key> keys = new ArrayList<Key>();
   private ArrayList<Value> values = new ArrayList<Value>();
-  
+
   private int currentPosition;
-  
+
   private int maxColumns;
-  
+
   private boolean propogateSuppression = false;
-  
+
   private Range range;
   private Collection<ByteSequence> columnFamilies;
   private boolean inclusive;
   private boolean dropEmptyColFams;
-  
+
   private boolean isSuppressionMarker(Key key, Value val) {
     return key.getColumnFamilyData().length() == 0 && key.getColumnQualifierData().length() == 0 && key.getColumnVisibilityData().length() == 0
         && val.equals(SUPPRESS_ROW_VALUE);
   }
-  
+
   private void reseek(Key key) throws IOException {
     if (range.afterEndKey(key)) {
       range = new Range(range.getEndKey(), true, range.getEndKey(), range.isEndKeyInclusive());
@@ -84,11 +84,11 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
       source.seek(range, columnFamilies, inclusive);
     }
   }
-  
+
   private void consumeRow(ByteSequence row) throws IOException {
     // try reading a few and if still not to next row, then seek
     int count = 0;
-    
+
     while (source.hasTop() && source.getTopKey().getRowData().equals(row)) {
       source.next();
       count++;
@@ -99,7 +99,7 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
       }
     }
   }
-  
+
   private void addKeyValue(Key k, Value v) {
     if (dropEmptyColFams && k.getColumnFamilyData().equals(EMPTY)) {
       return;
@@ -107,34 +107,34 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
     keys.add(new Key(k));
     values.add(new Value(v));
   }
-  
+
   private void bufferNextRow() throws IOException {
-    
+
     keys.clear();
     values.clear();
     currentPosition = 0;
-    
+
     while (source.hasTop() && keys.size() == 0) {
-      
+
       addKeyValue(source.getTopKey(), source.getTopValue());
-      
+
       if (isSuppressionMarker(source.getTopKey(), source.getTopValue())) {
-        
+
         consumeRow(source.getTopKey().getRowData());
-        
+
       } else {
-        
+
         ByteSequence currentRow = keys.get(0).getRowData();
         source.next();
-        
+
         while (source.hasTop() && source.getTopKey().getRowData().equals(currentRow)) {
-          
+
           addKeyValue(source.getTopKey(), source.getTopValue());
-          
+
           if (keys.size() > maxColumns) {
             keys.clear();
             values.clear();
-            
+
             // when the row is to big, just emit a suppression
             // marker
             addKeyValue(new Key(new Text(currentRow.toArray())), SUPPRESS_ROW_VALUE);
@@ -144,56 +144,56 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
           }
         }
       }
-      
+
     }
   }
-  
+
   private void readNextRow() throws IOException {
-    
+
     bufferNextRow();
-    
+
     while (!propogateSuppression && currentPosition < keys.size() && isSuppressionMarker(keys.get(0), values.get(0))) {
       bufferNextRow();
     }
   }
-  
+
   private LargeRowFilter(SortedKeyValueIterator<Key,Value> source, boolean propogateSuppression, int maxColumns) {
     this.source = source;
     this.propogateSuppression = propogateSuppression;
     this.maxColumns = maxColumns;
   }
-  
+
   public LargeRowFilter() {}
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     this.source = source;
     this.maxColumns = Integer.parseInt(options.get(MAX_COLUMNS));
     this.propogateSuppression = env.getIteratorScope() != IteratorScope.scan;
   }
-  
+
   @Override
   public boolean hasTop() {
     return currentPosition < keys.size();
   }
-  
+
   @Override
   public void next() throws IOException {
-    
+
     if (currentPosition >= keys.size()) {
       throw new IllegalStateException("Called next() when hasTop() is false");
     }
-    
+
     currentPosition++;
-    
+
     if (currentPosition == keys.size()) {
       readNextRow();
     }
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-    
+
     if (inclusive && !columnFamilies.contains(EMPTY)) {
       columnFamilies = new HashSet<ByteSequence>(columnFamilies);
       columnFamilies.add(EMPTY);
@@ -205,48 +205,48 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
     } else {
       dropEmptyColFams = false;
     }
-    
+
     this.range = range;
     this.columnFamilies = columnFamilies;
     this.inclusive = inclusive;
-    
+
     if (range.getStartKey() != null) {
       // seek to beginning of row to see if there is a suppression marker
       Range newRange = new Range(new Key(range.getStartKey().getRow()), true, range.getEndKey(), range.isEndKeyInclusive());
       source.seek(newRange, columnFamilies, inclusive);
-      
+
       readNextRow();
-      
+
       // it is possible that all or some of the data read for the current
       // row is before the start of the range
       while (currentPosition < keys.size() && range.beforeStartKey(keys.get(currentPosition)))
         currentPosition++;
-      
+
       if (currentPosition == keys.size())
         readNextRow();
-      
+
     } else {
       source.seek(range, columnFamilies, inclusive);
       readNextRow();
     }
-    
+
   }
-  
+
   @Override
   public Key getTopKey() {
     return keys.get(currentPosition);
   }
-  
+
   @Override
   public Value getTopValue() {
     return values.get(currentPosition);
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new LargeRowFilter(source.deepCopy(env), propogateSuppression, maxColumns);
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     String description = "This iterator suppresses rows that exceed a specified number of columns. Once\n"
@@ -255,13 +255,13 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
     return new IteratorOptions(this.getClass().getSimpleName(), description, Collections.singletonMap(MAX_COLUMNS, "Number Of Columns To Begin Suppression"),
         null);
   }
-  
+
   @Override
   public boolean validateOptions(Map<String,String> options) {
     if (options == null || options.size() < 1) {
       throw new IllegalArgumentException("Bad # of options, must supply: " + MAX_COLUMNS + " as value");
     }
-    
+
     if (!options.containsKey(MAX_COLUMNS))
       throw new IllegalArgumentException("Bad # of options, must supply: " + MAX_COLUMNS + " as value");
     try {
@@ -269,13 +269,13 @@ public class LargeRowFilter implements SortedKeyValueIterator<Key,Value>, Option
     } catch (Exception e) {
       throw new IllegalArgumentException("bad integer " + MAX_COLUMNS + ":" + options.get(MAX_COLUMNS));
     }
-    
+
     return true;
   }
-  
+
   /**
    * A convenience method for setting the maximum number of columns to keep.
-   * 
+   *
    * @param is
    *          IteratorSetting object to configure.
    * @param maxColumns

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/MaxCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/MaxCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/MaxCombiner.java
index 1b5a4e2..08ec483 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/MaxCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/MaxCombiner.java
@@ -35,7 +35,7 @@ public class MaxCombiner extends LongCombiner {
     }
     return max;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/iterators/user/MinCombiner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/iterators/user/MinCombiner.java b/core/src/main/java/org/apache/accumulo/core/iterators/user/MinCombiner.java
index 891806e..31be086 100644
--- a/core/src/main/java/org/apache/accumulo/core/iterators/user/MinCombiner.java
+++ b/core/src/main/java/org/apache/accumulo/core/iterators/user/MinCombiner.java
@@ -35,7 +35,7 @@ public class MinCombiner extends LongCombiner {
     }
     return min;
   }
-  
+
   @Override
   public IteratorOptions describeOptions() {
     IteratorOptions io = super.describeOptions();


[48/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/bloomfilter/DynamicBloomFilter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/bloomfilter/DynamicBloomFilter.java b/core/src/main/java/org/apache/accumulo/core/bloomfilter/DynamicBloomFilter.java
index 9b14f32..740bdda 100644
--- a/core/src/main/java/org/apache/accumulo/core/bloomfilter/DynamicBloomFilter.java
+++ b/core/src/main/java/org/apache/accumulo/core/bloomfilter/DynamicBloomFilter.java
@@ -2,30 +2,30 @@
  *
  * Copyright (c) 2005, European Commission project OneLab under contract 034819 (http://www.one-lab.org)
  * All rights reserved.
- * Redistribution and use in source and binary forms, with or 
- * without modification, are permitted provided that the following 
+ * Redistribution and use in source and binary forms, with or
+ * without modification, are permitted provided that the following
  * conditions are met:
- *  - Redistributions of source code must retain the above copyright 
+ *  - Redistributions of source code must retain the above copyright
  *    notice, this list of conditions and the following disclaimer.
- *  - Redistributions in binary form must reproduce the above copyright 
- *    notice, this list of conditions and the following disclaimer in 
+ *  - Redistributions in binary form must reproduce the above copyright
+ *    notice, this list of conditions and the following disclaimer in
  *    the documentation and/or other materials provided with the distribution.
  *  - Neither the name of the University Catholique de Louvain - UCL
- *    nor the names of its contributors may be used to endorse or 
- *    promote products derived from this software without specific prior 
+ *    nor the names of its contributors may be used to endorse or
+ *    promote products derived from this software without specific prior
  *    written permission.
- *    
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 
- * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE 
- * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 
- * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, 
- * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER 
- * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN 
- * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
+ * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
+ * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
+ * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
+ * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
+ * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
  * POSSIBILITY OF SUCH DAMAGE.
  */
 
@@ -71,10 +71,10 @@ import org.apache.hadoop.util.bloom.Key;
  * one. A given key is said to belong to the DBF if the <code>k</code> positions are set to one in one of the matrix rows.
  * <p>
  * Originally created by <a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.
- * 
+ *
  * @see Filter The general behavior of a filter
  * @see BloomFilter A Bloom filter
- * 
+ *
  * @see <a href="http://www.cse.fau.edu/~jie/research/publications/Publication_files/infocom2006.pdf">Theory and Network Applications of Dynamic Bloom
  *      Filters</a>
  */
@@ -83,27 +83,27 @@ public class DynamicBloomFilter extends Filter {
    * Threshold for the maximum number of key to record in a dynamic Bloom filter row.
    */
   private int nr;
-  
+
   /**
    * The number of keys recorded in the current standard active Bloom filter.
    */
   private int currentNbRecord;
-  
+
   /**
    * The matrix of Bloom filter.
    */
   private BloomFilter[] matrix;
-  
+
   /**
    * Zero-args constructor for the serialization.
    */
   public DynamicBloomFilter() {}
-  
+
   /**
    * Constructor.
    * <p>
    * Builds an empty Dynamic Bloom filter.
-   * 
+   *
    * @param vectorSize
    *          The number of bits in the vector.
    * @param nbHash
@@ -115,83 +115,83 @@ public class DynamicBloomFilter extends Filter {
    */
   public DynamicBloomFilter(final int vectorSize, final int nbHash, final int hashType, final int nr) {
     super(vectorSize, nbHash, hashType);
-    
+
     this.nr = nr;
     this.currentNbRecord = 0;
-    
+
     matrix = new BloomFilter[1];
     matrix[0] = new BloomFilter(this.vectorSize, this.nbHash, this.hashType);
   }
-  
+
   @Override
   public boolean add(final Key key) {
     if (key == null) {
       throw new NullPointerException("Key can not be null");
     }
-    
+
     BloomFilter bf = getActiveStandardBF();
-    
+
     if (bf == null) {
       addRow();
       bf = matrix[matrix.length - 1];
       currentNbRecord = 0;
     }
-    
+
     boolean added = bf.add(key);
-    
+
     if (added)
       currentNbRecord++;
-    
+
     return added;
   }
-  
+
   @Override
   public void and(final Filter filter) {
     if (filter == null || !(filter instanceof DynamicBloomFilter) || filter.vectorSize != this.vectorSize || filter.nbHash != this.nbHash) {
       throw new IllegalArgumentException("filters cannot be and-ed");
     }
-    
+
     DynamicBloomFilter dbf = (DynamicBloomFilter) filter;
-    
+
     if (dbf.matrix.length != this.matrix.length || dbf.nr != this.nr) {
       throw new IllegalArgumentException("filters cannot be and-ed");
     }
-    
+
     for (int i = 0; i < matrix.length; i++) {
       matrix[i].and(dbf.matrix[i]);
     }
   }
-  
+
   @Override
   public boolean membershipTest(final Key key) {
     if (key == null) {
       return true;
     }
-    
+
     for (int i = 0; i < matrix.length; i++) {
       if (matrix[i].membershipTest(key)) {
         return true;
       }
     }
-    
+
     return false;
   }
-  
+
   @Override
   public void not() {
     for (int i = 0; i < matrix.length; i++) {
       matrix[i].not();
     }
   }
-  
+
   @Override
   public void or(final Filter filter) {
     if (filter == null || !(filter instanceof DynamicBloomFilter) || filter.vectorSize != this.vectorSize || filter.nbHash != this.nbHash) {
       throw new IllegalArgumentException("filters cannot be or-ed");
     }
-    
+
     DynamicBloomFilter dbf = (DynamicBloomFilter) filter;
-    
+
     if (dbf.matrix.length != this.matrix.length || dbf.nr != this.nr) {
       throw new IllegalArgumentException("filters cannot be or-ed");
     }
@@ -199,36 +199,36 @@ public class DynamicBloomFilter extends Filter {
       matrix[i].or(dbf.matrix[i]);
     }
   }
-  
+
   @Override
   public void xor(final Filter filter) {
     if (filter == null || !(filter instanceof DynamicBloomFilter) || filter.vectorSize != this.vectorSize || filter.nbHash != this.nbHash) {
       throw new IllegalArgumentException("filters cannot be xor-ed");
     }
     DynamicBloomFilter dbf = (DynamicBloomFilter) filter;
-    
+
     if (dbf.matrix.length != this.matrix.length || dbf.nr != this.nr) {
       throw new IllegalArgumentException("filters cannot be xor-ed");
     }
-    
+
     for (int i = 0; i < matrix.length; i++) {
       matrix[i].xor(dbf.matrix[i]);
     }
   }
-  
+
   @Override
   public String toString() {
     StringBuilder res = new StringBuilder();
-    
+
     for (int i = 0; i < matrix.length; i++) {
       res.append(matrix[i]);
       res.append(Character.LINE_SEPARATOR);
     }
     return res.toString();
   }
-  
+
   // Writable
-  
+
   @Override
   public void write(final DataOutput out) throws IOException {
     super.write(out);
@@ -239,12 +239,12 @@ public class DynamicBloomFilter extends Filter {
       matrix[i].write(out);
     }
   }
-  
+
   @Override
   public void readFields(final DataInput in) throws IOException {
-    
+
     super.readFields(in);
-    
+
     nr = in.readInt();
     currentNbRecord = in.readInt();
     int len = in.readInt();
@@ -254,32 +254,32 @@ public class DynamicBloomFilter extends Filter {
       matrix[i].readFields(in);
     }
   }
-  
+
   /**
    * Adds a new row to <i>this</i> dynamic Bloom filter.
    */
   private void addRow() {
     BloomFilter[] tmp = new BloomFilter[matrix.length + 1];
-    
+
     for (int i = 0; i < matrix.length; i++) {
       tmp[i] = matrix[i];
     }
-    
+
     tmp[tmp.length - 1] = new BloomFilter(vectorSize, nbHash, hashType);
-    
+
     matrix = tmp;
   }
-  
+
   /**
    * Returns the active standard Bloom filter in <i>this</i> dynamic Bloom filter.
-   * 
+   *
    * @return BloomFilter The active standard Bloom filter. <code>Null</code> otherwise.
    */
   private BloomFilter getActiveStandardBF() {
     if (currentNbRecord >= nr) {
       return null;
     }
-    
+
     return matrix[matrix.length - 1];
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/bloomfilter/Filter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/bloomfilter/Filter.java b/core/src/main/java/org/apache/accumulo/core/bloomfilter/Filter.java
index eec1c51..12961a9 100644
--- a/core/src/main/java/org/apache/accumulo/core/bloomfilter/Filter.java
+++ b/core/src/main/java/org/apache/accumulo/core/bloomfilter/Filter.java
@@ -2,32 +2,32 @@
  *
  * Copyright (c) 2005, European Commission project OneLab under contract 034819
  * (http://www.one-lab.org)
- * 
+ *
  * All rights reserved.
- * Redistribution and use in source and binary forms, with or 
- * without modification, are permitted provided that the following 
+ * Redistribution and use in source and binary forms, with or
+ * without modification, are permitted provided that the following
  * conditions are met:
- *  - Redistributions of source code must retain the above copyright 
+ *  - Redistributions of source code must retain the above copyright
  *    notice, this list of conditions and the following disclaimer.
- *  - Redistributions in binary form must reproduce the above copyright 
- *    notice, this list of conditions and the following disclaimer in 
+ *  - Redistributions in binary form must reproduce the above copyright
+ *    notice, this list of conditions and the following disclaimer in
  *    the documentation and/or other materials provided with the distribution.
  *  - Neither the name of the University Catholique de Louvain - UCL
- *    nor the names of its contributors may be used to endorse or 
- *    promote products derived from this software without specific prior 
+ *    nor the names of its contributors may be used to endorse or
+ *    promote products derived from this software without specific prior
  *    written permission.
- *    
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 
- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 
- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 
- * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE 
- * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 
- * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, 
- * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER 
- * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 
- * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN 
- * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
+ * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
+ * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
+ * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
+ * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
+ * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
+ * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
  * POSSIBILITY OF SUCH DAMAGE.
  */
 
@@ -70,7 +70,7 @@ import org.apache.hadoop.util.hash.Hash;
  * Typically, a filter will be implemented as a Bloom filter (or a Bloom filter extension).
  * <p>
  * It must be extended in order to define the real behavior.
- * 
+ *
  * @see Key The general behavior of a key
  * @see HashFunction A hash function
  */
@@ -78,22 +78,22 @@ public abstract class Filter implements Writable {
   private static final int VERSION = -2; // negative to accommodate for old format
   /** The vector size of <i>this</i> filter. */
   protected int vectorSize;
-  
+
   private int rVersion;
   /** The hash function used to map a key to several positions in the vector. */
   protected HashFunction hash;
-  
+
   /** The number of hash function to consider. */
   protected int nbHash;
-  
+
   /** Type of hashing function to use. */
   protected int hashType;
-  
+
   protected Filter() {}
-  
+
   /**
    * Constructor.
-   * 
+   *
    * @param vectorSize
    *          The vector size of <i>this</i> filter.
    * @param nbHash
@@ -107,65 +107,65 @@ public abstract class Filter implements Writable {
     this.hashType = hashType;
     this.hash = new HashFunction(this.vectorSize, this.nbHash, this.hashType);
   }
-  
+
   /**
    * Adds a key to <i>this</i> filter.
-   * 
+   *
    * @param key
    *          The key to add.
    * @return true if the key was added, false otherwise.
    */
   public abstract boolean add(Key key);
-  
+
   /**
    * Determines whether a specified key belongs to <i>this</i> filter.
-   * 
+   *
    * @param key
    *          The key to test.
    * @return boolean True if the specified key belongs to <i>this</i> filter. False otherwise.
    */
   public abstract boolean membershipTest(Key key);
-  
+
   /**
    * Peforms a logical AND between <i>this</i> filter and a specified filter.
    * <p>
    * <b>Invariant</b>: The result is assigned to <i>this</i> filter.
-   * 
+   *
    * @param filter
    *          The filter to AND with.
    */
   public abstract void and(Filter filter);
-  
+
   /**
    * Peforms a logical OR between <i>this</i> filter and a specified filter.
    * <p>
    * <b>Invariant</b>: The result is assigned to <i>this</i> filter.
-   * 
+   *
    * @param filter
    *          The filter to OR with.
    */
   public abstract void or(Filter filter);
-  
+
   /**
    * Peforms a logical XOR between <i>this</i> filter and a specified filter.
    * <p>
    * <b>Invariant</b>: The result is assigned to <i>this</i> filter.
-   * 
+   *
    * @param filter
    *          The filter to XOR with.
    */
   public abstract void xor(Filter filter);
-  
+
   /**
    * Performs a logical NOT on <i>this</i> filter.
    * <p>
    * The result is assigned to <i>this</i> filter.
    */
   public abstract void not();
-  
+
   /**
    * Adds a list of keys to <i>this</i> filter.
-   * 
+   *
    * @param keys
    *          The list of keys.
    */
@@ -173,15 +173,15 @@ public abstract class Filter implements Writable {
     if (keys == null) {
       throw new IllegalArgumentException("ArrayList<Key> may not be null");
     }
-    
+
     for (Key key : keys) {
       add(key);
     }
   }// end add()
-  
+
   /**
    * Adds a collection of keys to <i>this</i> filter.
-   * 
+   *
    * @param keys
    *          The collection of keys.
    */
@@ -193,10 +193,10 @@ public abstract class Filter implements Writable {
       add(key);
     }
   }// end add()
-  
+
   /**
    * Adds an array of keys to <i>this</i> filter.
-   * 
+   *
    * @param keys
    *          The array of keys.
    */
@@ -208,31 +208,31 @@ public abstract class Filter implements Writable {
       add(keys[i]);
     }
   }// end add()
-  
+
   // Writable interface
-  
+
   public void write(final DataOutput out) throws IOException {
     out.writeInt(VERSION);
     out.writeInt(this.nbHash);
     out.writeByte(this.hashType);
     out.writeInt(this.vectorSize);
   }
-  
+
   protected int getSerialVersion() {
     return rVersion;
   }
-  
+
   protected int getVersion() {
     return VERSION;
   }
-  
+
   public void readFields(final DataInput in) throws IOException {
     final int ver = in.readInt();
     rVersion = ver;
     if (ver > 0) { // old unversioned format
       this.nbHash = ver;
       this.hashType = Hash.JENKINS_HASH;
-      
+
     } else if (ver == VERSION | ver == VERSION + 1) { // Support for directly serialzing the bitset
       this.nbHash = in.readInt();
       this.hashType = in.readByte();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/BatchScannerOpts.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/BatchScannerOpts.java b/core/src/main/java/org/apache/accumulo/core/cli/BatchScannerOpts.java
index 4503df4..2d5b51d 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/BatchScannerOpts.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/BatchScannerOpts.java
@@ -21,10 +21,10 @@ import org.apache.accumulo.core.cli.ClientOpts.TimeConverter;
 import com.beust.jcommander.Parameter;
 
 public class BatchScannerOpts {
-  @Parameter(names="--scanThreads", description="Number of threads to use when batch scanning")
+  @Parameter(names = "--scanThreads", description = "Number of threads to use when batch scanning")
   public Integer scanThreads = 10;
-  
-  @Parameter(names="--scanTimeout", converter=TimeConverter.class, description="timeout used to fail a batch scan")
+
+  @Parameter(names = "--scanTimeout", converter = TimeConverter.class, description = "timeout used to fail a batch scan")
   public Long scanTimeout = Long.MAX_VALUE;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/BatchWriterOpts.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/BatchWriterOpts.java b/core/src/main/java/org/apache/accumulo/core/cli/BatchWriterOpts.java
index 45a1bc7..02fb3da 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/BatchWriterOpts.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/BatchWriterOpts.java
@@ -26,19 +26,19 @@ import com.beust.jcommander.Parameter;
 
 public class BatchWriterOpts {
   private static final BatchWriterConfig BWDEFAULTS = new BatchWriterConfig();
-  
-  @Parameter(names="--batchThreads", description="Number of threads to use when writing large batches")
+
+  @Parameter(names = "--batchThreads", description = "Number of threads to use when writing large batches")
   public Integer batchThreads = BWDEFAULTS.getMaxWriteThreads();
 
-  @Parameter(names="--batchLatency", converter=TimeConverter.class, description="The maximum time to wait before flushing data to servers when writing")
+  @Parameter(names = "--batchLatency", converter = TimeConverter.class, description = "The maximum time to wait before flushing data to servers when writing")
   public Long batchLatency = BWDEFAULTS.getMaxLatency(TimeUnit.MILLISECONDS);
-  
-  @Parameter(names="--batchMemory", converter=MemoryConverter.class, description="memory used to batch data when writing")
+
+  @Parameter(names = "--batchMemory", converter = MemoryConverter.class, description = "memory used to batch data when writing")
   public Long batchMemory = BWDEFAULTS.getMaxMemory();
-  
-  @Parameter(names="--batchTimeout", converter=TimeConverter.class, description="timeout used to fail a batch write")
+
+  @Parameter(names = "--batchTimeout", converter = TimeConverter.class, description = "timeout used to fail a batch write")
   public Long batchTimeout = BWDEFAULTS.getTimeout(TimeUnit.MILLISECONDS);
-  
+
   public BatchWriterConfig getBatchWriterConfig() {
     BatchWriterConfig config = new BatchWriterConfig();
     config.setMaxWriteThreads(this.batchThreads);
@@ -47,5 +47,5 @@ public class BatchWriterOpts {
     config.setTimeout(this.batchTimeout, TimeUnit.MILLISECONDS);
     return config;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/ClientOnDefaultTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/ClientOnDefaultTable.java b/core/src/main/java/org/apache/accumulo/core/cli/ClientOnDefaultTable.java
index d12f4a5..42dec8f 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/ClientOnDefaultTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/ClientOnDefaultTable.java
@@ -25,7 +25,7 @@ public class ClientOnDefaultTable extends ClientOpts {
   public ClientOnDefaultTable(String table) {
     this.tableName = table;
   }
-  
+
   public String getTableName() {
     return tableName;
   }
@@ -33,5 +33,5 @@ public class ClientOnDefaultTable extends ClientOpts {
   public void setTableName(String tableName) {
     this.tableName = tableName;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/ClientOnRequiredTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/ClientOnRequiredTable.java b/core/src/main/java/org/apache/accumulo/core/cli/ClientOnRequiredTable.java
index e6d331c..e41e351 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/ClientOnRequiredTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/ClientOnRequiredTable.java
@@ -18,7 +18,6 @@ package org.apache.accumulo.core.cli;
 
 import com.beust.jcommander.Parameter;
 
-
 public class ClientOnRequiredTable extends ClientOpts {
   @Parameter(names = {"-t", "--table"}, required = true, description = "table to use")
   private String tableName;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java b/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
index f6ea934..fa690e6 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/ClientOpts.java
@@ -55,77 +55,77 @@ import com.beust.jcommander.IStringConverter;
 import com.beust.jcommander.Parameter;
 
 public class ClientOpts extends Help {
-  
+
   public static class TimeConverter implements IStringConverter<Long> {
     @Override
     public Long convert(String value) {
       return AccumuloConfiguration.getTimeInMillis(value);
     }
   }
-  
+
   public static class MemoryConverter implements IStringConverter<Long> {
     @Override
     public Long convert(String value) {
       return AccumuloConfiguration.getMemoryInBytes(value);
     }
   }
-  
+
   public static class AuthConverter implements IStringConverter<Authorizations> {
     @Override
     public Authorizations convert(String value) {
       return new Authorizations(value.split(","));
     }
   }
-  
+
   public static class Password {
     public byte[] value;
-    
+
     public Password(String dfault) {
       value = dfault.getBytes(UTF_8);
     }
-    
+
     @Override
     public String toString() {
       return new String(value, UTF_8);
     }
   }
-  
+
   public static class PasswordConverter implements IStringConverter<Password> {
     @Override
     public Password convert(String value) {
       return new Password(value);
     }
   }
-  
+
   public static class VisibilityConverter implements IStringConverter<ColumnVisibility> {
     @Override
     public ColumnVisibility convert(String value) {
       return new ColumnVisibility(value);
     }
   }
-  
+
   @Parameter(names = {"-u", "--user"}, description = "Connection user")
   public String principal = System.getProperty("user.name");
-  
+
   @Parameter(names = "-p", converter = PasswordConverter.class, description = "Connection password")
   public Password password = null;
-  
+
   @Parameter(names = "--password", converter = PasswordConverter.class, description = "Enter the connection password", password = true)
   public Password securePassword = null;
-  
+
   @Parameter(names = {"-tc", "--tokenClass"}, description = "Token class")
   public String tokenClassName = PasswordToken.class.getName();
-  
+
   @DynamicParameter(names = "-l",
       description = "login properties in the format key=value. Reuse -l for each property (prompt for properties if this option is missing")
   public Map<String,String> loginProps = new LinkedHashMap<String,String>();
-  
+
   public AuthenticationToken getToken() {
     if (!loginProps.isEmpty()) {
       Properties props = new Properties();
       for (Entry<String,String> loginOption : loginProps.entrySet())
         props.put(loginOption.getKey(), loginOption.getValue());
-      
+
       try {
         AuthenticationToken token = Class.forName(tokenClassName).asSubclass(AuthenticationToken.class).newInstance();
         token.init(props);
@@ -133,70 +133,72 @@ public class ClientOpts extends Help {
       } catch (Exception e) {
         throw new RuntimeException(e);
       }
-      
+
     }
-    
+
     if (securePassword != null)
       return new PasswordToken(securePassword.value);
-    
+
     if (password != null)
       return new PasswordToken(password.value);
-    
+
     return null;
   }
-  
+
   @Parameter(names = {"-z", "--keepers"}, description = "Comma separated list of zookeeper hosts (host:port,host:port)")
   public String zookeepers = "localhost:2181";
-  
+
   @Parameter(names = {"-i", "--instance"}, description = "The name of the accumulo instance")
   public String instance = null;
-  
+
   @Parameter(names = {"-auths", "--auths"}, converter = AuthConverter.class, description = "the authorizations to use when reading or writing")
   public Authorizations auths = Authorizations.EMPTY;
-  
+
   @Parameter(names = "--debug", description = "turn on TRACE-level log messages")
   public boolean debug = false;
-  
+
   @Parameter(names = {"-fake", "--mock"}, description = "Use a mock Instance")
   public boolean mock = false;
-  
+
   @Parameter(names = "--site-file", description = "Read the given accumulo site file to find the accumulo instance")
   public String siteFile = null;
-  
+
   @Parameter(names = "--ssl", description = "Connect to accumulo over SSL")
   public boolean sslEnabled = false;
 
-  @Parameter(names = "--config-file", description = "Read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
+  @Parameter(
+      names = "--config-file",
+      description = "Read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
   public String clientConfigFile = null;
 
   public void startDebugLogging() {
     if (debug)
       Logger.getLogger(Constants.CORE_PACKAGE_NAME).setLevel(Level.TRACE);
   }
-  
+
   @Parameter(names = "--trace", description = "turn on distributed tracing")
   public boolean trace = false;
-  
+
   public void startTracing(String applicationName) {
     if (trace) {
       Trace.on(applicationName);
     }
   }
-  
+
   public void stopTracing() {
     Trace.off();
   }
-  
+
   @Override
   public void parseArgs(String programName, String[] args, Object... others) {
     super.parseArgs(programName, args, others);
     startDebugLogging();
     startTracing(programName);
   }
-  
+
   protected Instance cachedInstance = null;
   protected ClientConfiguration cachedClientConfig = null;
-  
+
   synchronized public Instance getInstance() {
     if (cachedInstance != null)
       return cachedInstance;
@@ -232,7 +234,7 @@ public class ClientOpts extends Help {
         {
           xml.addResource(new Path(siteFile));
         }
-        
+
         @Override
         public void getProperties(Map<String,String> props, PropertyFilter filter) {
           for (Entry<String,String> prop : DefaultConfiguration.getInstance())
@@ -242,7 +244,7 @@ public class ClientOpts extends Help {
             if (filter.accept(prop.getKey()))
               props.put(prop.getKey(), prop.getValue());
         }
-        
+
         @Override
         public String get(Property property) {
           String value = xml.get(property.getKey());
@@ -262,5 +264,5 @@ public class ClientOpts extends Help {
     }
     return cachedClientConfig = clientConfig.withInstance(instance).withZkHosts(zookeepers);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/Help.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/Help.java b/core/src/main/java/org/apache/accumulo/core/cli/Help.java
index b5b16bc..68475e2 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/Help.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/Help.java
@@ -21,10 +21,10 @@ import com.beust.jcommander.Parameter;
 import com.beust.jcommander.ParameterException;
 
 public class Help {
-  @Parameter(names={"-h", "-?", "--help", "-help"}, help=true)
+  @Parameter(names = {"-h", "-?", "--help", "-help"}, help = true)
   public boolean help = false;
-  
-  public void parseArgs(String programName, String[] args, Object ... others) {
+
+  public void parseArgs(String programName, String[] args, Object... others) {
     JCommander commander = new JCommander();
     commander.addObject(this);
     for (Object other : others)
@@ -41,11 +41,11 @@ public class Help {
       exit(0);
     }
   }
-  
+
   public void exit(int status) {
     System.exit(status);
   }
-  
+
   public void exitWithError(String message, int status) {
     System.err.println(message);
     exit(status);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/cli/ScannerOpts.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/cli/ScannerOpts.java b/core/src/main/java/org/apache/accumulo/core/cli/ScannerOpts.java
index 54e3619..439b374 100644
--- a/core/src/main/java/org/apache/accumulo/core/cli/ScannerOpts.java
+++ b/core/src/main/java/org/apache/accumulo/core/cli/ScannerOpts.java
@@ -19,6 +19,6 @@ package org.apache.accumulo.core.cli;
 import com.beust.jcommander.Parameter;
 
 public class ScannerOpts {
-  @Parameter(names="--scanBatchSize", description="the number of key-values to pull during a scan")
-  public int scanBatchSize = 1000; 
+  @Parameter(names = "--scanBatchSize", description = "the number of key-values to pull during a scan")
+  public int scanBatchSize = 1000;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/AccumuloException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/AccumuloException.java b/core/src/main/java/org/apache/accumulo/core/client/AccumuloException.java
index 9de0b96..97bfebd 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/AccumuloException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/AccumuloException.java
@@ -18,12 +18,12 @@ package org.apache.accumulo.core.client;
 
 /**
  * A generic Accumulo Exception for general accumulo failures.
- * 
+ *
  */
 public class AccumuloException extends Exception {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   /**
    * @param why
    *          is the reason for the error being thrown
@@ -31,7 +31,7 @@ public class AccumuloException extends Exception {
   public AccumuloException(final String why) {
     super(why);
   }
-  
+
   /**
    * @param cause
    *          is the exception that this exception wraps
@@ -39,7 +39,7 @@ public class AccumuloException extends Exception {
   public AccumuloException(final Throwable cause) {
     super(cause);
   }
-  
+
   /**
    * @param why
    *          is the reason for the error being thrown
@@ -49,5 +49,5 @@ public class AccumuloException extends Exception {
   public AccumuloException(final String why, final Throwable cause) {
     super(why, cause);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/AccumuloSecurityException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/AccumuloSecurityException.java b/core/src/main/java/org/apache/accumulo/core/client/AccumuloSecurityException.java
index 35ea188..a614e8c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/AccumuloSecurityException.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/AccumuloSecurityException.java
@@ -22,11 +22,11 @@ import org.apache.commons.lang.StringUtils;
 
 /**
  * An Accumulo Exception for security violations, authentication failures, authorization failures, etc.
- * 
+ *
  */
 public class AccumuloSecurityException extends Exception {
   private static final long serialVersionUID = 1L;
-  
+
   private static String getDefaultErrorMessage(final SecurityErrorCode errorcode) {
     switch (errorcode == null ? SecurityErrorCode.DEFAULT_SECURITY_ERROR : errorcode) {
       case BAD_CREDENTIALS:
@@ -62,21 +62,23 @@ public class AccumuloSecurityException extends Exception {
         return "Unknown security exception";
     }
   }
-  
+
   private String user;
   private String tableInfo;
   private SecurityErrorCode errorCode;
-  
+
   /**
    * @return this exception as a thrift exception
    */
   public ThriftSecurityException asThriftException() {
     return new ThriftSecurityException(user, errorCode);
   }
-  
+
   /**
    * Construct a user-facing exception from a serialized version.
-   * @param thrift a serialized version
+   *
+   * @param thrift
+   *          a serialized version
    */
   public AccumuloSecurityException(final ThriftSecurityException thrift) {
     this(thrift.getUser(), thrift.getCode(), thrift);
@@ -95,7 +97,7 @@ public class AccumuloSecurityException extends Exception {
     this.user = user;
     this.errorCode = errorcode == null ? SecurityErrorCode.DEFAULT_SECURITY_ERROR : errorcode;
   }
-  
+
   /**
    * @param user
    *          the relevant user for the security violation
@@ -112,7 +114,7 @@ public class AccumuloSecurityException extends Exception {
     this.errorCode = errorcode == null ? SecurityErrorCode.DEFAULT_SECURITY_ERROR : errorcode;
     this.tableInfo = tableInfo;
   }
-  
+
   /**
    * @param user
    *          the relevant user for the security violation
@@ -124,7 +126,7 @@ public class AccumuloSecurityException extends Exception {
     this.user = user;
     this.errorCode = errorcode == null ? SecurityErrorCode.DEFAULT_SECURITY_ERROR : errorcode;
   }
-  
+
   /**
    * @param user
    *          the relevant user for the security violation
@@ -139,38 +141,38 @@ public class AccumuloSecurityException extends Exception {
     this.errorCode = errorcode == null ? SecurityErrorCode.DEFAULT_SECURITY_ERROR : errorcode;
     this.tableInfo = tableInfo;
   }
-  
+
   /**
    * @return the relevant user for the security violation
    */
   public String getUser() {
     return user;
   }
-  
+
   public void setUser(String s) {
     this.user = s;
   }
-  
+
   /**
    * @return the relevant tableInfo for the security violation
    */
   public String getTableInfo() {
     return tableInfo;
   }
-  
+
   public void setTableInfo(String tableInfo) {
     this.tableInfo = tableInfo;
   }
-  
+
   /**
    * @return the specific reason for this exception
    * @since 1.5.0
    */
-  
+
   public org.apache.accumulo.core.client.security.SecurityErrorCode getSecurityErrorCode() {
     return org.apache.accumulo.core.client.security.SecurityErrorCode.valueOf(errorCode.name());
   }
-  
+
   @Override
   public String getMessage() {
     StringBuilder message = new StringBuilder();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/BatchDeleter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/BatchDeleter.java b/core/src/main/java/org/apache/accumulo/core/client/BatchDeleter.java
index 2bfc347..54d49d1 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/BatchDeleter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/BatchDeleter.java
@@ -22,28 +22,28 @@ import org.apache.accumulo.core.data.Range;
 
 /**
  * Implementations of BatchDeleter support efficient deletion of ranges in accumulo.
- * 
+ *
  */
 
 public interface BatchDeleter extends ScannerBase {
   /**
    * Deletes the ranges specified by {@link #setRanges}.
-   * 
+   *
    * @throws MutationsRejectedException
    *           this can be thrown when deletion mutations fail
    * @throws TableNotFoundException
    *           when the table does not exist
    */
   void delete() throws MutationsRejectedException, TableNotFoundException;
-  
+
   /**
    * Allows deleting multiple ranges efficiently.
-   * 
+   *
    * @param ranges
    *          specifies the non-overlapping ranges to query
    */
   void setRanges(Collection<Range> ranges);
-  
+
   /**
    * Releases any resources.
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/BatchScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/BatchScanner.java b/core/src/main/java/org/apache/accumulo/core/client/BatchScanner.java
index aa57297..39f96f5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/BatchScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/BatchScanner.java
@@ -23,37 +23,37 @@ import org.apache.accumulo.core.data.Range;
 
 /**
  * Implementations of BatchScanner support efficient lookups of many ranges in accumulo.
- * 
+ *
  * Use this when looking up lots of ranges and you expect each range to contain a small amount of data. Also only use this when you do not care about the
  * returned data being in sorted order.
- * 
+ *
  * If you want to lookup a few ranges and expect those ranges to contain a lot of data, then use the Scanner instead. Also, the Scanner will return data in
  * sorted order, this will not.
  */
 
 public interface BatchScanner extends ScannerBase {
-  
+
   /**
    * Allows scanning over multiple ranges efficiently.
-   * 
+   *
    * @param ranges
    *          specifies the non-overlapping ranges to query
    */
   void setRanges(Collection<Range> ranges);
-  
+
   /**
    * Cleans up and finalizes the scanner
    */
   void close();
-  
+
   /**
    * Sets a timeout threshold for a server to respond. The batch scanner will accomplish as much work as possible before throwing an exception. BatchScanner
    * iterators will throw a {@link TimedOutException} when all needed servers timeout. Setting the timeout to zero or Long.MAX_VALUE and TimeUnit.MILLISECONDS
    * means no timeout.
-   * 
+   *
    * <p>
    * If not set, there is not timeout. The BatchScanner will retry forever.
-   * 
+   *
    * @param timeUnit
    *          determines how timeout is interpreted
    * @since 1.5.0

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/BatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/BatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/BatchWriter.java
index b321411..94d988d 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/BatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/BatchWriter.java
@@ -20,52 +20,49 @@ import org.apache.accumulo.core.data.Mutation;
 
 /**
  * Send Mutations to a single Table in Accumulo.
- * 
- * When the user uses a Connector to create a BatchWriter,
- * they specify how much memory and how many threads it should use. 
- * As the user adds mutations to the batch writer, it buffers them. 
- * Once the buffered mutations have used half of the user specified buffer, 
- * the mutations are dumped into the background to be written by a thread pool.
- * If the user specified memory completely fills up, then writes are held.
- * When a user calls flush, it does not return until all buffered mutations are written.
+ *
+ * When the user uses a Connector to create a BatchWriter, they specify how much memory and how many threads it should use. As the user adds mutations to the
+ * batch writer, it buffers them. Once the buffered mutations have used half of the user specified buffer, the mutations are dumped into the background to be
+ * written by a thread pool. If the user specified memory completely fills up, then writes are held. When a user calls flush, it does not return until all
+ * buffered mutations are written.
  */
 public interface BatchWriter {
-  
+
   /**
    * Queues one mutation to write.
-   * 
+   *
    * @param m
    *          the mutation to add
    * @throws MutationsRejectedException
    *           this could be thrown because current or previous mutations failed
    */
-  
+
   void addMutation(Mutation m) throws MutationsRejectedException;
-  
+
   /**
    * Queues several mutations to write.
-   * 
+   *
    * @param iterable
    *          allows adding any number of mutations iteratively
    * @throws MutationsRejectedException
    *           this could be thrown because current or previous mutations failed
    */
   void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException;
-  
+
   /**
    * Send any buffered mutations to Accumulo immediately.
-   * 
+   *
    * @throws MutationsRejectedException
    *           this could be thrown because current or previous mutations failed
    */
   void flush() throws MutationsRejectedException;
-  
+
   /**
    * Flush and release any resources.
-   * 
+   *
    * @throws MutationsRejectedException
    *           this could be thrown because current or previous mutations failed
    */
   void close() throws MutationsRejectedException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/BatchWriterConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/BatchWriterConfig.java b/core/src/main/java/org/apache/accumulo/core/client/BatchWriterConfig.java
index ecb031b..6ceefad 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/BatchWriterConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/BatchWriterConfig.java
@@ -180,11 +180,11 @@ public class BatchWriterConfig implements Writable {
   }
 
   /**
-   * Change the durability for the BatchWriter session. The default durability is "default" which
-   * is the table's durability setting.  If the durability is set to something other than the default,
-   * it will override the durability setting of the table.
+   * Change the durability for the BatchWriter session. The default durability is "default" which is the table's durability setting. If the durability is set to
+   * something other than the default, it will override the durability setting of the table.
    *
-   * @param durability the Durability to be used by the BatchWriter
+   * @param durability
+   *          the Durability to be used by the BatchWriter
    * @since 1.7.0
    *
    */
@@ -320,8 +320,7 @@ public class BatchWriterConfig implements Writable {
   public String toString() {
     StringBuilder sb = new StringBuilder(32);
     sb.append("[maxMemory=").append(getMaxMemory()).append(", maxLatency=").append(getMaxLatency(TimeUnit.MILLISECONDS)).append(", maxWriteThreads=")
-        .append(getMaxWriteThreads()).append(", timeout=").append(getTimeout(TimeUnit.MILLISECONDS))
-        .append(", durability=").append(durability).append("]");
+        .append(getMaxWriteThreads()).append(", timeout=").append(getTimeout(TimeUnit.MILLISECONDS)).append(", durability=").append(durability).append("]");
     return sb.toString();
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/ClientConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/ClientConfiguration.java b/core/src/main/java/org/apache/accumulo/core/client/ClientConfiguration.java
index 6fe61a5..df53645 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/ClientConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/ClientConfiguration.java
@@ -17,6 +17,7 @@
 package org.apache.accumulo.core.client;
 
 import static com.google.common.base.Preconditions.checkArgument;
+
 import java.io.File;
 import java.io.StringReader;
 import java.io.StringWriter;
@@ -224,9 +225,11 @@ public class ClientConfiguration extends CompositeConfiguration {
   /**
    * Gets all properties under the given prefix in this configuration.
    *
-   * @param property prefix property, must be of type PropertyType.PREFIX
+   * @param property
+   *          prefix property, must be of type PropertyType.PREFIX
    * @return a map of property keys to values
-   * @throws IllegalArgumentException if property is not a prefix
+   * @throws IllegalArgumentException
+   *           if property is not a prefix
    */
   public Map<String,String> getAllPropertiesWithPrefix(ClientProperty property) {
     checkType(property, PropertyType.PREFIX);
@@ -234,7 +237,7 @@ public class ClientConfiguration extends CompositeConfiguration {
     Map<String,String> propMap = new HashMap<String,String>();
     Iterator<?> iter = this.getKeys(property.getKey());
     while (iter.hasNext()) {
-      String p = (String)iter.next();
+      String p = (String) iter.next();
       propMap.put(p, getString(p));
     }
     return propMap;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/ClientSideIteratorScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/ClientSideIteratorScanner.java b/core/src/main/java/org/apache/accumulo/core/client/ClientSideIteratorScanner.java
index 4903656..d6261fb 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/ClientSideIteratorScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/ClientSideIteratorScanner.java
@@ -55,11 +55,11 @@ import org.apache.hadoop.io.Text;
  */
 public class ClientSideIteratorScanner extends ScannerOptions implements Scanner {
   private int size;
-  
+
   private Range range;
   private boolean isolated = false;
   private long readaheadThreshold = Constants.SCANNER_DEFAULT_READAHEAD_THRESHOLD;
-  
+
   /**
    * A class that wraps a Scanner in a SortedKeyValueIterator so that other accumulo iterators can use it as a source.
    */
@@ -67,27 +67,27 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
     protected Scanner scanner;
     Iterator<Entry<Key,Value>> iter;
     Entry<Key,Value> top = null;
-    
+
     /**
      * Constructs an accumulo iterator from a scanner.
-     * 
+     *
      * @param scanner
      *          the scanner to iterate over
      */
     public ScannerTranslator(final Scanner scanner) {
       this.scanner = scanner;
     }
-    
+
     @Override
     public void init(final SortedKeyValueIterator<Key,Value> source, final Map<String,String> options, final IteratorEnvironment env) throws IOException {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public boolean hasTop() {
       return top != null;
     }
-    
+
     @Override
     public void next() throws IOException {
       if (iter.hasNext())
@@ -95,7 +95,7 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
       else
         top = null;
     }
-    
+
     @Override
     public void seek(final Range range, final Collection<ByteSequence> columnFamilies, final boolean inclusive) throws IOException {
       if (!inclusive && columnFamilies.size() > 0) {
@@ -109,28 +109,28 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
       iter = scanner.iterator();
       next();
     }
-    
+
     @Override
     public Key getTopKey() {
       return top.getKey();
     }
-    
+
     @Override
     public Value getTopValue() {
       return top.getValue();
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> deepCopy(final IteratorEnvironment env) {
       return new ScannerTranslator(scanner);
     }
   }
-  
+
   private ScannerTranslator smi;
-  
+
   /**
    * Constructs a scanner that can execute client-side iterators.
-   * 
+   *
    * @param scanner
    *          the source scanner
    */
@@ -141,14 +141,14 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
     this.timeOut = scanner.getTimeout(TimeUnit.MILLISECONDS);
     this.readaheadThreshold = scanner.getReadaheadThreshold();
   }
-  
+
   /**
    * Sets the source Scanner.
    */
   public void setSource(final Scanner scanner) {
     smi = new ScannerTranslator(scanner);
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     smi.scanner.setBatchSize(size);
@@ -158,13 +158,13 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
       smi.scanner.enableIsolation();
     else
       smi.scanner.disableIsolation();
-    
+
     final TreeMap<Integer,IterInfo> tm = new TreeMap<Integer,IterInfo>();
-    
+
     for (IterInfo iterInfo : serverSideIteratorList) {
       tm.put(iterInfo.getPriority(), iterInfo);
     }
-    
+
     SortedKeyValueIterator<Key,Value> skvi;
     try {
       skvi = IteratorUtil.loadIterators(smi, tm.values(), serverSideIteratorOptions, new IteratorEnvironment() {
@@ -172,43 +172,43 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
         public SortedKeyValueIterator<Key,Value> reserveMapFileReader(final String mapFileName) throws IOException {
           return null;
         }
-        
+
         @Override
         public AccumuloConfiguration getConfig() {
           return null;
         }
-        
+
         @Override
         public IteratorScope getIteratorScope() {
           return null;
         }
-        
+
         @Override
         public boolean isFullMajorCompaction() {
           return false;
         }
-        
+
         @Override
         public void registerSideChannel(final SortedKeyValueIterator<Key,Value> iter) {}
       }, false, null);
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
     final Set<ByteSequence> colfs = new TreeSet<ByteSequence>();
     for (Column c : this.getFetchedColumns()) {
       colfs.add(new ArrayByteSequence(c.getColumnFamily()));
     }
-    
+
     try {
       skvi.seek(range, colfs, true);
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
     return new IteratorAdapter(skvi);
   }
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {
@@ -217,7 +217,7 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
     else
       setTimeout(timeOut, TimeUnit.SECONDS);
   }
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
@@ -226,32 +226,32 @@ public class ClientSideIteratorScanner extends ScannerOptions implements Scanner
       return Integer.MAX_VALUE;
     return (int) timeout;
   }
-  
+
   @Override
   public void setRange(final Range range) {
     this.range = range;
   }
-  
+
   @Override
   public Range getRange() {
     return range;
   }
-  
+
   @Override
   public void setBatchSize(final int size) {
     this.size = size;
   }
-  
+
   @Override
   public int getBatchSize() {
     return size;
   }
-  
+
   @Override
   public void enableIsolation() {
     this.isolated = true;
   }
-  
+
   @Override
   public void disableIsolation() {
     this.isolated = false;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriter.java b/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriter.java
index f9848c4..62244e6 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriter.java
@@ -25,23 +25,23 @@ import org.apache.accumulo.core.data.ConditionalMutation;
 /**
  * ConditionalWriter provides the ability to do efficient, atomic read-modify-write operations on rows. These operations are performed on the tablet server
  * while a row lock is held.
- * 
+ *
  * @since 1.6.0
  */
 public interface ConditionalWriter {
   class Result {
-    
+
     private Status status;
     private ConditionalMutation mutation;
     private String server;
     private Exception exception;
-    
+
     public Result(Status s, ConditionalMutation m, String server) {
       this.status = s;
       this.mutation = m;
       this.server = server;
     }
-    
+
     public Result(Exception e, ConditionalMutation cm, String server) {
       this.exception = e;
       this.mutation = cm;
@@ -51,7 +51,7 @@ public interface ConditionalWriter {
     /**
      * If this method throws an exception, then its possible the mutation is still being actively processed. Therefore if code chooses to continue after seeing
      * an exception it should take this into consideration.
-     * 
+     *
      * @return status of a conditional mutation
      */
 
@@ -62,31 +62,30 @@ public interface ConditionalWriter {
         if (exception instanceof AccumuloSecurityException) {
           AccumuloSecurityException ase = (AccumuloSecurityException) exception;
           throw new AccumuloSecurityException(ase.getUser(), SecurityErrorCode.valueOf(ase.getSecurityErrorCode().name()), ase.getTableInfo(), ase);
-        }
-        else
+        } else
           throw new AccumuloException(exception);
       }
 
       return status;
     }
-    
+
     /**
-     * 
+     *
      * @return A copy of the mutation previously submitted by a user. The mutation will reference the same data, but the object may be different.
      */
     public ConditionalMutation getMutation() {
       return mutation;
     }
-    
+
     /**
-     * 
+     *
      * @return The server this mutation was sent to. Returns null if was not sent to a server.
      */
     public String getTabletServer() {
       return server;
     }
   }
-  
+
   public static enum Status {
     /**
      * conditions were met and mutation was written
@@ -115,15 +114,15 @@ public interface ConditionalWriter {
   /**
    * This method returns one result for each mutation passed to it. This method is thread safe. Multiple threads can safely use a single conditional writer.
    * Sharing a conditional writer between multiple threads may result in batching of request to tablet servers.
-   * 
+   *
    * @return Result for each mutation submitted. The mutations may still be processing in the background when this method returns, if so the iterator will
    *         block.
    */
   Iterator<Result> write(Iterator<ConditionalMutation> mutations);
-  
+
   /**
    * This method has the same thread safety guarantees as @link {@link #write(Iterator)}
-   * 
+   *
    * @return Result for the submitted mutation
    */
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriterConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriterConfig.java b/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriterConfig.java
index 52c6a76..b5cb474 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriterConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/ConditionalWriterConfig.java
@@ -109,6 +109,7 @@ public class ConditionalWriterConfig {
    * Sets the Durability for the mutation, if applied.
    * <p>
    * <b>Default:</b> Durability.DEFAULT: use the table's durability configuration.
+   *
    * @return {@code this} to allow chaining of set methods
    * @since 1.7.0
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/Connector.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/Connector.java b/core/src/main/java/org/apache/accumulo/core/client/Connector.java
index 301577c..e36cc82 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/Connector.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/Connector.java
@@ -26,14 +26,14 @@ import org.apache.accumulo.core.security.Authorizations;
 /**
  * Connector connects to an Accumulo instance and allows the user to request readers and writers for the instance as well as various objects that permit
  * administrative operations.
- * 
+ *
  * The Connector enforces security on the client side by forcing all API calls to be accompanied by user credentials.
  */
 public abstract class Connector {
 
   /**
    * Factory method to create a BatchScanner connected to Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table to query
    * @param authorizations
@@ -42,7 +42,7 @@ public abstract class Connector {
    *          passed, then an exception will be thrown.
    * @param numQueryThreads
    *          the number of concurrent threads to spawn for querying
-   * 
+   *
    * @return BatchScanner object for configuring and querying
    * @throws TableNotFoundException
    *           when the specified table doesn't exist
@@ -51,7 +51,7 @@ public abstract class Connector {
 
   /**
    * Factory method to create a BatchDeleter connected to Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table to query and delete from
    * @param authorizations
@@ -66,7 +66,7 @@ public abstract class Connector {
    *          size in milliseconds; set to 0 or Long.MAX_VALUE to allow the maximum time to hold a batch before writing
    * @param maxWriteThreads
    *          the maximum number of threads to use for writing data to the tablet servers
-   * 
+   *
    * @return BatchDeleter object for configuring and deleting
    * @throws TableNotFoundException
    *           when the specified table doesn't exist
@@ -77,7 +77,7 @@ public abstract class Connector {
       int maxWriteThreads) throws TableNotFoundException;
 
   /**
-   * 
+   *
    * @param tableName
    *          the name of the table to query and delete from
    * @param authorizations
@@ -97,7 +97,7 @@ public abstract class Connector {
 
   /**
    * Factory method to create a BatchWriter connected to Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table to insert data into
    * @param maxMemory
@@ -106,7 +106,7 @@ public abstract class Connector {
    *          time in milliseconds; set to 0 or Long.MAX_VALUE to allow the maximum time to hold a batch before writing
    * @param maxWriteThreads
    *          the maximum number of threads to use for writing data to the tablet servers
-   * 
+   *
    * @return BatchWriter object for configuring and writing data to
    * @throws TableNotFoundException
    *           when the specified table doesn't exist
@@ -117,7 +117,7 @@ public abstract class Connector {
 
   /**
    * Factory method to create a BatchWriter connected to Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table to insert data into
    * @param config
@@ -131,14 +131,14 @@ public abstract class Connector {
   /**
    * Factory method to create a Multi-Table BatchWriter connected to Accumulo. Multi-table batch writers can queue data for multiple tables, which is good for
    * ingesting data into multiple tables from the same source
-   * 
+   *
    * @param maxMemory
    *          size in bytes of the maximum memory to batch before writing
    * @param maxLatency
    *          size in milliseconds; set to 0 or Long.MAX_VALUE to allow the maximum time to hold a batch before writing
    * @param maxWriteThreads
    *          the maximum number of threads to use for writing data to the tablet servers
-   * 
+   *
    * @return MultiTableBatchWriter object for configuring and writing data to
    * @deprecated since 1.5.0; Use {@link #createMultiTableBatchWriter(BatchWriterConfig)} instead.
    */
@@ -148,7 +148,7 @@ public abstract class Connector {
   /**
    * Factory method to create a Multi-Table BatchWriter connected to Accumulo. Multi-table batch writers can queue data for multiple tables. Also data for
    * multiple tables can be sent to a server in a single batch. Its an efficient way to ingest data into multiple tables from a single process.
-   * 
+   *
    * @param config
    *          configuration used to create multi-table batch writer
    * @return MultiTableBatchWriter object for configuring and writing data to
@@ -159,14 +159,14 @@ public abstract class Connector {
 
   /**
    * Factory method to create a Scanner connected to Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table to query data from
    * @param authorizations
    *          A set of authorization labels that will be checked against the column visibility of each key in order to filter data. The authorizations passed in
    *          must be a subset of the accumulo user's set of authorizations. If the accumulo user has authorizations (A1, A2) and authorizations (A2, A3) are
    *          passed, then an exception will be thrown.
-   * 
+   *
    * @return Scanner object for configuring and querying data with
    * @throws TableNotFoundException
    *           when the specified table doesn't exist
@@ -175,12 +175,12 @@ public abstract class Connector {
 
   /**
    * Factory method to create a ConditionalWriter connected to Accumulo.
-   * 
+   *
    * @param tableName
    *          the name of the table to query data from
    * @param config
    *          configuration used to create conditional writer
-   * 
+   *
    * @return ConditionalWriter object for writing ConditionalMutations
    * @throws TableNotFoundException
    *           when the specified table doesn't exist
@@ -190,49 +190,49 @@ public abstract class Connector {
 
   /**
    * Accessor method for internal instance object.
-   * 
+   *
    * @return the internal instance object
    */
   public abstract Instance getInstance();
 
   /**
    * Get the current user for this connector
-   * 
+   *
    * @return the user name
    */
   public abstract String whoami();
 
   /**
    * Retrieves a TableOperations object to perform table functions, such as create and delete.
-   * 
+   *
    * @return an object to manipulate tables
    */
   public abstract TableOperations tableOperations();
 
   /**
    * Retrieves a NamespaceOperations object to perform namespace functions, such as create and delete.
-   * 
+   *
    * @return an object to manipulate namespaces
    */
   public abstract NamespaceOperations namespaceOperations();
 
   /**
    * Retrieves a SecurityOperations object to perform user security operations, such as creating users.
-   * 
+   *
    * @return an object to modify users and permissions
    */
   public abstract SecurityOperations securityOperations();
 
   /**
    * Retrieves an InstanceOperations object to modify instance configuration.
-   * 
+   *
    * @return an object to modify instance configuration
    */
   public abstract InstanceOperations instanceOperations();
 
   /**
    * Retrieves a ReplicationOperations object to manage replication configuration.
-   * 
+   *
    * @return an object to modify replication configuration
    * @since 1.7.0
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/Durability.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/Durability.java b/core/src/main/java/org/apache/accumulo/core/client/Durability.java
index 3e69cb2..08b2092 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/Durability.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/Durability.java
@@ -18,6 +18,7 @@ package org.apache.accumulo.core.client;
 
 /**
  * The value for the durability of a BatchWriter or ConditionalWriter.
+ *
  * @since 1.7.0
  */
 public enum Durability {
@@ -42,4 +43,4 @@ public enum Durability {
    * Write mutations to the write-ahead log, and ensure the data is saved to persistent storage.
    */
   SYNC
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/IsolatedScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/IsolatedScanner.java b/core/src/main/java/org/apache/accumulo/core/client/IsolatedScanner.java
index 443596a..3546cdf 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/IsolatedScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/IsolatedScanner.java
@@ -34,33 +34,33 @@ import org.apache.hadoop.io.Text;
 /**
  * A scanner that presents a row isolated view of an accumulo table. Rows are buffered in memory on the client side. If you think your rows may not fit into
  * memory, then you can provide an alternative row buffer factory to the constructor. This would allow rows to be buffered to disk for example.
- * 
+ *
  */
 
 public class IsolatedScanner extends ScannerOptions implements Scanner {
-  
+
   private static class RowBufferingIterator implements Iterator<Entry<Key,Value>> {
-    
+
     private Iterator<Entry<Key,Value>> source;
     private RowBuffer buffer;
     private Entry<Key,Value> nextRowStart;
     private Iterator<Entry<Key,Value>> rowIter;
     private ByteSequence lastRow = null;
     private long timeout;
-    
+
     private final Scanner scanner;
     private ScannerOptions opts;
     private Range range;
     private int batchSize;
     private long readaheadThreshold;
-    
+
     private void readRow() {
-      
+
       ByteSequence row = null;
-      
+
       while (true) {
         buffer.clear();
-        
+
         try {
           if (nextRowStart != null) {
             buffer.add(nextRowStart);
@@ -71,10 +71,10 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
             buffer.add(entry);
             row = entry.getKey().getRowData();
           }
-          
+
           while (source.hasNext()) {
             Entry<Key,Value> entry = source.next();
-            
+
             if (entry.getKey().getRowData().equals(row)) {
               buffer.add(entry);
             } else {
@@ -82,16 +82,16 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
               break;
             }
           }
-          
+
           lastRow = row;
           rowIter = buffer.iterator();
           // System.out.println("lastRow <- "+lastRow + " "+buffer);
           return;
         } catch (IsolationException ie) {
           Range seekRange = null;
-          
+
           nextRowStart = null;
-          
+
           if (lastRow == null)
             seekRange = range;
           else {
@@ -103,21 +103,21 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
             }
             // System.out.println(seekRange);
           }
-          
+
           if (seekRange == null) {
             buffer.clear();
             rowIter = buffer.iterator();
             return;
           }
-          
+
           // wait a moment before retrying
           UtilWaitThread.sleep(100);
-          
+
           source = newIterator(seekRange);
         }
       }
     }
-    
+
     private Iterator<Entry<Key,Value>> newIterator(Range r) {
       synchronized (scanner) {
         scanner.enableIsolation();
@@ -126,101 +126,102 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
         scanner.setRange(r);
         scanner.setReadaheadThreshold(readaheadThreshold);
         setOptions((ScannerOptions) scanner, opts);
-        
+
         return scanner.iterator();
         // return new FaultyIterator(scanner.iterator());
       }
     }
-    
-    public RowBufferingIterator(Scanner scanner, ScannerOptions opts, Range range, long timeout, int batchSize, long readaheadThreshold, RowBufferFactory bufferFactory) {
+
+    public RowBufferingIterator(Scanner scanner, ScannerOptions opts, Range range, long timeout, int batchSize, long readaheadThreshold,
+        RowBufferFactory bufferFactory) {
       this.scanner = scanner;
       this.opts = new ScannerOptions(opts);
       this.range = range;
       this.timeout = timeout;
       this.batchSize = batchSize;
       this.readaheadThreshold = readaheadThreshold;
-      
+
       buffer = bufferFactory.newBuffer();
-      
+
       this.source = newIterator(range);
-      
+
       readRow();
     }
-    
+
     @Override
     public boolean hasNext() {
       return rowIter.hasNext();
     }
-    
+
     @Override
     public Entry<Key,Value> next() {
       Entry<Key,Value> next = rowIter.next();
       if (!rowIter.hasNext()) {
         readRow();
       }
-      
+
       return next;
     }
-    
+
     @Override
     public void remove() {
       throw new UnsupportedOperationException();
     }
-    
+
   }
-  
+
   interface RowBufferFactory {
     RowBuffer newBuffer();
   }
-  
+
   interface RowBuffer extends Iterable<Entry<Key,Value>> {
     void add(Entry<Key,Value> entry);
-    
+
     @Override
     Iterator<Entry<Key,Value>> iterator();
-    
+
     void clear();
   }
-  
+
   public static class MemoryRowBufferFactory implements RowBufferFactory {
-    
+
     @Override
     public RowBuffer newBuffer() {
       return new MemoryRowBuffer();
     }
   }
-  
+
   public static class MemoryRowBuffer implements RowBuffer {
-    
+
     private ArrayList<Entry<Key,Value>> buffer = new ArrayList<Entry<Key,Value>>();
-    
+
     @Override
     public void add(Entry<Key,Value> entry) {
       buffer.add(entry);
     }
-    
+
     @Override
     public Iterator<Entry<Key,Value>> iterator() {
       return buffer.iterator();
     }
-    
+
     @Override
     public void clear() {
       buffer.clear();
     }
-    
+
   }
-  
+
   private Scanner scanner;
   private Range range;
   private int batchSize;
   private long readaheadThreshold;
   private RowBufferFactory bufferFactory;
-  
+
   public IsolatedScanner(Scanner scanner) {
     this(scanner, new MemoryRowBufferFactory());
   }
-  
+
   public IsolatedScanner(Scanner scanner, RowBufferFactory bufferFactory) {
     this.scanner = scanner;
     this.range = scanner.getRange();
@@ -229,12 +230,12 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
     this.readaheadThreshold = scanner.getReadaheadThreshold();
     this.bufferFactory = bufferFactory;
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     return new RowBufferingIterator(scanner, this, range, timeOut, batchSize, readaheadThreshold, bufferFactory);
   }
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {
@@ -243,7 +244,7 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
     else
       setTimeout(timeOut, TimeUnit.SECONDS);
   }
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
@@ -252,32 +253,32 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
       return Integer.MAX_VALUE;
     return (int) timeout;
   }
-  
+
   @Override
   public void setRange(Range range) {
     this.range = range;
   }
-  
+
   @Override
   public Range getRange() {
     return range;
   }
-  
+
   @Override
   public void setBatchSize(int size) {
     this.batchSize = size;
   }
-  
+
   @Override
   public int getBatchSize() {
     return batchSize;
   }
-  
+
   @Override
   public void enableIsolation() {
     // aye aye captain, already done sir
   }
-  
+
   @Override
   public void disableIsolation() {
     throw new UnsupportedOperationException();
@@ -293,7 +294,7 @@ public class IsolatedScanner extends ScannerOptions implements Scanner {
     if (0 > batches) {
       throw new IllegalArgumentException("Number of batches before read-ahead must be non-negative");
     }
-    
+
     this.readaheadThreshold = batches;
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/IteratorSetting.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/IteratorSetting.java b/core/src/main/java/org/apache/accumulo/core/client/IteratorSetting.java
index b966a4a..baf9860 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/IteratorSetting.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/IteratorSetting.java
@@ -17,6 +17,7 @@
 package org.apache.accumulo.core.client;
 
 import static com.google.common.base.Preconditions.checkArgument;
+
 import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.IOException;
@@ -36,11 +37,11 @@ import org.apache.hadoop.io.WritableUtils;
 
 /**
  * Configure an iterator for minc, majc, and/or scan. By default, IteratorSetting will be configured for scan.
- * 
+ *
  * Every iterator has a priority, a name, a class, a set of scopes, and configuration parameters.
- * 
+ *
  * A typical use case configured for scan:
- * 
+ *
  * <pre>
  * IteratorSetting cfg = new IteratorSetting(priority, &quot;myIter&quot;, MyIterator.class);
  * MyIterator.addOption(cfg, 42);
@@ -52,19 +53,19 @@ public class IteratorSetting implements Writable {
   private String name;
   private String iteratorClass;
   private Map<String,String> properties;
-  
+
   /**
    * Get layer at which this iterator applies. See {@link #setPriority(int)} for how the priority is used.
-   * 
+   *
    * @return the priority of this Iterator
    */
   public int getPriority() {
     return priority;
   }
-  
+
   /**
    * Set layer at which this iterator applies.
-   * 
+   *
    * @param priority
    *          determines the order in which iterators are applied (system iterators are always applied first, then user-configured iterators, lowest priority
    *          first)
@@ -73,16 +74,16 @@ public class IteratorSetting implements Writable {
     checkArgument(priority > 0, "property must be strictly positive");
     this.priority = priority;
   }
-  
+
   /**
    * Get the iterator's name.
-   * 
+   *
    * @return the name of the iterator
    */
   public String getName() {
     return name;
   }
-  
+
   /**
    * Set the iterator's name. Must be a simple alphanumeric identifier.
    */
@@ -90,16 +91,16 @@ public class IteratorSetting implements Writable {
     checkArgument(name != null, "name is null");
     this.name = name;
   }
-  
+
   /**
    * Get the name of the class that implements the iterator.
-   * 
+   *
    * @return the iterator's class name
    */
   public String getIteratorClass() {
     return iteratorClass;
   }
-  
+
   /**
    * Set the name of the class that implements the iterator. The class does not have to be present on the client, but it must be available to all tablet
    * servers.
@@ -108,10 +109,10 @@ public class IteratorSetting implements Writable {
     checkArgument(iteratorClass != null, "iteratorClass is null");
     this.iteratorClass = iteratorClass;
   }
-  
+
   /**
    * Constructs an iterator setting configured for the scan scope with no parameters. (Parameters can be added later.)
-   * 
+   *
    * @param priority
    *          the priority for the iterator (see {@link #setPriority(int)})
    * @param name
@@ -122,10 +123,10 @@ public class IteratorSetting implements Writable {
   public IteratorSetting(int priority, String name, String iteratorClass) {
     this(priority, name, iteratorClass, new HashMap<String,String>());
   }
-  
+
   /**
    * Constructs an iterator setting configured for the specified scopes with the specified parameters.
-   * 
+   *
    * @param priority
    *          the priority for the iterator (see {@link #setPriority(int)})
    * @param name
@@ -142,11 +143,11 @@ public class IteratorSetting implements Writable {
     this.properties = new HashMap<String,String>();
     addOptions(properties);
   }
-  
+
   /**
    * Constructs an iterator setting using the given class's SimpleName for the iterator name. The iterator setting will be configured for the scan scope with no
    * parameters.
-   * 
+   *
    * @param priority
    *          the priority for the iterator (see {@link #setPriority(int)})
    * @param iteratorClass
@@ -155,12 +156,12 @@ public class IteratorSetting implements Writable {
   public IteratorSetting(int priority, Class<? extends SortedKeyValueIterator<Key,Value>> iteratorClass) {
     this(priority, iteratorClass.getSimpleName(), iteratorClass.getName());
   }
-  
+
   /**
-   * 
+   *
    * Constructs an iterator setting using the given class's SimpleName for the iterator name and configured for the specified scopes with the specified
    * parameters.
-   * 
+   *
    * @param priority
    *          the priority for the iterator (see {@link #setPriority(int)})
    * @param iteratorClass
@@ -171,10 +172,10 @@ public class IteratorSetting implements Writable {
   public IteratorSetting(int priority, Class<? extends SortedKeyValueIterator<Key,Value>> iteratorClass, Map<String,String> properties) {
     this(priority, iteratorClass.getSimpleName(), iteratorClass.getName(), properties);
   }
-  
+
   /**
    * Constructs an iterator setting configured for the scan scope with no parameters.
-   * 
+   *
    * @param priority
    *          the priority for the iterator (see {@link #setPriority(int)})
    * @param name
@@ -185,21 +186,25 @@ public class IteratorSetting implements Writable {
   public IteratorSetting(int priority, String name, Class<? extends SortedKeyValueIterator<Key,Value>> iteratorClass) {
     this(priority, name, iteratorClass.getName());
   }
-  
+
   /**
    * Constructs an iterator setting using the provided name and the provided class's name for the scan scope with the provided parameters.
-   * 
-   * @param priority The priority for the iterator (see {@link #setPriority(int)})
-   * @param name The distinguishing name for the iterator
-   * @param iteratorClass The class for the iterator
-   * @param properties Any properties for the iterator
-   * 
+   *
+   * @param priority
+   *          The priority for the iterator (see {@link #setPriority(int)})
+   * @param name
+   *          The distinguishing name for the iterator
+   * @param iteratorClass
+   *          The class for the iterator
+   * @param properties
+   *          Any properties for the iterator
+   *
    * @since 1.6.0
    */
   public IteratorSetting(int priority, String name, Class<? extends SortedKeyValueIterator<Key,Value>> iteratorClass, Map<String,String> properties) {
     this(priority, name, iteratorClass.getName(), properties);
   }
-  
+
   /**
    * @since 1.5.0
    */
@@ -207,10 +212,10 @@ public class IteratorSetting implements Writable {
     this.properties = new HashMap<String,String>();
     this.readFields(din);
   }
-  
+
   /**
    * Add another option to the iterator.
-   * 
+   *
    * @param option
    *          the name of the option
    * @param value
@@ -221,10 +226,10 @@ public class IteratorSetting implements Writable {
     checkArgument(value != null, "value is null");
     properties.put(option, value);
   }
-  
+
   /**
    * Remove an option from the iterator.
-   * 
+   *
    * @param option
    *          the name of the option
    * @return the value previously associated with the option, or null if no such option existed
@@ -233,10 +238,10 @@ public class IteratorSetting implements Writable {
     checkArgument(option != null, "option is null");
     return properties.remove(option);
   }
-  
+
   /**
    * Add many options to the iterator.
-   * 
+   *
    * @param propertyEntries
    *          a set of entries to add to the options
    */
@@ -246,10 +251,10 @@ public class IteratorSetting implements Writable {
       addOption(keyValue.getKey(), keyValue.getValue());
     }
   }
-  
+
   /**
    * Add many options to the iterator.
-   * 
+   *
    * @param properties
    *          a map of entries to add to the options
    */
@@ -257,23 +262,23 @@ public class IteratorSetting implements Writable {
     checkArgument(properties != null, "properties is null");
     addOptions(properties.entrySet());
   }
-  
+
   /**
    * Get the configuration parameters for this iterator.
-   * 
+   *
    * @return the properties
    */
   public Map<String,String> getOptions() {
     return Collections.unmodifiableMap(properties);
   }
-  
+
   /**
    * Remove all options from the iterator.
    */
   public void clearOptions() {
     properties.clear();
   }
-  
+
   /**
    * @see java.lang.Object#hashCode()
    */
@@ -287,7 +292,7 @@ public class IteratorSetting implements Writable {
     result = prime * result + ((properties == null) ? 0 : properties.hashCode());
     return result;
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     if (this == obj)
@@ -316,7 +321,7 @@ public class IteratorSetting implements Writable {
       return false;
     return true;
   }
-  
+
   /**
    * @see java.lang.Object#toString()
    */
@@ -333,38 +338,38 @@ public class IteratorSetting implements Writable {
     sb.append(properties);
     return sb.toString();
   }
-  
+
   /**
    * A convenience class for passing column family and column qualifiers to iterator configuration methods.
    */
   public static class Column extends Pair<Text,Text> {
-    
+
     public Column(Text columnFamily, Text columnQualifier) {
       super(columnFamily, columnQualifier);
     }
-    
+
     public Column(Text columnFamily) {
       super(columnFamily, null);
     }
-    
+
     public Column(String columnFamily, String columnQualifier) {
       super(new Text(columnFamily), new Text(columnQualifier));
     }
-    
+
     public Column(String columnFamily) {
       super(new Text(columnFamily), null);
     }
-    
+
     public Text getColumnFamily() {
       return getFirst();
     }
-    
+
     public Text getColumnQualifier() {
       return getSecond();
     }
-    
+
   }
-  
+
   /**
    * @since 1.5.0
    */
@@ -380,7 +385,7 @@ public class IteratorSetting implements Writable {
       size--;
     }
   }
-  
+
   /**
    * @since 1.5.0
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/MultiTableBatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/MultiTableBatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/MultiTableBatchWriter.java
index 39287e8..5495598 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/MultiTableBatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/MultiTableBatchWriter.java
@@ -19,13 +19,13 @@ package org.apache.accumulo.core.client;
 /**
  * This class enables efficient batch writing to multiple tables. When creating a batch writer for each table, each has its own memory and network resources.
  * Using this class these resources may be shared among multiple tables.
- * 
+ *
  */
 public interface MultiTableBatchWriter {
-  
+
   /**
    * Returns a BatchWriter for a particular table.
-   * 
+   *
    * @param table
    *          the name of a table whose batch writer you wish to retrieve
    * @return an instance of a batch writer for the specified table
@@ -36,28 +36,28 @@ public interface MultiTableBatchWriter {
    * @throws TableNotFoundException
    *           when the table does not exist
    */
-   BatchWriter getBatchWriter(String table) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
-  
+  BatchWriter getBatchWriter(String table) throws AccumuloException, AccumuloSecurityException, TableNotFoundException;
+
   /**
    * Send mutations for all tables to accumulo.
-   * 
+   *
    * @throws MutationsRejectedException
    *           when queued mutations are unable to be inserted
    */
   void flush() throws MutationsRejectedException;
-  
+
   /**
    * Flush and release all resources.
-   * 
+   *
    * @throws MutationsRejectedException
    *           when queued mutations are unable to be inserted
-   * 
+   *
    */
   void close() throws MutationsRejectedException;
-  
+
   /**
    * Returns true if this batch writer has been closed.
-   * 
+   *
    * @return true if this batch writer has been closed
    */
   boolean isClosed();


[02/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/RewriteTabletDirectoriesIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/RewriteTabletDirectoriesIT.java b/test/src/test/java/org/apache/accumulo/test/RewriteTabletDirectoriesIT.java
index cacee7a..f9ce176 100644
--- a/test/src/test/java/org/apache/accumulo/test/RewriteTabletDirectoriesIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/RewriteTabletDirectoriesIT.java
@@ -93,7 +93,7 @@ public class RewriteTabletDirectoriesIT extends ConfigurableMacIT {
     for (String split : "a,b,c,d,e,f,g,h,i,j,k,l,m,n,o,p,q,r,s,t,u,v,w,x,y,z".split(",")) {
       splits.add(new Text(split));
       Mutation m = new Mutation(new Text(split));
-      m.put(new byte[]{}, new byte[]{}, new byte[]{});
+      m.put(new byte[] {}, new byte[] {}, new byte[] {});
       bw.addMutation(m);
     }
     bw.close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/ShellServerIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/ShellServerIT.java b/test/src/test/java/org/apache/accumulo/test/ShellServerIT.java
index 44ddb50..9296548 100644
--- a/test/src/test/java/org/apache/accumulo/test/ShellServerIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/ShellServerIT.java
@@ -724,7 +724,7 @@ public class ShellServerIT extends SharedMiniClusterIT {
 
       if (entry.getKey().equals("table.custom.testProp"))
         Assert.assertTrue("Initial property was not set correctly", entry.getValue().equals("testProp"));
-      
+
       if (entry.getKey().equals(Property.TABLE_SPLIT_THRESHOLD.getKey()))
         Assert.assertTrue("Initial property was not set correctly", entry.getValue().equals("10K"));
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/TableConfigurationUpdateIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/TableConfigurationUpdateIT.java b/test/src/test/java/org/apache/accumulo/test/TableConfigurationUpdateIT.java
index 73e0e81..94e0a3c 100644
--- a/test/src/test/java/org/apache/accumulo/test/TableConfigurationUpdateIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/TableConfigurationUpdateIT.java
@@ -53,7 +53,8 @@ public class TableConfigurationUpdateIT extends AccumuloClusterIT {
     String table = getUniqueNames(1)[0];
     conn.tableOperations().create(table);
 
-    final NamespaceConfiguration defaultConf = new NamespaceConfiguration(Namespaces.DEFAULT_NAMESPACE_ID, inst, AccumuloConfiguration.getDefaultConfiguration());
+    final NamespaceConfiguration defaultConf = new NamespaceConfiguration(Namespaces.DEFAULT_NAMESPACE_ID, inst,
+        AccumuloConfiguration.getDefaultConfiguration());
 
     // Cache invalidates 25% of the time
     int randomMax = 4;
@@ -83,8 +84,8 @@ public class TableConfigurationUpdateIT extends AccumuloClusterIT {
     }
 
     long end = System.currentTimeMillis();
-    log.debug(tableConf + " with " + iterations + " iterations and " + numThreads + " threads and cache invalidates "
-        + ((1. / randomMax) * 100.) + "% took " + (end - start) / 1000 + " second(s)");
+    log.debug(tableConf + " with " + iterations + " iterations and " + numThreads + " threads and cache invalidates " + ((1. / randomMax) * 100.) + "% took "
+        + (end - start) / 1000 + " second(s)");
   }
 
   public static class TableConfRunner implements Callable<Exception> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/TabletServerGivesUpIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/TabletServerGivesUpIT.java b/test/src/test/java/org/apache/accumulo/test/TabletServerGivesUpIT.java
index 0a62746..ac24083 100644
--- a/test/src/test/java/org/apache/accumulo/test/TabletServerGivesUpIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/TabletServerGivesUpIT.java
@@ -16,6 +16,8 @@
  */
 package org.apache.accumulo.test;
 
+import static org.junit.Assert.assertEquals;
+
 import java.util.TreeSet;
 import java.util.concurrent.atomic.AtomicReference;
 
@@ -28,11 +30,9 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
-import static org.junit.Assert.assertEquals;
-
 // ACCUMULO-2480
 public class TabletServerGivesUpIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.useMiniDFS(true);
@@ -50,7 +50,7 @@ public class TabletServerGivesUpIT extends ConfigurableMacIT {
     // Kill dfs
     cluster.getMiniDfs().shutdown();
     // ask the tserver to do something
-    final AtomicReference<Exception> ex = new AtomicReference<>(); 
+    final AtomicReference<Exception> ex = new AtomicReference<>();
     Thread splitter = new Thread() {
       public void run() {
         try {
@@ -68,5 +68,5 @@ public class TabletServerGivesUpIT extends ConfigurableMacIT {
       UtilWaitThread.sleep(1000);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/TotalQueuedIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/TotalQueuedIT.java b/test/src/test/java/org/apache/accumulo/test/TotalQueuedIT.java
index ad9d43c..708d2d4 100644
--- a/test/src/test/java/org/apache/accumulo/test/TotalQueuedIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/TotalQueuedIT.java
@@ -42,14 +42,14 @@ import com.google.common.net.HostAndPort;
 
 // see ACCUMULO-1950
 public class TotalQueuedIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setNumTservers(1);
     cfg.setDefaultMemory(cfg.getDefaultMemory() * 2, MemoryUnit.BYTE);
     cfg.useMiniDFS();
   }
-  
+
   int SMALL_QUEUE_SIZE = 100000;
   int LARGE_QUEUE_SIZE = SMALL_QUEUE_SIZE * 10;
   static final long N = 1000000;
@@ -69,7 +69,7 @@ public class TotalQueuedIT extends ConfigurableMacIT {
     BatchWriterConfig cfg = new BatchWriterConfig();
     cfg.setMaxWriteThreads(10);
     cfg.setMaxLatency(1, TimeUnit.SECONDS);
-    cfg.setMaxMemory(1024*1024);
+    cfg.setMaxMemory(1024 * 1024);
     long realSyncs = getSyncs();
     BatchWriter bw = c.createBatchWriter(tableName, cfg);
     long now = System.currentTimeMillis();
@@ -86,11 +86,11 @@ public class TotalQueuedIT extends ConfigurableMacIT {
     double secs = diff / 1000.;
     double syncs = bytesSent / SMALL_QUEUE_SIZE;
     double syncsPerSec = syncs / secs;
-    System.out.println(String.format("Sent %d bytes in %f secs approximately %d syncs (%f syncs per sec)", bytesSent, secs, ((long)syncs), syncsPerSec));
+    System.out.println(String.format("Sent %d bytes in %f secs approximately %d syncs (%f syncs per sec)", bytesSent, secs, ((long) syncs), syncsPerSec));
     long update = getSyncs();
     System.out.println("Syncs " + (update - realSyncs));
     realSyncs = update;
-    
+
     // Now with a much bigger total queue
     c.instanceOperations().setProperty(Property.TSERV_TOTAL_MUTATION_QUEUE_MAX.getKey(), "" + LARGE_QUEUE_SIZE);
     c.tableOperations().flush(tableName, null, null, true);
@@ -110,7 +110,7 @@ public class TotalQueuedIT extends ConfigurableMacIT {
     secs = diff / 1000.;
     syncs = bytesSent / LARGE_QUEUE_SIZE;
     syncsPerSec = syncs / secs;
-    System.out.println(String.format("Sent %d bytes in %f secs approximately %d syncs (%f syncs per sec)", bytesSent, secs, ((long)syncs), syncsPerSec));
+    System.out.println(String.format("Sent %d bytes in %f secs approximately %d syncs (%f syncs per sec)", bytesSent, secs, ((long) syncs), syncsPerSec));
     update = getSyncs();
     System.out.println("Syncs " + (update - realSyncs));
     assertTrue(update - realSyncs < realSyncs);
@@ -127,5 +127,5 @@ public class TotalQueuedIT extends ConfigurableMacIT {
     }
     return 0;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/TraceRepoDeserializationTest.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/TraceRepoDeserializationTest.java b/test/src/test/java/org/apache/accumulo/test/TraceRepoDeserializationTest.java
index 5215d3e..413a6c9 100644
--- a/test/src/test/java/org/apache/accumulo/test/TraceRepoDeserializationTest.java
+++ b/test/src/test/java/org/apache/accumulo/test/TraceRepoDeserializationTest.java
@@ -16,33 +16,27 @@
  */
 package org.apache.accumulo.test;
 
+import static org.junit.Assert.fail;
+
 import java.io.ByteArrayInputStream;
 import java.io.InvalidClassException;
 import java.io.ObjectInputStream;
 
 import org.apache.accumulo.core.util.Base64;
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class TraceRepoDeserializationTest {
-  
+
   // Zookeeper data for a merge request
-  static private final String oldValue = 
-      "rO0ABXNyAC1vcmcuYXBhY2hlLmFjY3VtdWxvLm1hc3Rlci50YWJsZU9wcy5UcmFjZVJlc" +
-      "G8AAAAAAAAAAQIAAkwABHJlcG90AB9Mb3JnL2FwYWNoZS9hY2N1bXVsby9mYXRlL1Jl" +
-      "cG87TAAFdGluZm90AChMb3JnL2FwYWNoZS9hY2N1bXVsby90cmFjZS90aHJpZnQvVEl" +
-      "uZm87eHBzcgAwb3JnLmFwYWNoZS5hY2N1bXVsby5tYXN0ZXIudGFibGVPcHMuVGFibG" +
-      "VSYW5nZU9wAAAAAAAAAAECAAVbAAZlbmRSb3d0AAJbQkwAC25hbWVzcGFjZUlkdAAST" +
-      "GphdmEvbGFuZy9TdHJpbmc7TAACb3B0AD1Mb3JnL2FwYWNoZS9hY2N1bXVsby9zZXJ2" +
-      "ZXIvbWFzdGVyL3N0YXRlL01lcmdlSW5mbyRPcGVyYXRpb247WwAIc3RhcnRSb3dxAH4A" +
-      "BUwAB3RhYmxlSWRxAH4ABnhyAC5vcmcuYXBhY2hlLmFjY3VtdWxvLm1hc3Rlci50YWJs" +
-      "ZU9wcy5NYXN0ZXJSZXBvAAAAAAAAAAECAAB4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAAA" +
-      "dAAIK2RlZmF1bHR+cgA7b3JnLmFwYWNoZS5hY2N1bXVsby5zZXJ2ZXIubWFzdGVyLnN0" +
-      "YXRlLk1lcmdlSW5mbyRPcGVyYXRpb24AAAAAAAAAABIAAHhyAA5qYXZhLmxhbmcuRW51" +
-      "bQAAAAAAAAAAEgAAeHB0AAVNRVJHRXEAfgALdAABMnNyACZvcmcuYXBhY2hlLmFjY3Vt" +
-      "dWxvLnRyYWNlLnRocmlmdC5USW5mb79UcL31bhZ9AwADQgAQX19pc3NldF9iaXRmaWVs" +
-      "ZEoACHBhcmVudElkSgAHdHJhY2VJZHhwdwUWABYAAHg=";
-  
+  static private final String oldValue = "rO0ABXNyAC1vcmcuYXBhY2hlLmFjY3VtdWxvLm1hc3Rlci50YWJsZU9wcy5UcmFjZVJlc"
+      + "G8AAAAAAAAAAQIAAkwABHJlcG90AB9Mb3JnL2FwYWNoZS9hY2N1bXVsby9mYXRlL1Jl" + "cG87TAAFdGluZm90AChMb3JnL2FwYWNoZS9hY2N1bXVsby90cmFjZS90aHJpZnQvVEl"
+      + "uZm87eHBzcgAwb3JnLmFwYWNoZS5hY2N1bXVsby5tYXN0ZXIudGFibGVPcHMuVGFibG" + "VSYW5nZU9wAAAAAAAAAAECAAVbAAZlbmRSb3d0AAJbQkwAC25hbWVzcGFjZUlkdAAST"
+      + "GphdmEvbGFuZy9TdHJpbmc7TAACb3B0AD1Mb3JnL2FwYWNoZS9hY2N1bXVsby9zZXJ2" + "ZXIvbWFzdGVyL3N0YXRlL01lcmdlSW5mbyRPcGVyYXRpb247WwAIc3RhcnRSb3dxAH4A"
+      + "BUwAB3RhYmxlSWRxAH4ABnhyAC5vcmcuYXBhY2hlLmFjY3VtdWxvLm1hc3Rlci50YWJs" + "ZU9wcy5NYXN0ZXJSZXBvAAAAAAAAAAECAAB4cHVyAAJbQqzzF/gGCFTgAgAAeHAAAAAA"
+      + "dAAIK2RlZmF1bHR+cgA7b3JnLmFwYWNoZS5hY2N1bXVsby5zZXJ2ZXIubWFzdGVyLnN0" + "YXRlLk1lcmdlSW5mbyRPcGVyYXRpb24AAAAAAAAAABIAAHhyAA5qYXZhLmxhbmcuRW51"
+      + "bQAAAAAAAAAAEgAAeHB0AAVNRVJHRXEAfgALdAABMnNyACZvcmcuYXBhY2hlLmFjY3Vt" + "dWxvLnRyYWNlLnRocmlmdC5USW5mb79UcL31bhZ9AwADQgAQX19pc3NldF9iaXRmaWVs"
+      + "ZEoACHBhcmVudElkSgAHdHJhY2VJZHhwdwUWABYAAHg=";
+
   @Test(expected = InvalidClassException.class)
   public void test() throws Exception {
     byte bytes[] = Base64.decodeBase64(oldValue);
@@ -51,5 +45,5 @@ public class TraceRepoDeserializationTest {
     ois.readObject();
     fail("did not throw exception");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/VerifySerialRecoveryIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/VerifySerialRecoveryIT.java b/test/src/test/java/org/apache/accumulo/test/VerifySerialRecoveryIT.java
index 525a29a..d1d4681 100644
--- a/test/src/test/java/org/apache/accumulo/test/VerifySerialRecoveryIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/VerifySerialRecoveryIT.java
@@ -14,7 +14,11 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
- package org.apache.accumulo.test;
+package org.apache.accumulo.test;
+
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
 
 import java.util.Map.Entry;
 import java.util.SortedSet;
@@ -39,19 +43,15 @@ import org.apache.hadoop.fs.RawLocalFileSystem;
 import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
-
 public class VerifySerialRecoveryIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setNumTservers(1);
     cfg.setProperty(Property.TSERV_ASSIGNMENT_MAXCONCURRENT, "20");
     hadoopCoreSite.set("fs.file.impl", RawLocalFileSystem.class.getName());
   }
-  
+
   @Test(timeout = 4 * 60 * 1000)
   public void testSerializedRecovery() throws Exception {
     // make a table with many splits
@@ -75,9 +75,10 @@ public class VerifySerialRecoveryIT extends ConfigurableMacIT {
     for (ProcessReference ref : getCluster().getProcesses().get(ServerType.TABLET_SERVER))
       getCluster().killProcess(ServerType.TABLET_SERVER, ref);
     final Process ts = cluster.exec(TabletServer.class);
-    
+
     // wait for recovery
-    for (@SuppressWarnings("unused") Entry<Key,Value> entry : c.createScanner(tableName, Authorizations.EMPTY))
+    for (@SuppressWarnings("unused")
+    Entry<Key,Value> entry : c.createScanner(tableName, Authorizations.EMPTY))
       ;
     assertEquals(0, cluster.exec(Admin.class, "stopAll").waitFor());
     ts.waitFor();
@@ -104,5 +105,5 @@ public class VerifySerialRecoveryIT extends ConfigurableMacIT {
     }
     assertFalse(started);
     assertTrue(recoveries > 0);
-  }  
+  }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/BackupMasterIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/BackupMasterIT.java b/test/src/test/java/org/apache/accumulo/test/functional/BackupMasterIT.java
index bf2d2e1..efed7a4 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/BackupMasterIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/BackupMasterIT.java
@@ -39,7 +39,7 @@ public class BackupMasterIT extends ConfigurableMacIT {
     // create a backup
     Process backup = exec(Master.class);
     try {
-      ZooReaderWriter writer = new ZooReaderWriter(cluster.getZooKeepers(), 30*1000, "digest", "accumulo:DONTTELL".getBytes());
+      ZooReaderWriter writer = new ZooReaderWriter(cluster.getZooKeepers(), 30 * 1000, "digest", "accumulo:DONTTELL".getBytes());
       String root = "/accumulo/" + getConnector().getInstance().getInstanceID();
       List<String> children = Collections.emptyList();
       // wait for 2 lock entries

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/BalanceInPresenceOfOfflineTableIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/BalanceInPresenceOfOfflineTableIT.java b/test/src/test/java/org/apache/accumulo/test/functional/BalanceInPresenceOfOfflineTableIT.java
index a9ea5db..114d79b 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/BalanceInPresenceOfOfflineTableIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/BalanceInPresenceOfOfflineTableIT.java
@@ -64,10 +64,10 @@ public class BalanceInPresenceOfOfflineTableIT extends AccumuloClusterIT {
 
   @Override
   public void configureMiniCluster(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
-    Map<String,String> siteConfig = new HashMap<String, String>();
+    Map<String,String> siteConfig = new HashMap<String,String>();
     siteConfig.put(Property.TSERV_MAXMEM.getKey(), "10K");
     siteConfig.put(Property.TSERV_MAJC_DELAY.getKey(), "0");
-    cfg.setSiteConfig(siteConfig );
+    cfg.setSiteConfig(siteConfig);
     // ensure we have two tservers
     if (cfg.getNumTservers() < 2) {
       cfg.setNumTservers(2);
@@ -130,10 +130,10 @@ public class BalanceInPresenceOfOfflineTableIT extends AccumuloClusterIT {
 
     log.debug("waiting for balancing, up to ~5 minutes to allow for migration cleanup.");
     final long startTime = System.currentTimeMillis();
-    long currentWait = 10*1000;
+    long currentWait = 10 * 1000;
     boolean balancingWorked = false;
 
-    while (!balancingWorked && (System.currentTimeMillis() - startTime) < ((5*60 + 15)*1000)) {
+    while (!balancingWorked && (System.currentTimeMillis() - startTime) < ((5 * 60 + 15) * 1000)) {
       Thread.sleep(currentWait);
       currentWait *= 2;
 
@@ -167,7 +167,7 @@ public class BalanceInPresenceOfOfflineTableIT extends AccumuloClusterIT {
       long[] tabletsPerServer = new long[stats.getTServerInfoSize()];
       Arrays.fill(tabletsPerServer, 0l);
       for (int i = 0; i < stats.getTServerInfoSize(); i++) {
-        for (Map.Entry<String, TableInfo> entry : stats.getTServerInfo().get(i).getTableMap().entrySet()) {
+        for (Map.Entry<String,TableInfo> entry : stats.getTServerInfo().get(i).getTableMap().entrySet()) {
           tabletsPerServer[i] += entry.getValue().getTablets();
         }
       }
@@ -176,7 +176,7 @@ public class BalanceInPresenceOfOfflineTableIT extends AccumuloClusterIT {
         log.debug("We should have > 10 tablets. sleeping for " + currentWait + "ms");
         continue;
       }
-      if ((NumberUtils.min(tabletsPerServer) / ((double)NumberUtils.max(tabletsPerServer))) < 0.5) {
+      if ((NumberUtils.min(tabletsPerServer) / ((double) NumberUtils.max(tabletsPerServer))) < 0.5) {
         log.debug("ratio of min to max tablets per server should be roughly even. sleeping for " + currentWait + "ms");
         continue;
       }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/BinaryStressIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/BinaryStressIT.java b/test/src/test/java/org/apache/accumulo/test/functional/BinaryStressIT.java
index 29f1f57..5a41d32 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/BinaryStressIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/BinaryStressIT.java
@@ -54,7 +54,7 @@ public class BinaryStressIT extends AccumuloClusterIT {
     Map<String,String> siteConfig = new HashMap<String,String>();
     siteConfig.put(Property.TSERV_MAXMEM.getKey(), "50K");
     siteConfig.put(Property.TSERV_MAJC_DELAY.getKey(), "0");
-    cfg.setSiteConfig(siteConfig );
+    cfg.setSiteConfig(siteConfig);
   }
 
   private String majcDelay, maxMem;
@@ -83,7 +83,7 @@ public class BinaryStressIT extends AccumuloClusterIT {
       InstanceOperations iops = getConnector().instanceOperations();
       iops.setProperty(Property.TSERV_MAJC_DELAY.getKey(), majcDelay);
       iops.setProperty(Property.TSERV_MAXMEM.getKey(), maxMem);
-      
+
       getClusterControl().stopAllServers(ServerType.TABLET_SERVER);
       getClusterControl().startAllServers(ServerType.TABLET_SERVER);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/BloomFilterIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/BloomFilterIT.java b/test/src/test/java/org/apache/accumulo/test/functional/BloomFilterIT.java
index 3319247..a4ed85d 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/BloomFilterIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/BloomFilterIT.java
@@ -57,9 +57,9 @@ public class BloomFilterIT extends AccumuloClusterIT {
   public void configureMiniCluster(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setDefaultMemory(1, MemoryUnit.GIGABYTE);
     cfg.setNumTservers(1);
-    Map<String,String> siteConfig = new HashMap<String, String>();
+    Map<String,String> siteConfig = new HashMap<String,String>();
     siteConfig.put(Property.TSERV_TOTAL_MUTATION_QUEUE_MAX.getKey(), "10M");
-    cfg.setSiteConfig(siteConfig );
+    cfg.setSiteConfig(siteConfig);
   }
 
   @Override
@@ -161,7 +161,7 @@ public class BloomFilterIT extends AccumuloClusterIT {
   private void timeCheck(long t1, long t2) throws Exception {
     double improvement = (t1 - t2) * 1.0 / t1;
     if (improvement < .1) {
-      throw new Exception("Queries had less than 10% improvement (old: " + t1 + " new: " + t2 + " improvement: " + (improvement*100) + "%)");
+      throw new Exception("Queries had less than 10% improvement (old: " + t1 + " new: " + t2 + " improvement: " + (improvement * 100) + "%)");
     }
     log.info(String.format("Improvement: %.2f%% (%d vs %d)", (improvement * 100), t1, t2));
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/BulkIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/BulkIT.java b/test/src/test/java/org/apache/accumulo/test/functional/BulkIT.java
index 7d4b226..39299da 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/BulkIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/BulkIT.java
@@ -74,7 +74,7 @@ public class BulkIT extends AccumuloIT {
     cluster = harness.create(getToken());
     cluster.start();
   }
-  
+
   @After
   public void stopMiniCluster() throws Exception {
     cluster.stop();
@@ -85,8 +85,8 @@ public class BulkIT extends AccumuloIT {
     runTest(getConnector(), getUniqueNames(1)[0], this.getClass().getName(), testName.getMethodName());
   }
 
-  static void runTest(Connector c, String tableName, String filePrefix, String dirSuffix) throws AccumuloException, AccumuloSecurityException, TableExistsException, IOException, TableNotFoundException,
-      MutationsRejectedException {
+  static void runTest(Connector c, String tableName, String filePrefix, String dirSuffix) throws AccumuloException, AccumuloSecurityException,
+      TableExistsException, IOException, TableNotFoundException, MutationsRejectedException {
     c.tableOperations().create(tableName);
     FileSystem fs = FileSystem.get(CachedConfiguration.getInstance());
     String base = "target/accumulo-maven-plugin";
@@ -101,7 +101,7 @@ public class BulkIT extends AccumuloIT {
     opts.instance = c.getInstance().getInstanceName();
     opts.cols = 1;
     opts.setTableName(tableName);
-    String fileFormat = "/testrf/"+filePrefix+"rf%02d";
+    String fileFormat = "/testrf/" + filePrefix + "rf%02d";
     for (int i = 0; i < COUNT; i++) {
       opts.outputFile = base + String.format(fileFormat, i);
       opts.startRow = N * i;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/ChaoticBalancerIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/ChaoticBalancerIT.java b/test/src/test/java/org/apache/accumulo/test/functional/ChaoticBalancerIT.java
index 8ffdbd2..e80f823 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/ChaoticBalancerIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/ChaoticBalancerIT.java
@@ -38,10 +38,10 @@ public class ChaoticBalancerIT extends AccumuloClusterIT {
 
   @Override
   public void configureMiniCluster(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
-    Map<String,String> siteConfig = new HashMap<String, String>();
+    Map<String,String> siteConfig = new HashMap<String,String>();
     siteConfig.put(Property.TSERV_MAXMEM.getKey(), "10K");
     siteConfig.put(Property.TSERV_MAJC_DELAY.getKey(), "0");
-    cfg.setSiteConfig(siteConfig );
+    cfg.setSiteConfig(siteConfig);
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/DeleteEverythingIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/DeleteEverythingIT.java b/test/src/test/java/org/apache/accumulo/test/functional/DeleteEverythingIT.java
index 68911ca..2de5eda 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/DeleteEverythingIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/DeleteEverythingIT.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.test.functional;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
 
 import java.util.Collections;
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/DeleteTableDuringSplitIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/DeleteTableDuringSplitIT.java b/test/src/test/java/org/apache/accumulo/test/functional/DeleteTableDuringSplitIT.java
index 155c865..bc0503f 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/DeleteTableDuringSplitIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/DeleteTableDuringSplitIT.java
@@ -52,7 +52,7 @@ public class DeleteTableDuringSplitIT extends AccumuloClusterIT {
     }
     final SortedSet<Text> splits = new TreeSet<Text>();
     for (byte i = 0; i < 100; i++) {
-      splits.add(new Text(new byte[]{0, 0, i}));
+      splits.add(new Text(new byte[] {0, 0, i}));
     }
 
     List<Future<?>> results = new ArrayList<Future<?>>();
@@ -65,8 +65,7 @@ public class DeleteTableDuringSplitIT extends AccumuloClusterIT {
         public void run() {
           try {
             getConnector().tableOperations().addSplits(finalName, splits);
-          } catch (TableNotFoundException ex) {
-          } catch (Exception ex) {
+          } catch (TableNotFoundException ex) {} catch (Exception ex) {
             throw new RuntimeException(finalName, ex);
           }
         }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/DeletedTablesDontFlushIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/DeletedTablesDontFlushIT.java b/test/src/test/java/org/apache/accumulo/test/functional/DeletedTablesDontFlushIT.java
index 97c696e..292ad63 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/DeletedTablesDontFlushIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/DeletedTablesDontFlushIT.java
@@ -45,7 +45,7 @@ public class DeletedTablesDontFlushIT extends SharedMiniClusterIT {
 
     Mutation m = new Mutation("xyzzy");
     for (int i = 0; i < 100; i++) {
-      m.put("cf", "" + i, new Value(new byte[]{}));
+      m.put("cf", "" + i, new Value(new byte[] {}));
     }
     BatchWriter bw = c.createBatchWriter(tableName, new BatchWriterConfig());
     bw.addMutation(m);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/DurabilityIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/DurabilityIT.java b/test/src/test/java/org/apache/accumulo/test/functional/DurabilityIT.java
index 9c68f5f..7507290 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/DurabilityIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/DurabilityIT.java
@@ -81,13 +81,17 @@ public class DurabilityIT extends ConfigurableMacIT {
     TableOperations tableOps = getConnector().tableOperations();
     String tableNames[] = init();
     // write some gunk, delete the table to keep that table from messing with the performance numbers of successive calls
-    long t0 = writeSome(tableNames[0], N); tableOps.delete(tableNames[0]);
-    long t1 = writeSome(tableNames[1], N); tableOps.delete(tableNames[1]);
-    long t2 = writeSome(tableNames[2], N); tableOps.delete(tableNames[2]);
-    long t3 = writeSome(tableNames[3], N); tableOps.delete(tableNames[3]);
+    long t0 = writeSome(tableNames[0], N);
+    tableOps.delete(tableNames[0]);
+    long t1 = writeSome(tableNames[1], N);
+    tableOps.delete(tableNames[1]);
+    long t2 = writeSome(tableNames[2], N);
+    tableOps.delete(tableNames[2]);
+    long t3 = writeSome(tableNames[3], N);
+    tableOps.delete(tableNames[3]);
     System.out.println(String.format("sync %d flush %d log %d none %d", t0, t1, t2, t3));
-    assertTrue("flush-only should be faster than sync",   t0 > t1);
-    assertTrue("sync should be faster than log",          t1 > t2);
+    assertTrue("flush-only should be faster than sync", t0 > t1);
+    assertTrue("sync should be faster than log", t1 > t2);
     assertTrue("no durability should be faster than log", t2 > t3);
   }
 
@@ -147,8 +151,8 @@ public class DurabilityIT extends ConfigurableMacIT {
     assertTrue(N == readSome(tableName));
   }
 
-  private static Map<String, String> map(Iterable<Entry<String, String>> entries) {
-    Map<String, String> result = new HashMap<String,String>();
+  private static Map<String,String> map(Iterable<Entry<String,String>> entries) {
+    Map<String,String> result = new HashMap<String,String>();
     for (Entry<String,String> entry : entries) {
       result.put(entry.getKey(), entry.getValue());
     }
@@ -160,7 +164,7 @@ public class DurabilityIT extends ConfigurableMacIT {
     Connector c = getConnector();
     String tableName = getUniqueNames(1)[0];
     c.instanceOperations().setProperty(Property.TABLE_DURABILITY.getKey(), "none");
-    Map<String, String> props = map(c.tableOperations().getProperties(MetadataTable.NAME));
+    Map<String,String> props = map(c.tableOperations().getProperties(MetadataTable.NAME));
     assertEquals("sync", props.get(Property.TABLE_DURABILITY.getKey()));
     c.tableOperations().create(tableName);
     props = map(c.tableOperations().getProperties(tableName));
@@ -188,13 +192,13 @@ public class DurabilityIT extends ConfigurableMacIT {
       Mutation m = new Mutation("" + i);
       m.put("", "", "");
       bw.addMutation(m);
-      if (i % (Math.max(1, count/100)) == 0) {
+      if (i % (Math.max(1, count / 100)) == 0) {
         bw.flush();
       }
     }
     bw.close();
     long result = System.currentTimeMillis() - now;
-    //c.tableOperations().flush(table, null, null, true);
+    // c.tableOperations().flush(table, null, null, true);
     return result;
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/ExamplesIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/ExamplesIT.java b/test/src/test/java/org/apache/accumulo/test/functional/ExamplesIT.java
index b47c9ca..6a3e7d1 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/ExamplesIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/ExamplesIT.java
@@ -203,11 +203,8 @@ public class ExamplesIT extends AccumuloClusterIT {
         new String[] {"-i", instance, "-z", keepers, "-u", user, "-p", passwd, "--dirTable", dirTable, "--indexTable", indexTable, "--dataTable", dataTable,
             "--vis", visibility, "--chunkSize", Integer.toString(10000), getUsableDir()});
     assertEquals("Got non-zero return code. Stdout=" + entry.getValue(), 0, entry.getKey().intValue());
-    entry = getClusterControl()
-        .execWithStdout(
-            QueryUtil.class,
-        new String[] {"-i", instance, "-z", keepers, "-p", passwd, "-u", user, "-t", indexTable, "--auths", auths, "--search", "--path",
-                "accumulo-site.xml"});
+    entry = getClusterControl().execWithStdout(QueryUtil.class,
+        new String[] {"-i", instance, "-z", keepers, "-p", passwd, "-u", user, "-t", indexTable, "--auths", auths, "--search", "--path", "accumulo-site.xml"});
     if (ClusterType.MINI == getClusterType()) {
       MiniAccumuloClusterImpl impl = (MiniAccumuloClusterImpl) cluster;
       for (LogWriter writer : impl.getLogWriters()) {
@@ -360,8 +357,7 @@ public class ExamplesIT extends AccumuloClusterIT {
   @Test
   public void testTeraSortAndRead() throws Exception {
     String tableName = getUniqueNames(1)[0];
-    goodExec(TeraSortIngest.class, "--count", (1000 * 1000) + "", "-nk", "10", "-xk", "10", "-nv", "10", "-xv", "10", "-t", tableName, "-i", instance,
-        "-z",
+    goodExec(TeraSortIngest.class, "--count", (1000 * 1000) + "", "-nk", "10", "-xk", "10", "-nv", "10", "-xv", "10", "-t", tableName, "-i", instance, "-z",
         keepers, "-u", user, "-p", passwd, "--splits", "4");
     Path output = new Path(dir, "tmp/nines");
     if (fs.exists(output)) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/FunctionalTestUtils.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/FunctionalTestUtils.java b/test/src/test/java/org/apache/accumulo/test/functional/FunctionalTestUtils.java
index e4e7229..5f59a19 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/FunctionalTestUtils.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/FunctionalTestUtils.java
@@ -16,6 +16,8 @@
  */
 package org.apache.accumulo.test.functional;
 
+import static org.junit.Assert.assertFalse;
+
 import java.io.FileInputStream;
 import java.io.IOException;
 import java.io.InputStream;
@@ -47,8 +49,6 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 
-import static org.junit.Assert.assertFalse;
-
 public class FunctionalTestUtils {
 
   public static int countRFiles(Connector c, String tableName) throws Exception {
@@ -182,7 +182,8 @@ public class FunctionalTestUtils {
 
   public static int count(Iterable<?> i) {
     int count = 0;
-    for (@SuppressWarnings("unused") Object entry : i)
+    for (@SuppressWarnings("unused")
+    Object entry : i)
       count++;
     return count;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/HalfDeadTServerIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/HalfDeadTServerIT.java b/test/src/test/java/org/apache/accumulo/test/functional/HalfDeadTServerIT.java
index 0315891..a7b65cd 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/HalfDeadTServerIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/HalfDeadTServerIT.java
@@ -102,9 +102,9 @@ public class HalfDeadTServerIT extends ConfigurableMacIT {
   public void testTimeout() throws Exception {
     String results = test(20, true);
     if (results != null) {
-    	if (!results.contains("Session expired")) {
+      if (!results.contains("Session expired")) {
         log.info("Failed to find 'Session expired' in output, but TServer did die which is expected");
-    	}
+      }
     }
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/LateLastContactIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/LateLastContactIT.java b/test/src/test/java/org/apache/accumulo/test/functional/LateLastContactIT.java
index abc8182..7b8cb2b 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/LateLastContactIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/LateLastContactIT.java
@@ -26,11 +26,10 @@ import org.apache.hadoop.conf.Configuration;
 import org.junit.Test;
 
 /**
- * Fake the "tablet stops talking but holds its lock" problem we see when hard drives and NFS fail. 
- * Start a ZombieTServer, and see that master stops it.
+ * Fake the "tablet stops talking but holds its lock" problem we see when hard drives and NFS fail. Start a ZombieTServer, and see that master stops it.
  */
 public class LateLastContactIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setSiteConfig(Collections.singletonMap(Property.GENERAL_RPC_TIMEOUT.getKey(), "2s"));
@@ -46,5 +45,5 @@ public class LateLastContactIT extends ConfigurableMacIT {
     Process zombie = cluster.exec(ZombieTServer.class);
     assertEquals(0, zombie.waitFor());
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/LogicalTimeIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/LogicalTimeIT.java b/test/src/test/java/org/apache/accumulo/test/functional/LogicalTimeIT.java
index 6076868..ca0c62a 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/LogicalTimeIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/LogicalTimeIT.java
@@ -69,7 +69,8 @@ public class LogicalTimeIT extends AccumuloClusterIT {
 
   }
 
-  private void runMergeTest(Connector conn, String table, String[] splits, String[] inserts, String start, String end, String last, long expected) throws Exception {
+  private void runMergeTest(Connector conn, String table, String[] splits, String[] inserts, String start, String end, String last, long expected)
+      throws Exception {
     log.info("table " + table);
     conn.tableOperations().create(table, new NewTableConfiguration().setTimeType(TimeType.LOGICAL));
     TreeSet<Text> splitSet = new TreeSet<Text>();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/MaxOpenIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/MaxOpenIT.java b/test/src/test/java/org/apache/accumulo/test/functional/MaxOpenIT.java
index 729ec46..27c4010 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/MaxOpenIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/MaxOpenIT.java
@@ -48,7 +48,7 @@ public class MaxOpenIT extends AccumuloClusterIT {
 
   @Override
   public void configureMiniCluster(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
-    Map<String, String> conf = new HashMap<String, String>();
+    Map<String,String> conf = new HashMap<String,String>();
     conf.put(Property.TSERV_SCAN_MAX_OPENFILES.getKey(), "4");
     conf.put(Property.TSERV_MAJC_MAXCONCURRENT.getKey(), "1");
     conf.put(Property.TSERV_MAJC_THREAD_MAXOPEN.getKey(), "2");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/MetadataMaxFilesIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/MetadataMaxFilesIT.java b/test/src/test/java/org/apache/accumulo/test/functional/MetadataMaxFilesIT.java
index ed08cad..19977d3 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/MetadataMaxFilesIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/MetadataMaxFilesIT.java
@@ -46,7 +46,7 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class MetadataMaxFilesIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     Map<String,String> siteConfig = new HashMap<String,String>();
@@ -55,7 +55,7 @@ public class MetadataMaxFilesIT extends ConfigurableMacIT {
     cfg.setSiteConfig(siteConfig);
     hadoopCoreSite.set("fs.file.impl", RawLocalFileSystem.class.getName());
   }
-  
+
   @Override
   protected int defaultTimeoutSeconds() {
     return 4 * 60;
@@ -85,9 +85,9 @@ public class MetadataMaxFilesIT extends ConfigurableMacIT {
     cluster.stop();
     log.info("starting up");
     cluster.start();
-    
+
     UtilWaitThread.sleep(30 * 1000);
-    
+
     while (true) {
       MasterMonitorInfo stats = null;
       Credentials creds = new Credentials("root", new PasswordToken(ROOT_PASSWORD));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/MetadataSplitIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/MetadataSplitIT.java b/test/src/test/java/org/apache/accumulo/test/functional/MetadataSplitIT.java
index e0c8d0e..3930cda 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/MetadataSplitIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/MetadataSplitIT.java
@@ -30,12 +30,12 @@ import org.apache.hadoop.conf.Configuration;
 import org.junit.Test;
 
 public class MetadataSplitIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setSiteConfig(Collections.singletonMap(Property.TSERV_MAJC_DELAY.getKey(), "100ms"));
   }
- 
+
   @Override
   protected int defaultTimeoutSeconds() {
     return 2 * 60;
@@ -50,7 +50,7 @@ public class MetadataSplitIT extends ConfigurableMacIT {
       c.tableOperations().create("table" + i);
       c.tableOperations().flush(MetadataTable.NAME, null, null, true);
     }
-    UtilWaitThread.sleep(10*1000);
+    UtilWaitThread.sleep(10 * 1000);
     assertTrue(c.tableOperations().listSplits(MetadataTable.NAME).size() > 2);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/MonitorLoggingIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/MonitorLoggingIT.java b/test/src/test/java/org/apache/accumulo/test/functional/MonitorLoggingIT.java
index 803c6c3..cb87946 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/MonitorLoggingIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/MonitorLoggingIT.java
@@ -94,8 +94,8 @@ public class MonitorLoggingIT extends ConfigurableMacIT {
     }
 
     String result = "";
-    while(true) {
-      Thread.sleep(LOCATION_DELAY_SECS * 1000);  // extra precaution to ensure monitor has opportunity to log
+    while (true) {
+      Thread.sleep(LOCATION_DELAY_SECS * 1000); // extra precaution to ensure monitor has opportunity to log
 
       // Verify messages were received at the monitor.
       URL url = new URL("http://" + monitorLocation + "/log");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/NativeMapIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/NativeMapIT.java b/test/src/test/java/org/apache/accumulo/test/functional/NativeMapIT.java
index f4101de..9175379 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/NativeMapIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/NativeMapIT.java
@@ -319,7 +319,7 @@ public class NativeMapIT {
     tm.put(new Key(new Text("foo")), new Value(new byte[] {'1'}));
     tm.put(new Key(new Text("foo1")), new Value(new byte[] {'2'}));
     tm.put(new Key(new Text("foo2")), new Value(new byte[] {'3'}));
- 
+
     for (Entry<Key,Value> entry : tm.entrySet()) {
       nm.put(entry.getKey(), entry.getValue());
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/RecoveryWithEmptyRFileIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/RecoveryWithEmptyRFileIT.java b/test/src/test/java/org/apache/accumulo/test/functional/RecoveryWithEmptyRFileIT.java
index b2ee4ce..5edc137 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/RecoveryWithEmptyRFileIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/RecoveryWithEmptyRFileIT.java
@@ -39,16 +39,12 @@ import org.apache.log4j.Logger;
 import org.junit.Test;
 
 /**
-  XXX As a part of verifying lossy recovery via inserting an empty rfile,
-  this test deletes test table tablets. This will require write access to
-  the backing files of the test Accumulo mini cluster.
-
-  This test should read the file location from the test harness and that
-  file should be on the local filesystem. If you want to take a paranoid
-  approach just make sure the test user doesn't have write access to the
-  HDFS files of any colocated live Accumulo instance or any important
-  local filesystem files..
-*/
+ * XXX As a part of verifying lossy recovery via inserting an empty rfile, this test deletes test table tablets. This will require write access to the backing
+ * files of the test Accumulo mini cluster.
+ *
+ * This test should read the file location from the test harness and that file should be on the local filesystem. If you want to take a paranoid approach just
+ * make sure the test user doesn't have write access to the HDFS files of any colocated live Accumulo instance or any important local filesystem files..
+ */
 public class RecoveryWithEmptyRFileIT extends ConfigurableMacIT {
   private static final Logger log = Logger.getLogger(RecoveryWithEmptyRFileIT.class);
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/RestartStressIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/RestartStressIT.java b/test/src/test/java/org/apache/accumulo/test/functional/RestartStressIT.java
index f4ff912..253a59c 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/RestartStressIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/RestartStressIT.java
@@ -110,8 +110,8 @@ public class RestartStressIT extends AccumuloClusterIT {
       public Integer call() {
         try {
           return control.exec(TestIngest.class,
-              new String[] {"-u", "root", "-p", new String(token.getPassword(), Charsets.UTF_8), "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(),
-                  "--rows", "" + IOPTS.rows, "--table", tableName});
+              new String[] {"-u", "root", "-p", new String(token.getPassword(), Charsets.UTF_8), "-i", cluster.getInstanceName(), "-z",
+                  cluster.getZooKeepers(), "--rows", "" + IOPTS.rows, "--table", tableName});
         } catch (Exception e) {
           log.error("Error running TestIngest", e);
           return -1;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/RowDeleteIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/RowDeleteIT.java b/test/src/test/java/org/apache/accumulo/test/functional/RowDeleteIT.java
index bb681c0..b43a92a 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/RowDeleteIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/RowDeleteIT.java
@@ -58,9 +58,9 @@ public class RowDeleteIT extends AccumuloClusterIT {
     Connector c = getConnector();
     String tableName = getUniqueNames(1)[0];
     c.tableOperations().create(tableName);
-    Map<String,Set<Text>> groups = new HashMap<String, Set<Text>>();
+    Map<String,Set<Text>> groups = new HashMap<String,Set<Text>>();
     groups.put("lg1", Collections.singleton(new Text("foo")));
-    groups.put("dg", Collections.<Text>emptySet());
+    groups.put("dg", Collections.<Text> emptySet());
     c.tableOperations().setLocalityGroups(tableName, groups);
     IteratorSetting setting = new IteratorSetting(30, RowDeletingIterator.class);
     c.tableOperations().attachIterator(tableName, setting, EnumSet.of(IteratorScope.majc));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/ScanIdIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/ScanIdIT.java b/test/src/test/java/org/apache/accumulo/test/functional/ScanIdIT.java
index fe2e8cb..de4a47e 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/ScanIdIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/ScanIdIT.java
@@ -16,6 +16,10 @@
  */
 package org.apache.accumulo.test.functional;
 
+import static com.google.common.base.Charsets.UTF_8;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
 import java.util.EnumSet;
 import java.util.HashSet;
 import java.util.List;
@@ -52,17 +56,13 @@ import org.junit.Test;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import static com.google.common.base.Charsets.UTF_8;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-
 /**
  * ACCUMULO-2641 Integration test. ACCUMULO-2641 Adds scan id to thrift protocol so that {@code org.apache.accumulo.core.client.admin.ActiveScan.getScanid()}
- * returns a unique scan id.<p>
+ * returns a unique scan id.
+ * <p>
  * <p/>
- * The test uses the Minicluster and the {@code org.apache.accumulo.test.functional.SlowIterator} to create multiple scan sessions.
- * The test exercises multiple tablet servers with splits and multiple ranges to force the scans to occur across multiple tablet servers
- * for completeness.
+ * The test uses the Minicluster and the {@code org.apache.accumulo.test.functional.SlowIterator} to create multiple scan sessions. The test exercises multiple
+ * tablet servers with splits and multiple ranges to force the scans to occur across multiple tablet servers for completeness.
  * <p/>
  * This patch modified thrift, the TraceRepoDeserializationTest test seems to fail unless the following be added:
  * <p/>
@@ -92,7 +92,8 @@ public class ScanIdIT extends AccumuloClusterIT {
   }
 
   /**
-   * @throws Exception any exception is a test failure.
+   * @throws Exception
+   *           any exception is a test failure.
    */
   @Test
   public void testScanId() throws Exception {
@@ -174,15 +175,14 @@ public class ScanIdIT extends AccumuloClusterIT {
     }
 
     /**
-     * execute the scan across the sample data and put scan result into result map until
-     * testInProgress flag is set to false.
+     * execute the scan across the sample data and put scan result into result map until testInProgress flag is set to false.
      */
-    @Override public void run() {
+    @Override
+    public void run() {
 
       /*
-      * set random initial delay of up to to
-      * allow scanners to proceed to different points.
-      */
+       * set random initial delay of up to to allow scanners to proceed to different points.
+       */
 
       long delay = random.nextInt(5000);
 
@@ -245,10 +245,10 @@ public class ScanIdIT extends AccumuloClusterIT {
   }
 
   /**
-   * Create splits on table and force migration by taking table offline and then bring back
-   * online for test.
+   * Create splits on table and force migration by taking table offline and then bring back online for test.
    *
-   * @param conn Accumulo connector Accumulo connector to test cluster or MAC instance.
+   * @param conn
+   *          Accumulo connector Accumulo connector to test cluster or MAC instance.
    */
   private void addSplits(final Connector conn, final String tableName) {
 
@@ -296,11 +296,11 @@ public class ScanIdIT extends AccumuloClusterIT {
   /**
    * Generate some sample data using random row id to distribute across splits.
    * <p/>
-   * The primary goal is to determine that each scanner is assigned a unique scan id.
-   * This test does check that the count value  for fam1 increases if a scanner reads multiple value, but this is
-   * secondary consideration for this test, that is included for completeness.
+   * The primary goal is to determine that each scanner is assigned a unique scan id. This test does check that the count value for fam1 increases if a scanner
+   * reads multiple value, but this is secondary consideration for this test, that is included for completeness.
    *
-   * @param connector Accumulo connector Accumulo connector to test cluster or MAC instance.
+   * @param connector
+   *          Accumulo connector Accumulo connector to test cluster or MAC instance.
    */
   private void generateSampleData(Connector connector, final String tablename) {
 
@@ -333,11 +333,11 @@ public class ScanIdIT extends AccumuloClusterIT {
   }
 
   /**
-   * Attach the test slow iterator so that we have time to read the scan id without creating a large dataset. Uses a
-   * fairly large sleep and delay times because we are not concerned with how much data is read and we do not read
-   * all of the data - the test stops once each scanner reports a scan id.
+   * Attach the test slow iterator so that we have time to read the scan id without creating a large dataset. Uses a fairly large sleep and delay times because
+   * we are not concerned with how much data is read and we do not read all of the data - the test stops once each scanner reports a scan id.
    *
-   * @param connector Accumulo connector Accumulo connector to test cluster or MAC instance.
+   * @param connector
+   *          Accumulo connector Accumulo connector to test cluster or MAC instance.
    */
   private void attachSlowIterator(Connector connector, final String tablename) {
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/SessionDurabilityIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/SessionDurabilityIT.java b/test/src/test/java/org/apache/accumulo/test/functional/SessionDurabilityIT.java
index 58e6007..f0df2d0 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/SessionDurabilityIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/SessionDurabilityIT.java
@@ -135,7 +135,7 @@ public class SessionDurabilityIT extends ConfigurableMacIT {
     Connector c = getConnector();
     ConditionalWriter cw = c.createConditionalWriter(tableName, cfg);
     for (int i = 0; i < n; i++) {
-      ConditionalMutation m = new ConditionalMutation((CharSequence)(i + ""), new Condition("", ""));
+      ConditionalMutation m = new ConditionalMutation((CharSequence) (i + ""), new Condition("", ""));
       m.put("", "", "X");
       assertEquals(Status.ACCEPTED, cw.write(m).getStatus());
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/ShutdownIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/ShutdownIT.java b/test/src/test/java/org/apache/accumulo/test/functional/ShutdownIT.java
index 1b9aa21..c3042d7 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/ShutdownIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/ShutdownIT.java
@@ -41,7 +41,8 @@ public class ShutdownIT extends ConfigurableMacIT {
 
   @Test
   public void shutdownDuringIngest() throws Exception {
-    Process ingest = cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD, "--createTable");
+    Process ingest = cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD,
+        "--createTable");
     UtilWaitThread.sleep(100);
     assertEquals(0, cluster.exec(Admin.class, "stopAll").waitFor());
     ingest.destroy();
@@ -49,7 +50,9 @@ public class ShutdownIT extends ConfigurableMacIT {
 
   @Test
   public void shutdownDuringQuery() throws Exception {
-    assertEquals(0, cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root","-p", ROOT_PASSWORD, "--createTable").waitFor());
+    assertEquals(0,
+        cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD, "--createTable")
+            .waitFor());
     Process verify = cluster.exec(VerifyIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD);
     UtilWaitThread.sleep(100);
     assertEquals(0, cluster.exec(Admin.class, "stopAll").waitFor());
@@ -58,18 +61,19 @@ public class ShutdownIT extends ConfigurableMacIT {
 
   @Test
   public void shutdownDuringDelete() throws Exception {
-    assertEquals(0, cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD, "--createTable").waitFor());
+    assertEquals(0,
+        cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD, "--createTable")
+            .waitFor());
     Process deleter = cluster.exec(TestRandomDeletes.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD);
     UtilWaitThread.sleep(100);
     assertEquals(0, cluster.exec(Admin.class, "stopAll").waitFor());
     deleter.destroy();
   }
 
-
   @Test
   public void shutdownDuringDeleteTable() throws Exception {
     final Connector c = getConnector();
-    for (int i = 0; i < 10 ; i++) {
+    for (int i = 0; i < 10; i++) {
       c.tableOperations().create("table" + i);
     }
     final AtomicReference<Exception> ref = new AtomicReference<Exception>();
@@ -102,7 +106,9 @@ public class ShutdownIT extends ConfigurableMacIT {
   }
 
   static void runAdminStopTest(Connector c, MiniAccumuloClusterImpl cluster) throws InterruptedException, IOException {
-    assertEquals(0, cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD, "--createTable").waitFor());
+    assertEquals(0,
+        cluster.exec(TestIngest.class, "-i", cluster.getInstanceName(), "-z", cluster.getZooKeepers(), "-u", "root", "-p", ROOT_PASSWORD, "--createTable")
+            .waitFor());
     List<String> tabletServers = c.instanceOperations().getTabletServers();
     assertEquals(2, tabletServers.size());
     String doomed = tabletServers.get(0);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/SplitRecoveryIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/SplitRecoveryIT.java b/test/src/test/java/org/apache/accumulo/test/functional/SplitRecoveryIT.java
index d639090..7439e23 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/SplitRecoveryIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/SplitRecoveryIT.java
@@ -69,16 +69,16 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class SplitRecoveryIT extends ConfigurableMacIT {
-  
+
   @Override
   protected int defaultTimeoutSeconds() {
     return 60;
   }
-  
+
   private KeyExtent nke(String table, String endRow, String prevEndRow) {
     return new KeyExtent(new Text(table), endRow == null ? null : new Text(endRow), prevEndRow == null ? null : new Text(prevEndRow));
   }
-  
+
   private void run() throws Exception {
     Instance inst = HdfsZooInstance.getInstance();
     AccumuloServerContext c = new AccumuloServerContext(new ServerConfigurationFactory(inst));
@@ -87,61 +87,61 @@ public class SplitRecoveryIT extends ConfigurableMacIT {
     zoo.putPersistentData(zPath, new byte[0], NodeExistsPolicy.OVERWRITE);
     ZooLock zl = new ZooLock(zPath);
     boolean gotLock = zl.tryLock(new LockWatcher() {
-      
+
       @Override
       public void lostLock(LockLossReason reason) {
         System.exit(-1);
-        
+
       }
-      
+
       @Override
       public void unableToMonitorLockNode(Throwable e) {
         System.exit(-1);
       }
     }, "foo".getBytes(UTF_8));
-    
+
     if (!gotLock) {
       System.err.println("Failed to get lock " + zPath);
     }
-    
+
     // run test for a table with one tablet
     runSplitRecoveryTest(c, 0, "sp", 0, zl, nke("foo0", null, null));
     runSplitRecoveryTest(c, 1, "sp", 0, zl, nke("foo1", null, null));
-    
+
     // run test for tables with two tablets, run test on first and last tablet
     runSplitRecoveryTest(c, 0, "k", 0, zl, nke("foo2", "m", null), nke("foo2", null, "m"));
     runSplitRecoveryTest(c, 1, "k", 0, zl, nke("foo3", "m", null), nke("foo3", null, "m"));
     runSplitRecoveryTest(c, 0, "o", 1, zl, nke("foo4", "m", null), nke("foo4", null, "m"));
     runSplitRecoveryTest(c, 1, "o", 1, zl, nke("foo5", "m", null), nke("foo5", null, "m"));
-    
+
     // run test for table w/ three tablets, run test on middle tablet
     runSplitRecoveryTest(c, 0, "o", 1, zl, nke("foo6", "m", null), nke("foo6", "r", "m"), nke("foo6", null, "r"));
     runSplitRecoveryTest(c, 1, "o", 1, zl, nke("foo7", "m", null), nke("foo7", "r", "m"), nke("foo7", null, "r"));
-    
+
     // run test for table w/ three tablets, run test on first
     runSplitRecoveryTest(c, 0, "g", 0, zl, nke("foo8", "m", null), nke("foo8", "r", "m"), nke("foo8", null, "r"));
     runSplitRecoveryTest(c, 1, "g", 0, zl, nke("foo9", "m", null), nke("foo9", "r", "m"), nke("foo9", null, "r"));
-    
+
     // run test for table w/ three tablets, run test on last tablet
     runSplitRecoveryTest(c, 0, "w", 2, zl, nke("fooa", "m", null), nke("fooa", "r", "m"), nke("fooa", null, "r"));
     runSplitRecoveryTest(c, 1, "w", 2, zl, nke("foob", "m", null), nke("foob", "r", "m"), nke("foob", null, "r"));
   }
-  
+
   private void runSplitRecoveryTest(AccumuloServerContext context, int failPoint, String mr, int extentToSplit, ZooLock zl, KeyExtent... extents)
       throws Exception {
-    
+
     Text midRow = new Text(mr);
-    
+
     SortedMap<FileRef,DataFileValue> splitMapFiles = null;
-    
+
     for (int i = 0; i < extents.length; i++) {
       KeyExtent extent = extents[i];
-      
+
       String tdir = ServerConstants.getTablesDirs()[0] + "/" + extent.getTableId().toString() + "/dir_" + i;
       MetadataTableUtil.addTablet(extent, tdir, context, TabletTime.LOGICAL_TIME_ID, zl);
       SortedMap<FileRef,DataFileValue> mapFiles = new TreeMap<FileRef,DataFileValue>();
       mapFiles.put(new FileRef(tdir + "/" + RFile.EXTENSION + "_000_000"), new DataFileValue(1000017 + i, 10000 + i));
-      
+
       if (i == extentToSplit) {
         splitMapFiles = mapFiles;
       }
@@ -149,25 +149,25 @@ public class SplitRecoveryIT extends ConfigurableMacIT {
       TransactionWatcher.ZooArbitrator.start(Constants.BULK_ARBITRATOR_TYPE, tid);
       MetadataTableUtil.updateTabletDataFile(tid, extent, mapFiles, "L0", context, zl);
     }
-    
+
     KeyExtent extent = extents[extentToSplit];
-    
+
     KeyExtent high = new KeyExtent(extent.getTableId(), extent.getEndRow(), midRow);
     KeyExtent low = new KeyExtent(extent.getTableId(), midRow, extent.getPrevEndRow());
-    
+
     splitPartiallyAndRecover(context, extent, high, low, .4, splitMapFiles, midRow, "localhost:1234", failPoint, zl);
   }
-  
+
   private void splitPartiallyAndRecover(AccumuloServerContext context, KeyExtent extent, KeyExtent high, KeyExtent low, double splitRatio,
       SortedMap<FileRef,DataFileValue> mapFiles, Text midRow, String location, int steps, ZooLock zl) throws Exception {
-    
+
     SortedMap<FileRef,DataFileValue> lowDatafileSizes = new TreeMap<FileRef,DataFileValue>();
     SortedMap<FileRef,DataFileValue> highDatafileSizes = new TreeMap<FileRef,DataFileValue>();
     List<FileRef> highDatafilesToRemove = new ArrayList<FileRef>();
-    
+
     MetadataTableUtil.splitDatafiles(extent.getTableId(), midRow, splitRatio, new HashMap<FileRef,FileUtil.FileInfo>(), mapFiles, lowDatafileSizes,
         highDatafileSizes, highDatafilesToRemove);
-    
+
     MetadataTableUtil.splitTablet(high, extent.getPrevEndRow(), splitRatio, context, zl);
     TServerInstance instance = new TServerInstance(location, zl.getSessionId());
     Writer writer = MetadataTableUtil.getMetadataTable(context);
@@ -175,27 +175,27 @@ public class SplitRecoveryIT extends ConfigurableMacIT {
     Mutation m = new Mutation(assignment.tablet.getMetadataEntry());
     assignment.server.putFutureLocation(m);
     writer.update(m);
-    
+
     if (steps >= 1) {
       Map<FileRef,Long> bulkFiles = MetadataTableUtil.getBulkFilesLoaded(context, extent);
       MasterMetadataUtil.addNewTablet(context, low, "/lowDir", instance, lowDatafileSizes, bulkFiles, TabletTime.LOGICAL_TIME_ID + "0", -1l, -1l, zl);
     }
     if (steps >= 2)
       MetadataTableUtil.finishSplit(high, highDatafileSizes, highDatafilesToRemove, context, zl);
-    
+
     TabletServer.verifyTabletInformation(context, high, instance, null, "127.0.0.1:0", zl);
-    
+
     if (steps >= 1) {
       ensureTabletHasNoUnexpectedMetadataEntries(context, low, lowDatafileSizes);
       ensureTabletHasNoUnexpectedMetadataEntries(context, high, highDatafileSizes);
-      
+
       Map<FileRef,Long> lowBulkFiles = MetadataTableUtil.getBulkFilesLoaded(context, low);
       Map<FileRef,Long> highBulkFiles = MetadataTableUtil.getBulkFilesLoaded(context, high);
-      
+
       if (!lowBulkFiles.equals(highBulkFiles)) {
         throw new Exception(" " + lowBulkFiles + " != " + highBulkFiles + " " + low + " " + high);
       }
-      
+
       if (lowBulkFiles.size() == 0) {
         throw new Exception(" no bulk files " + low);
       }
@@ -203,62 +203,62 @@ public class SplitRecoveryIT extends ConfigurableMacIT {
       ensureTabletHasNoUnexpectedMetadataEntries(context, extent, mapFiles);
     }
   }
-  
+
   private void ensureTabletHasNoUnexpectedMetadataEntries(AccumuloServerContext context, KeyExtent extent, SortedMap<FileRef,DataFileValue> expectedMapFiles)
       throws Exception {
     Scanner scanner = new ScannerImpl(context, MetadataTable.ID, Authorizations.EMPTY);
     scanner.setRange(extent.toMetadataRange());
-    
+
     HashSet<ColumnFQ> expectedColumns = new HashSet<ColumnFQ>();
     expectedColumns.add(TabletsSection.ServerColumnFamily.DIRECTORY_COLUMN);
     expectedColumns.add(TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN);
     expectedColumns.add(TabletsSection.ServerColumnFamily.TIME_COLUMN);
     expectedColumns.add(TabletsSection.ServerColumnFamily.LOCK_COLUMN);
-    
+
     HashSet<Text> expectedColumnFamilies = new HashSet<Text>();
     expectedColumnFamilies.add(DataFileColumnFamily.NAME);
     expectedColumnFamilies.add(TabletsSection.FutureLocationColumnFamily.NAME);
     expectedColumnFamilies.add(TabletsSection.CurrentLocationColumnFamily.NAME);
     expectedColumnFamilies.add(TabletsSection.LastLocationColumnFamily.NAME);
     expectedColumnFamilies.add(TabletsSection.BulkFileColumnFamily.NAME);
-    
+
     Iterator<Entry<Key,Value>> iter = scanner.iterator();
     while (iter.hasNext()) {
       Key key = iter.next().getKey();
-      
+
       if (!key.getRow().equals(extent.getMetadataEntry())) {
         throw new Exception("Tablet " + extent + " contained unexpected " + MetadataTable.NAME + " entry " + key);
       }
-      
+
       if (expectedColumnFamilies.contains(key.getColumnFamily())) {
         continue;
       }
-      
+
       if (expectedColumns.remove(new ColumnFQ(key))) {
         continue;
       }
-      
+
       throw new Exception("Tablet " + extent + " contained unexpected " + MetadataTable.NAME + " entry " + key);
     }
     System.out.println("expectedColumns " + expectedColumns);
     if (expectedColumns.size() > 1 || (expectedColumns.size() == 1)) {
       throw new Exception("Not all expected columns seen " + extent + " " + expectedColumns);
     }
-    
+
     SortedMap<FileRef,DataFileValue> fixedMapFiles = MetadataTableUtil.getDataFileSizes(extent, context);
     verifySame(expectedMapFiles, fixedMapFiles);
   }
-  
+
   private void verifySame(SortedMap<FileRef,DataFileValue> datafileSizes, SortedMap<FileRef,DataFileValue> fixedDatafileSizes) throws Exception {
-    
+
     if (!datafileSizes.keySet().containsAll(fixedDatafileSizes.keySet()) || !fixedDatafileSizes.keySet().containsAll(datafileSizes.keySet())) {
       throw new Exception("Key sets not the same " + datafileSizes.keySet() + " !=  " + fixedDatafileSizes.keySet());
     }
-    
+
     for (Entry<FileRef,DataFileValue> entry : datafileSizes.entrySet()) {
       DataFileValue dfv = entry.getValue();
       DataFileValue otherDfv = fixedDatafileSizes.get(entry.getKey());
-      
+
       if (!dfv.equals(otherDfv)) {
         throw new Exception(entry.getKey() + " dfv not equal  " + dfv + "  " + otherDfv);
       }
@@ -268,7 +268,7 @@ public class SplitRecoveryIT extends ConfigurableMacIT {
   public static void main(String[] args) throws Exception {
     new SplitRecoveryIT().run();
   }
-  
+
   @Test
   public void test() throws Exception {
     assertEquals(0, exec(SplitRecoveryIT.class).waitFor());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/WatchTheWatchCountIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/WatchTheWatchCountIT.java b/test/src/test/java/org/apache/accumulo/test/functional/WatchTheWatchCountIT.java
index 1291204..bd0555b 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/WatchTheWatchCountIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/WatchTheWatchCountIT.java
@@ -34,7 +34,7 @@ public class WatchTheWatchCountIT extends ConfigurableMacIT {
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
     cfg.setNumTservers(3);
   }
-  
+
   @Test(timeout = 30 * 1000)
   public void test() throws Exception {
     Connector c = getConnector();
@@ -57,6 +57,5 @@ public class WatchTheWatchCountIT extends ConfigurableMacIT {
       socket.close();
     }
   }
-  
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/WriteAheadLogIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/WriteAheadLogIT.java b/test/src/test/java/org/apache/accumulo/test/functional/WriteAheadLogIT.java
index 2f41323..008476f 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/WriteAheadLogIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/WriteAheadLogIT.java
@@ -35,7 +35,7 @@ public class WriteAheadLogIT extends AccumuloClusterIT {
 
   @Override
   public void configureMiniCluster(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
-    Map<String, String> siteConfig = new HashMap<String, String>();
+    Map<String,String> siteConfig = new HashMap<String,String>();
     siteConfig.put(Property.TSERV_WALOG_MAX_SIZE.getKey(), "2M");
     siteConfig.put(Property.GC_CYCLE_DELAY.getKey(), "1");
     siteConfig.put(Property.GC_CYCLE_START.getKey(), "1");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/functional/ZookeeperRestartIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/functional/ZookeeperRestartIT.java b/test/src/test/java/org/apache/accumulo/test/functional/ZookeeperRestartIT.java
index 9becfc2..f852901 100644
--- a/test/src/test/java/org/apache/accumulo/test/functional/ZookeeperRestartIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/functional/ZookeeperRestartIT.java
@@ -41,10 +41,10 @@ import org.apache.hadoop.conf.Configuration;
 import org.junit.Test;
 
 public class ZookeeperRestartIT extends ConfigurableMacIT {
-  
+
   @Override
   public void configure(MiniAccumuloConfigImpl cfg, Configuration hadoopCoreSite) {
-    Map<String,String> siteConfig = new HashMap<String, String>();
+    Map<String,String> siteConfig = new HashMap<String,String>();
     siteConfig.put(Property.INSTANCE_ZK_TIMEOUT.getKey(), "3s");
     cfg.setSiteConfig(siteConfig);
   }
@@ -63,17 +63,17 @@ public class ZookeeperRestartIT extends ConfigurableMacIT {
     m.put("cf", "cq", "value");
     bw.addMutation(m);
     bw.close();
-    
+
     // kill zookeeper
     for (ProcessReference proc : cluster.getProcesses().get(ServerType.ZOOKEEPER))
       cluster.killProcess(ServerType.ZOOKEEPER, proc);
-    
+
     // give the servers time to react
     UtilWaitThread.sleep(1000);
-    
+
     // start zookeeper back up
     cluster.start();
-    
+
     // use the tservers
     Scanner s = c.createScanner("test_ingest", Authorizations.EMPTY);
     Iterator<Entry<Key,Value>> i = s.iterator();
@@ -83,5 +83,5 @@ public class ZookeeperRestartIT extends ConfigurableMacIT {
     // use the master
     c.tableOperations().delete("test_ingest");
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/iterator/RegExTest.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/iterator/RegExTest.java b/test/src/test/java/org/apache/accumulo/test/iterator/RegExTest.java
index 3583894..82c89be 100644
--- a/test/src/test/java/org/apache/accumulo/test/iterator/RegExTest.java
+++ b/test/src/test/java/org/apache/accumulo/test/iterator/RegExTest.java
@@ -40,23 +40,23 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class RegExTest {
-  
+
   Instance inst = new MockInstance();
   Connector conn;
-  
+
   @Test
   public void runTest() throws Exception {
     conn = inst.getConnector("user", new PasswordToken("pass"));
     conn.tableOperations().create("ret");
     BatchWriter bw = conn.createBatchWriter("ret", new BatchWriterConfig());
-    
+
     ArrayList<Character> chars = new ArrayList<Character>();
     for (char c = 'a'; c <= 'z'; c++)
       chars.add(c);
-    
+
     for (char c = '0'; c <= '9'; c++)
       chars.add(c);
-    
+
     // insert some data into accumulo
     for (Character rc : chars) {
       Mutation m = new Mutation(new Text("r" + rc));
@@ -66,76 +66,76 @@ public class RegExTest {
           m.put(new Text("cf" + cfc), new Text("cq" + cqc), v);
         }
       }
-      
+
       bw.addMutation(m);
     }
-    
+
     bw.close();
-    
+
     runTest1();
     runTest2();
     runTest3();
     runTest4();
     runTest5();
   }
-  
+
   private void check(String regex, String val) throws Exception {
     if (regex != null && !val.matches(regex))
       throw new Exception(" " + val + " does not match " + regex);
   }
-  
+
   private void check(String regex, Text val) throws Exception {
     check(regex, val.toString());
   }
-  
+
   private void check(String regex, Value val) throws Exception {
     check(regex, val.toString());
   }
-  
+
   private void runTest1() throws Exception {
     // try setting all regex
     Range range = new Range(new Text("rf"), true, new Text("rl"), true);
     runTest(range, "r[g-k]", "cf[1-5]", "cq[x-z]", "v[g-k][1-5][t-y]", 5 * 5 * (3 - 1));
   }
-  
+
   private void runTest2() throws Exception {
     // try setting only a row regex
     Range range = new Range(new Text("rf"), true, new Text("rl"), true);
     runTest(range, "r[g-k]", null, null, null, 5 * 36 * 36);
   }
-  
+
   private void runTest3() throws Exception {
     // try setting only a col fam regex
     Range range = new Range((Key) null, (Key) null);
     runTest(range, null, "cf[a-f]", null, null, 36 * 6 * 36);
   }
-  
+
   private void runTest4() throws Exception {
     // try setting only a col qual regex
     Range range = new Range((Key) null, (Key) null);
     runTest(range, null, null, "cq[1-7]", null, 36 * 36 * 7);
   }
-  
+
   private void runTest5() throws Exception {
     // try setting only a value regex
     Range range = new Range((Key) null, (Key) null);
     runTest(range, null, null, null, "v[a-c][d-f][g-i]", 3 * 3 * 3);
   }
-  
+
   private void runTest(Range range, String rowRegEx, String cfRegEx, String cqRegEx, String valRegEx, int expected) throws Exception {
-    
+
     Scanner s = conn.createScanner("ret", Authorizations.EMPTY);
     s.setRange(range);
     setRegexs(s, rowRegEx, cfRegEx, cqRegEx, valRegEx);
     runTest(s, rowRegEx, cfRegEx, cqRegEx, valRegEx, expected);
-    
+
     BatchScanner bs = conn.createBatchScanner("ret", Authorizations.EMPTY, 1);
     bs.setRanges(Collections.singletonList(range));
     setRegexs(bs, rowRegEx, cfRegEx, cqRegEx, valRegEx);
     runTest(bs, rowRegEx, cfRegEx, cqRegEx, valRegEx, expected);
     bs.close();
   }
-  
+
   private void setRegexs(ScannerBase scanner, String rowRegEx, String cfRegEx, String cqRegEx, String valRegEx) {
     IteratorSetting regex = new IteratorSetting(50, "regex", RegExFilter.class);
     if (rowRegEx != null)
@@ -148,22 +148,22 @@ public class RegExTest {
       regex.addOption(RegExFilter.VALUE_REGEX, valRegEx);
     scanner.addScanIterator(regex);
   }
-  
+
   private void runTest(Iterable<Entry<Key,Value>> scanner, String rowRegEx, String cfRegEx, String cqRegEx, String valRegEx, int expected) throws Exception {
-    
+
     int counter = 0;
-    
+
     for (Entry<Key,Value> entry : scanner) {
       Key k = entry.getKey();
-      
+
       check(rowRegEx, k.getRow());
       check(cfRegEx, k.getColumnFamily());
       check(cqRegEx, k.getColumnQualifier());
       check(valRegEx, entry.getValue());
-      
+
       counter++;
     }
-    
+
     if (counter != expected) {
       throw new Exception("scan did not return the expected number of entries " + counter + " " + expected);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/replication/CyclicReplicationIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/replication/CyclicReplicationIT.java b/test/src/test/java/org/apache/accumulo/test/replication/CyclicReplicationIT.java
index 73c8896..f9cff98 100644
--- a/test/src/test/java/org/apache/accumulo/test/replication/CyclicReplicationIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/replication/CyclicReplicationIT.java
@@ -151,7 +151,7 @@ public class CyclicReplicationIT {
     MiniAccumuloConfigImpl master1Cfg;
     MiniAccumuloClusterImpl master1Cluster;
     while (true) {
-      master1Cfg= new MiniAccumuloConfigImpl(master1Dir, password);
+      master1Cfg = new MiniAccumuloConfigImpl(master1Dir, password);
       master1Cfg.setNumTservers(1);
       master1Cfg.setInstanceName("master1");
 
@@ -238,13 +238,11 @@ public class CyclicReplicationIT {
 
       // Replicate master1 in the master1 cluster to master2 in the master2 cluster
       connMaster1.tableOperations().setProperty(master1Table, Property.TABLE_REPLICATION.getKey(), "true");
-      connMaster1.tableOperations().setProperty(master1Table,
-          Property.TABLE_REPLICATION_TARGET.getKey() + master2Cluster.getInstanceName(), master2TableId);
+      connMaster1.tableOperations().setProperty(master1Table, Property.TABLE_REPLICATION_TARGET.getKey() + master2Cluster.getInstanceName(), master2TableId);
 
       // Replicate master2 in the master2 cluster to master1 in the master2 cluster
       connMaster2.tableOperations().setProperty(master2Table, Property.TABLE_REPLICATION.getKey(), "true");
-      connMaster2.tableOperations().setProperty(master2Table,
-          Property.TABLE_REPLICATION_TARGET.getKey() + master1Cluster.getInstanceName(), master1TableId);
+      connMaster2.tableOperations().setProperty(master2Table, Property.TABLE_REPLICATION_TARGET.getKey() + master1Cluster.getInstanceName(), master1TableId);
 
       // Give our replication user the ability to write to the respective table
       connMaster1.securityOperations().grantTablePermission(master1UserName, master1Table, TablePermission.WRITE);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/replication/GarbageCollectorCommunicatesWithTServersIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/replication/GarbageCollectorCommunicatesWithTServersIT.java b/test/src/test/java/org/apache/accumulo/test/replication/GarbageCollectorCommunicatesWithTServersIT.java
index 9d39325..1ef47e5 100644
--- a/test/src/test/java/org/apache/accumulo/test/replication/GarbageCollectorCommunicatesWithTServersIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/replication/GarbageCollectorCommunicatesWithTServersIT.java
@@ -378,7 +378,7 @@ public class GarbageCollectorCommunicatesWithTServersIT extends ConfigurableMacI
     bw = conn.createBatchWriter(otherTable, null);
     // 500k
     byte[] bigValue = new byte[1024 * 500];
-    Arrays.fill(bigValue, (byte)1);
+    Arrays.fill(bigValue, (byte) 1);
     // 500k * 50
     for (int i = 0; i < 50; i++) {
       Mutation m = new Mutation(Integer.toString(i));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/replication/MultiTserverReplicationIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/replication/MultiTserverReplicationIT.java b/test/src/test/java/org/apache/accumulo/test/replication/MultiTserverReplicationIT.java
index 554b08e..6b24e99 100644
--- a/test/src/test/java/org/apache/accumulo/test/replication/MultiTserverReplicationIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/replication/MultiTserverReplicationIT.java
@@ -41,7 +41,7 @@ import com.google.common.collect.Iterables;
 import com.google.common.net.HostAndPort;
 
 /**
- * 
+ *
  */
 public class MultiTserverReplicationIT extends ConfigurableMacIT {
   private static final Logger log = LoggerFactory.getLogger(MultiTserverReplicationIT.class);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/replication/UnusedWalDoesntCloseReplicationStatusIT.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/replication/UnusedWalDoesntCloseReplicationStatusIT.java b/test/src/test/java/org/apache/accumulo/test/replication/UnusedWalDoesntCloseReplicationStatusIT.java
index 145e986..6f6ab8a 100644
--- a/test/src/test/java/org/apache/accumulo/test/replication/UnusedWalDoesntCloseReplicationStatusIT.java
+++ b/test/src/test/java/org/apache/accumulo/test/replication/UnusedWalDoesntCloseReplicationStatusIT.java
@@ -175,8 +175,7 @@ public class UnusedWalDoesntCloseReplicationStatusIT extends ConfigurableMacIT {
     KeyExtent extent = new KeyExtent(new Text(tableId), null, null);
     bw = conn.createBatchWriter(MetadataTable.NAME, new BatchWriterConfig());
     m = new Mutation(extent.getMetadataEntry());
-    m.put(MetadataSchema.TabletsSection.LogColumnFamily.NAME, new Text("localhost:12345/" + walUri),
-        new Value((walUri + "|1").getBytes(UTF_8)));
+    m.put(MetadataSchema.TabletsSection.LogColumnFamily.NAME, new Text("localhost:12345/" + walUri), new Value((walUri + "|1").getBytes(UTF_8)));
     bw.addMutation(m);
 
     // Add a replication entry for our fake WAL

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/test/java/org/apache/accumulo/test/util/CertUtils.java
----------------------------------------------------------------------
diff --git a/test/src/test/java/org/apache/accumulo/test/util/CertUtils.java b/test/src/test/java/org/apache/accumulo/test/util/CertUtils.java
index 552a332..395870b 100644
--- a/test/src/test/java/org/apache/accumulo/test/util/CertUtils.java
+++ b/test/src/test/java/org/apache/accumulo/test/util/CertUtils.java
@@ -46,8 +46,8 @@ import org.apache.accumulo.core.cli.Help;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 import org.apache.accumulo.core.conf.DefaultConfiguration;
-import org.apache.accumulo.core.conf.SiteConfiguration;
 import org.apache.accumulo.core.conf.Property;
+import org.apache.accumulo.core.conf.SiteConfiguration;
 import org.apache.commons.io.FileExistsException;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.Path;
@@ -177,7 +177,8 @@ public class CertUtils {
     CertUtils certUtils = new CertUtils(opts.keystoreType, opts.issuerDirString, opts.encryptionAlg, opts.keysize, opts.signingAlg);
 
     if ("generate-all".equals(operation)) {
-      certUtils.createAll(new File(opts.rootKeystore), new File(opts.localKeystore), new File(opts.truststore), opts.keyNamePrefix, rootKeyPassword, keyPassword, opts.truststorePassword);
+      certUtils.createAll(new File(opts.rootKeystore), new File(opts.localKeystore), new File(opts.truststore), opts.keyNamePrefix, rootKeyPassword,
+          keyPassword, opts.truststorePassword);
     } else if ("generate-local".equals(operation)) {
       certUtils.createSignedCert(new File(opts.localKeystore), opts.keyNamePrefix + "-local", keyPassword, opts.rootKeystore, rootKeyPassword);
     } else if ("generate-self-trusted".equals(operation)) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/trace/src/main/java/org/apache/accumulo/trace/instrument/CloudtraceSpan.java
----------------------------------------------------------------------
diff --git a/trace/src/main/java/org/apache/accumulo/trace/instrument/CloudtraceSpan.java b/trace/src/main/java/org/apache/accumulo/trace/instrument/CloudtraceSpan.java
index 3d75917..7110134 100644
--- a/trace/src/main/java/org/apache/accumulo/trace/instrument/CloudtraceSpan.java
+++ b/trace/src/main/java/org/apache/accumulo/trace/instrument/CloudtraceSpan.java
@@ -24,49 +24,49 @@ import java.util.Map;
 @Deprecated
 public interface CloudtraceSpan {
   static final long ROOT_SPAN_ID = 0;
-  
+
   /** Begin gathering timing information */
   void start();
-  
+
   /** The block has completed, stop the clock */
   void stop();
-  
+
   /** Get the start time, in milliseconds */
   long getStartTimeMillis();
-  
+
   /** Get the stop time, in milliseconds */
   long getStopTimeMillis();
-  
+
   /** Return the total amount of time elapsed since start was called, if running, or difference between stop and start */
   long accumulatedMillis();
-  
+
   /** Has the span been started and not yet stopped? */
   boolean running();
-  
+
   /** Return a textual description of this span */
   String description();
-  
+
   /** A pseudo-unique (random) number assigned to this span instance */
   long spanId();
-  
+
   /** The parent span: returns null if this is the root span */
   Span parent();
-  
+
   /** A pseudo-unique (random) number assigned to the trace associated with this span */
   long traceId();
-  
+
   /** Create a child span of this span with the given description */
   Span child(String description);
-  
+
   @Override
   String toString();
-  
+
   /** Return the pseudo-unique (random) number of the parent span, returns ROOT_SPAN_ID if this is the root span */
   long parentId();
-  
+
   /** Add data associated with this span */
   void data(String key, String value);
-  
+
   /** Get data associated with this span (read only) */
   Map<String,String> getData();
 }


[65/66] [abbrv] accumulo git commit: ACCUMULO-3451 fixes checkstyle violations for 1.6

Posted by ct...@apache.org.
ACCUMULO-3451 fixes checkstyle violations for 1.6

  fixes additional checkstyle rule violations introduced in the 1.6 branch


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/9ca1ff02
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/9ca1ff02
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/9ca1ff02

Branch: refs/heads/1.6
Commit: 9ca1ff02ef732a8c3727049615ea005bffb7778a
Parents: 1368d09
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 21:29:50 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 21:29:50 2015 -0500

----------------------------------------------------------------------
 .../accumulo/core/security/crypto/CryptoModuleParameters.java  | 2 +-
 .../security/crypto/NonCachingSecretKeyEncryptionStrategy.java | 3 ++-
 .../org/apache/accumulo/core/util/shell/ShellOptionsJC.java    | 6 +++---
 pom.xml                                                        | 2 +-
 .../main/java/org/apache/accumulo/server/init/Initialize.java  | 3 ++-
 .../apache/accumulo/server/util/CustomNonBlockingServer.java   | 4 ++--
 6 files changed, 11 insertions(+), 9 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
index 573d64b..b9bf253 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/CryptoModuleParameters.java
@@ -280,7 +280,7 @@ public class CryptoModuleParameters {
    * <li>the code reading an encrypted stream and coming across the encrypted version of one of these keys, OR
    * <li>the {@link CryptoModuleParameters#getKeyEncryptionStrategyClass()} that encrypted the plaintext key (see
    * {@link CryptoModuleParameters#getPlaintextKey()}).
-   * <ul>
+   * </ul>
    * <p>
    * For <b>encryption</b>, this value is generally not required, but is usually set by the underlying module during encryption. <br>
    * For <b>decryption</b>, this value is <b>usually required</b>.

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java b/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
index f42f9ff..500627c 100644
--- a/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
+++ b/core/src/main/java/org/apache/accumulo/core/security/crypto/NonCachingSecretKeyEncryptionStrategy.java
@@ -47,7 +47,8 @@ public class NonCachingSecretKeyEncryptionStrategy implements SecretKeyEncryptio
       if (!fs.exists(pathToKey)) {
 
         if (encryptionMode == Cipher.UNWRAP_MODE) {
-          log.error("There was a call to decrypt the session key but no key encryption key exists.  Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
+          log.error("There was a call to decrypt the session key but no key encryption key exists. "
+              + "Either restore it, reconfigure the conf file to point to it in HDFS, or throw the affected data away and begin again.");
           throw new RuntimeException("Could not find key encryption key file in configured location in HDFS (" + pathToKeyName + ")");
         } else {
           DataOutputStream out = null;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java b/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
index 4787693..dfa24c5 100644
--- a/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
+++ b/core/src/main/java/org/apache/accumulo/core/util/shell/ShellOptionsJC.java
@@ -169,9 +169,9 @@ public class ShellOptionsJC {
   @Parameter(names = {"--ssl"}, description = "use ssl to connect to accumulo")
   private boolean useSsl = false;
 
-  @Parameter(
-      names = "--config-file",
-      description = "read the given client config file.  If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
+  @Parameter(names = "--config-file", description = "read the given client config file. "
+      + "If omitted, the path searched can be specified with $ACCUMULO_CLIENT_CONF_PATH, "
+      + "which defaults to ~/.accumulo/config:$ACCUMULO_CONF_DIR/client.conf:/etc/accumulo/client.conf")
   private String clientConfigFile = null;
 
   @Parameter(names = {"-zi", "--zooKeeperInstanceName"}, description = "use a zookeeper instance with the given instance name")

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 6a90918..8ada8fc 100644
--- a/pom.xml
+++ b/pom.xml
@@ -954,7 +954,7 @@
           </checkstyleRules>
           <violationSeverity>warning</violationSeverity>
           <includeTestSourceDirectory>true</includeTestSourceDirectory>
-          <excludes>**/thrift/*.java</excludes>
+          <excludes>**/thrift/*.java,**/HelpMojo.java</excludes>
         </configuration>
         <dependencies>
           <dependency>

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java b/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
index fae9397..e0a3797 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/init/Initialize.java
@@ -170,7 +170,8 @@ public class Initialize {
       c.println();
       c.println("You can change the instance secret in accumulo by using:");
       c.println("   bin/accumulo " + org.apache.accumulo.server.util.ChangeSecret.class.getName() + " oldPassword newPassword.");
-      c.println("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. Without this accumulo will not operate correctly");
+      c.println("You will also need to edit your secret in your configuration file by adding the property instance.secret to your conf/accumulo-site.xml. "
+          + "Without this accumulo will not operate correctly");
     }
     try {
       if (isInitialized(fs)) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/9ca1ff02/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java b/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
index ca53399..472bcb9 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/CustomNonBlockingServer.java
@@ -250,8 +250,8 @@ public class CustomNonBlockingServer extends THsHaServer {
         clientKey = client.registerSelector(selector, SelectionKey.OP_READ);
 
         // add this key to the map
-          FrameBuffer frameBuffer = processorFactory_.isAsyncProcessor() ?
-                  new CustomAsyncFrameBuffer(client, clientKey,SelectAcceptThread.this) :
+          FrameBuffer frameBuffer =
+              processorFactory_.isAsyncProcessor() ? new CustomAsyncFrameBuffer(client, clientKey,SelectAcceptThread.this) :
                   new CustomFrameBuffer(client, clientKey,SelectAcceptThread.this);
 
           clientKey.attach(frameBuffer);


[20/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataTableScanner.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataTableScanner.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataTableScanner.java
index d318ccc..e2294cd 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataTableScanner.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/MetaDataTableScanner.java
@@ -47,14 +47,14 @@ import org.apache.log4j.Logger;
 
 public class MetaDataTableScanner implements ClosableIterator<TabletLocationState> {
   private static final Logger log = Logger.getLogger(MetaDataTableScanner.class);
-  
+
   BatchScanner mdScanner = null;
   Iterator<Entry<Key,Value>> iter = null;
-  
+
   public MetaDataTableScanner(ClientContext context, Range range, CurrentState state) {
     this(context, range, state, MetadataTable.NAME);
   }
-  
+
   MetaDataTableScanner(ClientContext context, Range range, CurrentState state, String tableName) {
     // scan over metadata table, looking for tablets in the wrong state based on the live servers and online tables
     try {
@@ -71,7 +71,7 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
       throw new RuntimeException(ex);
     }
   }
-  
+
   static public void configureScanner(ScannerBase scanner, CurrentState state) {
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.fetch(scanner);
     scanner.fetchColumnFamily(TabletsSection.CurrentLocationColumnFamily.NAME);
@@ -88,15 +88,15 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
     }
     scanner.addScanIterator(tabletChange);
   }
-  
+
   public MetaDataTableScanner(ClientContext context, Range range) {
     this(context, range, MetadataTable.NAME);
   }
-  
+
   public MetaDataTableScanner(ClientContext context, Range range, String tableName) {
     this(context, range, null, tableName);
   }
-  
+
   @Override
   public void close() {
     if (iter != null) {
@@ -104,12 +104,12 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
       iter = null;
     }
   }
-  
+
   @Override
   protected void finalize() {
     close();
   }
-  
+
   @Override
   public boolean hasNext() {
     if (iter == null)
@@ -120,12 +120,12 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
     }
     return result;
   }
-  
+
   @Override
   public TabletLocationState next() {
-      return fetch();
+    return fetch();
   }
-  
+
   public static TabletLocationState createTabletLocationState(Key k, Value v) throws IOException, BadLocationStateException {
     final SortedMap<Key,Value> decodedRow = WholeRowIterator.decodeRow(k, v);
     KeyExtent extent = null;
@@ -135,13 +135,13 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
     long lastTimestamp = 0;
     List<Collection<String>> walogs = new ArrayList<Collection<String>>();
     boolean chopped = false;
-    
+
     for (Entry<Key,Value> entry : decodedRow.entrySet()) {
       Key key = entry.getKey();
       Text row = key.getRow();
       Text cf = key.getColumnFamily();
       Text cq = key.getColumnQualifier();
-      
+
       if (cf.compareTo(TabletsSection.FutureLocationColumnFamily.NAME) == 0) {
         TServerInstance location = new TServerInstance(entry.getValue(), cq);
         if (future != null) {
@@ -174,7 +174,7 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
     }
     return new TabletLocationState(extent, future, current, last, walogs, chopped);
   }
-  
+
   private TabletLocationState fetch() {
     try {
       Entry<Key,Value> e = iter.next();
@@ -185,7 +185,7 @@ public class MetaDataTableScanner implements ClosableIterator<TabletLocationStat
       throw new RuntimeException(ex);
     }
   }
-  
+
   @Override
   public void remove() {
     throw new RuntimeException("Unimplemented");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/RootTabletStateStore.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/RootTabletStateStore.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/RootTabletStateStore.java
index 5c8e102..ecc246b 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/RootTabletStateStore.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/RootTabletStateStore.java
@@ -22,20 +22,20 @@ import org.apache.accumulo.core.metadata.schema.MetadataSchema;
 import org.apache.accumulo.server.AccumuloServerContext;
 
 public class RootTabletStateStore extends MetaDataStateStore {
-  
+
   public RootTabletStateStore(ClientContext context, CurrentState state) {
     super(context, state, RootTable.NAME);
   }
-  
+
   public RootTabletStateStore(AccumuloServerContext context) {
     super(context, RootTable.NAME);
   }
-  
+
   @Override
   public ClosableIterator<TabletLocationState> iterator() {
     return new MetaDataTableScanner(context, MetadataSchema.TabletsSection.getRange(), state, RootTable.NAME);
   }
-  
+
   @Override
   public String name() {
     return "Metadata Tablets";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/TServerInstance.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/TServerInstance.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/TServerInstance.java
index f473ba3..c0c71e6 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/TServerInstance.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/TServerInstance.java
@@ -22,6 +22,7 @@ import java.io.IOException;
 import java.io.ObjectInputStream;
 import java.io.ObjectOutputStream;
 import java.io.Serializable;
+
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.metadata.schema.MetadataSchema.TabletsSection;
@@ -33,51 +34,51 @@ import com.google.common.net.HostAndPort;
 /**
  * A tablet is assigned to a tablet server at the given address as long as it is alive and well. When the tablet server is restarted, the instance information
  * it advertises will change. Therefore tablet assignments can be considered out-of-date if the tablet server instance information has been changed.
- * 
+ *
  */
 public class TServerInstance implements Comparable<TServerInstance>, Serializable {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   // HostAndPort is not Serializable
   private transient HostAndPort location;
   private String session;
   private String cachedStringRepresentation;
-  
+
   public TServerInstance(HostAndPort address, String session) {
     this.location = address;
     this.session = session;
     this.cachedStringRepresentation = hostPort() + "[" + session + "]";
   }
-  
+
   public TServerInstance(HostAndPort address, long session) {
     this(address, Long.toHexString(session));
   }
-  
+
   public TServerInstance(String address, long session) {
     this(AddressUtil.parseAddress(address, false), Long.toHexString(session));
   }
-  
+
   public TServerInstance(Value address, Text session) {
     this(AddressUtil.parseAddress(new String(address.get(), UTF_8), false), session.toString());
   }
-  
+
   public void putLocation(Mutation m) {
     m.put(TabletsSection.CurrentLocationColumnFamily.NAME, asColumnQualifier(), asMutationValue());
   }
-  
+
   public void putFutureLocation(Mutation m) {
     m.put(TabletsSection.FutureLocationColumnFamily.NAME, asColumnQualifier(), asMutationValue());
   }
-  
+
   public void putLastLocation(Mutation m) {
     m.put(TabletsSection.LastLocationColumnFamily.NAME, asColumnQualifier(), asMutationValue());
   }
-  
+
   public void clearLastLocation(Mutation m) {
     m.putDelete(TabletsSection.LastLocationColumnFamily.NAME, asColumnQualifier());
   }
-  
+
   public void clearFutureLocation(Mutation m) {
     m.putDelete(TabletsSection.FutureLocationColumnFamily.NAME, asColumnQualifier());
   }
@@ -85,19 +86,19 @@ public class TServerInstance implements Comparable<TServerInstance>, Serializabl
   public void clearLocation(Mutation m) {
     m.putDelete(TabletsSection.CurrentLocationColumnFamily.NAME, asColumnQualifier());
   }
-  
+
   @Override
   public int compareTo(TServerInstance other) {
     if (this == other)
       return 0;
     return this.toString().compareTo(other.toString());
   }
-  
+
   @Override
   public int hashCode() {
     return toString().hashCode();
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     if (obj instanceof TServerInstance) {
@@ -105,45 +106,45 @@ public class TServerInstance implements Comparable<TServerInstance>, Serializabl
     }
     return false;
   }
-  
+
   @Override
   public String toString() {
     return cachedStringRepresentation;
   }
-  
+
   public int port() {
     return getLocation().getPort();
   }
-  
+
   public String host() {
     return getLocation().getHostText();
   }
-  
+
   public String hostPort() {
     return getLocation().toString();
   }
-  
+
   private Text asColumnQualifier() {
     return new Text(this.getSession());
   }
-  
+
   private Value asMutationValue() {
     return new Value(getLocation().toString().getBytes(UTF_8));
   }
-  
+
   public HostAndPort getLocation() {
     return location;
   }
-  
+
   public String getSession() {
     return session;
   }
-  
+
   private void writeObject(ObjectOutputStream out) throws IOException {
     out.defaultWriteObject();
     out.writeObject(location.toString());
   }
-  
+
   private void readObject(ObjectInputStream in) throws IOException, ClassNotFoundException {
     in.defaultReadObject();
     location = HostAndPort.fromString(in.readObject().toString());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletLocationState.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletLocationState.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletLocationState.java
index 5432d32..b24b562 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletLocationState.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletLocationState.java
@@ -26,21 +26,26 @@ import org.apache.hadoop.io.Text;
 /**
  * When a tablet is assigned, we mark its future location. When the tablet is opened, we set its current location. A tablet should never have both a future and
  * current location.
- * 
+ *
  * A tablet server is always associated with a unique session id. If the current tablet server has a different session, we know the location information is
  * out-of-date.
  */
 public class TabletLocationState {
-  
+
   static public class BadLocationStateException extends Exception {
     private static final long serialVersionUID = 1L;
     private Text metadataTableEntry;
 
-    BadLocationStateException(String msg, Text row) { super(msg); this.metadataTableEntry = row; }
+    BadLocationStateException(String msg, Text row) {
+      super(msg);
+      this.metadataTableEntry = row;
+    }
 
-    public Text getEncodedEndRow() { return metadataTableEntry; }
+    public Text getEncodedEndRow() {
+      return metadataTableEntry;
+    }
   }
-  
+
   public TabletLocationState(KeyExtent extent, TServerInstance future, TServerInstance current, TServerInstance last, Collection<Collection<String>> walogs,
       boolean chopped) throws BadLocationStateException {
     this.extent = extent;
@@ -55,18 +60,18 @@ public class TabletLocationState {
       throw new BadLocationStateException(extent + " is both assigned and hosted, which should never happen: " + this, extent.getMetadataEntry());
     }
   }
-  
+
   final public KeyExtent extent;
   final public TServerInstance future;
   final public TServerInstance current;
   final public TServerInstance last;
   final public Collection<Collection<String>> walogs;
   final public boolean chopped;
-  
+
   public String toString() {
     return extent + "@(" + future + "," + current + "," + last + ")" + (chopped ? " chopped" : "");
   }
-  
+
   public TServerInstance getServer() {
     TServerInstance result = null;
     if (current != null) {
@@ -78,7 +83,7 @@ public class TabletLocationState {
     }
     return result;
   }
-  
+
   public TabletState getState(Set<TServerInstance> liveServers) {
     TServerInstance server = getServer();
     if (server == null)
@@ -97,5 +102,5 @@ public class TabletLocationState {
     // server == last
     return TabletState.UNASSIGNED;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletMigration.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletMigration.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletMigration.java
index f0a3664..e852af3 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletMigration.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletMigration.java
@@ -22,13 +22,13 @@ public class TabletMigration {
   public KeyExtent tablet;
   public TServerInstance oldServer;
   public TServerInstance newServer;
-  
+
   public TabletMigration(KeyExtent extent, TServerInstance before, TServerInstance after) {
     this.tablet = extent;
     this.oldServer = before;
     this.newServer = after;
   }
-  
+
   public String toString() {
     return tablet + ": " + oldServer + " -> " + newServer;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletServerState.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletServerState.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletServerState.java
index 23f16e3..dde9807 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletServerState.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletServerState.java
@@ -24,13 +24,13 @@ import java.util.Set;
 public enum TabletServerState {
   // not a valid state, reserved for internal use only
   RESERVED((byte) (-1)),
-  
+
   // the following are normally functioning states
   NEW((byte) 0),
   ONLINE((byte) 1),
   UNRESPONSIVE((byte) 2),
   DOWN((byte) 3),
-  
+
   // the following are bad states and cause tservers to be ignored by the master
   BAD_SYSTEM_PASSWORD((byte) 101),
   BAD_VERSION((byte) 102),
@@ -40,12 +40,12 @@ public enum TabletServerState {
   BAD_VERSION_AND_CONFIG((byte) 106),
   BAD_VERSION_AND_INSTANCE_AND_CONFIG((byte) 107),
   BAD_INSTANCE_AND_CONFIG((byte) 108);
-  
+
   private byte id;
-  
+
   private static HashMap<Byte,TabletServerState> mapping;
   private static HashSet<TabletServerState> badStates;
-  
+
   static {
     mapping = new HashMap<Byte,TabletServerState>(TabletServerState.values().length);
     badStates = new HashSet<TabletServerState>();
@@ -55,21 +55,21 @@ public enum TabletServerState {
         badStates.add(state);
     }
   }
-  
+
   private TabletServerState(byte id) {
     this.id = id;
   }
-  
+
   public byte getId() {
     return this.id;
   }
-  
+
   public static TabletServerState getStateById(byte id) {
     if (mapping.containsKey(id))
       return mapping.get(id);
     throw new IndexOutOfBoundsException("No such state");
   }
-  
+
   public static Set<TabletServerState> getBadStates() {
     return Collections.unmodifiableSet(badStates);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateChangeIterator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateChangeIterator.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateChangeIterator.java
index 2a84e70..9351cd0 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateChangeIterator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateChangeIterator.java
@@ -45,16 +45,16 @@ import org.apache.hadoop.io.Text;
 import com.google.common.base.Joiner;
 
 public class TabletStateChangeIterator extends SkippingIterator {
-  
+
   private static final String SERVERS_OPTION = "servers";
   private static final String TABLES_OPTION = "tables";
   private static final String MERGES_OPTION = "merges";
   // private static final Logger log = Logger.getLogger(TabletStateChangeIterator.class);
-  
+
   Set<TServerInstance> current;
   Set<String> onlineTables;
   Map<Text,MergeInfo> merges;
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
@@ -62,7 +62,7 @@ public class TabletStateChangeIterator extends SkippingIterator {
     onlineTables = parseTables(options.get(TABLES_OPTION));
     merges = parseMerges(options.get(MERGES_OPTION));
   }
-  
+
   private Set<String> parseTables(String tables) {
     if (tables == null)
       return null;
@@ -71,7 +71,7 @@ public class TabletStateChangeIterator extends SkippingIterator {
       result.add(table);
     return result;
   }
-  
+
   private Set<TServerInstance> parseServers(String servers) {
     if (servers == null)
       return null;
@@ -89,7 +89,7 @@ public class TabletStateChangeIterator extends SkippingIterator {
     }
     return result;
   }
-  
+
   private Map<Text,MergeInfo> parseMerges(String merges) {
     if (merges == null)
       return null;
@@ -108,16 +108,16 @@ public class TabletStateChangeIterator extends SkippingIterator {
       throw new RuntimeException(ex);
     }
   }
-  
+
   @Override
   protected void consume() throws IOException {
     while (getSource().hasTop()) {
       Key k = getSource().getTopKey();
       Value v = getSource().getTopValue();
-      
+
       if (onlineTables == null || current == null)
         return;
-      
+
       TabletLocationState tls;
       try {
         tls = MetaDataTableScanner.createTabletLocationState(k, v);
@@ -134,7 +134,7 @@ public class TabletStateChangeIterator extends SkippingIterator {
       }
       // is the table supposed to be online or offline?
       boolean shouldBeOnline = onlineTables.contains(tls.extent.getTableId().toString());
-      
+
       switch (tls.getState(current)) {
         case ASSIGNED:
           // we always want data about assigned tablets
@@ -152,12 +152,12 @@ public class TabletStateChangeIterator extends SkippingIterator {
       getSource().next();
     }
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   public static void setCurrentServers(IteratorSetting cfg, Set<TServerInstance> goodServers) {
     if (goodServers != null) {
       List<String> servers = new ArrayList<String>();
@@ -166,12 +166,12 @@ public class TabletStateChangeIterator extends SkippingIterator {
       cfg.addOption(SERVERS_OPTION, Joiner.on(",").join(servers));
     }
   }
-  
+
   public static void setOnlineTables(IteratorSetting cfg, Set<String> onlineTables) {
     if (onlineTables != null)
       cfg.addOption(TABLES_OPTION, Joiner.on(",").join(onlineTables));
   }
-  
+
   public static void setMerges(IteratorSetting cfg, Collection<MergeInfo> merges) {
     DataOutputBuffer buffer = new DataOutputBuffer();
     try {
@@ -187,5 +187,5 @@ public class TabletStateChangeIterator extends SkippingIterator {
     String encoded = Base64.encodeBase64String(Arrays.copyOf(buffer.getData(), buffer.getLength()));
     cfg.addOption(MERGES_OPTION, encoded);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateStore.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateStore.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateStore.java
index 41de64a..5413e31 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateStore.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/TabletStateStore.java
@@ -23,42 +23,42 @@ import org.apache.accumulo.server.AccumuloServerContext;
 
 /**
  * Interface for storing information about tablet assignments. There are three implementations:
- * 
+ *
  * ZooTabletStateStore: information about the root tablet is stored in ZooKeeper MetaDataStateStore: information about the other tablets are stored in the
  * metadata table
- * 
+ *
  */
 public abstract class TabletStateStore implements Iterable<TabletLocationState> {
-  
+
   /**
    * Identifying name for this tablet state store.
    */
   abstract public String name();
-  
+
   /**
    * Scan the information about the tablets covered by this store
    */
   @Override
   abstract public ClosableIterator<TabletLocationState> iterator();
-  
+
   /**
    * Store the assigned locations in the data store.
    */
   abstract public void setFutureLocations(Collection<Assignment> assignments) throws DistributedStoreException;
-  
+
   /**
    * Tablet servers will update the data store with the location when they bring the tablet online
    */
   abstract public void setLocations(Collection<Assignment> assignments) throws DistributedStoreException;
-  
+
   /**
    * Mark the tablets as having no known or future location.
-   * 
+   *
    * @param tablets
    *          the tablets' current information
    */
   abstract public void unassign(Collection<TabletLocationState> tablets) throws DistributedStoreException;
-  
+
   public static void unassign(AccumuloServerContext context, TabletLocationState tls) throws DistributedStoreException {
     TabletStateStore store;
     if (tls.extent.isRootTablet()) {
@@ -70,7 +70,7 @@ public abstract class TabletStateStore implements Iterable<TabletLocationState>
     }
     store.unassign(Collections.singletonList(tls));
   }
-  
+
   public static void setLocation(AccumuloServerContext context, Assignment assignment) throws DistributedStoreException {
     TabletStateStore store;
     if (assignment.tablet.isRootTablet()) {
@@ -82,5 +82,5 @@ public abstract class TabletStateStore implements Iterable<TabletLocationState>
     }
     store.setLocations(Collections.singletonList(assignment));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooStore.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooStore.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooStore.java
index 1bcf482..2b7cb4c 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooStore.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooStore.java
@@ -31,23 +31,23 @@ import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.apache.log4j.Logger;
 
 public class ZooStore implements DistributedStore {
-  
+
   private static final Logger log = Logger.getLogger(ZooStore.class);
-  
+
   String basePath;
-  
+
   ZooCache cache = new ZooCache();
-  
+
   public ZooStore(String basePath) throws IOException {
     if (basePath.endsWith("/"))
       basePath = basePath.substring(0, basePath.length() - 1);
     this.basePath = basePath;
   }
-  
+
   public ZooStore() throws IOException {
     this(ZooUtil.getRoot(HdfsZooInstance.getInstance().getInstanceID()));
   }
-  
+
   @Override
   public byte[] get(String path) throws DistributedStoreException {
     try {
@@ -56,11 +56,11 @@ public class ZooStore implements DistributedStore {
       throw new DistributedStoreException(ex);
     }
   }
-  
+
   private String relative(String path) {
     return basePath + path;
   }
-  
+
   @Override
   public List<String> getChildren(String path) throws DistributedStoreException {
     try {
@@ -69,7 +69,7 @@ public class ZooStore implements DistributedStore {
       throw new DistributedStoreException(ex);
     }
   }
-  
+
   @Override
   public void put(String path, byte[] bs) throws DistributedStoreException {
     try {
@@ -81,7 +81,7 @@ public class ZooStore implements DistributedStore {
       throw new DistributedStoreException(ex);
     }
   }
-  
+
   @Override
   public void remove(String path) throws DistributedStoreException {
     try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooTabletStateStore.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooTabletStateStore.java b/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooTabletStateStore.java
index 5481531..58b8446 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooTabletStateStore.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/state/ZooTabletStateStore.java
@@ -32,14 +32,14 @@ import org.apache.log4j.Logger;
 import com.google.common.net.HostAndPort;
 
 public class ZooTabletStateStore extends TabletStateStore {
-  
+
   private static final Logger log = Logger.getLogger(ZooTabletStateStore.class);
   final private DistributedStore store;
-  
+
   public ZooTabletStateStore(DistributedStore store) {
     this.store = store;
   }
-  
+
   public ZooTabletStateStore() throws DistributedStoreException {
     try {
       store = new ZooStore();
@@ -47,17 +47,17 @@ public class ZooTabletStateStore extends TabletStateStore {
       throw new DistributedStoreException(ex);
     }
   }
-  
+
   @Override
   public ClosableIterator<TabletLocationState> iterator() {
     return new ClosableIterator<TabletLocationState>() {
       boolean finished = false;
-      
+
       @Override
       public boolean hasNext() {
         return !finished;
       }
-      
+
       @Override
       public TabletLocationState next() {
         finished = true;
@@ -65,17 +65,17 @@ public class ZooTabletStateStore extends TabletStateStore {
           byte[] future = store.get(RootTable.ZROOT_TABLET_FUTURE_LOCATION);
           byte[] current = store.get(RootTable.ZROOT_TABLET_LOCATION);
           byte[] last = store.get(RootTable.ZROOT_TABLET_LAST_LOCATION);
-          
+
           TServerInstance currentSession = null;
           TServerInstance futureSession = null;
           TServerInstance lastSession = null;
-          
+
           if (future != null)
             futureSession = parse(future);
-          
+
           if (last != null)
             lastSession = parse(last);
-          
+
           if (current != null) {
             currentSession = parse(current);
             futureSession = null;
@@ -97,18 +97,17 @@ public class ZooTabletStateStore extends TabletStateStore {
           throw new RuntimeException(ex);
         }
       }
-      
+
       @Override
       public void remove() {
         throw new NotImplementedException();
       }
 
       @Override
-      public void close() {
-      }
+      public void close() {}
     };
   }
-  
+
   protected TServerInstance parse(byte[] current) {
     String str = new String(current, UTF_8);
     String[] parts = str.split("[|]", 2);
@@ -120,7 +119,7 @@ public class ZooTabletStateStore extends TabletStateStore {
       return null;
     }
   }
-  
+
   @Override
   public void setFutureLocations(Collection<Assignment> assignments) throws DistributedStoreException {
     if (assignments.size() != 1)
@@ -136,7 +135,7 @@ public class ZooTabletStateStore extends TabletStateStore {
     }
     store.put(RootTable.ZROOT_TABLET_FUTURE_LOCATION, value.getBytes(UTF_8));
   }
-  
+
   @Override
   public void setLocations(Collection<Assignment> assignments) throws DistributedStoreException {
     if (assignments.size() != 1)
@@ -159,7 +158,7 @@ public class ZooTabletStateStore extends TabletStateStore {
     store.remove(RootTable.ZROOT_TABLET_FUTURE_LOCATION);
     log.debug("Put down root tablet location");
   }
-  
+
   @Override
   public void unassign(Collection<TabletLocationState> tablets) throws DistributedStoreException {
     if (tablets.size() != 1)
@@ -171,10 +170,10 @@ public class ZooTabletStateStore extends TabletStateStore {
     store.remove(RootTable.ZROOT_TABLET_FUTURE_LOCATION);
     log.debug("unassign root tablet location");
   }
-  
+
   @Override
   public String name() {
     return "Root Table";
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/master/tableOps/UserCompactionConfig.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/master/tableOps/UserCompactionConfig.java b/server/base/src/main/java/org/apache/accumulo/server/master/tableOps/UserCompactionConfig.java
index 6314105..98d7fd7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/master/tableOps/UserCompactionConfig.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/master/tableOps/UserCompactionConfig.java
@@ -23,10 +23,9 @@ import java.util.ArrayList;
 import java.util.Collections;
 import java.util.List;
 
+import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.admin.CompactionStrategyConfig;
 import org.apache.accumulo.core.client.impl.CompactionStrategyConfigUtil;
-
-import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.io.Writable;
 
@@ -117,4 +116,4 @@ public class UserCompactionConfig implements Writable {
   public CompactionStrategyConfig getCompactionStrategy() {
     return compactionStrategy;
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/metrics/AbstractMetricsImpl.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/metrics/AbstractMetricsImpl.java b/server/base/src/main/java/org/apache/accumulo/server/metrics/AbstractMetricsImpl.java
index 657fc31..93db9c8 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/metrics/AbstractMetricsImpl.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/metrics/AbstractMetricsImpl.java
@@ -85,7 +85,8 @@ public abstract class AbstractMetricsImpl implements Metrics {
 
     @Override
     public String toString() {
-      return new ToStringBuilder(this, ToStringStyle.SHORT_PREFIX_STYLE).append("count", count).append("average", avg).append("minimum", min).append("maximum", max).toString();
+      return new ToStringBuilder(this, ToStringStyle.SHORT_PREFIX_STYLE).append("count", count).append("average", avg).append("minimum", min)
+          .append("maximum", max).toString();
     }
 
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/metrics/MetricsConfiguration.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/metrics/MetricsConfiguration.java b/server/base/src/main/java/org/apache/accumulo/server/metrics/MetricsConfiguration.java
index 087ca12..b0ffd64 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/metrics/MetricsConfiguration.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/metrics/MetricsConfiguration.java
@@ -32,50 +32,50 @@ import org.apache.commons.configuration.reloading.FileChangedReloadingStrategy;
 import org.apache.commons.lang.builder.ToStringBuilder;
 
 public class MetricsConfiguration {
-  
+
   private static final org.apache.log4j.Logger log = org.apache.log4j.Logger.getLogger(MetricsConfiguration.class);
-  
+
   private static final String metricsFileName = "accumulo-metrics.xml";
-  
+
   private static boolean alreadyWarned = false;
-  
+
   private boolean notFound = false;
-  
+
   private int notFoundCount = 0;
-  
+
   private static SystemConfiguration sysConfig = null;
-  
+
   private static EnvironmentConfiguration envConfig = null;
-  
+
   private XMLConfiguration xConfig = null;
-  
+
   private Configuration config = null;
-  
+
   private final Object lock = new Object();
-  
+
   private boolean needsReloading = false;
-  
+
   private long lastCheckTime = 0;
-  
+
   private static long CONFIG_FILE_CHECK_INTERVAL = 1000 * 60 * 10; // 10 minutes
-  
+
   private static int CONFIG_FILE_CHECK_COUNTER = 100;
-  
+
   public final static long CONFIG_FILE_RELOAD_DELAY = 60000;
-  
+
   private MetricsConfigWatcher watcher = null;
-  
+
   private boolean enabled = false;
-  
+
   private String enabledName = null;
-  
+
   /**
    * Background thread that pokes the XMLConfiguration file to see if it has changed. If it has, then the Configuration Listener will get an event.
-   * 
+   *
    */
   private class MetricsConfigWatcher extends Daemon {
     public MetricsConfigWatcher() {}
-    
+
     public void run() {
       while (this.isAlive()) {
         try {
@@ -87,7 +87,7 @@ public class MetricsConfiguration {
       }
     }
   }
-  
+
   /**
    * ConfigurationListener that sets a flag to reload the XML config file
    */
@@ -97,7 +97,7 @@ public class MetricsConfiguration {
         needsReloading = true;
     }
   }
-  
+
   public MetricsConfiguration(String name) {
     // We are going to store the "enabled" parameter for this
     // name as a shortcut so that it doesn't have to be looked
@@ -105,7 +105,7 @@ public class MetricsConfiguration {
     this.enabledName = name + ".enabled";
     getMetricsConfiguration();
   }
-  
+
   public Configuration getEnvironmentConfiguration() {
     synchronized (MetricsConfiguration.class) {
       if (null == envConfig)
@@ -113,7 +113,7 @@ public class MetricsConfiguration {
       return envConfig;
     }
   }
-  
+
   public Configuration getSystemConfiguration() {
     synchronized (MetricsConfiguration.class) {
       if (null == sysConfig)
@@ -121,7 +121,7 @@ public class MetricsConfiguration {
       return sysConfig;
     }
   }
-  
+
   public Configuration getMetricsConfiguration() {
     if (notFound) {
       if (notFoundCount <= CONFIG_FILE_CHECK_COUNTER) {
@@ -145,7 +145,7 @@ public class MetricsConfiguration {
       }
     return config;
   }
-  
+
   private void loadConfiguration() {
     // Check to see if ACCUMULO_HOME environment variable is set.
     String ACUHOME = getEnvironmentConfiguration().getString("ACCUMULO_CONF_DIR");
@@ -160,7 +160,7 @@ public class MetricsConfiguration {
           xConfig.append(getEnvironmentConfiguration());
           xConfig.addConfigurationListener(new MetricsConfigListener());
           xConfig.setReloadingStrategy(new FileChangedReloadingStrategy());
-          
+
           // Start a background Thread that checks a property from the XMLConfiguration
           // every so often to force the FileChangedReloadingStrategy to fire.
           if (null == watcher || !watcher.isAlive()) {
@@ -197,16 +197,16 @@ public class MetricsConfiguration {
     } else {
       enabled = false;
     }
-    
+
   }
-  
+
   public boolean isEnabled() {
     // Force reload if necessary
     if (null == getMetricsConfiguration())
       return false;
     return enabled;
   }
-  
+
   public static String toStringValue(Configuration config) {
     ToStringBuilder tsb = new ToStringBuilder(MetricsConfiguration.class);
     Iterator<?> keys = config.getKeys();
@@ -220,7 +220,7 @@ public class MetricsConfiguration {
     }
     return tsb.toString();
   }
-  
+
   public static void main(String[] args) throws Exception {
     MetricsConfiguration mc = new MetricsConfiguration("master");
     while (true) {
@@ -232,5 +232,5 @@ public class MetricsConfiguration {
       Thread.sleep(1000);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/metrics/ThriftMetrics.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/metrics/ThriftMetrics.java b/server/base/src/main/java/org/apache/accumulo/server/metrics/ThriftMetrics.java
index d87d055..869f3c2 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/metrics/ThriftMetrics.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/metrics/ThriftMetrics.java
@@ -18,7 +18,6 @@ package org.apache.accumulo.server.metrics;
 
 import javax.management.ObjectName;
 
-
 public class ThriftMetrics extends AbstractMetricsImpl implements ThriftMetricsMBean {
 
   static final org.apache.log4j.Logger log = org.apache.log4j.Logger.getLogger(ThriftMetrics.class);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/monitor/DedupedLogEvent.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/monitor/DedupedLogEvent.java b/server/base/src/main/java/org/apache/accumulo/server/monitor/DedupedLogEvent.java
index 4acb1a9..96966ec 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/monitor/DedupedLogEvent.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/monitor/DedupedLogEvent.java
@@ -19,32 +19,32 @@ package org.apache.accumulo.server.monitor;
 import org.apache.log4j.spi.LoggingEvent;
 
 public class DedupedLogEvent {
-  
+
   private LoggingEvent event;
   private int count = 0;
   private int hash = -1;
-  
+
   public DedupedLogEvent(LoggingEvent event) {
     this(event, 1);
   }
-  
+
   public DedupedLogEvent(LoggingEvent event, int count) {
     this.event = event;
     this.count = count;
   }
-  
+
   public LoggingEvent getEvent() {
     return event;
   }
-  
+
   public int getCount() {
     return count;
   }
-  
+
   public void setCount(int count) {
     this.count = count;
   }
-  
+
   @Override
   public int hashCode() {
     if (hash == -1) {
@@ -53,14 +53,14 @@ public class DedupedLogEvent {
     }
     return hash;
   }
-  
+
   @Override
   public boolean equals(Object obj) {
     if (obj instanceof DedupedLogEvent)
       return this.event.equals(((DedupedLogEvent) obj).event);
     return false;
   }
-  
+
   @Override
   public String toString() {
     return event.getMDC("application").toString() + ":" + event.getLevel().toString() + ":" + event.getMessage().toString();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/monitor/LogService.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/monitor/LogService.java b/server/base/src/main/java/org/apache/accumulo/server/monitor/LogService.java
index 930b634..8f0f13e 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/monitor/LogService.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/monitor/LogService.java
@@ -40,7 +40,7 @@ import org.apache.zookeeper.KeeperException;
 
 /**
  * Hijack log4j and capture log events for display.
- * 
+ *
  */
 public class LogService extends org.apache.log4j.AppenderSkeleton {
 
@@ -48,7 +48,7 @@ public class LogService extends org.apache.log4j.AppenderSkeleton {
 
   /**
    * Read logging events forward to us over the net.
-   * 
+   *
    */
   static class SocketServer implements Runnable {
     private ServerSocket server = null;
@@ -83,7 +83,7 @@ public class LogService extends org.apache.log4j.AppenderSkeleton {
 
   /**
    * Place the host:port advertisement for the Monitor's Log4j listener in ZooKeeper
-   * 
+   *
    * @param conf
    *          configuration for the instance
    * @param instanceId

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReport.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReport.java b/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReport.java
index 0dfe9c0..2f2d0c4 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReport.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReport.java
@@ -50,32 +50,32 @@ public class ProblemReport {
   private String exception;
   private String server;
   private long creationTime;
-  
+
   public ProblemReport(String table, ProblemType problemType, String resource, String server, Throwable e, long creationTime) {
     checkNotNull(table, "table is null");
     checkNotNull(problemType, "problemType is null");
     checkNotNull(resource, "resource is null");
     this.tableName = table;
-    
+
     this.problemType = problemType;
     this.resource = resource;
-    
+
     if (e != null) {
       this.exception = e.getMessage();
     }
-    
+
     if (server == null) {
       try {
         server = InetAddress.getLocalHost().getHostAddress();
       } catch (UnknownHostException e1) {
-        
+
       }
     }
-    
+
     this.server = server;
     this.creationTime = creationTime;
   }
-  
+
   public ProblemReport(String table, ProblemType problemType, String resource, String server, Throwable e) {
     this(table, problemType, resource, server, e, System.currentTimeMillis());
   }
@@ -83,7 +83,7 @@ public class ProblemReport {
   public ProblemReport(String table, ProblemType problemType, String resource, Throwable e) {
     this(table, problemType, resource, null, e);
   }
-  
+
   private ProblemReport(String table, ProblemType problemType, String resource, byte enc[]) throws IOException {
     checkNotNull(table, "table is null");
     checkNotNull(problemType, "problemType is null");
@@ -91,63 +91,63 @@ public class ProblemReport {
     this.tableName = table;
     this.problemType = problemType;
     this.resource = resource;
-    
+
     decode(enc);
   }
-  
+
   private byte[] encode() throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
-    
+
     dos.writeLong(creationTime);
-    
+
     dos.writeBoolean(server != null);
     if (server != null) {
       dos.writeUTF(server);
     }
-    
+
     dos.writeBoolean(exception != null);
     if (exception != null) {
       dos.writeUTF(exception);
     }
-    
+
     dos.close();
     baos.close();
-    
+
     return baos.toByteArray();
   }
-  
+
   private void decode(byte enc[]) throws IOException {
     ByteArrayInputStream bais = new ByteArrayInputStream(enc);
     DataInputStream dis = new DataInputStream(bais);
-    
+
     creationTime = dis.readLong();
-    
+
     if (dis.readBoolean()) {
       server = dis.readUTF();
     } else {
       server = null;
     }
-    
+
     if (dis.readBoolean()) {
       exception = dis.readUTF();
     } else {
       exception = null;
     }
   }
-  
+
   void removeFromMetadataTable(AccumuloServerContext context) throws Exception {
     Mutation m = new Mutation(new Text("~err_" + tableName));
     m.putDelete(new Text(problemType.name()), new Text(resource));
     MetadataTableUtil.getMetadataTable(context).update(m);
   }
-  
+
   void saveToMetadataTable(AccumuloServerContext context) throws Exception {
     Mutation m = new Mutation(new Text("~err_" + tableName));
     m.put(new Text(problemType.name()), new Text(resource), new Value(encode()));
     MetadataTableUtil.getMetadataTable(context).update(m);
   }
-  
+
   void removeFromZooKeeper() throws Exception {
     removeFromZooKeeper(ZooReaderWriter.getInstance(), HdfsZooInstance.getInstance());
   }
@@ -156,7 +156,7 @@ public class ProblemReport {
     String zpath = getZPath(instance);
     zoorw.recursiveDelete(zpath, NodeMissingPolicy.SKIP);
   }
-  
+
   void saveToZooKeeper() throws Exception {
     saveToZooKeeper(ZooReaderWriter.getInstance(), HdfsZooInstance.getInstance());
   }
@@ -164,7 +164,7 @@ public class ProblemReport {
   void saveToZooKeeper(ZooReaderWriter zoorw, Instance instance) throws IOException, KeeperException, InterruptedException {
     zoorw.putPersistentData(getZPath(instance), encode(), NodeExistsPolicy.OVERWRITE);
   }
-  
+
   private String getZPath(Instance instance) throws IOException {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream dos = new DataOutputStream(baos);
@@ -173,69 +173,69 @@ public class ProblemReport {
     dos.writeUTF(getResource());
     dos.close();
     baos.close();
-    
+
     String zpath = ZooUtil.getRoot(instance) + Constants.ZPROBLEMS + "/" + Encoding.encodeAsBase64FileName(new Text(baos.toByteArray()));
     return zpath;
   }
-  
+
   static ProblemReport decodeZooKeeperEntry(String node) throws Exception {
     return decodeZooKeeperEntry(node, ZooReaderWriter.getInstance(), HdfsZooInstance.getInstance());
   }
 
   static ProblemReport decodeZooKeeperEntry(String node, ZooReaderWriter zoorw, Instance instance) throws IOException, KeeperException, InterruptedException {
     byte bytes[] = Encoding.decodeBase64FileName(node);
-    
+
     ByteArrayInputStream bais = new ByteArrayInputStream(bytes);
     DataInputStream dis = new DataInputStream(bais);
-    
+
     String tableName = dis.readUTF();
     String problemType = dis.readUTF();
     String resource = dis.readUTF();
-    
+
     String zpath = ZooUtil.getRoot(instance) + Constants.ZPROBLEMS + "/" + node;
     byte[] enc = zoorw.getData(zpath, null);
-    
+
     return new ProblemReport(tableName, ProblemType.valueOf(problemType), resource, enc);
-    
+
   }
-  
+
   public static ProblemReport decodeMetadataEntry(Entry<Key,Value> entry) throws IOException {
     String tableName = entry.getKey().getRow().toString().substring("~err_".length());
     String problemType = entry.getKey().getColumnFamily().toString();
     String resource = entry.getKey().getColumnQualifier().toString();
-    
+
     return new ProblemReport(tableName, ProblemType.valueOf(problemType), resource, entry.getValue().get());
   }
-  
+
   public String getTableName() {
     return tableName;
   }
-  
+
   public ProblemType getProblemType() {
     return problemType;
   }
-  
+
   public String getResource() {
     return resource;
   }
-  
+
   public String getException() {
     return exception;
   }
-  
+
   public String getServer() {
     return server;
   }
-  
+
   public long getTime() {
     return creationTime;
   }
-  
+
   @Override
   public int hashCode() {
     return tableName.hashCode() + problemType.hashCode() + resource.hashCode();
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof ProblemReport) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReportingIterator.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReportingIterator.java b/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReportingIterator.java
index 51ac12f..349ed20 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReportingIterator.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReportingIterator.java
@@ -37,7 +37,7 @@ public class ProblemReportingIterator implements InterruptibleIterator {
   private String resource;
   private String table;
   private final AccumuloServerContext context;
-  
+
   public ProblemReportingIterator(AccumuloServerContext context, String table, String resource, boolean continueOnError,
       SortedKeyValueIterator<Key,Value> source) {
     this.context = context;
@@ -46,22 +46,22 @@ public class ProblemReportingIterator implements InterruptibleIterator {
     this.continueOnError = continueOnError;
     this.source = source;
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     return new ProblemReportingIterator(context, table, resource, continueOnError, source.deepCopy(env));
   }
-  
+
   @Override
   public Key getTopKey() {
     return source.getTopKey();
   }
-  
+
   @Override
   public Value getTopValue() {
     return source.getTopValue();
   }
-  
+
   @Override
   public boolean hasTop() {
     if (sawError) {
@@ -69,12 +69,12 @@ public class ProblemReportingIterator implements InterruptibleIterator {
     }
     return source.hasTop();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void next() throws IOException {
     try {
@@ -87,13 +87,13 @@ public class ProblemReportingIterator implements InterruptibleIterator {
       }
     }
   }
-  
+
   @Override
   public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
     if (continueOnError && sawError) {
       return;
     }
-    
+
     try {
       source.seek(range, columnFamilies, inclusive);
     } catch (IOException ioe) {
@@ -104,15 +104,15 @@ public class ProblemReportingIterator implements InterruptibleIterator {
       }
     }
   }
-  
+
   public boolean sawError() {
     return sawError;
   }
-  
+
   public String getResource() {
     return resource;
   }
-  
+
   @Override
   public void setInterruptFlag(AtomicBoolean flag) {
     ((InterruptibleIterator) source).setInterruptFlag(flag);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReports.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReports.java b/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReports.java
index c1da89a..440ee81 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReports.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/problems/ProblemReports.java
@@ -61,16 +61,16 @@ public class ProblemReports implements Iterable<ProblemReport> {
   private static final Logger log = LoggerFactory.getLogger(ProblemReports.class);
 
   private final LRUMap problemReports = new LRUMap(1000);
-  
+
   /*
    * use a thread pool so that reporting a problem never blocks
-   * 
+   *
    * make the thread pool use a bounded queue to avoid the case where problem reports are not being processed because the whole system is in a really bad state
    * (like HDFS is down) and everything is reporting lots of problems, but problem reports can not be processed
    */
   private ExecutorService reportExecutor = new ThreadPoolExecutor(0, 1, 60, TimeUnit.SECONDS, new LinkedBlockingQueue<Runnable>(500), new NamingThreadFactory(
       "acu-problem-reporter"));
-  
+
   private final AccumuloServerContext context;
 
   public ProblemReports(AccumuloServerContext context) {
@@ -78,22 +78,22 @@ public class ProblemReports implements Iterable<ProblemReport> {
   }
 
   public void report(final ProblemReport pr) {
-    
+
     synchronized (problemReports) {
       if (problemReports.containsKey(pr)) {
         return;
       }
-      
+
       problemReports.put(pr, System.currentTimeMillis());
     }
-    
+
     Runnable r = new Runnable() {
-      
+
       @Override
       public void run() {
-        
+
         log.debug("Filing problem report " + pr.getTableName() + " " + pr.getProblemType() + " " + pr.getResource());
-        
+
         try {
           if (isMeta(pr.getTableName())) {
             // file report in zookeeper
@@ -106,33 +106,33 @@ public class ProblemReports implements Iterable<ProblemReport> {
           log.error("Failed to file problem report " + pr.getTableName() + " " + pr.getProblemType() + " " + pr.getResource(), e);
         }
       }
-      
+
     };
-    
+
     try {
       reportExecutor.execute(new LoggingRunnable(log, r));
     } catch (RejectedExecutionException ree) {
       log.error("Failed to report problem " + pr.getTableName() + " " + pr.getProblemType() + " " + pr.getResource() + "  " + ree.getMessage());
     }
-    
+
   }
-  
+
   public void printProblems() throws Exception {
     for (ProblemReport pr : this) {
       System.out.println(pr.getTableName() + " " + pr.getProblemType() + " " + pr.getResource() + " " + pr.getException());
     }
   }
-  
+
   public void deleteProblemReport(String table, ProblemType pType, String resource) {
     final ProblemReport pr = new ProblemReport(table, pType, resource, null);
-    
+
     Runnable r = new Runnable() {
-      
+
       @Override
       public void run() {
         try {
           if (isMeta(pr.getTableName())) {
-           // file report in zookeeper
+            // file report in zookeeper
             pr.removeFromZooKeeper();
           } else {
             // file report in metadata table
@@ -143,18 +143,18 @@ public class ProblemReports implements Iterable<ProblemReport> {
         }
       }
     };
-    
+
     try {
       reportExecutor.execute(new LoggingRunnable(log, r));
     } catch (RejectedExecutionException ree) {
       log.error("Failed to delete problem report " + pr.getTableName() + " " + pr.getProblemType() + " " + pr.getResource() + "  " + ree.getMessage());
     }
   }
-  
+
   private static ProblemReports instance;
-  
+
   public void deleteProblemReports(String table) throws Exception {
-    
+
     if (isMeta(table)) {
       Iterator<ProblemReport> pri = iterator(table);
       while (pri.hasNext()) {
@@ -162,38 +162,38 @@ public class ProblemReports implements Iterable<ProblemReport> {
       }
       return;
     }
-    
+
     Connector connector = context.getConnector();
     Scanner scanner = connector.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     scanner.addScanIterator(new IteratorSetting(1, "keys-only", SortedKeyIterator.class));
-    
+
     scanner.setRange(new Range(new Text("~err_" + table)));
-    
+
     Mutation delMut = new Mutation(new Text("~err_" + table));
-    
+
     boolean hasProblems = false;
     for (Entry<Key,Value> entry : scanner) {
       hasProblems = true;
       delMut.putDelete(entry.getKey().getColumnFamily(), entry.getKey().getColumnQualifier());
     }
-    
+
     if (hasProblems)
       MetadataTableUtil.getMetadataTable(context).update(delMut);
   }
-  
+
   private static boolean isMeta(String tableId) {
     return tableId.equals(MetadataTable.ID) || tableId.equals(RootTable.ID);
   }
-  
+
   public Iterator<ProblemReport> iterator(final String table) {
     try {
-      
+
       return new Iterator<ProblemReport>() {
-        
+
         IZooReaderWriter zoo = ZooReaderWriter.getInstance();
         private int iter1Count = 0;
         private Iterator<String> iter1;
-        
+
         private Iterator<String> getIter1() {
           if (iter1 == null) {
             try {
@@ -208,29 +208,29 @@ public class ProblemReports implements Iterable<ProblemReport> {
               throw new RuntimeException(e);
             }
           }
-          
+
           return iter1;
         }
-        
+
         private Iterator<Entry<Key,Value>> iter2;
-        
+
         private Iterator<Entry<Key,Value>> getIter2() {
           if (iter2 == null) {
             try {
               if ((table == null || !isMeta(table)) && iter1Count == 0) {
                 Connector connector = context.getConnector();
                 Scanner scanner = connector.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
-                
+
                 scanner.setTimeout(3, TimeUnit.SECONDS);
-                
+
                 if (table == null) {
                   scanner.setRange(new Range(new Text("~err_"), false, new Text("~err`"), false));
                 } else {
                   scanner.setRange(new Range(new Text("~err_" + table)));
                 }
-                
+
                 iter2 = scanner.iterator();
-                
+
               } else {
                 Map<Key,Value> m = Collections.emptyMap();
                 iter2 = m.entrySet().iterator();
@@ -239,23 +239,23 @@ public class ProblemReports implements Iterable<ProblemReport> {
               throw new RuntimeException(e);
             }
           }
-          
+
           return iter2;
         }
-        
+
         @Override
         public boolean hasNext() {
           if (getIter1().hasNext()) {
             return true;
           }
-          
+
           if (getIter2().hasNext()) {
             return true;
           }
-          
+
           return false;
         }
-        
+
         @Override
         public ProblemReport next() {
           try {
@@ -263,66 +263,66 @@ public class ProblemReports implements Iterable<ProblemReport> {
               iter1Count++;
               return ProblemReport.decodeZooKeeperEntry(getIter1().next());
             }
-            
+
             if (getIter2().hasNext()) {
               return ProblemReport.decodeMetadataEntry(getIter2().next());
             }
           } catch (Exception e) {
             throw new RuntimeException(e);
           }
-          
+
           throw new NoSuchElementException();
         }
-        
+
         @Override
         public void remove() {
           throw new UnsupportedOperationException();
         }
-        
+
       };
-      
+
     } catch (Exception e) {
       throw new RuntimeException(e);
     }
   }
-  
+
   @Override
   public Iterator<ProblemReport> iterator() {
     return iterator(null);
   }
-  
+
   public static synchronized ProblemReports getInstance(AccumuloServerContext context) {
     if (instance == null) {
       instance = new ProblemReports(context);
     }
-    
+
     return instance;
   }
-  
+
   public static void main(String args[]) throws Exception {
     getInstance(new AccumuloServerContext(new ServerConfigurationFactory(HdfsZooInstance.getInstance()))).printProblems();
   }
-  
+
   public Map<String,Map<ProblemType,Integer>> summarize() {
-    
+
     TreeMap<String,Map<ProblemType,Integer>> summary = new TreeMap<String,Map<ProblemType,Integer>>();
-    
+
     for (ProblemReport pr : this) {
       Map<ProblemType,Integer> tableProblems = summary.get(pr.getTableName());
       if (tableProblems == null) {
         tableProblems = new EnumMap<ProblemType,Integer>(ProblemType.class);
         summary.put(pr.getTableName(), tableProblems);
       }
-      
+
       Integer count = tableProblems.get(pr.getProblemType());
       if (count == null) {
         count = 0;
       }
-      
+
       tableProblems.put(pr.getProblemType(), count + 1);
     }
-    
+
     return summary;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/replication/DistributedWorkQueueWorkAssignerHelper.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/replication/DistributedWorkQueueWorkAssignerHelper.java b/server/base/src/main/java/org/apache/accumulo/server/replication/DistributedWorkQueueWorkAssignerHelper.java
index baa8383..2494ee2 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/replication/DistributedWorkQueueWorkAssignerHelper.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/replication/DistributedWorkQueueWorkAssignerHelper.java
@@ -25,16 +25,15 @@ import com.google.common.base.Preconditions;
 import com.google.common.collect.Maps;
 
 /**
- * 
+ *
  */
 public class DistributedWorkQueueWorkAssignerHelper {
 
-
   public static final String KEY_SEPARATOR = "|";
 
   /**
    * Serialize a filename and a {@link ReplicationTarget} into the expected key format for use with the {@link DistributedWorkQueue}
-   * 
+   *
    * @param filename
    *          Filename for data to be replicated
    * @param replTarget

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/replication/ReplicationUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/replication/ReplicationUtil.java b/server/base/src/main/java/org/apache/accumulo/server/replication/ReplicationUtil.java
index edd2642..e1bbe3d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/replication/ReplicationUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/replication/ReplicationUtil.java
@@ -98,8 +98,7 @@ public class ReplicationUtil {
     for (Entry<String,String> property : context.getConfiguration().getAllPropertiesWithPrefix(Property.REPLICATION_PEERS).entrySet()) {
       String key = property.getKey();
       // Filter out cruft that we don't want
-      if (!key.startsWith(Property.REPLICATION_PEER_USER.getKey())
-          && !key.startsWith(Property.REPLICATION_PEER_PASSWORD.getKey())) {
+      if (!key.startsWith(Property.REPLICATION_PEER_USER.getKey()) && !key.startsWith(Property.REPLICATION_PEER_PASSWORD.getKey())) {
         String peerName = property.getKey().substring(Property.REPLICATION_PEERS.getKey().length());
         ReplicaSystem replica;
         try {
@@ -135,12 +134,12 @@ public class ReplicationUtil {
       TableConfiguration tableConf = context.getServerConfigurationFactory().getTableConfiguration(localId);
       for (Entry<String,String> prop : tableConf.getAllPropertiesWithPrefix(Property.TABLE_REPLICATION_TARGET).entrySet()) {
         String peerName = prop.getKey().substring(Property.TABLE_REPLICATION_TARGET.getKey().length());
-          String remoteIdentifier = prop.getValue();
-          ReplicationTarget target = new ReplicationTarget(peerName, remoteIdentifier, localId);
+        String remoteIdentifier = prop.getValue();
+        ReplicationTarget target = new ReplicationTarget(peerName, remoteIdentifier, localId);
 
-          allConfiguredTargets.add(target);
-        }
+        allConfiguredTargets.add(target);
       }
+    }
 
     return allConfiguredTargets;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/replication/StatusCombiner.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/replication/StatusCombiner.java b/server/base/src/main/java/org/apache/accumulo/server/replication/StatusCombiner.java
index ecca99e..5317c4d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/replication/StatusCombiner.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/replication/StatusCombiner.java
@@ -35,7 +35,7 @@ import com.google.protobuf.InvalidProtocolBufferException;
 
 /**
  * Defines the rules for combining multiple {@link Status} messages
- * 
+ *
  * Messages that are "closed", stay closed. "Begin" and "end" always choose the maximum of the two.
  */
 public class StatusCombiner extends TypedValueCombiner<Status> {
@@ -113,7 +113,7 @@ public class StatusCombiner extends TypedValueCombiner<Status> {
 
   /**
    * Update a {@link Builder} with another {@link Status}
-   * 
+   *
    * @param combined
    *          The Builder to combine into
    * @param status

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/replication/WorkAssigner.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/replication/WorkAssigner.java b/server/base/src/main/java/org/apache/accumulo/server/replication/WorkAssigner.java
index 1069835..da52354 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/replication/WorkAssigner.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/replication/WorkAssigner.java
@@ -19,7 +19,6 @@ package org.apache.accumulo.server.replication;
 import org.apache.accumulo.core.client.Connector;
 import org.apache.accumulo.core.conf.AccumuloConfiguration;
 
-
 /**
  * Interface to allow for multiple implementations that assign replication work
  */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/replication/ZooKeeperInitialization.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/replication/ZooKeeperInitialization.java b/server/base/src/main/java/org/apache/accumulo/server/replication/ZooKeeperInitialization.java
index a97853d..9a0825d 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/replication/ZooKeeperInitialization.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/replication/ZooKeeperInitialization.java
@@ -23,8 +23,7 @@ import org.apache.zookeeper.KeeperException;
 /**
  * We don't want to introduce an upgrade path to 1.7 only for some new nodes within ZooKeeper
  * <p>
- * We can take the penalty of embedding this logic into the server processes, but alleviate
- * users/developers from having to worry about the zookeeper state.
+ * We can take the penalty of embedding this logic into the server processes, but alleviate users/developers from having to worry about the zookeeper state.
  */
 public class ZooKeeperInitialization {
   /**

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/rpc/ClientInfoProcessorFactory.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/rpc/ClientInfoProcessorFactory.java b/server/base/src/main/java/org/apache/accumulo/server/rpc/ClientInfoProcessorFactory.java
index 5f630c2..cbd719b 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/rpc/ClientInfoProcessorFactory.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/rpc/ClientInfoProcessorFactory.java
@@ -50,4 +50,4 @@ public class ClientInfoProcessorFactory extends TProcessorFactory {
     }
     return super.getProcessor(trans);
   }
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/rpc/RpcWrapper.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/rpc/RpcWrapper.java b/server/base/src/main/java/org/apache/accumulo/server/rpc/RpcWrapper.java
index a488da9..b28ecb7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/rpc/RpcWrapper.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/rpc/RpcWrapper.java
@@ -40,6 +40,7 @@ public class RpcWrapper {
   public static <T> T service(final T instance) {
     InvocationHandler handler = new RpcServerInvocationHandler<T>(instance) {
       private final Logger log = LoggerFactory.getLogger(instance.getClass());
+
       @Override
       public Object invoke(Object obj, Method method, Object[] args) throws Throwable {
         try {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/rpc/TNonblockingServerSocket.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/rpc/TNonblockingServerSocket.java b/server/base/src/main/java/org/apache/accumulo/server/rpc/TNonblockingServerSocket.java
index 3afe149..d035862 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/rpc/TNonblockingServerSocket.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/rpc/TNonblockingServerSocket.java
@@ -19,11 +19,6 @@
 
 package org.apache.accumulo.server.rpc;
 
-import org.apache.log4j.Logger;
-import org.apache.thrift.transport.TNonblockingServerTransport;
-import org.apache.thrift.transport.TNonblockingSocket;
-import org.apache.thrift.transport.TTransportException;
-
 import java.io.IOException;
 import java.net.InetSocketAddress;
 import java.net.ServerSocket;
@@ -34,11 +29,16 @@ import java.nio.channels.Selector;
 import java.nio.channels.ServerSocketChannel;
 import java.nio.channels.SocketChannel;
 
+import org.apache.log4j.Logger;
+import org.apache.thrift.transport.TNonblockingServerTransport;
+import org.apache.thrift.transport.TNonblockingSocket;
+import org.apache.thrift.transport.TTransportException;
+
 /**
  * Wrapper around ServerSocketChannel.
  *
- * This class is copied from org.apache.thrift.transport.TNonblockingServerSocket version 0.9.
- * The only change (apart from the logging statements) is the addition of the {@link #getPort()} method to retrieve the port used by the ServerSocket.
+ * This class is copied from org.apache.thrift.transport.TNonblockingServerSocket version 0.9. The only change (apart from the logging statements) is the
+ * addition of the {@link #getPort()} method to retrieve the port used by the ServerSocket.
  */
 public class TNonblockingServerSocket extends TNonblockingServerTransport {
   private static final Logger log = Logger.getLogger(TNonblockingServerTransport.class.getName());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/AuditedSecurityOperation.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/AuditedSecurityOperation.java b/server/base/src/main/java/org/apache/accumulo/server/security/AuditedSecurityOperation.java
index a2afeac..cc7a7cd 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/AuditedSecurityOperation.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/AuditedSecurityOperation.java
@@ -121,7 +121,8 @@ public class AuditedSecurityOperation extends SecurityOperation {
   private void audit(TCredentials credentials, boolean permitted, String template, Object... args) {
     if (shouldAudit(credentials)) {
       String prefix = permitted ? "permitted" : "denied";
-      audit.info("operation: " + prefix + "; user: " + credentials.getPrincipal() + "; client: "  + TServerUtils.clientAddress.get() + "; " + String.format(template, args));
+      audit.info("operation: " + prefix + "; user: " + credentials.getPrincipal() + "; client: " + TServerUtils.clientAddress.get() + "; "
+          + String.format(template, args));
     }
   }
 
@@ -166,8 +167,8 @@ public class AuditedSecurityOperation extends SecurityOperation {
   public static final String CAN_SCAN_BATCH_AUDIT_TEMPLATE = "action: scan; targetTable: %s; authorizations: %s; range: %s; columns: %s; iterators: %s; iteratorOptions: %s;";
 
   @Override
-  public boolean canScan(TCredentials credentials, String tableId, String namespaceId, Map<TKeyExtent,List<TRange>> tbatch, List<TColumn> tcolumns, List<IterInfo> ssiList,
-      Map<String,Map<String,String>> ssio, List<ByteBuffer> authorizations) throws ThriftSecurityException {
+  public boolean canScan(TCredentials credentials, String tableId, String namespaceId, Map<TKeyExtent,List<TRange>> tbatch, List<TColumn> tcolumns,
+      List<IterInfo> ssiList, Map<String,Map<String,String>> ssio, List<ByteBuffer> authorizations) throws ThriftSecurityException {
     if (shouldAudit(credentials, tableId)) {
       @SuppressWarnings({"unchecked", "rawtypes"})
       Map<KeyExtent,List<Range>> convertedBatch = Translator.translate(tbatch, new Translator.TKeyExtentTranslator(), new Translator.ListTranslator(
@@ -278,7 +279,8 @@ public class AuditedSecurityOperation extends SecurityOperation {
   public static final String CAN_CLONE_TABLE_AUDIT_TEMPLATE = "action: cloneTable; targetTable: %s; newTableName: %s";
 
   @Override
-  public boolean canCloneTable(TCredentials c, String tableId, String tableName, String destinationNamespaceId, String sourceNamespaceId) throws ThriftSecurityException {
+  public boolean canCloneTable(TCredentials c, String tableId, String tableName, String destinationNamespaceId, String sourceNamespaceId)
+      throws ThriftSecurityException {
     String oldTableName = getTableName(tableId);
     try {
       boolean result = super.canCloneTable(c, tableId, tableName, destinationNamespaceId, sourceNamespaceId);
@@ -293,9 +295,10 @@ public class AuditedSecurityOperation extends SecurityOperation {
   public static final String CAN_DELETE_RANGE_AUDIT_TEMPLATE = "action: deleteData; targetTable: %s; startRange: %s; endRange: %s;";
 
   @Override
-  public boolean canDeleteRange(TCredentials c, String tableId, String tableName, Text startRow, Text endRow, String namespaceId) throws ThriftSecurityException {
+  public boolean canDeleteRange(TCredentials c, String tableId, String tableName, Text startRow, Text endRow, String namespaceId)
+      throws ThriftSecurityException {
     try {
-      boolean result = super.canDeleteRange(c, tableId, tableName, startRow, endRow,namespaceId);
+      boolean result = super.canDeleteRange(c, tableId, tableName, startRow, endRow, namespaceId);
       audit(c, result, CAN_DELETE_RANGE_AUDIT_TEMPLATE, tableName, startRow.toString(), endRow.toString());
       return result;
     } catch (ThriftSecurityException ex) {
@@ -377,7 +380,8 @@ public class AuditedSecurityOperation extends SecurityOperation {
   public static final String GRANT_TABLE_PERMISSION_AUDIT_TEMPLATE = "action: grantTablePermission; permission: %s; targetTable: %s; targetUser: %s;";
 
   @Override
-  public void grantTablePermission(TCredentials credentials, String user, String tableId, TablePermission permission, String namespaceId) throws ThriftSecurityException {
+  public void grantTablePermission(TCredentials credentials, String user, String tableId, TablePermission permission, String namespaceId)
+      throws ThriftSecurityException {
     String tableName = getTableName(tableId);
     try {
       super.grantTablePermission(credentials, user, tableId, permission, namespaceId);
@@ -405,7 +409,8 @@ public class AuditedSecurityOperation extends SecurityOperation {
   public static final String REVOKE_TABLE_PERMISSION_AUDIT_TEMPLATE = "action: revokeTablePermission; permission: %s; targetTable: %s; targetUser: %s;";
 
   @Override
-  public void revokeTablePermission(TCredentials credentials, String user, String tableId, TablePermission permission, String namespaceId) throws ThriftSecurityException {
+  public void revokeTablePermission(TCredentials credentials, String user, String tableId, TablePermission permission, String namespaceId)
+      throws ThriftSecurityException {
     String tableName = getTableName(tableId);
     try {
       super.revokeTablePermission(credentials, user, tableId, permission, namespaceId);
@@ -437,7 +442,7 @@ public class AuditedSecurityOperation extends SecurityOperation {
   }
 
   // The audit log is already logging the principal, so we don't have anything else to audit
-  public static final String AUTHENICATE_AUDIT_TEMPLATE =  "";
+  public static final String AUTHENICATE_AUDIT_TEMPLATE = "";
 
   @Override
   protected void authenticate(TCredentials credentials) throws ThriftSecurityException {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/base/src/main/java/org/apache/accumulo/server/security/SecurityOperation.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/security/SecurityOperation.java b/server/base/src/main/java/org/apache/accumulo/server/security/SecurityOperation.java
index 5e81018..5fe57b7 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/security/SecurityOperation.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/security/SecurityOperation.java
@@ -104,8 +104,8 @@ public class SecurityOperation {
   }
 
   protected static PermissionHandler getPermHandler(String instanceId, boolean initialize) {
-    PermissionHandler toRet = SiteConfiguration.getInstance().instantiateClassProperty(Property.INSTANCE_SECURITY_PERMISSION_HANDLER,
-        PermissionHandler.class, ZKPermHandler.getInstance());
+    PermissionHandler toRet = SiteConfiguration.getInstance().instantiateClassProperty(Property.INSTANCE_SECURITY_PERMISSION_HANDLER, PermissionHandler.class,
+        ZKPermHandler.getInstance());
     toRet.initialize(instanceId, initialize);
     return toRet;
   }
@@ -241,7 +241,7 @@ public class SecurityOperation {
 
   /**
    * Checks if a user has a system permission
-   * 
+   *
    * @return true if a user exists and has permission; false otherwise
    */
   private boolean hasSystemPermissionWithNamespaceId(TCredentials credentials, SystemPermission permission, String namespaceId, boolean useCached)
@@ -261,7 +261,7 @@ public class SecurityOperation {
   /**
    * Checks if a user has a system permission<br/>
    * This cannot check if a system user has permission.
-   * 
+   *
    * @return true if a user exists and has permission; false otherwise
    */
   private boolean _hasSystemPermission(String user, SystemPermission permission, boolean useCached) throws ThriftSecurityException {
@@ -281,10 +281,11 @@ public class SecurityOperation {
 
   /**
    * Checks if a user has a table permission
-   * 
+   *
    * @return true if a user exists and has permission; false otherwise
    */
-  protected boolean hasTablePermission(TCredentials credentials, String tableId, String namespaceId, TablePermission permission, boolean useCached) throws ThriftSecurityException {
+  protected boolean hasTablePermission(TCredentials credentials, String tableId, String namespaceId, TablePermission permission, boolean useCached)
+      throws ThriftSecurityException {
     if (isSystemUser(credentials))
       return true;
     return _hasTablePermission(credentials.getPrincipal(), tableId, permission, useCached)
@@ -294,7 +295,7 @@ public class SecurityOperation {
   /**
    * Checks if a user has a table permission<br/>
    * This cannot check if a system user has permission.
-   * 
+   *
    * @return true if a user exists and has permission; false otherwise
    */
   protected boolean _hasTablePermission(String user, String table, TablePermission permission, boolean useCached) throws ThriftSecurityException {
@@ -317,7 +318,7 @@ public class SecurityOperation {
   /**
    * Checks if a user has a namespace permission<br/>
    * This cannot check if a system user has permission.
-   * 
+   *
    * @return true if a user exists and has permission; false otherwise
    */
   protected boolean _hasNamespacePermission(String user, String namespace, NamespacePermission permission, boolean useCached) throws ThriftSecurityException {
@@ -369,8 +370,8 @@ public class SecurityOperation {
     return canScan(credentials, tableId, namespaceId);
   }
 
-  public boolean canScan(TCredentials credentials, String table, String namespaceId, Map<TKeyExtent,List<TRange>> tbatch, List<TColumn> tcolumns, List<IterInfo> ssiList,
-      Map<String,Map<String,String>> ssio, List<ByteBuffer> authorizations) throws ThriftSecurityException {
+  public boolean canScan(TCredentials credentials, String table, String namespaceId, Map<TKeyExtent,List<TRange>> tbatch, List<TColumn> tcolumns,
+      List<IterInfo> ssiList, Map<String,Map<String,String>> ssio, List<ByteBuffer> authorizations) throws ThriftSecurityException {
     return canScan(credentials, table, namespaceId);
   }
 
@@ -379,11 +380,13 @@ public class SecurityOperation {
     return hasTablePermission(credentials, tableId, namespaceId, TablePermission.WRITE, true);
   }
 
-  public boolean canConditionallyUpdate(TCredentials credentials, String tableID, String namespaceId, List<ByteBuffer> authorizations) throws ThriftSecurityException {
+  public boolean canConditionallyUpdate(TCredentials credentials, String tableID, String namespaceId, List<ByteBuffer> authorizations)
+      throws ThriftSecurityException {
 
     authenticate(credentials);
 
-    return hasTablePermission(credentials, tableID, namespaceId, TablePermission.WRITE, true) && hasTablePermission(credentials, tableID, namespaceId, TablePermission.READ, true);
+    return hasTablePermission(credentials, tableID, namespaceId, TablePermission.WRITE, true)
+        && hasTablePermission(credentials, tableID, namespaceId, TablePermission.READ, true);
   }
 
   public boolean canSplitTablet(TCredentials credentials, String tableId, String namespaceId) throws ThriftSecurityException {
@@ -403,7 +406,8 @@ public class SecurityOperation {
 
   public boolean canFlush(TCredentials c, String tableId, String namespaceId) throws ThriftSecurityException {
     authenticate(c);
-    return hasTablePermission(c, tableId, namespaceId, TablePermission.WRITE, false) || hasTablePermission(c, tableId, namespaceId, TablePermission.ALTER_TABLE, false);
+    return hasTablePermission(c, tableId, namespaceId, TablePermission.WRITE, false)
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.ALTER_TABLE, false);
   }
 
   public boolean canAlterTable(TCredentials c, String tableId, String namespaceId) throws ThriftSecurityException {
@@ -423,14 +427,17 @@ public class SecurityOperation {
         || hasTablePermission(c, tableId, namespaceId, TablePermission.ALTER_TABLE, false);
   }
 
-  public boolean canCloneTable(TCredentials c, String tableId, String tableName, String destinationNamespaceId, String srcNamespaceId) throws ThriftSecurityException {
+  public boolean canCloneTable(TCredentials c, String tableId, String tableName, String destinationNamespaceId, String srcNamespaceId)
+      throws ThriftSecurityException {
     authenticate(c);
-    return hasSystemPermissionWithNamespaceId(c, SystemPermission.CREATE_TABLE, destinationNamespaceId, false) && hasTablePermission(c, tableId, srcNamespaceId, TablePermission.READ, false);
+    return hasSystemPermissionWithNamespaceId(c, SystemPermission.CREATE_TABLE, destinationNamespaceId, false)
+        && hasTablePermission(c, tableId, srcNamespaceId, TablePermission.READ, false);
   }
 
   public boolean canDeleteTable(TCredentials c, String tableId, String namespaceId) throws ThriftSecurityException {
     authenticate(c);
-    return hasSystemPermissionWithNamespaceId(c, SystemPermission.DROP_TABLE, namespaceId, false) || hasTablePermission(c, tableId, namespaceId, TablePermission.DROP_TABLE, false);
+    return hasSystemPermissionWithNamespaceId(c, SystemPermission.DROP_TABLE, namespaceId, false)
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.DROP_TABLE, false);
   }
 
   public boolean canOnlineOfflineTable(TCredentials c, String tableId, FateOperation op, String namespaceId) throws ThriftSecurityException {
@@ -447,9 +454,11 @@ public class SecurityOperation {
         || hasTablePermission(c, tableId, namespaceId, TablePermission.ALTER_TABLE, false);
   }
 
-  public boolean canDeleteRange(TCredentials c, String tableId, String tableName, Text startRow, Text endRow, String namespaceId) throws ThriftSecurityException {
+  public boolean canDeleteRange(TCredentials c, String tableId, String tableName, Text startRow, Text endRow, String namespaceId)
+      throws ThriftSecurityException {
     authenticate(c);
-    return hasSystemPermissionWithNamespaceId(c, SystemPermission.SYSTEM, namespaceId, false) || hasTablePermission(c, tableId, namespaceId, TablePermission.WRITE, false);
+    return hasSystemPermissionWithNamespaceId(c, SystemPermission.SYSTEM, namespaceId, false)
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.WRITE, false);
   }
 
   public boolean canBulkImport(TCredentials c, String tableId, String tableName, String dir, String failDir, String namespaceId) throws ThriftSecurityException {
@@ -464,7 +473,8 @@ public class SecurityOperation {
   public boolean canCompact(TCredentials c, String tableId, String namespaceId) throws ThriftSecurityException {
     authenticate(c);
     return hasSystemPermissionWithNamespaceId(c, SystemPermission.ALTER_TABLE, namespaceId, false)
-        || hasTablePermission(c, tableId, namespaceId, TablePermission.ALTER_TABLE, false) || hasTablePermission(c, tableId, namespaceId, TablePermission.WRITE, false);
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.ALTER_TABLE, false)
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.WRITE, false);
   }
 
   public boolean canChangeAuthorizations(TCredentials c, String user) throws ThriftSecurityException {
@@ -499,7 +509,8 @@ public class SecurityOperation {
 
   public boolean canGrantTable(TCredentials c, String user, String tableId, String namespaceId) throws ThriftSecurityException {
     authenticate(c);
-    return hasSystemPermissionWithNamespaceId(c, SystemPermission.ALTER_TABLE, namespaceId, false) || hasTablePermission(c, tableId, namespaceId, TablePermission.GRANT, false);
+    return hasSystemPermissionWithNamespaceId(c, SystemPermission.ALTER_TABLE, namespaceId, false)
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.GRANT, false);
   }
 
   public boolean canGrantNamespace(TCredentials c, String user, String namespace) throws ThriftSecurityException {
@@ -510,7 +521,7 @@ public class SecurityOperation {
     authenticate(c);
     // The one case where Table/SystemPermission -> NamespacePermission breaks down. The alternative is to make SystemPermission.ALTER_NAMESPACE provide
     // NamespacePermission.GRANT & ALTER_NAMESPACE, but then it would cause some permission checks to succeed with GRANT when they shouldn't
-    
+
     // This is a bit hackier then I (vines) wanted, but I think this one hackiness makes the overall SecurityOperations more succinct.
     return hasSystemPermissionWithNamespaceId(c, SystemPermission.ALTER_NAMESPACE, namespace, false)
         || hasNamespacePermission(c, c.principal, namespace, NamespacePermission.GRANT);
@@ -531,7 +542,8 @@ public class SecurityOperation {
 
   public boolean canRevokeTable(TCredentials c, String user, String tableId, String namespaceId) throws ThriftSecurityException {
     authenticate(c);
-    return hasSystemPermissionWithNamespaceId(c, SystemPermission.ALTER_TABLE, namespaceId, false) || hasTablePermission(c, tableId, namespaceId, TablePermission.GRANT, false);
+    return hasSystemPermissionWithNamespaceId(c, SystemPermission.ALTER_TABLE, namespaceId, false)
+        || hasTablePermission(c, tableId, namespaceId, TablePermission.GRANT, false);
   }
 
   public boolean canRevokeNamespace(TCredentials c, String user, String namespace) throws ThriftSecurityException {


[52/66] [abbrv] accumulo git commit: Merge branch '1.6' with -sours

Posted by ct...@apache.org.
Merge branch '1.6' with -sours


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/b693b6da
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/b693b6da
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/b693b6da

Branch: refs/heads/master
Commit: b693b6da8be9608372684cfa64fe34fc45ea4f6a
Parents: 6bc6760 dff686b
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:22:11 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:22:11 2015 -0500

----------------------------------------------------------------------

----------------------------------------------------------------------



[60/66] [abbrv] accumulo git commit: Merge branch '1.5' into 1.6

Posted by ct...@apache.org.
Merge branch '1.5' into 1.6

Conflicts:
	pom.xml


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/627e525b
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/627e525b
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/627e525b

Branch: refs/heads/master
Commit: 627e525b6b881873a1b31d63ca22df26f3b7776f
Parents: dff686b d2c116f
Author: Christopher Tubbs <ct...@apache.org>
Authored: Thu Jan 8 20:35:48 2015 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:35:48 2015 -0500

----------------------------------------------------------------------
 pom.xml | 108 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 108 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/627e525b/pom.xml
----------------------------------------------------------------------
diff --cc pom.xml
index 22d5450,711a21d..6a90918
--- a/pom.xml
+++ b/pom.xml
@@@ -496,21 -367,14 +496,26 @@@
      <pluginManagement>
        <plugins>
          <plugin>
 +          <groupId>org.codehaus.mojo</groupId>
 +          <artifactId>findbugs-maven-plugin</artifactId>
 +          <version>2.5.5</version>
 +          <configuration>
 +            <xmlOutput>true</xmlOutput>
 +            <effort>Max</effort>
 +            <failOnError>true</failOnError>
 +            <includeTests>true</includeTests>
 +            <maxRank>6</maxRank>
 +          </configuration>
 +        </plugin>
 +        <plugin>
+           <groupId>org.apache.maven.plugins</groupId>
+           <artifactId>maven-checkstyle-plugin</artifactId>
+           <version>2.13</version>
+         </plugin>
+         <plugin>
            <groupId>com.google.code.sortpom</groupId>
            <artifactId>maven-sortpom-plugin</artifactId>
 -          <version>2.1.0</version>
 +          <version>2.3.0</version>
            <configuration>
              <predefinedSortOrder>recommended_2008_06</predefinedSortOrder>
              <lineSeparator>\n</lineSeparator>
@@@ -676,20 -571,19 +681,33 @@@
                  <pluginExecution>
                    <pluginExecutionFilter>
                      <groupId>org.apache.maven.plugins</groupId>
 +                    <artifactId>maven-plugin-plugin</artifactId>
 +                    <versionRange>[3.2,)</versionRange>
 +                    <goals>
 +                      <goal>helpmojo</goal>
 +                      <goal>descriptor</goal>
 +                    </goals>
 +                  </pluginExecutionFilter>
 +                  <action>
 +                    <ignore />
 +                  </action>
 +                </pluginExecution>
 +                <pluginExecution>
 +                  <pluginExecutionFilter>
 +                    <groupId>org.apache.maven.plugins</groupId>
+                     <artifactId>maven-checkstyle-plugin</artifactId>
+                     <versionRange>[2.13,)</versionRange>
+                     <goals>
+                       <goal>check</goal>
+                     </goals>
+                   </pluginExecutionFilter>
+                   <action>
+                     <ignore />
+                   </action>
+                 </pluginExecution>
+                 <pluginExecution>
+                   <pluginExecutionFilter>
+                     <groupId>org.apache.maven.plugins</groupId>
                      <artifactId>maven-dependency-plugin</artifactId>
                      <versionRange>[2.0,)</versionRange>
                      <goals>


[54/66] [abbrv] accumulo git commit: ACCUMULO-3451 Add minimal checkstyle enforcement

Posted by ct...@apache.org.
ACCUMULO-3451 Add minimal checkstyle enforcement


Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/d2c116ff
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/d2c116ff
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/d2c116ff

Branch: refs/heads/master
Commit: d2c116ffae59672ff995033e81be73260c4d5c25
Parents: c2155f4
Author: Christopher Tubbs <ct...@apache.org>
Authored: Tue Dec 23 18:51:04 2014 -0500
Committer: Christopher Tubbs <ct...@apache.org>
Committed: Thu Jan 8 20:23:36 2015 -0500

----------------------------------------------------------------------
 pom.xml | 108 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 108 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo/blob/d2c116ff/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 8316a33..711a21d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -367,6 +367,11 @@
     <pluginManagement>
       <plugins>
         <plugin>
+          <groupId>org.apache.maven.plugins</groupId>
+          <artifactId>maven-checkstyle-plugin</artifactId>
+          <version>2.13</version>
+        </plugin>
+        <plugin>
           <groupId>com.google.code.sortpom</groupId>
           <artifactId>maven-sortpom-plugin</artifactId>
           <version>2.1.0</version>
@@ -566,6 +571,19 @@
                 <pluginExecution>
                   <pluginExecutionFilter>
                     <groupId>org.apache.maven.plugins</groupId>
+                    <artifactId>maven-checkstyle-plugin</artifactId>
+                    <versionRange>[2.13,)</versionRange>
+                    <goals>
+                      <goal>check</goal>
+                    </goals>
+                  </pluginExecutionFilter>
+                  <action>
+                    <ignore />
+                  </action>
+                </pluginExecution>
+                <pluginExecution>
+                  <pluginExecutionFilter>
+                    <groupId>org.apache.maven.plugins</groupId>
                     <artifactId>maven-dependency-plugin</artifactId>
                     <versionRange>[2.0,)</versionRange>
                     <goals>
@@ -709,6 +727,96 @@
         </executions>
       </plugin>
       <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-checkstyle-plugin</artifactId>
+        <configuration>
+          <checkstyleRules>
+            <module name="Checker">
+              <property name="charset" value="UTF-8" />
+              <property name="severity" value="warning" />
+              <!-- Checks for whitespace                               -->
+              <!-- See http://checkstyle.sf.net/config_whitespace.html -->
+              <module name="FileTabCharacter">
+                <property name="eachLine" value="true" />
+              </module>
+              <module name="TreeWalker">
+                <module name="OuterTypeFilename" />
+                <module name="LineLength">
+                  <!-- needs extra, because Eclipse formatter ignores the ending left brace -->
+                  <property name="max" value="200"/>
+                  <property name="ignorePattern" value="^package.*|^import.*|a href|href|http://|https://|ftp://"/>
+                </module>
+                <module name="AvoidStarImport" />
+                <module name="NoLineWrap" />
+                <module name="LeftCurly">
+                  <property name="maxLineLength" value="160" />
+                </module>
+                <module name="RightCurly" />
+                <module name="RightCurly">
+                  <property name="option" value="alone" />
+                  <property name="tokens" value="CLASS_DEF, METHOD_DEF, CTOR_DEF, LITERAL_FOR, LITERAL_WHILE, LITERAL_DO, STATIC_INIT, INSTANCE_INIT" />
+                </module>
+                <module name="SeparatorWrap">
+                  <property name="tokens" value="DOT" />
+                  <property name="option" value="nl" />
+                </module>
+                <module name="SeparatorWrap">
+                  <property name="tokens" value="COMMA" />
+                  <property name="option" value="EOL" />
+                </module>
+                <module name="PackageName">
+                  <property name="format" value="^[a-z]+(\.[a-z][a-zA-Z0-9]*)*$" />
+                </module>
+                <module name="MethodTypeParameterName">
+                  <property name="format" value="(^[A-Z][0-9]?)$|([A-Z][a-zA-Z0-9]*[T]$)" />
+                </module>
+                <module name="MethodParamPad" />
+                <module name="OperatorWrap">
+                  <property name="option" value="NL" />
+                  <property name="tokens" value="BAND, BOR, BSR, BXOR, DIV, EQUAL, GE, GT, LAND, LE, LITERAL_INSTANCEOF, LOR, LT, MINUS, MOD, NOT_EQUAL, QUESTION, SL, SR, STAR " />
+                </module>
+                <module name="AnnotationLocation">
+                  <property name="tokens" value="CLASS_DEF, INTERFACE_DEF, ENUM_DEF, METHOD_DEF, CTOR_DEF" />
+                </module>
+                <module name="AnnotationLocation">
+                  <property name="tokens" value="VARIABLE_DEF" />
+                  <property name="allowSamelineMultipleAnnotations" value="true" />
+                </module>
+                <module name="NonEmptyAtclauseDescription" />
+                <module name="JavadocTagContinuationIndentation" />
+                <module name="JavadocMethod">
+                  <property name="allowMissingJavadoc" value="true" />
+                  <property name="allowMissingParamTags" value="true" />
+                  <property name="allowMissingThrowsTags" value="true" />
+                  <property name="allowMissingReturnTag" value="true" />
+                  <property name="allowedAnnotations" value="Override,Test,BeforeClass,AfterClass,Before,After" />
+                  <property name="allowThrowsTagsForSubclasses" value="true" />
+                </module>
+                <module name="SingleLineJavadoc" />
+              </module>
+            </module>
+          </checkstyleRules>
+          <violationSeverity>warning</violationSeverity>
+          <includeTestSourceDirectory>true</includeTestSourceDirectory>
+          <excludes>**/thrift/*.java</excludes>
+        </configuration>
+        <dependencies>
+          <dependency>
+            <groupId>com.puppycrawl.tools</groupId>
+            <artifactId>checkstyle</artifactId>
+            <version>6.1.1</version>
+          </dependency>
+        </dependencies>
+        <executions>
+          <execution>
+            <id>check-style</id>
+            <goals>
+              <goal>check</goal>
+            </goals>
+          </execution>
+        </executions>
+      </plugin>
+      <plugin>
         <groupId>com.github.koraktor</groupId>
         <artifactId>mavanagaiata</artifactId>
         <executions>


[26/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/IntersectingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/IntersectingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/IntersectingIteratorTest.java
index 17e1c1d..365cee4 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/IntersectingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/IntersectingIteratorTest.java
@@ -51,11 +51,11 @@ import org.apache.log4j.Level;
 import org.apache.log4j.Logger;
 
 public class IntersectingIteratorTest extends TestCase {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
   private static final Logger log = Logger.getLogger(IntersectingIterator.class);
   private static IteratorEnvironment env = new DefaultIteratorEnvironment();
-  
+
   TreeMap<Key,Value> map;
   HashSet<Text> docs = new HashSet<Text>();
   Text[] columnFamilies;
@@ -63,20 +63,20 @@ public class IntersectingIteratorTest extends TestCase {
   Text[] otherColumnFamilies;
   Text[] searchFamilies;
   boolean[] notFlags;
-  
+
   int docid = 0;
-  
+
   static {
     log.setLevel(Level.OFF);
   }
-  
+
   private TreeMap<Key,Value> createSortedMap(float hitRatio, int numRows, int numDocsPerRow, Text[] columnFamilies, Text[] otherColumnFamilies,
       HashSet<Text> docs, Text[] negatedColumns) {
     Random r = new Random();
     Value v = new Value(new byte[0]);
     TreeMap<Key,Value> map = new TreeMap<Key,Value>();
     boolean[] negateMask = new boolean[columnFamilies.length];
-    
+
     for (int i = 0; i < columnFamilies.length; i++) {
       negateMask[i] = false;
       if (negatedColumns.length > 0)
@@ -113,33 +113,33 @@ public class IntersectingIteratorTest extends TestCase {
     }
     return map;
   }
-  
+
   private SortedKeyValueIterator<Key,Value> createIteratorStack(float hitRatio, int numRows, int numDocsPerRow, Text[] columnFamilies,
       Text[] otherColumnFamilies, HashSet<Text> docs) throws IOException {
     Text nullText[] = new Text[0];
     return createIteratorStack(hitRatio, numRows, numDocsPerRow, columnFamilies, otherColumnFamilies, docs, nullText);
   }
-  
+
   private SortedKeyValueIterator<Key,Value> createIteratorStack(float hitRatio, int numRows, int numDocsPerRow, Text[] columnFamilies,
       Text[] otherColumnFamilies, HashSet<Text> docs, Text[] negatedColumns) throws IOException {
     TreeMap<Key,Value> inMemoryMap = createSortedMap(hitRatio, numRows, numDocsPerRow, columnFamilies, otherColumnFamilies, docs, negatedColumns);
     return new SortedMapIterator(inMemoryMap);
   }
-  
+
   private void cleanup() throws IOException {
     docid = 0;
   }
-  
+
   public void testNull() {}
-  
+
   @Override
   public void setUp() {
     Logger.getRootLogger().setLevel(Level.ERROR);
   }
-  
+
   private static final int NUM_ROWS = 10;
   private static final int NUM_DOCIDS = 1000;
-  
+
   public void test1() throws IOException {
     columnFamilies = new Text[2];
     columnFamilies[0] = new Text("C");
@@ -149,7 +149,7 @@ public class IntersectingIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("B");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
     IteratorSetting is = new IteratorSetting(1, IntersectingIterator.class);
@@ -167,7 +167,7 @@ public class IntersectingIteratorTest extends TestCase {
     assertTrue(hitCount == docs.size());
     cleanup();
   }
-  
+
   public void test2() throws IOException {
     columnFamilies = new Text[3];
     columnFamilies[0] = new Text("A");
@@ -178,7 +178,7 @@ public class IntersectingIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("C");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
     IteratorSetting is = new IteratorSetting(1, IntersectingIterator.class);
@@ -196,7 +196,7 @@ public class IntersectingIteratorTest extends TestCase {
     assertTrue(hitCount == docs.size());
     cleanup();
   }
-  
+
   public void test3() throws IOException {
     columnFamilies = new Text[6];
     columnFamilies[0] = new Text("C");
@@ -210,7 +210,7 @@ public class IntersectingIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("B");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
     SortedKeyValueIterator<Key,Value> source2 = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
@@ -233,7 +233,7 @@ public class IntersectingIteratorTest extends TestCase {
     assertTrue(hitCount == docs.size());
     cleanup();
   }
-  
+
   public void test4() throws IOException {
     columnFamilies = new Text[3];
     notFlags = new boolean[3];
@@ -251,7 +251,7 @@ public class IntersectingIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("C");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs, negatedColumns);
     IteratorSetting is = new IteratorSetting(1, IntersectingIterator.class);
@@ -269,7 +269,7 @@ public class IntersectingIteratorTest extends TestCase {
     assertTrue(hitCount == docs.size());
     cleanup();
   }
-  
+
   public void test6() throws IOException {
     columnFamilies = new Text[1];
     columnFamilies[0] = new Text("C");
@@ -278,7 +278,7 @@ public class IntersectingIteratorTest extends TestCase {
     otherColumnFamilies[1] = new Text("B");
     otherColumnFamilies[2] = new Text("D");
     otherColumnFamilies[3] = new Text("F");
-    
+
     float hitRatio = 0.5f;
     SortedKeyValueIterator<Key,Value> source = createIteratorStack(hitRatio, NUM_ROWS, NUM_DOCIDS, columnFamilies, otherColumnFamilies, docs);
     IteratorSetting is = new IteratorSetting(1, IntersectingIterator.class);
@@ -296,7 +296,7 @@ public class IntersectingIteratorTest extends TestCase {
     assertTrue(hitCount == docs.size());
     cleanup();
   }
-  
+
   public void testWithBatchScanner() throws Exception {
     Value empty = new Value(new byte[] {});
     MockInstance inst = new MockInstance("mockabye");
@@ -308,7 +308,7 @@ public class IntersectingIteratorTest extends TestCase {
     m.put("15qh", "5000000000000000", empty);
     bw.addMutation(m);
     bw.close();
-    
+
     BatchScanner bs = connector.createBatchScanner("index", Authorizations.EMPTY, 10);
     IteratorSetting ii = new IteratorSetting(20, IntersectingIterator.class);
     IntersectingIterator.setColumnFamilies(ii, new Text[] {new Text("rvy"), new Text("15qh")});

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/LargeRowFilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/LargeRowFilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/LargeRowFilterTest.java
index bb59c84..af610ca 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/LargeRowFilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/LargeRowFilterTest.java
@@ -35,27 +35,27 @@ import org.apache.accumulo.core.iterators.system.ColumnFamilySkippingIterator;
 import org.apache.accumulo.core.util.LocalityGroupUtil;
 
 public class LargeRowFilterTest extends TestCase {
-  
+
   private String genRow(int r) {
     return String.format("row%03d", r);
   }
-  
+
   private String genCQ(int cq) {
     return String.format("cf%03d", cq);
   }
-  
+
   private void genRow(TreeMap<Key,Value> testData, int row, int startCQ, int stopCQ) {
     for (int cq = startCQ; cq < stopCQ; cq++) {
       testData.put(new Key(genRow(row), "cf001", genCQ(cq), 5), new Value(("v" + row + "_" + cq).getBytes()));
     }
   }
-  
+
   private void genTestData(TreeMap<Key,Value> testData, int numRows) {
     for (int i = 1; i <= numRows; i++) {
       genRow(testData, i, 0, i);
     }
   }
-  
+
   private LargeRowFilter setupIterator(TreeMap<Key,Value> testData, int maxColumns, IteratorScope scope) throws IOException {
     SortedMapIterator smi = new SortedMapIterator(testData);
     LargeRowFilter lrfi = new LargeRowFilter();
@@ -64,133 +64,133 @@ public class LargeRowFilterTest extends TestCase {
     lrfi.init(new ColumnFamilySkippingIterator(smi), is.getOptions(), new RowDeletingIteratorTest.TestIE(scope, false));
     return lrfi;
   }
-  
+
   public void testBasic() throws Exception {
     TreeMap<Key,Value> testData = new TreeMap<Key,Value>();
-    
+
     genTestData(testData, 20);
-    
+
     for (int i = 1; i <= 20; i++) {
       TreeMap<Key,Value> expectedData = new TreeMap<Key,Value>();
       genTestData(expectedData, i);
-      
+
       LargeRowFilter lrfi = setupIterator(testData, i, IteratorScope.scan);
       lrfi.seek(new Range(), LocalityGroupUtil.EMPTY_CF_SET, false);
-      
+
       TreeMap<Key,Value> filteredData = new TreeMap<Key,Value>();
-      
+
       while (lrfi.hasTop()) {
         filteredData.put(lrfi.getTopKey(), lrfi.getTopValue());
         lrfi.next();
       }
-      
+
       assertEquals(expectedData, filteredData);
     }
   }
-  
+
   public void testSeek() throws Exception {
     TreeMap<Key,Value> testData = new TreeMap<Key,Value>();
-    
+
     genTestData(testData, 20);
-    
+
     for (int i = 1; i <= 20; i++) {
       TreeMap<Key,Value> expectedData = new TreeMap<Key,Value>();
       genTestData(expectedData, i);
-      
+
       LargeRowFilter lrfi = setupIterator(testData, i, IteratorScope.scan);
-      
+
       TreeMap<Key,Value> filteredData = new TreeMap<Key,Value>();
-      
+
       // seek to each row... rows that exceed max columns should be filtered
       for (int j = 1; j <= i; j++) {
         lrfi.seek(new Range(genRow(j), genRow(j)), LocalityGroupUtil.EMPTY_CF_SET, false);
-        
+
         while (lrfi.hasTop()) {
           assertEquals(genRow(j), lrfi.getTopKey().getRow().toString());
           filteredData.put(lrfi.getTopKey(), lrfi.getTopValue());
           lrfi.next();
         }
       }
-      
+
       assertEquals(expectedData, filteredData);
     }
   }
-  
+
   public void testSeek2() throws Exception {
     TreeMap<Key,Value> testData = new TreeMap<Key,Value>();
-    
+
     genTestData(testData, 20);
-    
+
     LargeRowFilter lrfi = setupIterator(testData, 13, IteratorScope.scan);
-    
+
     // test seeking to the middle of a row
     lrfi.seek(new Range(new Key(genRow(15), "cf001", genCQ(4), 5), true, new Key(genRow(15)).followingKey(PartialKey.ROW), false),
         LocalityGroupUtil.EMPTY_CF_SET, false);
     assertFalse(lrfi.hasTop());
-    
+
     lrfi.seek(new Range(new Key(genRow(10), "cf001", genCQ(4), 5), true, new Key(genRow(10)).followingKey(PartialKey.ROW), false),
         LocalityGroupUtil.EMPTY_CF_SET, false);
     TreeMap<Key,Value> expectedData = new TreeMap<Key,Value>();
     genRow(expectedData, 10, 4, 10);
-    
+
     TreeMap<Key,Value> filteredData = new TreeMap<Key,Value>();
     while (lrfi.hasTop()) {
       filteredData.put(lrfi.getTopKey(), lrfi.getTopValue());
       lrfi.next();
     }
-    
+
     assertEquals(expectedData, filteredData);
   }
-  
+
   public void testCompaction() throws Exception {
     TreeMap<Key,Value> testData = new TreeMap<Key,Value>();
-    
+
     genTestData(testData, 20);
-    
+
     LargeRowFilter lrfi = setupIterator(testData, 13, IteratorScope.majc);
     lrfi.seek(new Range(), LocalityGroupUtil.EMPTY_CF_SET, false);
-    
+
     TreeMap<Key,Value> compactedData = new TreeMap<Key,Value>();
     while (lrfi.hasTop()) {
       compactedData.put(lrfi.getTopKey(), lrfi.getTopValue());
       lrfi.next();
     }
-    
+
     // compacted data should now contain suppression markers
     // add column to row that should be suppressed\
     genRow(compactedData, 15, 15, 16);
-    
+
     // scanning over data w/ higher max columns should not change behavior
     // because there are suppression markers.. if there was a bug and data
     // was not suppressed, increasing the threshold would expose the bug
     lrfi = setupIterator(compactedData, 20, IteratorScope.scan);
     lrfi.seek(new Range(), LocalityGroupUtil.EMPTY_CF_SET, false);
-    
+
     // only expect to see 13 rows
     TreeMap<Key,Value> expectedData = new TreeMap<Key,Value>();
     genTestData(expectedData, 13);
-    
+
     TreeMap<Key,Value> filteredData = new TreeMap<Key,Value>();
     while (lrfi.hasTop()) {
       filteredData.put(lrfi.getTopKey(), lrfi.getTopValue());
       lrfi.next();
     }
-    
+
     assertEquals(expectedData.size() + 8, compactedData.size());
     assertEquals(expectedData, filteredData);
-    
+
     // try seeking to the middle of row 15... row has data and suppression marker... this seeks past the marker but before the column
     lrfi.seek(new Range(new Key(genRow(15), "cf001", genCQ(4), 5), true, new Key(genRow(15)).followingKey(PartialKey.ROW), false),
         LocalityGroupUtil.EMPTY_CF_SET, false);
     assertFalse(lrfi.hasTop());
-    
+
     // test seeking w/ column families
     HashSet<ByteSequence> colfams = new HashSet<ByteSequence>();
     colfams.add(new ArrayByteSequence("cf001"));
     lrfi.seek(new Range(new Key(genRow(15), "cf001", genCQ(4), 5), true, new Key(genRow(15)).followingKey(PartialKey.ROW), false), colfams, true);
     assertFalse(lrfi.hasTop());
   }
-  
+
   // in other test data is generated in such a way that once a row
   // is suppressed, all subsequent rows are suppressed
   public void testSuppressInner() throws Exception {
@@ -199,21 +199,21 @@ public class LargeRowFilterTest extends TestCase {
     genRow(testData, 2, 0, 50);
     genRow(testData, 3, 0, 15);
     genRow(testData, 4, 0, 5);
-    
+
     TreeMap<Key,Value> expectedData = new TreeMap<Key,Value>();
     genRow(expectedData, 1, 0, 2);
     genRow(expectedData, 4, 0, 5);
-    
+
     LargeRowFilter lrfi = setupIterator(testData, 13, IteratorScope.scan);
     lrfi.seek(new Range(), LocalityGroupUtil.EMPTY_CF_SET, false);
-    
+
     TreeMap<Key,Value> filteredData = new TreeMap<Key,Value>();
     while (lrfi.hasTop()) {
       filteredData.put(lrfi.getTopKey(), lrfi.getTopValue());
       lrfi.next();
     }
-    
+
     assertEquals(expectedData, filteredData);
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/RowDeletingIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/RowDeletingIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/RowDeletingIteratorTest.java
index 4521e55..d2dd6da 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/RowDeletingIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/RowDeletingIteratorTest.java
@@ -37,101 +37,101 @@ import org.apache.accumulo.core.iterators.system.ColumnFamilySkippingIterator;
 import org.apache.hadoop.io.Text;
 
 public class RowDeletingIteratorTest extends TestCase {
-  
+
   public static class TestIE implements IteratorEnvironment {
-    
+
     private IteratorScope scope;
     private boolean fmc;
-    
+
     public TestIE(IteratorScope scope, boolean fmc) {
       this.scope = scope;
       this.fmc = fmc;
     }
-    
+
     @Override
     public AccumuloConfiguration getConfig() {
       return null;
     }
-    
+
     @Override
     public IteratorScope getIteratorScope() {
       return scope;
     }
-    
+
     @Override
     public boolean isFullMajorCompaction() {
       return fmc;
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
       return null;
     }
-    
+
     @Override
     public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {}
-    
+
   }
-  
+
   Key nk(String row, String cf, String cq, long time) {
     return new Key(new Text(row), new Text(cf), new Text(cq), time);
   }
-  
+
   void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, Value val) {
     tm.put(nk(row, cf, cq, time), val);
   }
-  
+
   void put(TreeMap<Key,Value> tm, String row, String cf, String cq, long time, String val) {
     put(tm, row, cf, cq, time, new Value(val.getBytes()));
   }
-  
+
   private void ane(RowDeletingIterator rdi, String row, String cf, String cq, long time, String val) {
     assertTrue(rdi.hasTop());
     assertEquals(nk(row, cf, cq, time), rdi.getTopKey());
     assertEquals(val, rdi.getTopValue().toString());
   }
-  
+
   public void test1() throws Exception {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE);
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq3", 5, "v1");
     put(tm1, "r2", "cf1", "cq1", 5, "v1");
-    
+
     RowDeletingIterator rdi = new RowDeletingIterator();
     rdi.init(new SortedMapIterator(tm1), null, new TestIE(IteratorScope.scan, false));
-    
+
     rdi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     for (int i = 0; i < 5; i++) {
       rdi.seek(new Range(nk("r1", "cf1", "cq" + i, 5), null), new ArrayList<ByteSequence>(), false);
       ane(rdi, "r2", "cf1", "cq1", 5, "v1");
     }
-    
+
     rdi.seek(new Range(nk("r11", "cf1", "cq1", 5), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     put(tm1, "r2", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE);
     rdi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     assertFalse(rdi.hasTop());
-    
+
     for (int i = 0; i < 5; i++) {
       rdi.seek(new Range(nk("r1", "cf1", "cq" + i, 5), null), new ArrayList<ByteSequence>(), false);
       assertFalse(rdi.hasTop());
     }
-    
+
     put(tm1, "r0", "cf1", "cq1", 5, "v1");
     rdi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r0", "cf1", "cq1", 5, "v1");
     rdi.next();
     assertFalse(rdi.hasTop());
-    
+
   }
-  
+
   public void test2() throws Exception {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE);
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
@@ -139,99 +139,99 @@ public class RowDeletingIteratorTest extends TestCase {
     put(tm1, "r1", "cf1", "cq4", 5, "v1");
     put(tm1, "r1", "cf1", "cq5", 15, "v1");
     put(tm1, "r2", "cf1", "cq1", 5, "v1");
-    
+
     RowDeletingIterator rdi = new RowDeletingIterator();
     rdi.init(new SortedMapIterator(tm1), null, new TestIE(IteratorScope.scan, false));
-    
+
     rdi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "cf1", "cq3", 15, "v1");
     rdi.next();
     ane(rdi, "r1", "cf1", "cq5", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     rdi.seek(new Range(nk("r1", "cf1", "cq1", 5), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "cf1", "cq3", 15, "v1");
     rdi.next();
     ane(rdi, "r1", "cf1", "cq5", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     rdi.seek(new Range(nk("r1", "cf1", "cq4", 5), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "cf1", "cq5", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     rdi.seek(new Range(nk("r1", "cf1", "cq5", 20), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "cf1", "cq5", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     rdi.seek(new Range(nk("r1", "cf1", "cq9", 20), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
   }
-  
+
   public void test3() throws Exception {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE);
     put(tm1, "r1", "", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r2", "", "cq1", 5, "v1");
     put(tm1, "r2", "cf1", "cq1", 5, "v1");
-    
+
     RowDeletingIterator rdi = new RowDeletingIterator();
     rdi.init(new ColumnFamilySkippingIterator(new SortedMapIterator(tm1)), null, new TestIE(IteratorScope.scan, false));
-    
+
     HashSet<ByteSequence> cols = new HashSet<ByteSequence>();
     cols.add(new ArrayByteSequence("cf1".getBytes()));
-    
+
     rdi.seek(new Range(), cols, true);
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     cols.clear();
     cols.add(new ArrayByteSequence("".getBytes()));
     rdi.seek(new Range(), cols, false);
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     cols.clear();
     rdi.seek(new Range(), cols, false);
     ane(rdi, "r2", "", "cq1", 5, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
   }
-  
+
   public void test4() throws Exception {
-    
+
     TreeMap<Key,Value> tm1 = new TreeMap<Key,Value>();
     put(tm1, "r1", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE);
     put(tm1, "r1", "cf1", "cq1", 5, "v1");
     put(tm1, "r1", "cf1", "cq3", 15, "v1");
     put(tm1, "r1", "cf1", "cq4", 5, "v1");
     put(tm1, "r2", "cf1", "cq1", 5, "v1");
-    
+
     RowDeletingIterator rdi = new RowDeletingIterator();
     rdi.init(new SortedMapIterator(tm1), null, new TestIE(IteratorScope.minc, false));
-    
+
     rdi.seek(new Range(), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE.toString());
     rdi.next();
     ane(rdi, "r1", "cf1", "cq3", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     rdi.seek(new Range(nk("r1", "cf1", "cq3", 20), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "cf1", "cq3", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
     rdi.seek(new Range(nk("r1", "", "", 42), null), new ArrayList<ByteSequence>(), false);
     ane(rdi, "r1", "", "", 10, RowDeletingIterator.DELETE_ROW_VALUE.toString());
     rdi.next();
     ane(rdi, "r1", "cf1", "cq3", 15, "v1");
     rdi.next();
     ane(rdi, "r2", "cf1", "cq1", 5, "v1");
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/RowFilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/RowFilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/RowFilterTest.java
index 958ac18..461f609 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/RowFilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/RowFilterTest.java
@@ -50,7 +50,7 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 
 public class RowFilterTest {
@@ -67,7 +67,7 @@ public class RowFilterTest {
       if (rowIterator.hasTop()) {
         firstKey = new Key(rowIterator.getTopKey());
       }
-      
+
       while (rowIterator.hasTop()) {
         sum += Integer.parseInt(rowIterator.getTopValue().toString());
         rowIterator.next();
@@ -198,7 +198,7 @@ public class RowFilterTest {
     }
     IteratorSetting is = new IteratorSetting(40, SummingRowFilter.class);
     conn.tableOperations().attachIterator("table1", is);
-    
+
     Scanner scanner = conn.createScanner("table1", Authorizations.EMPTY);
     assertEquals(new HashSet<String>(Arrays.asList("2", "3")), getRows(scanner));
 
@@ -226,7 +226,7 @@ public class RowFilterTest {
     scanner.fetchColumn(new Text("cf1"), new Text("cq2"));
     scanner.fetchColumn(new Text("cf1"), new Text("cq4"));
     assertEquals(new HashSet<String>(Arrays.asList("4")), getRows(scanner));
-    
+
   }
 
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/VersioningIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/VersioningIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/VersioningIteratorTest.java
index 04190d1..730b4cb 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/VersioningIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/VersioningIteratorTest.java
@@ -42,7 +42,7 @@ public class VersioningIteratorTest {
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
   private static final Encoder<Long> encoder = LongCombiner.FIXED_LEN_ENCODER;
   private static final Logger log = Logger.getLogger(VersioningIteratorTest.class);
-  
+
   void createTestData(TreeMap<Key,Value> tm, Text colf, Text colq) {
     for (int i = 0; i < 2; i++) {
       for (long j = 0; j < 20; j++) {
@@ -50,38 +50,38 @@ public class VersioningIteratorTest {
         tm.put(k, new Value(encoder.encode(j)));
       }
     }
-    
+
     assertTrue("Initial size was " + tm.size(), tm.size() == 40);
   }
-  
+
   TreeMap<Key,Value> iteratorOverTestData(VersioningIterator it) throws IOException {
     TreeMap<Key,Value> tmOut = new TreeMap<Key,Value>();
     while (it.hasTop()) {
       tmOut.put(it.getTopKey(), it.getTopValue());
       it.next();
     }
-    
+
     return tmOut;
   }
-  
+
   @Test
   public void test1() {
     Text colf = new Text("a");
     Text colq = new Text("b");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     createTestData(tm, colf, colq);
-    
+
     try {
       VersioningIterator it = new VersioningIterator();
       IteratorSetting is = new IteratorSetting(1, VersioningIterator.class);
       VersioningIterator.setMaxVersions(is, 3);
       it.init(new SortedMapIterator(tm), is.getOptions(), null);
       it.seek(new Range(), EMPTY_COL_FAMS, false);
-      
+
       TreeMap<Key,Value> tmOut = iteratorOverTestData(it);
-      
+
       for (Entry<Key,Value> e : tmOut.entrySet()) {
         assertTrue(e.getValue().get().length == 8);
         assertTrue(16 < encoder.decode(e.getValue().get()));
@@ -94,30 +94,30 @@ public class VersioningIteratorTest {
       assertFalse(true);
     }
   }
-  
+
   @Test
   public void test2() {
     Text colf = new Text("a");
     Text colq = new Text("b");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     createTestData(tm, colf, colq);
-    
+
     try {
       VersioningIterator it = new VersioningIterator();
       IteratorSetting is = new IteratorSetting(1, VersioningIterator.class);
       VersioningIterator.setMaxVersions(is, 3);
       it.init(new SortedMapIterator(tm), is.getOptions(), null);
-      
+
       // after doing this seek, should only get two keys for row 1
       // since we are seeking to the middle of the most recent
       // three keys
       Key seekKey = new Key(new Text(String.format("%03d", 1)), colf, colq, 18);
       it.seek(new Range(seekKey, null), EMPTY_COL_FAMS, false);
-      
+
       TreeMap<Key,Value> tmOut = iteratorOverTestData(it);
-      
+
       for (Entry<Key,Value> e : tmOut.entrySet()) {
         assertTrue(e.getValue().get().length == 8);
         assertTrue(16 < encoder.decode(e.getValue().get()));
@@ -130,48 +130,48 @@ public class VersioningIteratorTest {
       assertFalse(true);
     }
   }
-  
+
   @Test
   public void test3() {
     Text colf = new Text("a");
     Text colq = new Text("b");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     createTestData(tm, colf, colq);
-    
+
     try {
       VersioningIterator it = new VersioningIterator();
       IteratorSetting is = new IteratorSetting(1, VersioningIterator.class);
       VersioningIterator.setMaxVersions(is, 3);
       it.init(new SortedMapIterator(tm), is.getOptions(), null);
-      
+
       // after doing this seek, should get zero keys for row 1
       Key seekKey = new Key(new Text(String.format("%03d", 1)), colf, colq, 15);
       it.seek(new Range(seekKey, null), EMPTY_COL_FAMS, false);
-      
+
       TreeMap<Key,Value> tmOut = iteratorOverTestData(it);
-      
+
       for (Entry<Key,Value> e : tmOut.entrySet()) {
         assertTrue(e.getValue().get().length == 8);
         assertTrue(16 < encoder.decode(e.getValue().get()));
       }
-      
+
       assertTrue("size after seeking past versions was " + tmOut.size(), tmOut.size() == 0);
-      
+
       // after doing this seek, should get zero keys for row 0 and 3 keys for row 1
       seekKey = new Key(new Text(String.format("%03d", 0)), colf, colq, 15);
       it.seek(new Range(seekKey, null), EMPTY_COL_FAMS, false);
-      
+
       tmOut = iteratorOverTestData(it);
-      
+
       for (Entry<Key,Value> e : tmOut.entrySet()) {
         assertTrue(e.getValue().get().length == 8);
         assertTrue(16 < encoder.decode(e.getValue().get()));
       }
-      
+
       assertTrue("size after seeking past versions was " + tmOut.size(), tmOut.size() == 3);
-      
+
     } catch (IOException e) {
       assertFalse(true);
     } catch (Exception e) {
@@ -179,16 +179,16 @@ public class VersioningIteratorTest {
       assertFalse(true);
     }
   }
-  
+
   @Test
   public void test4() {
     Text colf = new Text("a");
     Text colq = new Text("b");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     createTestData(tm, colf, colq);
-    
+
     for (int i = 1; i <= 30; i++) {
       try {
         VersioningIterator it = new VersioningIterator();
@@ -196,9 +196,9 @@ public class VersioningIteratorTest {
         VersioningIterator.setMaxVersions(is, i);
         it.init(new SortedMapIterator(tm), is.getOptions(), null);
         it.seek(new Range(), EMPTY_COL_FAMS, false);
-        
+
         TreeMap<Key,Value> tmOut = iteratorOverTestData(it);
-        
+
         assertTrue("size after keeping " + i + " versions was " + tmOut.size(), tmOut.size() == Math.min(40, 2 * i));
       } catch (IOException e) {
         assertFalse(true);
@@ -208,51 +208,51 @@ public class VersioningIteratorTest {
       }
     }
   }
-  
+
   @Test
   public void test5() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     createTestData(tm, colf, colq);
-    
+
     VersioningIterator it = new VersioningIterator();
     IteratorSetting is = new IteratorSetting(1, VersioningIterator.class);
     VersioningIterator.setMaxVersions(is, 3);
     it.init(new SortedMapIterator(tm), is.getOptions(), null);
-    
+
     Key seekKey = new Key(new Text(String.format("%03d", 1)), colf, colq, 19);
     it.seek(new Range(seekKey, false, null, true), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(it.hasTop());
     assertTrue(it.getTopKey().getTimestamp() == 18);
-    
+
   }
-  
+
   @Test
   public void test6() throws IOException {
     Text colf = new Text("a");
     Text colq = new Text("b");
-    
+
     TreeMap<Key,Value> tm = new TreeMap<Key,Value>();
-    
+
     createTestData(tm, colf, colq);
-    
+
     VersioningIterator it = new VersioningIterator();
     IteratorSetting is = new IteratorSetting(1, VersioningIterator.class);
     VersioningIterator.setMaxVersions(is, 3);
     it.init(new SortedMapIterator(tm), is.getOptions(), null);
     VersioningIterator it2 = it.deepCopy(null);
-    
+
     Key seekKey = new Key(new Text(String.format("%03d", 1)), colf, colq, 19);
     it.seek(new Range(seekKey, false, null, true), EMPTY_COL_FAMS, false);
     it2.seek(new Range(seekKey, false, null, true), EMPTY_COL_FAMS, false);
-    
+
     assertTrue(it.hasTop());
     assertTrue(it.getTopKey().getTimestamp() == 18);
-    
+
     assertTrue(it2.hasTop());
     assertTrue(it2.getTopKey().getTimestamp() == 18);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/VisibilityFilterTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/VisibilityFilterTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/VisibilityFilterTest.java
index 1804f6f..810c355 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/VisibilityFilterTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/VisibilityFilterTest.java
@@ -38,9 +38,9 @@ import org.apache.hadoop.io.Text;
 import org.junit.Test;
 
 public class VisibilityFilterTest {
-  
+
   private static final Collection<ByteSequence> EMPTY_COL_FAMS = new ArrayList<ByteSequence>();
-  
+
   private static final Text BAD = new Text("bad");
   private static final Text GOOD = new Text("good");
   private static final Text EMPTY_VIS = new Text("");
@@ -48,7 +48,7 @@ public class VisibilityFilterTest {
   private static final Text HIDDEN_VIS = new Text("abc&def&ghi");
   private static final Text BAD_VIS = new Text("&");
   private static final Value EMPTY_VALUE = new Value(new byte[0]);
-  
+
   private TreeMap<Key,Value> createUnprotectedSource(int numPublic, int numHidden) {
     TreeMap<Key,Value> source = new TreeMap<Key,Value>();
     for (int i = 0; i < numPublic; i++)
@@ -57,7 +57,7 @@ public class VisibilityFilterTest {
       source.put(new Key(new Text(String.format("%03d", i)), BAD, BAD, GOOD_VIS), EMPTY_VALUE);
     return source;
   }
-  
+
   private TreeMap<Key,Value> createPollutedSource(int numGood, int numBad) {
     TreeMap<Key,Value> source = new TreeMap<Key,Value>();
     for (int i = 0; i < numGood; i++)
@@ -66,7 +66,7 @@ public class VisibilityFilterTest {
       source.put(new Key(new Text(String.format("%03d", i)), BAD, BAD, BAD_VIS), EMPTY_VALUE);
     return source;
   }
-  
+
   private TreeMap<Key,Value> createSourceWithHiddenData(int numViewable, int numHidden) {
     TreeMap<Key,Value> source = new TreeMap<Key,Value>();
     for (int i = 0; i < numViewable; i++)
@@ -75,15 +75,15 @@ public class VisibilityFilterTest {
       source.put(new Key(new Text(String.format("%03d", i)), BAD, BAD, HIDDEN_VIS), EMPTY_VALUE);
     return source;
   }
-  
+
   private void verify(TreeMap<Key,Value> source, int expectedSourceSize, Map<String,String> options, Text expectedCF, Text expectedCQ, Text expectedCV,
       int expectedFinalCount) throws IOException {
     assertEquals(expectedSourceSize, source.size());
-    
+
     Filter filter = new VisibilityFilter();
     filter.init(new SortedMapIterator(source), options, null);
     filter.seek(new Range(), EMPTY_COL_FAMS, false);
-    
+
     int count = 0;
     while (filter.hasTop()) {
       count++;
@@ -96,82 +96,82 @@ public class VisibilityFilterTest {
     }
     assertEquals(expectedFinalCount, count);
   }
-  
+
   @Test
   public void testAllowValidLabelsOnly() throws IOException {
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.filterInvalidLabelsOnly(is, true);
-    
+
     TreeMap<Key,Value> source = createPollutedSource(1, 2);
     verify(source, 3, is.getOptions(), GOOD, GOOD, GOOD_VIS, 1);
-    
+
     source = createPollutedSource(30, 500);
     verify(source, 530, is.getOptions(), GOOD, GOOD, GOOD_VIS, 30);
-    
+
     source = createPollutedSource(1000, 500);
     verify(source, 1500, is.getOptions(), GOOD, GOOD, GOOD_VIS, 1000);
   }
-  
+
   @Test
   public void testAllowBadLabelsOnly() throws IOException {
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.setNegate(is, true);
     VisibilityFilter.filterInvalidLabelsOnly(is, true);
-    
+
     TreeMap<Key,Value> source = createPollutedSource(1, 2);
     verify(source, 3, is.getOptions(), BAD, BAD, BAD_VIS, 2);
-    
+
     source = createPollutedSource(30, 500);
     verify(source, 530, is.getOptions(), BAD, BAD, BAD_VIS, 500);
-    
+
     source = createPollutedSource(1000, 500);
     verify(source, 1500, is.getOptions(), BAD, BAD, BAD_VIS, 500);
   }
-  
+
   @Test
   public void testAllowAuthorizedLabelsOnly() throws IOException {
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.setAuthorizations(is, new Authorizations("def"));
-    
+
     TreeMap<Key,Value> source = createSourceWithHiddenData(1, 2);
     verify(source, 3, is.getOptions(), GOOD, GOOD, GOOD_VIS, 1);
-    
+
     source = createSourceWithHiddenData(30, 500);
     verify(source, 530, is.getOptions(), GOOD, GOOD, GOOD_VIS, 30);
-    
+
     source = createSourceWithHiddenData(1000, 500);
     verify(source, 1500, is.getOptions(), GOOD, GOOD, GOOD_VIS, 1000);
   }
-  
+
   @Test
   public void testAllowUnauthorizedLabelsOnly() throws IOException {
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.setNegate(is, true);
     VisibilityFilter.setAuthorizations(is, new Authorizations("def"));
-    
+
     TreeMap<Key,Value> source = createSourceWithHiddenData(1, 2);
     verify(source, 3, is.getOptions(), BAD, BAD, HIDDEN_VIS, 2);
-    
+
     source = createSourceWithHiddenData(30, 500);
     verify(source, 530, is.getOptions(), BAD, BAD, HIDDEN_VIS, 500);
-    
+
     source = createSourceWithHiddenData(1000, 500);
     verify(source, 1500, is.getOptions(), BAD, BAD, HIDDEN_VIS, 500);
   }
-  
+
   @Test
   public void testNoLabels() throws IOException {
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.setNegate(is, false);
     VisibilityFilter.setAuthorizations(is, new Authorizations());
-    
+
     TreeMap<Key,Value> source = createUnprotectedSource(5, 2);
     verify(source, 7, is.getOptions(), GOOD, GOOD, EMPTY_VIS, 5);
-    
+
     VisibilityFilter.setNegate(is, true);
     verify(source, 7, is.getOptions(), BAD, BAD, GOOD_VIS, 2);
   }
-  
+
   @Test
   public void testFilterUnauthorizedAndBad() throws IOException {
     /*
@@ -179,53 +179,53 @@ public class VisibilityFilterTest {
      */
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.setAuthorizations(is, new Authorizations("def"));
-    
+
     TreeMap<Key,Value> source = createSourceWithHiddenData(1, 5);
     for (Entry<Key,Value> entry : createPollutedSource(0, 1).entrySet())
       source.put(entry.getKey(), entry.getValue());
-    
+
     verify(source, 7, is.getOptions(), GOOD, GOOD, GOOD_VIS, 1);
   }
-  
+
   @Test
   public void testCommaSeparatedAuthorizations() throws IOException {
     Map<String,String> options = Collections.singletonMap("auths", "x,def,y");
-    
+
     TreeMap<Key,Value> source = createSourceWithHiddenData(1, 2);
     verify(source, 3, options, GOOD, GOOD, GOOD_VIS, 1);
-    
+
     source = createSourceWithHiddenData(30, 500);
     verify(source, 530, options, GOOD, GOOD, GOOD_VIS, 30);
-    
+
     source = createSourceWithHiddenData(1000, 500);
     verify(source, 1500, options, GOOD, GOOD, GOOD_VIS, 1000);
   }
-  
+
   @Test
   public void testSerializedAuthorizations() throws IOException {
     Map<String,String> options = Collections.singletonMap("auths", new Authorizations("x", "def", "y").serialize());
-    
+
     TreeMap<Key,Value> source = createSourceWithHiddenData(1, 2);
     verify(source, 3, options, GOOD, GOOD, GOOD_VIS, 1);
-    
+
     source = createSourceWithHiddenData(30, 500);
     verify(source, 530, options, GOOD, GOOD, GOOD_VIS, 30);
-    
+
     source = createSourceWithHiddenData(1000, 500);
     verify(source, 1500, options, GOOD, GOOD, GOOD_VIS, 1000);
   }
-  
+
   @Test
   public void testStaticConfigurators() {
     IteratorSetting is = new IteratorSetting(1, VisibilityFilter.class);
     VisibilityFilter.filterInvalidLabelsOnly(is, false);
     VisibilityFilter.setNegate(is, true);
     VisibilityFilter.setAuthorizations(is, new Authorizations("abc", "def"));
-    
+
     Map<String,String> opts = is.getOptions();
     assertEquals("false", opts.get("filterInvalid"));
     assertEquals("true", opts.get("negate"));
     assertEquals(new Authorizations("abc", "def").serialize(), opts.get("auths"));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIteratorTest.java
index dddab41..f5440cd 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeColumnFamilyIteratorTest.java
@@ -54,7 +54,7 @@ import org.apache.hadoop.io.Text;
 import org.junit.Assert;
 
 public class WholeColumnFamilyIteratorTest extends TestCase {
-  
+
   public void testEmptyStuff() throws IOException {
     SortedMap<Key,Value> map = new TreeMap<Key,Value>();
     SortedMap<Key,Value> map2 = new TreeMap<Key,Value>();
@@ -98,12 +98,12 @@ public class WholeColumnFamilyIteratorTest extends TestCase {
       resultMap.putAll(WholeColumnFamilyIterator.decodeColumnFamily(rowKey, rowValue));
       iter.next();
     }
-    
+
     // we have 7 groups of row key/cf
     Assert.assertEquals(7, numRows);
-    
+
     assertEquals(resultMap, map);
-    
+
     WholeColumnFamilyIterator iter2 = new WholeColumnFamilyIterator(source) {
       @Override
       public boolean filter(Text row, List<Key> keys, List<Value> values) {
@@ -123,81 +123,81 @@ public class WholeColumnFamilyIteratorTest extends TestCase {
     assertTrue(numRows == trueCount);
     assertEquals(resultMap, map2);
   }
-  
+
   private void pkv(SortedMap<Key,Value> map, String row, String cf, String cq, String cv, long ts, String val) {
     map.put(new Key(new Text(row), new Text(cf), new Text(cq), new Text(cv), ts), new Value(val.getBytes()));
   }
-  
+
   public void testContinue() throws Exception {
     SortedMap<Key,Value> map1 = new TreeMap<Key,Value>();
     pkv(map1, "row1", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map1, "row1", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map2 = new TreeMap<Key,Value>();
     pkv(map2, "row2", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map2, "row2", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map3 = new TreeMap<Key,Value>();
     pkv(map3, "row3", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map3, "row3", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map = new TreeMap<Key,Value>();
     map.putAll(map1);
     map.putAll(map2);
     map.putAll(map3);
-    
+
     SortedMapIterator source = new SortedMapIterator(map);
     WholeColumnFamilyIterator iter = new WholeColumnFamilyIterator(source);
-    
+
     Range range = new Range(new Text("row1"), true, new Text("row2"), true);
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertTrue(iter.hasTop());
     assertEquals(map1, WholeColumnFamilyIterator.decodeColumnFamily(iter.getTopKey(), iter.getTopValue()));
-    
+
     // simulate something continuing using the last key from the iterator
     // this is what client and server code will do
     range = new Range(iter.getTopKey(), false, range.getEndKey(), range.isEndKeyInclusive());
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertTrue(iter.hasTop());
     assertEquals(map2, WholeColumnFamilyIterator.decodeColumnFamily(iter.getTopKey(), iter.getTopValue()));
-    
+
     iter.next();
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
-  
+
   public void testBug1() throws Exception {
     SortedMap<Key,Value> map1 = new TreeMap<Key,Value>();
     pkv(map1, "row1", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map1, "row1", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map2 = new TreeMap<Key,Value>();
     pkv(map2, "row2", "cf1", "cq1", "cv1", 5, "foo");
-    
+
     SortedMap<Key,Value> map = new TreeMap<Key,Value>();
     map.putAll(map1);
     map.putAll(map2);
-    
+
     MultiIterator source = new MultiIterator(Collections.singletonList((SortedKeyValueIterator<Key,Value>) new SortedMapIterator(map)), new Range(null, true,
         new Text("row1"), true));
     WholeColumnFamilyIterator iter = new WholeColumnFamilyIterator(source);
-    
+
     Range range = new Range(new Text("row1"), true, new Text("row2"), true);
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertTrue(iter.hasTop());
     assertEquals(map1, WholeColumnFamilyIterator.decodeColumnFamily(iter.getTopKey(), iter.getTopValue()));
-    
+
     // simulate something continuing using the last key from the iterator
     // this is what client and server code will do
     range = new Range(iter.getTopKey(), false, range.getEndKey(), range.isEndKeyInclusive());
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeRowIteratorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeRowIteratorTest.java b/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeRowIteratorTest.java
index 8ad97a9..b47ef3e 100644
--- a/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeRowIteratorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/iterators/user/WholeRowIteratorTest.java
@@ -16,7 +16,9 @@
  */
 package org.apache.accumulo.core.iterators.user;
 
-import static org.junit.Assert.*;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -42,7 +44,7 @@ import com.google.common.collect.ImmutableList;
 
 public class WholeRowIteratorTest {
 
-  @Test(expected=IOException.class)
+  @Test(expected = IOException.class)
   public void testBadDecodeRow() throws IOException {
     Key k = new Key(new Text("r1"), new Text("cf1234567890"));
     Value v = new Value("v1".getBytes());
@@ -97,7 +99,7 @@ public class WholeRowIteratorTest {
     }
     assertTrue(numRows == 5);
     assertEquals(resultMap, map);
-    
+
     WholeRowIterator iter2 = new WholeRowIterator(source) {
       @Override
       public boolean filter(Text row, List<Key> keys, List<Value> values) {
@@ -117,83 +119,83 @@ public class WholeRowIteratorTest {
     assertTrue(numRows == trueCount);
     assertEquals(resultMap, map2);
   }
-  
+
   private void pkv(SortedMap<Key,Value> map, String row, String cf, String cq, String cv, long ts, String val) {
     map.put(new Key(new Text(row), new Text(cf), new Text(cq), new Text(cv), ts), new Value(val.getBytes()));
   }
-  
+
   @Test
   public void testContinue() throws Exception {
     SortedMap<Key,Value> map1 = new TreeMap<Key,Value>();
     pkv(map1, "row1", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map1, "row1", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map2 = new TreeMap<Key,Value>();
     pkv(map2, "row2", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map2, "row2", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map3 = new TreeMap<Key,Value>();
     pkv(map3, "row3", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map3, "row3", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map = new TreeMap<Key,Value>();
     map.putAll(map1);
     map.putAll(map2);
     map.putAll(map3);
-    
+
     SortedMapIterator source = new SortedMapIterator(map);
     WholeRowIterator iter = new WholeRowIterator(source);
-    
+
     Range range = new Range(new Text("row1"), true, new Text("row2"), true);
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertTrue(iter.hasTop());
     assertEquals(map1, WholeRowIterator.decodeRow(iter.getTopKey(), iter.getTopValue()));
-    
+
     // simulate something continuing using the last key from the iterator
     // this is what client and server code will do
     range = new Range(iter.getTopKey(), false, range.getEndKey(), range.isEndKeyInclusive());
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertTrue(iter.hasTop());
     assertEquals(map2, WholeRowIterator.decodeRow(iter.getTopKey(), iter.getTopValue()));
-    
+
     iter.next();
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
-  
+
   @Test
   public void testBug1() throws Exception {
     SortedMap<Key,Value> map1 = new TreeMap<Key,Value>();
     pkv(map1, "row1", "cf1", "cq1", "cv1", 5, "foo");
     pkv(map1, "row1", "cf1", "cq2", "cv1", 6, "bar");
-    
+
     SortedMap<Key,Value> map2 = new TreeMap<Key,Value>();
     pkv(map2, "row2", "cf1", "cq1", "cv1", 5, "foo");
-    
+
     SortedMap<Key,Value> map = new TreeMap<Key,Value>();
     map.putAll(map1);
     map.putAll(map2);
-    
+
     MultiIterator source = new MultiIterator(Collections.singletonList((SortedKeyValueIterator<Key,Value>) new SortedMapIterator(map)), new Range(null, true,
         new Text("row1"), true));
     WholeRowIterator iter = new WholeRowIterator(source);
-    
+
     Range range = new Range(new Text("row1"), true, new Text("row2"), true);
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertTrue(iter.hasTop());
     assertEquals(map1, WholeRowIterator.decodeRow(iter.getTopKey(), iter.getTopValue()));
-    
+
     // simulate something continuing using the last key from the iterator
     // this is what client and server code will do
     range = new Range(iter.getTopKey(), false, range.getEndKey(), range.isEndKeyInclusive());
     iter.seek(range, new ArrayList<ByteSequence>(), false);
-    
+
     assertFalse(iter.hasTop());
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/metadata/MetadataServicerTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/metadata/MetadataServicerTest.java b/core/src/test/java/org/apache/accumulo/core/metadata/MetadataServicerTest.java
index 7501c31..b8ef346 100644
--- a/core/src/test/java/org/apache/accumulo/core/metadata/MetadataServicerTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/metadata/MetadataServicerTest.java
@@ -34,13 +34,13 @@ import org.apache.accumulo.core.security.Credentials;
 import org.junit.Test;
 
 public class MetadataServicerTest {
-  
+
   @Test
   public void checkSystemTableIdentifiers() {
     assertNotEquals(RootTable.ID, MetadataTable.ID);
     assertNotEquals(RootTable.NAME, MetadataTable.NAME);
   }
-  
+
   @Test
   public void testGetCorrectServicer() throws AccumuloException, AccumuloSecurityException, TableExistsException, TableNotFoundException {
     String userTableName = "A";
@@ -50,35 +50,35 @@ public class MetadataServicerTest {
     String userTableId = connector.tableOperations().tableIdMap().get(userTableName);
     Credentials credentials = new Credentials("root", new PasswordToken(""));
     ClientContext context = new ClientContext(instance, credentials, new ClientConfiguration());
-    
+
     MetadataServicer ms = MetadataServicer.forTableId(context, RootTable.ID);
     assertTrue(ms instanceof ServicerForRootTable);
     assertFalse(ms instanceof TableMetadataServicer);
     assertEquals(RootTable.ID, ms.getServicedTableId());
-    
+
     ms = MetadataServicer.forTableId(context, MetadataTable.ID);
     assertTrue(ms instanceof ServicerForMetadataTable);
     assertTrue(ms instanceof TableMetadataServicer);
     assertEquals(RootTable.NAME, ((TableMetadataServicer) ms).getServicingTableName());
     assertEquals(MetadataTable.ID, ms.getServicedTableId());
-    
+
     ms = MetadataServicer.forTableId(context, userTableId);
     assertTrue(ms instanceof ServicerForUserTables);
     assertTrue(ms instanceof TableMetadataServicer);
     assertEquals(MetadataTable.NAME, ((TableMetadataServicer) ms).getServicingTableName());
     assertEquals(userTableId, ms.getServicedTableId());
-    
+
     ms = MetadataServicer.forTableName(context, RootTable.NAME);
     assertTrue(ms instanceof ServicerForRootTable);
     assertFalse(ms instanceof TableMetadataServicer);
     assertEquals(RootTable.ID, ms.getServicedTableId());
-    
+
     ms = MetadataServicer.forTableName(context, MetadataTable.NAME);
     assertTrue(ms instanceof ServicerForMetadataTable);
     assertTrue(ms instanceof TableMetadataServicer);
     assertEquals(RootTable.NAME, ((TableMetadataServicer) ms).getServicingTableName());
     assertEquals(MetadataTable.ID, ms.getServicedTableId());
-    
+
     ms = MetadataServicer.forTableName(context, userTableName);
     assertTrue(ms instanceof ServicerForUserTables);
     assertTrue(ms instanceof TableMetadataServicer);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtilTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtilTest.java b/core/src/test/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtilTest.java
index cb40de9..b2b91d7 100644
--- a/core/src/test/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtilTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/replication/ReplicationConfigurationUtilTest.java
@@ -31,7 +31,7 @@ import org.junit.Before;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ReplicationConfigurationUtilTest {
 
@@ -59,13 +59,15 @@ public class ReplicationConfigurationUtilTest {
   @Test
   public void rootTableExtentEmptyConf() {
     KeyExtent extent = new KeyExtent(new Text(RootTable.ID), null, null);
-    Assert.assertFalse("The root table should never be replicated", ReplicationConfigurationUtil.isEnabled(extent, new ConfigurationCopy(new HashMap<String,String>())));
+    Assert.assertFalse("The root table should never be replicated",
+        ReplicationConfigurationUtil.isEnabled(extent, new ConfigurationCopy(new HashMap<String,String>())));
   }
 
   @Test
   public void metadataTableExtentEmptyConf() {
     KeyExtent extent = new KeyExtent(new Text(MetadataTable.ID), null, null);
-    Assert.assertFalse("The metadata table should never be replicated", ReplicationConfigurationUtil.isEnabled(extent, new ConfigurationCopy(new HashMap<String,String>())));
+    Assert.assertFalse("The metadata table should never be replicated",
+        ReplicationConfigurationUtil.isEnabled(extent, new ConfigurationCopy(new HashMap<String,String>())));
   }
 
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/replication/ReplicationTargetTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/replication/ReplicationTargetTest.java b/core/src/test/java/org/apache/accumulo/core/replication/ReplicationTargetTest.java
index 6b754bf..88e6304 100644
--- a/core/src/test/java/org/apache/accumulo/core/replication/ReplicationTargetTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/replication/ReplicationTargetTest.java
@@ -23,7 +23,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class ReplicationTargetTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/replication/StatusUtilTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/replication/StatusUtilTest.java b/core/src/test/java/org/apache/accumulo/core/replication/StatusUtilTest.java
index 5a35d65..024e05f 100644
--- a/core/src/test/java/org/apache/accumulo/core/replication/StatusUtilTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/replication/StatusUtilTest.java
@@ -21,7 +21,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class StatusUtilTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/replication/proto/StatusTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/replication/proto/StatusTest.java b/core/src/test/java/org/apache/accumulo/core/replication/proto/StatusTest.java
index 8906fb4..d59a04f 100644
--- a/core/src/test/java/org/apache/accumulo/core/replication/proto/StatusTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/replication/proto/StatusTest.java
@@ -21,7 +21,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class StatusTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/security/AuthenticationTokenTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/security/AuthenticationTokenTest.java b/core/src/test/java/org/apache/accumulo/core/security/AuthenticationTokenTest.java
index 7cd8c42..f2253e4 100644
--- a/core/src/test/java/org/apache/accumulo/core/security/AuthenticationTokenTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/security/AuthenticationTokenTest.java
@@ -31,7 +31,7 @@ import org.apache.accumulo.core.client.security.tokens.PasswordToken;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class AuthenticationTokenTest {
   @Test
@@ -43,14 +43,14 @@ public class AuthenticationTokenTest {
     for (byte b : randomBytes)
       allZero = allZero && b == 0;
     assertFalse(allZero);
-    
+
     byte[] serialized = AuthenticationTokenSerializer.serialize(new PasswordToken(randomBytes));
     PasswordToken passwordToken = AuthenticationTokenSerializer.deserialize(PasswordToken.class, serialized);
     assertArrayEquals(randomBytes, passwordToken.getPassword());
-    
+
     serialized = AuthenticationTokenSerializer.serialize(new NullToken());
     AuthenticationToken nullToken = AuthenticationTokenSerializer.deserialize(NullToken.class, serialized);
     assertEquals(new NullToken(), nullToken);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/security/AuthorizationsTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/security/AuthorizationsTest.java b/core/src/test/java/org/apache/accumulo/core/security/AuthorizationsTest.java
index 855d778..4fb2139 100644
--- a/core/src/test/java/org/apache/accumulo/core/security/AuthorizationsTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/security/AuthorizationsTest.java
@@ -26,75 +26,75 @@ import org.apache.accumulo.core.util.ByteArraySet;
 import org.junit.Test;
 
 public class AuthorizationsTest {
-  
+
   @Test
   public void testSetOfByteArrays() {
     assertTrue(ByteArraySet.fromStrings("a", "b", "c").contains("a".getBytes()));
   }
-  
+
   @Test
   public void testEncodeDecode() {
     Authorizations a = new Authorizations("a", "abcdefg", "hijklmno", ",");
     byte[] array = a.getAuthorizationsArray();
     Authorizations b = new Authorizations(array);
     assertEquals(a, b);
-    
+
     // test encoding empty auths
     a = new Authorizations();
     array = a.getAuthorizationsArray();
     b = new Authorizations(array);
     assertEquals(a, b);
-    
+
     // test encoding multi-byte auths
     a = new Authorizations("五", "b", "c", "九");
     array = a.getAuthorizationsArray();
     b = new Authorizations(array);
     assertEquals(a, b);
   }
-  
+
   @Test
   public void testSerialization() {
     Authorizations a1 = new Authorizations("a", "b");
     Authorizations a2 = new Authorizations("b", "a");
-    
+
     assertEquals(a1, a2);
     assertEquals(a1.serialize(), a2.serialize());
   }
-  
+
   @Test
   public void testDefensiveAccess() {
     Authorizations expected = new Authorizations("foo", "a");
     Authorizations actual = new Authorizations("foo", "a");
-    
+
     // foo to goo; test defensive iterator
     for (byte[] bytes : actual) {
       bytes[0]++;
     }
     assertArrayEquals(expected.getAuthorizationsArray(), actual.getAuthorizationsArray());
-    
+
     // test defensive getter and serializer
     actual.getAuthorizations().get(0)[0]++;
     assertArrayEquals(expected.getAuthorizationsArray(), actual.getAuthorizationsArray());
     assertEquals(expected.serialize(), actual.serialize());
   }
-  
+
   // This should throw ReadOnlyBufferException, but THRIFT-883 requires that the ByteBuffers themselves not be read-only
   // @Test(expected = ReadOnlyBufferException.class)
   @Test
   public void testReadOnlyByteBuffer() {
     Authorizations expected = new Authorizations("foo");
     Authorizations actual = new Authorizations("foo");
-    
+
     assertArrayEquals(expected.getAuthorizationsArray(), actual.getAuthorizationsArray());
     actual.getAuthorizationsBB().get(0).array()[0]++;
     assertArrayEquals(expected.getAuthorizationsArray(), actual.getAuthorizationsArray());
   }
-  
+
   @Test(expected = UnsupportedOperationException.class)
   public void testUnmodifiableList() {
     Authorizations expected = new Authorizations("foo");
     Authorizations actual = new Authorizations("foo");
-    
+
     assertArrayEquals(expected.getAuthorizationsArray(), actual.getAuthorizationsArray());
     actual.getAuthorizationsBB().add(ByteBuffer.wrap(new byte[] {'a'}));
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/security/CredentialsTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/security/CredentialsTest.java b/core/src/test/java/org/apache/accumulo/core/security/CredentialsTest.java
index c5746bc..a399310 100644
--- a/core/src/test/java/org/apache/accumulo/core/security/CredentialsTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/security/CredentialsTest.java
@@ -38,10 +38,10 @@ import org.apache.accumulo.core.security.thrift.TCredentials;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class CredentialsTest {
-  
+
   @Test
   public void testToThrift() throws DestroyFailedException {
     // verify thrift serialization
@@ -50,7 +50,7 @@ public class CredentialsTest {
     assertEquals("test", tCreds.getPrincipal());
     assertEquals(PasswordToken.class.getName(), tCreds.getTokenClassName());
     assertArrayEquals(AuthenticationTokenSerializer.serialize(new PasswordToken("testing")), tCreds.getToken());
-    
+
     // verify that we can't serialize if it's destroyed
     creds.getToken().destroy();
     try {
@@ -70,14 +70,14 @@ public class CredentialsTest {
     Credentials roundtrip = Credentials.fromThrift(tCreds);
     assertEquals("Roundtrip through thirft changed credentials equality", creds, roundtrip);
   }
-  
+
   @Test
   public void testMockConnector() throws AccumuloException, DestroyFailedException, AccumuloSecurityException {
     Instance inst = new MockInstance();
     Connector rootConnector = inst.getConnector("root", new PasswordToken());
     PasswordToken testToken = new PasswordToken("testPass");
     rootConnector.securityOperations().createLocalUser("testUser", testToken);
-    
+
     assertFalse(testToken.isDestroyed());
     testToken.destroy();
     assertTrue(testToken.isDestroyed());
@@ -88,20 +88,20 @@ public class CredentialsTest {
       assertTrue(e.getSecurityErrorCode().equals(SecurityErrorCode.TOKEN_EXPIRED));
     }
   }
-  
+
   @Test
   public void testEqualsAndHashCode() {
     Credentials nullNullCreds = new Credentials(null, null);
     Credentials abcNullCreds = new Credentials("abc", new NullToken());
     Credentials cbaNullCreds = new Credentials("cba", new NullToken());
     Credentials abcBlahCreds = new Credentials("abc", new PasswordToken("blah"));
-    
+
     // check hash codes
     assertEquals(0, nullNullCreds.hashCode());
     assertEquals("abc".hashCode(), abcNullCreds.hashCode());
     assertEquals(abcNullCreds.hashCode(), abcBlahCreds.hashCode());
     assertFalse(abcNullCreds.hashCode() == cbaNullCreds.hashCode());
-    
+
     // identity
     assertEquals(abcNullCreds, abcNullCreds);
     assertEquals(new Credentials("abc", new NullToken()), abcNullCreds);
@@ -112,7 +112,7 @@ public class CredentialsTest {
     assertFalse(nullNullCreds.equals(abcNullCreds));
     assertFalse(abcNullCreds.equals(abcBlahCreds));
   }
-  
+
   @Test
   public void testCredentialsSerialization() throws AccumuloSecurityException {
     Credentials creds = new Credentials("a:b-c", new PasswordToken("d-e-f".getBytes(UTF_8)));
@@ -121,14 +121,14 @@ public class CredentialsTest {
     assertEquals(creds, result);
     assertEquals("a:b-c", result.getPrincipal());
     assertEquals(new PasswordToken("d-e-f"), result.getToken());
-    
+
     Credentials nullNullCreds = new Credentials(null, null);
     serialized = nullNullCreds.serialize();
     result = Credentials.deserialize(serialized);
     assertEquals(null, result.getPrincipal());
     assertEquals(null, result.getToken());
   }
-  
+
   @Test
   public void testToString() {
     Credentials creds = new Credentials(null, null);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/security/VisibilityEvaluatorTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/security/VisibilityEvaluatorTest.java b/core/src/test/java/org/apache/accumulo/core/security/VisibilityEvaluatorTest.java
index ee4d2ee..2996970 100644
--- a/core/src/test/java/org/apache/accumulo/core/security/VisibilityEvaluatorTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/security/VisibilityEvaluatorTest.java
@@ -27,34 +27,34 @@ import org.apache.accumulo.core.util.ByteArraySet;
 import org.junit.Test;
 
 public class VisibilityEvaluatorTest {
-  
+
   @Test
   public void testVisibilityEvaluator() throws VisibilityParseException {
     VisibilityEvaluator ct = new VisibilityEvaluator(new Authorizations(ByteArraySet.fromStrings("one", "two", "three", "four")));
-    
+
     // test for empty vis
     assertTrue(ct.evaluate(new ColumnVisibility(new byte[0])));
-    
+
     // test for and
     assertTrue("'and' test", ct.evaluate(new ColumnVisibility("one&two")));
-    
+
     // test for or
     assertTrue("'or' test", ct.evaluate(new ColumnVisibility("foor|four")));
-    
+
     // test for and and or
     assertTrue("'and' and 'or' test", ct.evaluate(new ColumnVisibility("(one&two)|(foo&bar)")));
-    
+
     // test for false negatives
     for (String marking : new String[] {"one", "one|five", "five|one", "(one)", "(one&two)|(foo&bar)", "(one|foo)&three", "one|foo|bar", "(one|foo)|bar",
         "((one|foo)|bar)&two"}) {
       assertTrue(marking, ct.evaluate(new ColumnVisibility(marking)));
     }
-    
+
     // test for false positives
     for (String marking : new String[] {"five", "one&five", "five&one", "((one|foo)|bar)&goober"}) {
       assertFalse(marking, ct.evaluate(new ColumnVisibility(marking)));
     }
-    
+
     // test missing separators; these should throw an exception
     for (String marking : new String[] {"one(five)", "(five)one", "(one)(two)", "a|(b(c))"}) {
       try {
@@ -64,7 +64,7 @@ public class VisibilityEvaluatorTest {
         // all is good
       }
     }
-    
+
     // test unexpected separator
     for (String marking : new String[] {"&(five)", "|(five)", "(five)&", "five|", "a|(b)&", "(&five)", "(five|)"}) {
       try {
@@ -74,7 +74,7 @@ public class VisibilityEvaluatorTest {
         // all is good
       }
     }
-    
+
     // test mismatched parentheses
     for (String marking : new String[] {"(", ")", "(a&b", "b|a)"}) {
       try {
@@ -85,23 +85,23 @@ public class VisibilityEvaluatorTest {
       }
     }
   }
-  
+
   @Test
   public void testQuotedExpressions() throws VisibilityParseException {
     VisibilityEvaluator ct = new VisibilityEvaluator(new Authorizations("A#C", "A\"C", "A\\C", "AC"));
-    
+
     assertTrue(ct.evaluate(new ColumnVisibility(quote("A#C") + "|" + quote("A?C"))));
     assertTrue(ct.evaluate(new ColumnVisibility(new ColumnVisibility(quote("A#C") + "|" + quote("A?C")).flatten())));
     assertTrue(ct.evaluate(new ColumnVisibility(quote("A\"C") + "&" + quote("A\\C"))));
     assertTrue(ct.evaluate(new ColumnVisibility(new ColumnVisibility(quote("A\"C") + "&" + quote("A\\C")).flatten())));
     assertTrue(ct.evaluate(new ColumnVisibility("(" + quote("A\"C") + "|B)&(" + quote("A#C") + "|D)")));
-    
+
     assertFalse(ct.evaluate(new ColumnVisibility(quote("A#C") + "&B")));
-    
+
     assertTrue(ct.evaluate(new ColumnVisibility(quote("A#C"))));
     assertTrue(ct.evaluate(new ColumnVisibility("(" + quote("A#C") + ")")));
   }
-  
+
   @Test
   public void testQuote() {
     assertEquals("\"A#C\"", quote("A#C"));
@@ -111,11 +111,11 @@ public class VisibilityEvaluatorTest {
     assertEquals("\"九\"", quote("九"));
     assertEquals("\"五十\"", quote("五十"));
   }
-  
+
   @Test
   public void testNonAscii() throws VisibilityParseException {
     VisibilityEvaluator ct = new VisibilityEvaluator(new Authorizations("五", "六", "八", "九", "五十"));
-    
+
     assertTrue(ct.evaluate(new ColumnVisibility(quote("五") + "|" + quote("四"))));
     assertFalse(ct.evaluate(new ColumnVisibility(quote("五") + "&" + quote("四"))));
     assertTrue(ct.evaluate(new ColumnVisibility(quote("五") + "&(" + quote("四") + "|" + quote("九") + ")")));

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/security/crypto/CryptoTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/security/crypto/CryptoTest.java b/core/src/test/java/org/apache/accumulo/core/security/crypto/CryptoTest.java
index fe16c0e..5588029 100644
--- a/core/src/test/java/org/apache/accumulo/core/security/crypto/CryptoTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/security/crypto/CryptoTest.java
@@ -53,47 +53,47 @@ import org.junit.rules.ExpectedException;
 import com.google.common.primitives.Bytes;
 
 public class CryptoTest {
-  
+
   private static final int MARKER_INT = 0xCADEFEDD;
   private static final String MARKER_STRING = "1 2 3 a b c";
   public static final String CONFIG_FILE_SYSTEM_PROP = "org.apache.accumulo.config.file";
   public static final String CRYPTO_ON_CONF = "crypto-on-accumulo-site.xml";
   public static final String CRYPTO_OFF_CONF = "crypto-off-accumulo-site.xml";
-  public static final String CRYPTO_ON_KEK_OFF_CONF = "crypto-on-no-key-encryption-accumulo-site.xml"; 
-  
+  public static final String CRYPTO_ON_KEK_OFF_CONF = "crypto-on-no-key-encryption-accumulo-site.xml";
+
   @Rule
   public ExpectedException exception = ExpectedException.none();
-  
+
   @Test
   public void testNoCryptoStream() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_OFF_CONF);    
-    
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_OFF_CONF);
+
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
-    
+
     assertNotNull(params);
     assertEquals("NullCipher", params.getAlgorithmName());
     assertNull(params.getEncryptionMode());
     assertNull(params.getPadding());
-    
+
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
     assertNotNull(cryptoModule);
     assertTrue(cryptoModule instanceof CryptoModuleFactory.NullCryptoModule);
-    
+
     ByteArrayOutputStream out = new ByteArrayOutputStream();
-    
+
     params.setPlaintextOutputStream(out);
-    
+
     params = cryptoModule.getEncryptingOutputStream(params);
     assertNotNull(params.getEncryptedOutputStream());
     assertEquals(out, params.getEncryptedOutputStream());
   }
-  
+
   @Test
   public void testCryptoModuleParamsParsing() {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
 
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
-    
+
     assertNotNull(params);
     assertEquals("AES", params.getAlgorithmName());
     assertEquals("CFB", params.getEncryptionMode());
@@ -103,10 +103,10 @@ public class CryptoTest {
     assertEquals("SUN", params.getRandomNumberGeneratorProvider());
     assertEquals("org.apache.accumulo.core.security.crypto.CachingHDFSSecretKeyEncryptionStrategy", params.getKeyEncryptionStrategyClass());
   }
-  
+
   @Test
   public void testCryptoModuleDoesntLeakSensitive() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
 
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
@@ -116,20 +116,21 @@ public class CryptoTest {
 
     cryptoModule.getEncryptingOutputStream(params);
     params.getEncryptedOutputStream().close();
-    
+
     // If we get here, we have encrypted bytes
     byte[] streamBytes = baos.toByteArray();
     for (Property prop : Property.values()) {
       if (prop.isSensitive()) {
         byte[] toCheck = prop.getKey().getBytes();
-        assertEquals(-1, Bytes.indexOf(streamBytes, toCheck));  }
-    }    
+        assertEquals(-1, Bytes.indexOf(streamBytes, toCheck));
+      }
+    }
 
   }
-  
+
   @Test
   public void testCryptoModuleParamsValidation1() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
 
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
@@ -142,8 +143,8 @@ public class CryptoTest {
 
   @Test
   public void testCryptoModuleParamsValidation2() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
-   
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
+
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
 
@@ -152,123 +153,122 @@ public class CryptoTest {
     exception.expect(RuntimeException.class);
     cryptoModule.getDecryptingInputStream(params);
   }
-  
+
   private String getStringifiedBytes(String s) throws IOException {
     ByteArrayOutputStream out = new ByteArrayOutputStream();
     DataOutputStream dataOut = new DataOutputStream(out);
-    
+
     dataOut.writeUTF(s);
     dataOut.close();
     byte[] stringMarkerBytes = out.toByteArray();
     return Arrays.toString(stringMarkerBytes);
-    
+
   }
-  
+
   private String getStringifiedBytes(int i) throws IOException {
     ByteArrayOutputStream out = new ByteArrayOutputStream();
     DataOutputStream dataOut = new DataOutputStream(out);
-    
+
     dataOut.writeInt(i);
     dataOut.close();
     byte[] stringMarkerBytes = out.toByteArray();
     return Arrays.toString(stringMarkerBytes);
-    
+
   }
 
   @Test
   public void testCryptoModuleBasicReadWrite() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_KEK_OFF_CONF);    
-  
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_KEK_OFF_CONF);
+
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
-    
+
     assertTrue(cryptoModule instanceof DefaultCryptoModule);
-    
+
     byte[] resultingBytes = setUpSampleEncryptedBytes(cryptoModule, params);
-    
+
     // If we get here, we have encrypted bytes
     ByteArrayInputStream in = new ByteArrayInputStream(resultingBytes);
-    
+
     params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     params.setEncryptedInputStream(in);
-    
+
     params = cryptoModule.getDecryptingInputStream(params);
-    
+
     InputStream plaintextIn = params.getPlaintextInputStream();
-    
+
     assertNotNull(plaintextIn);
     assertTrue(plaintextIn != in);
     DataInputStream dataIn = new DataInputStream(plaintextIn);
     String markerString = dataIn.readUTF();
     int markerInt = dataIn.readInt();
-    
+
     assertEquals(MARKER_STRING, markerString);
     assertEquals(MARKER_INT, markerInt);
   }
 
   private byte[] setUpSampleEncryptedBytes(CryptoModule cryptoModule, CryptoModuleParameters params) throws IOException {
     ByteArrayOutputStream out = new ByteArrayOutputStream();
-    
+
     params.setPlaintextOutputStream(new NoFlushOutputStream(out));
-    
+
     params = cryptoModule.getEncryptingOutputStream(params);
-    
+
     assertNotNull(params.getEncryptedOutputStream());
     assertTrue(params.getEncryptedOutputStream() != out);
-    
+
     DataOutputStream dataOut = new DataOutputStream(params.getEncryptedOutputStream());
     dataOut.writeUTF(MARKER_STRING);
     dataOut.writeInt(MARKER_INT);
     dataOut.close();
-    
+
     byte[] resultingBytes = out.toByteArray();
     String stringifiedBytes = Arrays.toString(resultingBytes);
-    
+
     String stringifiedMarkerBytes = getStringifiedBytes(MARKER_STRING);
     String stringifiedOtherBytes = getStringifiedBytes(MARKER_INT);
-    
-    
+
     // OK, let's make sure it's encrypted
     assertTrue(!stringifiedBytes.contains(stringifiedMarkerBytes));
     assertTrue(!stringifiedBytes.contains(stringifiedOtherBytes));
     return resultingBytes;
   }
-  
+
   @Test
   public void testKeyEncryptionAndCheckThatFileCannotBeReadWithoutKEK() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
-  
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
+
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
 
     assertTrue(cryptoModule instanceof DefaultCryptoModule);
     assertNotNull(params.getKeyEncryptionStrategyClass());
     assertEquals("org.apache.accumulo.core.security.crypto.CachingHDFSSecretKeyEncryptionStrategy", params.getKeyEncryptionStrategyClass());
-    
+
     byte[] resultingBytes = setUpSampleEncryptedBytes(cryptoModule, params);
 
     // So now that we have bytes encrypted by a key encrypted to a KEK, turn off the KEK configuration and try
-    // to decrypt.  We expect this to fail.  This also tests our ability to override the key encryption strategy.
+    // to decrypt. We expect this to fail. This also tests our ability to override the key encryption strategy.
     conf = setAndGetAccumuloConfig(CRYPTO_ON_KEK_OFF_CONF);
     params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     params.setOverrideStreamsSecretKeyEncryptionStrategy(true);
-    
+
     ByteArrayInputStream in = new ByteArrayInputStream(resultingBytes);
     params.setEncryptedInputStream(in);
-    
+
     params = cryptoModule.getDecryptingInputStream(params);
-    
+
     assertNotNull(params.getPlaintextInputStream());
     DataInputStream dataIn = new DataInputStream(params.getPlaintextInputStream());
     // We expect the following operation to fail and throw an exception
     exception.expect(IOException.class);
     @SuppressWarnings("unused")
     String markerString = dataIn.readUTF();
- }
+  }
 
   @Test
   public void testKeyEncryptionNormalPath() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
 
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
@@ -276,30 +276,30 @@ public class CryptoTest {
     assertTrue(cryptoModule instanceof DefaultCryptoModule);
     assertNotNull(params.getKeyEncryptionStrategyClass());
     assertEquals("org.apache.accumulo.core.security.crypto.CachingHDFSSecretKeyEncryptionStrategy", params.getKeyEncryptionStrategyClass());
-    
+
     byte[] resultingBytes = setUpSampleEncryptedBytes(cryptoModule, params);
 
     params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     params.setOverrideStreamsSecretKeyEncryptionStrategy(true);
-    
+
     ByteArrayInputStream in = new ByteArrayInputStream(resultingBytes);
     params.setEncryptedInputStream(in);
-    
+
     params = cryptoModule.getDecryptingInputStream(params);
-    
+
     assertNotNull(params.getPlaintextInputStream());
     DataInputStream dataIn = new DataInputStream(params.getPlaintextInputStream());
 
     String markerString = dataIn.readUTF();
     int markerInt = dataIn.readInt();
-    
+
     assertEquals(MARKER_STRING, markerString);
     assertEquals(MARKER_INT, markerInt);
   }
-  
+
   @Test
   public void testChangingCryptoParamsAndCanStillDecryptPreviouslyEncryptedFiles() throws IOException {
-    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);    
+    AccumuloConfiguration conf = setAndGetAccumuloConfig(CRYPTO_ON_CONF);
 
     CryptoModule cryptoModule = CryptoModuleFactory.getCryptoModule(conf);
     CryptoModuleParameters params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
@@ -307,30 +307,30 @@ public class CryptoTest {
     assertTrue(cryptoModule instanceof DefaultCryptoModule);
     assertNotNull(params.getKeyEncryptionStrategyClass());
     assertEquals("org.apache.accumulo.core.security.crypto.CachingHDFSSecretKeyEncryptionStrategy", params.getKeyEncryptionStrategyClass());
-    
+
     byte[] resultingBytes = setUpSampleEncryptedBytes(cryptoModule, params);
 
     // Now we're going to create a params object and set its algorithm and key length different
-    // from those configured within the site configuration.  After doing this, we should
+    // from those configured within the site configuration. After doing this, we should
     // still be able to read the file that was created with a different set of parameters.
     params = CryptoModuleFactory.createParamsObjectFromAccumuloConfiguration(conf);
     params.setAlgorithmName("DESede");
     params.setKeyLength(24 * 8);
-    
+
     ByteArrayInputStream in = new ByteArrayInputStream(resultingBytes);
     params.setEncryptedInputStream(in);
-    
+
     params = cryptoModule.getDecryptingInputStream(params);
-    
+
     assertNotNull(params.getPlaintextInputStream());
     DataInputStream dataIn = new DataInputStream(params.getPlaintextInputStream());
     String markerString = dataIn.readUTF();
     int markerInt = dataIn.readInt();
-    
+
     assertEquals(MARKER_STRING, markerString);
     assertEquals(MARKER_INT, markerInt);
   }
-  
+
   private AccumuloConfiguration setAndGetAccumuloConfig(String cryptoConfSetting) {
     ConfigurationCopy result = new ConfigurationCopy(AccumuloConfiguration.getDefaultConfiguration());
     Configuration conf = new Configuration(false);
@@ -340,33 +340,33 @@ public class CryptoTest {
     }
     return result;
   }
-  
+
   @Test
-  public void testKeyWrapAndUnwrap() throws NoSuchAlgorithmException, NoSuchPaddingException, NoSuchProviderException, InvalidKeyException, IllegalBlockSizeException, BadPaddingException {
+  public void testKeyWrapAndUnwrap() throws NoSuchAlgorithmException, NoSuchPaddingException, NoSuchProviderException, InvalidKeyException,
+      IllegalBlockSizeException, BadPaddingException {
     Cipher keyWrapCipher = Cipher.getInstance("AES/ECB/NoPadding");
     SecureRandom random = SecureRandom.getInstance("SHA1PRNG", "SUN");
-    
+
     byte[] kek = new byte[16];
     random.nextBytes(kek);
     byte[] randomKey = new byte[16];
     random.nextBytes(randomKey);
-    
+
     keyWrapCipher.init(Cipher.WRAP_MODE, new SecretKeySpec(kek, "AES"));
-    
+
     Key randKey = new SecretKeySpec(randomKey, "AES");
-    
+
     byte[] wrappedKey = keyWrapCipher.wrap(randKey);
-    
-    assert(wrappedKey != null);
-    assert(wrappedKey.length == randomKey.length);
 
-    
+    assert (wrappedKey != null);
+    assert (wrappedKey.length == randomKey.length);
+
     Cipher keyUnwrapCipher = Cipher.getInstance("AES/ECB/NoPadding");
     keyUnwrapCipher.init(Cipher.UNWRAP_MODE, new SecretKeySpec(kek, "AES"));
     Key unwrappedKey = keyUnwrapCipher.unwrap(wrappedKey, "AES", Cipher.SECRET_KEY);
-    
+
     byte[] unwrappedKeyBytes = unwrappedKey.getEncoded();
-    assert(Arrays.equals(unwrappedKeyBytes, randomKey));
-    
+    assert (Arrays.equals(unwrappedKeyBytes, randomKey));
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/test/java/org/apache/accumulo/core/trace/PerformanceTest.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/accumulo/core/trace/PerformanceTest.java b/core/src/test/java/org/apache/accumulo/core/trace/PerformanceTest.java
index 7fe5233..3d3b39e 100644
--- a/core/src/test/java/org/apache/accumulo/core/trace/PerformanceTest.java
+++ b/core/src/test/java/org/apache/accumulo/core/trace/PerformanceTest.java
@@ -19,14 +19,14 @@ package org.apache.accumulo.core.trace;
 import org.junit.Test;
 
 public class PerformanceTest {
-  
+
   @Test
   public void test() {
-    
+
   }
-  
+
   public static void main(String[] args) {
-    
+
     long now = System.currentTimeMillis();
     for (long i = 0; i < 1000 * 1000; i++) {
       @SuppressWarnings("unused")


[39/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/BloomFilterLayer.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/BloomFilterLayer.java b/core/src/main/java/org/apache/accumulo/core/file/BloomFilterLayer.java
index f78ae66..61e6b5c 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/BloomFilterLayer.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/BloomFilterLayer.java
@@ -64,44 +64,44 @@ import org.slf4j.LoggerFactory;
 
 /**
  * A class that sits on top of different accumulo file formats and provides bloom filter functionality.
- * 
+ *
  */
 public class BloomFilterLayer {
   private static final Logger LOG = LoggerFactory.getLogger(BloomFilterLayer.class);
   public static final String BLOOM_FILE_NAME = "acu_bloom";
   public static final int HASH_COUNT = 5;
-  
+
   private static ExecutorService loadThreadPool = null;
-  
+
   private static synchronized ExecutorService getLoadThreadPool(int maxLoadThreads) {
     if (loadThreadPool != null) {
       return loadThreadPool;
     }
-    
+
     if (maxLoadThreads > 0) {
       BlockingQueue<Runnable> q = new LinkedBlockingQueue<Runnable>();
       loadThreadPool = new ThreadPoolExecutor(0, maxLoadThreads, 60, TimeUnit.SECONDS, q, new NamingThreadFactory("bloom-loader"));
     }
-    
+
     return loadThreadPool;
   }
-  
+
   public static class Writer implements FileSKVWriter {
     private DynamicBloomFilter bloomFilter;
     private int numKeys;
     private int vectorSize;
-    
+
     private FileSKVWriter writer;
     private KeyFunctor transformer = null;
     private boolean closed = false;
-    
+
     Writer(FileSKVWriter writer, AccumuloConfiguration acuconf) {
       this.writer = writer;
       initBloomFilter(acuconf);
     }
-    
+
     private synchronized void initBloomFilter(AccumuloConfiguration acuconf) {
-      
+
       numKeys = acuconf.getCount(Property.TABLE_BLOOM_SIZE);
       // vector size should be <code>-kn / (ln(1 - c^(1/k)))</code> bits for
       // single key, where <code> is the number of hash functions,
@@ -111,7 +111,7 @@ public class BloomFilterLayer {
       double errorRate = acuconf.getFraction(Property.TABLE_BLOOM_ERRORRATE);
       vectorSize = (int) Math.ceil(-HASH_COUNT * numKeys / Math.log(1.0 - Math.pow(errorRate, 1.0 / HASH_COUNT)));
       bloomFilter = new DynamicBloomFilter(vectorSize, HASH_COUNT, Hash.parseHashType(acuconf.get(Property.TABLE_BLOOM_HASHTYPE)), numKeys);
-      
+
       /**
        * load KeyFunctor
        */
@@ -125,15 +125,15 @@ public class BloomFilterLayer {
           clazz = AccumuloVFSClassLoader.loadClass(classname, KeyFunctor.class);
 
         transformer = clazz.newInstance();
-        
+
       } catch (Exception e) {
         LOG.error("Failed to find KeyFunctor: " + acuconf.get(Property.TABLE_BLOOM_KEY_FUNCTOR), e);
         throw new IllegalArgumentException("Failed to find KeyFunctor: " + acuconf.get(Property.TABLE_BLOOM_KEY_FUNCTOR));
-        
+
       }
-      
+
     }
-    
+
     @Override
     public synchronized void append(org.apache.accumulo.core.data.Key key, Value val) throws IOException {
       writer.append(key, val);
@@ -141,13 +141,13 @@ public class BloomFilterLayer {
       if (bloomKey.getBytes().length > 0)
         bloomFilter.add(bloomKey);
     }
-    
+
     @Override
     public synchronized void close() throws IOException {
-      
+
       if (closed)
         return;
-      
+
       DataOutputStream out = writer.createMetaStore(BLOOM_FILE_NAME);
       out.writeUTF(transformer.getClass().getName());
       bloomFilter.write(out);
@@ -156,31 +156,31 @@ public class BloomFilterLayer {
       writer.close();
       closed = true;
     }
-    
+
     @Override
     public DataOutputStream createMetaStore(String name) throws IOException {
       return writer.createMetaStore(name);
     }
-    
+
     @Override
     public void startDefaultLocalityGroup() throws IOException {
       writer.startDefaultLocalityGroup();
-      
+
     }
-    
+
     @Override
     public void startNewLocalityGroup(String name, Set<ByteSequence> columnFamilies) throws IOException {
       writer.startNewLocalityGroup(name, columnFamilies);
     }
-    
+
     @Override
     public boolean supportsLocalityGroups() {
       return writer.supportsLocalityGroups();
     }
   }
-  
+
   static class BloomFilterLoader {
-    
+
     private volatile DynamicBloomFilter bloomFilter;
     private int loadRequest = 0;
     private int loadThreshold = 1;
@@ -188,33 +188,33 @@ public class BloomFilterLayer {
     private Runnable loadTask;
     private volatile KeyFunctor transformer = null;
     private volatile boolean closed = false;
-    
+
     BloomFilterLoader(final FileSKVIterator reader, AccumuloConfiguration acuconf) {
-      
+
       maxLoadThreads = acuconf.getCount(Property.TSERV_BLOOM_LOAD_MAXCONCURRENT);
-      
+
       loadThreshold = acuconf.getCount(Property.TABLE_BLOOM_LOAD_THRESHOLD);
-      
+
       final String context = acuconf.get(Property.TABLE_CLASSPATH);
 
       loadTask = new Runnable() {
         @Override
         public void run() {
-          
+
           // no need to load the bloom filter if the map file is closed
           if (closed)
             return;
           String ClassName = null;
           DataInputStream in = null;
-          
+
           try {
             in = reader.getMetaStore(BLOOM_FILE_NAME);
             DynamicBloomFilter tmpBloomFilter = new DynamicBloomFilter();
-            
+
             // check for closed again after open but before reading the bloom filter in
             if (closed)
               return;
-            
+
             /**
              * Load classname for keyFunctor
              */
@@ -226,11 +226,11 @@ public class BloomFilterLayer {
             else
               clazz = AccumuloVFSClassLoader.loadClass(ClassName, KeyFunctor.class);
             transformer = clazz.newInstance();
-            
+
             /**
              * read in bloom filter
              */
-            
+
             tmpBloomFilter.readFields(in);
             // only set the bloom filter after it is fully constructed
             bloomFilter = tmpBloomFilter;
@@ -241,7 +241,7 @@ public class BloomFilterLayer {
               LOG.warn("Can't open BloomFilter", ioe);
             else
               LOG.debug("Can't open BloomFilter, file closed : " + ioe.getMessage());
-            
+
             bloomFilter = null;
           } catch (ClassNotFoundException e) {
             LOG.error("Failed to find KeyFunctor in config: " + ClassName, e);
@@ -268,11 +268,11 @@ public class BloomFilterLayer {
           }
         }
       };
-      
+
       initiateLoad(maxLoadThreads);
-      
+
     }
-    
+
     private synchronized void initiateLoad(int maxLoadThreads) {
       // ensure only one thread initiates loading of bloom filter by
       // only taking action when loadTask != null
@@ -291,14 +291,14 @@ public class BloomFilterLayer {
           loadTask = null;
         }
       }
-      
+
       loadRequest++;
     }
-    
+
     /**
      * Checks if this {@link RFile} contains keys from this range. The membership test is performed using a Bloom filter, so the result has always non-zero
      * probability of false positives.
-     * 
+     *
      * @param range
      *          range of keys to check
      * @return false iff key doesn't exist, true if key probably exists.
@@ -309,45 +309,45 @@ public class BloomFilterLayer {
         if (bloomFilter == null)
           return true;
       }
-      
+
       Key bloomKey = transformer.transform(range);
-      
+
       if (bloomKey == null || bloomKey.getBytes().length == 0)
         return true;
-      
+
       return bloomFilter.membershipTest(bloomKey);
     }
-    
+
     public void close() {
       this.closed = true;
     }
   }
-  
+
   public static class Reader implements FileSKVIterator {
-    
+
     private BloomFilterLoader bfl;
     private FileSKVIterator reader;
-    
+
     public Reader(FileSKVIterator reader, AccumuloConfiguration acuconf) {
       this.reader = reader;
       bfl = new BloomFilterLoader(reader, acuconf);
     }
-    
+
     private Reader(FileSKVIterator src, BloomFilterLoader bfl) {
       this.reader = src;
       this.bfl = bfl;
     }
-    
+
     private boolean checkSuper = true;
-    
+
     @Override
     public boolean hasTop() {
       return checkSuper ? reader.hasTop() : false;
     }
-    
+
     @Override
     public void seek(Range range, Collection<ByteSequence> columnFamilies, boolean inclusive) throws IOException {
-      
+
       if (!bfl.probablyHasKey(range)) {
         checkSuper = false;
       } else {
@@ -355,118 +355,118 @@ public class BloomFilterLayer {
         checkSuper = true;
       }
     }
-    
+
     @Override
     public synchronized void close() throws IOException {
       bfl.close();
       reader.close();
     }
-    
+
     @Override
     public org.apache.accumulo.core.data.Key getFirstKey() throws IOException {
       return reader.getFirstKey();
     }
-    
+
     @Override
     public org.apache.accumulo.core.data.Key getLastKey() throws IOException {
       return reader.getLastKey();
     }
-    
+
     @Override
     public SortedKeyValueIterator<org.apache.accumulo.core.data.Key,Value> deepCopy(IteratorEnvironment env) {
       return new BloomFilterLayer.Reader((FileSKVIterator) reader.deepCopy(env), bfl);
     }
-    
+
     @Override
     public org.apache.accumulo.core.data.Key getTopKey() {
       return reader.getTopKey();
     }
-    
+
     @Override
     public Value getTopValue() {
       return reader.getTopValue();
     }
-    
+
     @Override
     public void init(SortedKeyValueIterator<org.apache.accumulo.core.data.Key,Value> source, Map<String,String> options, IteratorEnvironment env)
         throws IOException {
       throw new UnsupportedOperationException();
-      
+
     }
-    
+
     @Override
     public void next() throws IOException {
       reader.next();
     }
-    
+
     @Override
     public DataInputStream getMetaStore(String name) throws IOException {
       return reader.getMetaStore(name);
     }
-    
+
     @Override
     public void closeDeepCopies() throws IOException {
       reader.closeDeepCopies();
     }
-    
+
     @Override
     public void setInterruptFlag(AtomicBoolean flag) {
       reader.setInterruptFlag(flag);
     }
-    
+
   }
-  
+
   public static void main(String[] args) throws IOException {
     PrintStream out = System.out;
-    
+
     Random r = new Random();
-    
+
     HashSet<Integer> valsSet = new HashSet<Integer>();
-    
+
     for (int i = 0; i < 100000; i++) {
       valsSet.add(r.nextInt(Integer.MAX_VALUE));
     }
-    
+
     ArrayList<Integer> vals = new ArrayList<Integer>(valsSet);
     Collections.sort(vals);
-    
+
     ConfigurationCopy acuconf = new ConfigurationCopy(AccumuloConfiguration.getDefaultConfiguration());
     acuconf.set(Property.TABLE_BLOOM_ENABLED, "true");
     acuconf.set(Property.TABLE_BLOOM_KEY_FUNCTOR, "accumulo.core.file.keyfunctor.ColumnFamilyFunctor");
     acuconf.set(Property.TABLE_FILE_TYPE, RFile.EXTENSION);
     acuconf.set(Property.TABLE_BLOOM_LOAD_THRESHOLD, "1");
     acuconf.set(Property.TSERV_BLOOM_LOAD_MAXCONCURRENT, "1");
-    
+
     Configuration conf = CachedConfiguration.getInstance();
     FileSystem fs = FileSystem.get(conf);
-    
+
     String suffix = FileOperations.getNewFileExtension(acuconf);
     String fname = "/tmp/test." + suffix;
     FileSKVWriter bmfw = FileOperations.getInstance().openWriter(fname, fs, conf, acuconf);
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     bmfw.startDefaultLocalityGroup();
-    
+
     for (Integer i : vals) {
       String fi = String.format("%010d", i);
       bmfw.append(new org.apache.accumulo.core.data.Key(new Text("r" + fi), new Text("cf1")), new Value(("v" + fi).getBytes(UTF_8)));
       bmfw.append(new org.apache.accumulo.core.data.Key(new Text("r" + fi), new Text("cf2")), new Value(("v" + fi).getBytes(UTF_8)));
     }
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     out.printf("write rate %6.2f%n", vals.size() / ((t2 - t1) / 1000.0));
-    
+
     bmfw.close();
-    
+
     t1 = System.currentTimeMillis();
     FileSKVIterator bmfr = FileOperations.getInstance().openReader(fname, false, fs, conf, acuconf);
     t2 = System.currentTimeMillis();
     out.println("Opened " + fname + " in " + (t2 - t1));
-    
+
     t1 = System.currentTimeMillis();
-    
+
     int hits = 0;
     for (int i = 0; i < 5000; i++) {
       int row = r.nextInt(Integer.MAX_VALUE);
@@ -481,36 +481,36 @@ public class BloomFilterLayer {
         }
       }
     }
-    
+
     t2 = System.currentTimeMillis();
-    
+
     out.printf("random lookup rate : %6.2f%n", 5000 / ((t2 - t1) / 1000.0));
     out.println("hits = " + hits);
-    
+
     int count = 0;
-    
+
     t1 = System.currentTimeMillis();
-    
+
     for (Integer row : valsSet) {
       String fi = String.format("%010d", row);
       // bmfr.seek(new Range(new Text("r"+fi)));
-      
+
       org.apache.accumulo.core.data.Key k1 = new org.apache.accumulo.core.data.Key(new Text("r" + fi), new Text("cf1"));
       bmfr.seek(new Range(k1, true, k1.followingKey(PartialKey.ROW_COLFAM), false), new ArrayList<ByteSequence>(), false);
-      
+
       if (!bmfr.hasTop()) {
         out.println("ERROR 2 " + row);
       }
-      
+
       count++;
-      
+
       if (count >= 500) {
         break;
       }
     }
-    
+
     t2 = System.currentTimeMillis();
-    
+
     out.printf("existant lookup rate %6.2f%n", 500 / ((t2 - t1) / 1000.0));
     out.println("expected hits 500.  Receive hits: " + count);
     bmfr.close();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/FileOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/FileOperations.java b/core/src/main/java/org/apache/accumulo/core/file/FileOperations.java
index 17e540b..78d0407 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/FileOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/FileOperations.java
@@ -35,23 +35,23 @@ import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 
 class DispatchingFileFactory extends FileOperations {
-  
+
   private FileOperations findFileFactory(String file) {
-    
+
     Path p = new Path(file);
     String name = p.getName();
-    
+
     if (name.startsWith(Constants.MAPFILE_EXTENSION + "_")) {
       return new MapFileOperations();
     }
     String[] sp = name.split("\\.");
-    
+
     if (sp.length < 2) {
       throw new IllegalArgumentException("File name " + name + " has no extension");
     }
-    
+
     String extension = sp[sp.length - 1];
-    
+
     if (extension.equals(Constants.MAPFILE_EXTENSION) || extension.equals(Constants.MAPFILE_EXTENSION + "_tmp")) {
       return new MapFileOperations();
     } else if (extension.equals(RFile.EXTENSION) || extension.equals(RFile.EXTENSION + "_tmp")) {
@@ -60,12 +60,12 @@ class DispatchingFileFactory extends FileOperations {
       throw new IllegalArgumentException("File type " + extension + " not supported");
     }
   }
-  
+
   @Override
   public FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return findFileFactory(file).openIndex(file, fs, conf, acuconf, null, null);
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     FileSKVIterator iter = findFileFactory(file).openReader(file, seekToBeginning, fs, conf, acuconf, null, null);
@@ -74,7 +74,7 @@ class DispatchingFileFactory extends FileOperations {
     }
     return iter;
   }
-  
+
   @Override
   public FileSKVWriter openWriter(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     FileSKVWriter writer = findFileFactory(file).openWriter(file, fs, conf, acuconf);
@@ -83,107 +83,107 @@ class DispatchingFileFactory extends FileOperations {
     }
     return writer;
   }
-  
+
   @Override
   public long getFileSize(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException {
     return findFileFactory(file).getFileSize(file, fs, conf, acuconf);
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf) throws IOException {
     return findFileFactory(file).openReader(file, range, columnFamilies, inclusive, fs, conf, tableConf, null, null);
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf, BlockCache dataCache, BlockCache indexCache) throws IOException {
-    
+
     if (!tableConf.getBoolean(Property.TABLE_INDEXCACHE_ENABLED))
       indexCache = null;
     if (!tableConf.getBoolean(Property.TABLE_BLOCKCACHE_ENABLED))
       dataCache = null;
-    
+
     return findFileFactory(file).openReader(file, range, columnFamilies, inclusive, fs, conf, tableConf, dataCache, indexCache);
   }
-  
+
   @Override
   public FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf,
       BlockCache dataCache, BlockCache indexCache) throws IOException {
-    
+
     if (!acuconf.getBoolean(Property.TABLE_INDEXCACHE_ENABLED))
       indexCache = null;
     if (!acuconf.getBoolean(Property.TABLE_BLOCKCACHE_ENABLED))
       dataCache = null;
-    
+
     FileSKVIterator iter = findFileFactory(file).openReader(file, seekToBeginning, fs, conf, acuconf, dataCache, indexCache);
     if (acuconf.getBoolean(Property.TABLE_BLOOM_ENABLED)) {
       return new BloomFilterLayer.Reader(iter, acuconf);
     }
     return iter;
   }
-  
+
   @Override
   public FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf, BlockCache dCache, BlockCache iCache)
       throws IOException {
-    
+
     if (!acuconf.getBoolean(Property.TABLE_INDEXCACHE_ENABLED))
       iCache = null;
     if (!acuconf.getBoolean(Property.TABLE_BLOCKCACHE_ENABLED))
       dCache = null;
-    
+
     return findFileFactory(file).openIndex(file, fs, conf, acuconf, dCache, iCache);
   }
-  
+
 }
 
 public abstract class FileOperations {
-  
+
   private static final HashSet<String> validExtensions = new HashSet<String>(Arrays.asList(Constants.MAPFILE_EXTENSION, RFile.EXTENSION));
-  
+
   public static Set<String> getValidExtensions() {
     return validExtensions;
   }
-  
+
   public static String getNewFileExtension(AccumuloConfiguration acuconf) {
     return acuconf.get(Property.TABLE_FILE_TYPE);
   }
-  
+
   public static FileOperations getInstance() {
     return new DispatchingFileFactory();
   }
-  
+
   /**
    * Open a reader that will not be seeked giving an initial seek location. This is useful for file operations that only need to scan data within a range and do
    * not need to seek. Therefore file metadata such as indexes does not need to be kept in memory while the file is scanned. Also seek optimizations like bloom
    * filters do not need to be loaded.
-   * 
+   *
    */
-  
+
   public abstract FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf) throws IOException;
-  
+
   public abstract FileSKVIterator openReader(String file, Range range, Set<ByteSequence> columnFamilies, boolean inclusive, FileSystem fs, Configuration conf,
       AccumuloConfiguration tableConf, BlockCache dataCache, BlockCache indexCache) throws IOException;
-  
+
   /**
    * Open a reader that fully support seeking and also enable any optimizations related to seeking, like bloom filters.
-   * 
+   *
    */
-  
+
   public abstract FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf)
       throws IOException;
-  
+
   public abstract FileSKVIterator openReader(String file, boolean seekToBeginning, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf,
       BlockCache dataCache, BlockCache indexCache) throws IOException;
-  
+
   public abstract FileSKVWriter openWriter(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException;
-  
+
   public abstract FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException;
-  
+
   public abstract FileSKVIterator openIndex(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf, BlockCache dCache, BlockCache iCache)
       throws IOException;
-  
+
   public abstract long getFileSize(String file, FileSystem fs, Configuration conf, AccumuloConfiguration acuconf) throws IOException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/FileSKVIterator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/FileSKVIterator.java b/core/src/main/java/org/apache/accumulo/core/file/FileSKVIterator.java
index 2de5cfc..60970e2 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/FileSKVIterator.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/FileSKVIterator.java
@@ -24,12 +24,12 @@ import org.apache.accumulo.core.iterators.system.InterruptibleIterator;
 
 public interface FileSKVIterator extends InterruptibleIterator {
   Key getFirstKey() throws IOException;
-  
+
   Key getLastKey() throws IOException;
-  
+
   DataInputStream getMetaStore(String name) throws IOException, NoSuchMetaStoreException;
-  
+
   void closeDeepCopies() throws IOException;
-  
+
   void close() throws IOException;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/FileSKVWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/FileSKVWriter.java b/core/src/main/java/org/apache/accumulo/core/file/FileSKVWriter.java
index 8718515..f4aa888 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/FileSKVWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/FileSKVWriter.java
@@ -26,14 +26,14 @@ import org.apache.accumulo.core.data.Value;
 
 public interface FileSKVWriter {
   boolean supportsLocalityGroups();
-  
+
   void startNewLocalityGroup(String name, Set<ByteSequence> columnFamilies) throws IOException;
-  
+
   void startDefaultLocalityGroup() throws IOException;
-  
+
   void append(Key key, Value value) throws IOException;
-  
+
   DataOutputStream createMetaStore(String name) throws IOException;
-  
+
   void close() throws IOException;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/NoSuchMetaStoreException.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/NoSuchMetaStoreException.java b/core/src/main/java/org/apache/accumulo/core/file/NoSuchMetaStoreException.java
index 7de78b7..a7b9801 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/NoSuchMetaStoreException.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/NoSuchMetaStoreException.java
@@ -19,18 +19,15 @@ package org.apache.accumulo.core.file;
 import java.io.IOException;
 
 public class NoSuchMetaStoreException extends IOException {
-  
+
   public NoSuchMetaStoreException(String msg, Throwable e) {
     super(msg, e);
   }
-  
+
   public NoSuchMetaStoreException(String msg) {
     super(msg);
   }
-  
-  /**
-	 * 
-	 */
+
   private static final long serialVersionUID = 1L;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockReader.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockReader.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockReader.java
index 592d325..8df2469 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockReader.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockReader.java
@@ -21,30 +21,31 @@ import java.io.DataInputStream;
 import java.io.IOException;
 
 /*
- * Minimal interface to read a block from a 
+ * Minimal interface to read a block from a
  * block based file
- * 
+ *
  */
 
 public interface ABlockReader extends DataInput {
-  
+
   long getRawSize();
-  
+
   DataInputStream getStream() throws IOException;
-  
+
   void close() throws IOException;
-  
+
   /**
    * An indexable block supports seeking, getting a position, and associating an arbitrary index with the block
-   * 
+   *
    * @return true, if the block is indexable; otherwise false.
    */
   boolean isIndexable();
 
   void seek(int position);
 
-  /** Get the file position.
-
+  /**
+   * Get the file position.
+   *
    * @return the file position.
    */
   int getPosition();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockWriter.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockWriter.java
index 19b6f0c..ece0a5e 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/ABlockWriter.java
@@ -21,21 +21,21 @@ import java.io.DataOutputStream;
 import java.io.IOException;
 
 /*
- * Minimal interface to write a block to a 
+ * Minimal interface to write a block to a
  * block based file
- * 
+ *
  */
 
 public interface ABlockWriter extends DataOutput {
-  
+
   long getCompressedSize() throws IOException;
-  
+
   void close() throws IOException;
-  
+
   long getRawSize() throws IOException;
-  
+
   long getStartPos() throws IOException;
-  
+
   DataOutputStream getStream() throws IOException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileReader.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileReader.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileReader.java
index 2c918aa..6d2014a 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileReader.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileReader.java
@@ -19,23 +19,23 @@ package org.apache.accumulo.core.file.blockfile;
 import java.io.IOException;
 
 /**
- * 
+ *
  * Provides a generic interface for a Reader for a BlockBaseFile format. Supports the minimal interface required.
- * 
+ *
  * Read a metaBlock and a dataBlock
- * 
+ *
  */
 
 public interface BlockFileReader {
-  
+
   ABlockReader getMetaBlock(String name) throws IOException;
-  
+
   ABlockReader getDataBlock(int blockIndex) throws IOException;
-  
+
   void close() throws IOException;
-  
+
   ABlockReader getMetaBlock(long offset, long compressedSize, long rawSize) throws IOException;
-  
+
   ABlockReader getDataBlock(long offset, long compressedSize, long rawSize) throws IOException;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileWriter.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileWriter.java
index cf86006..3bdbea3 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/BlockFileWriter.java
@@ -19,20 +19,20 @@ package org.apache.accumulo.core.file.blockfile;
 import java.io.IOException;
 
 /**
- * 
+ *
  * Provides a generic interface for a Writer for a BlockBaseFile format. Supports the minimal interface required.
- * 
+ *
  * Write a metaBlock and a dataBlock.
- * 
+ *
  */
 
 public interface BlockFileWriter {
-  
+
   ABlockWriter prepareMetaBlock(String name, String compressionName) throws IOException;
-  
+
   ABlockWriter prepareMetaBlock(String name) throws IOException;
-  
+
   ABlockWriter prepareDataBlock() throws IOException;
-  
+
   void close() throws IOException;
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/BlockCache.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/BlockCache.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/BlockCache.java
index 8673385..a6c08ff 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/BlockCache.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/BlockCache.java
@@ -25,7 +25,7 @@ package org.apache.accumulo.core.file.blockfile.cache;
 public interface BlockCache {
   /**
    * Add block to cache.
-   * 
+   *
    * @param blockName
    *          Zero-based file block number.
    * @param buf
@@ -34,31 +34,31 @@ public interface BlockCache {
    *          Whether block should be treated as in-memory
    */
   CacheEntry cacheBlock(String blockName, byte buf[], boolean inMemory);
-  
+
   /**
    * Add block to cache (defaults to not in-memory).
-   * 
+   *
    * @param blockName
    *          Zero-based file block number.
    * @param buf
    *          The block contents wrapped in a ByteBuffer.
    */
   CacheEntry cacheBlock(String blockName, byte buf[]);
-  
+
   /**
    * Fetch block from cache.
-   * 
+   *
    * @param blockName
    *          Block number to fetch.
    * @return Block or null if block is not in the cache.
    */
   CacheEntry getBlock(String blockName);
-  
+
   /**
    * Shutdown the cache.
    */
   void shutdown();
-  
+
   /**
    * Get the maximum size of this cache.
    *

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CacheEntry.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CacheEntry.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CacheEntry.java
index 0e628d3..2faf696 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CacheEntry.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CacheEntry.java
@@ -18,9 +18,9 @@ package org.apache.accumulo.core.file.blockfile.cache;
 
 public interface CacheEntry {
   byte[] getBuffer();
-  
+
   Object getIndex();
-  
+
   void setIndex(Object idx);
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlock.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlock.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlock.java
index 4ff0d22..b6d6d41 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlock.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlock.java
@@ -19,19 +19,18 @@
  */
 package org.apache.accumulo.core.file.blockfile.cache;
 
-
 /**
  * Represents an entry in the {@link LruBlockCache}.
- * 
+ *
  * <p>
  * Makes the block memory-aware with {@link HeapSize} and Comparable to sort by access time for the LRU. It also takes care of priority by either instantiating
  * as in-memory or handling the transition from single to multiple access.
  */
 public class CachedBlock implements HeapSize, Comparable<CachedBlock>, CacheEntry {
-  
+
   public final static long PER_BLOCK_OVERHEAD = ClassSize.align(ClassSize.OBJECT + (3 * ClassSize.REFERENCE) + (2 * SizeConstants.SIZEOF_LONG)
       + ClassSize.STRING + ClassSize.BYTE_BUFFER);
-  
+
   static enum BlockPriority {
     /**
      * Accessed a single time (used for scan-resistance)
@@ -46,18 +45,18 @@ public class CachedBlock implements HeapSize, Comparable<CachedBlock>, CacheEntr
      */
     MEMORY
   };
-  
+
   private final String blockName;
   private final byte buf[];
   private volatile long accessTime;
   private long size;
   private BlockPriority priority;
   private Object index;
-  
+
   public CachedBlock(String blockName, byte buf[], long accessTime) {
     this(blockName, buf, accessTime, false);
   }
-  
+
   public CachedBlock(String blockName, byte buf[], long accessTime, boolean inMemory) {
     this.blockName = blockName;
     this.buf = buf;
@@ -69,7 +68,7 @@ public class CachedBlock implements HeapSize, Comparable<CachedBlock>, CacheEntr
       this.priority = BlockPriority.SINGLE;
     }
   }
-  
+
   /**
    * Block has been accessed. Update its local access time.
    */
@@ -79,35 +78,35 @@ public class CachedBlock implements HeapSize, Comparable<CachedBlock>, CacheEntr
       this.priority = BlockPriority.MULTI;
     }
   }
-  
+
   public long heapSize() {
     return size;
   }
-  
+
   public int compareTo(CachedBlock that) {
     if (this.accessTime == that.accessTime)
       return 0;
     return this.accessTime < that.accessTime ? 1 : -1;
   }
-  
+
   @Override
   public byte[] getBuffer() {
     return this.buf;
   }
-  
+
   public String getName() {
     return this.blockName;
   }
-  
+
   public BlockPriority getPriority() {
     return this.priority;
   }
-  
+
   @Override
   public Object getIndex() {
     return index;
   }
-  
+
   @Override
   public void setIndex(Object idx) {
     this.index = idx;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlockQueue.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlockQueue.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlockQueue.java
index d281d65..d08fee1 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlockQueue.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/CachedBlockQueue.java
@@ -25,21 +25,21 @@ import java.util.PriorityQueue;
 /**
  * A memory-bound queue that will grow until an element brings total size >= maxSize. From then on, only entries that are sorted larger than the smallest
  * current entry will be inserted/replaced.
- * 
+ *
  * <p>
  * Use this when you want to find the largest elements (according to their ordering, not their heap size) that consume as close to the specified maxSize as
  * possible. Default behavior is to grow just above rather than just below specified max.
- * 
+ *
  * <p>
  * Object used in this queue must implement {@link HeapSize} as well as {@link Comparable}.
  */
 public class CachedBlockQueue implements HeapSize {
-  
+
   private PriorityQueue<CachedBlock> queue;
-  
+
   private long heapSize;
   private long maxSize;
-  
+
   /**
    * @param maxSize
    *          the target size of elements in the queue
@@ -54,14 +54,14 @@ public class CachedBlockQueue implements HeapSize {
     heapSize = 0;
     this.maxSize = maxSize;
   }
-  
+
   /**
    * Attempt to add the specified cached block to this queue.
-   * 
+   *
    * <p>
    * If the queue is smaller than the max size, or if the specified element is ordered before the smallest element in the queue, the element will be added to
    * the queue. Otherwise, there is no side effect of this call.
-   * 
+   *
    * @param cb
    *          block to try to add to the queue
    */
@@ -83,10 +83,10 @@ public class CachedBlockQueue implements HeapSize {
       }
     }
   }
-  
+
   /**
    * Get a sorted List of all elements in this queue, in descending order.
-   * 
+   *
    * @return list of cached elements in descending order
    */
   public CachedBlock[] get() {
@@ -96,10 +96,10 @@ public class CachedBlockQueue implements HeapSize {
     }
     return blocks.toArray(new CachedBlock[blocks.size()]);
   }
-  
+
   /**
    * Get a sorted List of all elements in this queue, in descending order.
-   * 
+   *
    * @return list of cached elements in descending order
    */
   public LinkedList<CachedBlock> getList() {
@@ -109,10 +109,10 @@ public class CachedBlockQueue implements HeapSize {
     }
     return blocks;
   }
-  
+
   /**
    * Total size of all elements in this queue.
-   * 
+   *
    * @return size of all elements currently in queue, in bytes
    */
   public long heapSize() {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/ClassSize.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/ClassSize.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/ClassSize.java
index 8abf425..2d7586f 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/ClassSize.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/ClassSize.java
@@ -29,76 +29,76 @@ import org.apache.commons.logging.LogFactory;
 
 /**
  * Class for determining the "size" of a class, an attempt to calculate the actual bytes that an object of this class will occupy in memory
- * 
+ *
  * The core of this class is taken from the Derby project
  */
 public class ClassSize {
   static final Log LOG = LogFactory.getLog(ClassSize.class);
-  
+
   private static int nrOfRefsPerObj = 2;
-  
+
   /** Array overhead */
   public static int ARRAY = 0;
-  
+
   /** Overhead for ArrayList(0) */
   public static int ARRAYLIST = 0;
-  
+
   /** Overhead for ByteBuffer */
   public static int BYTE_BUFFER = 0;
-  
+
   /** Overhead for an Integer */
   public static int INTEGER = 0;
-  
+
   /** Overhead for entry in map */
   public static int MAP_ENTRY = 0;
-  
+
   /** Object overhead is minimum 2 * reference size (8 bytes on 64-bit) */
   public static int OBJECT = 0;
-  
+
   /** Reference size is 8 bytes on 64-bit, 4 bytes on 32-bit */
   public static int REFERENCE = 0;
-  
+
   /** String overhead */
   public static int STRING = 0;
-  
+
   /** Overhead for TreeMap */
   public static int TREEMAP = 0;
-  
+
   /** Overhead for ConcurrentHashMap */
   public static int CONCURRENT_HASHMAP = 0;
-  
+
   /** Overhead for ConcurrentHashMap.Entry */
   public static int CONCURRENT_HASHMAP_ENTRY = 0;
-  
+
   /** Overhead for ConcurrentHashMap.Segment */
   public static int CONCURRENT_HASHMAP_SEGMENT = 0;
-  
+
   /** Overhead for ConcurrentSkipListMap */
   public static int CONCURRENT_SKIPLISTMAP = 0;
-  
+
   /** Overhead for ConcurrentSkipListMap Entry */
   public static int CONCURRENT_SKIPLISTMAP_ENTRY = 0;
-  
+
   /** Overhead for ReentrantReadWriteLock */
   public static int REENTRANT_LOCK = 0;
-  
+
   /** Overhead for AtomicLong */
   public static int ATOMIC_LONG = 0;
-  
+
   /** Overhead for AtomicInteger */
   public static int ATOMIC_INTEGER = 0;
-  
+
   /** Overhead for AtomicBoolean */
   public static int ATOMIC_BOOLEAN = 0;
-  
+
   /** Overhead for CopyOnWriteArraySet */
   public static int COPYONWRITE_ARRAYSET = 0;
-  
+
   /** Overhead for CopyOnWriteArrayList */
   public static int COPYONWRITE_ARRAYLIST = 0;
-  
+
   private static final String THIRTY_TWO = "32";
-  
+
   /**
    * Method for reading the arc settings and setting overheads according to 32-bit or 64-bit architecture.
    */
@@ -106,60 +106,60 @@ public class ClassSize {
     // Figure out whether this is a 32 or 64 bit machine.
     Properties sysProps = System.getProperties();
     String arcModel = sysProps.getProperty("sun.arch.data.model");
-    
+
     // Default value is set to 8, covering the case when arcModel is unknown
     REFERENCE = 8;
     if (arcModel.equals(THIRTY_TWO)) {
       REFERENCE = 4;
     }
-    
+
     OBJECT = 2 * REFERENCE;
-    
+
     ARRAY = 3 * REFERENCE;
-    
+
     ARRAYLIST = align(OBJECT + align(REFERENCE) + align(ARRAY) + (2 * SizeConstants.SIZEOF_INT));
-    
+
     BYTE_BUFFER = align(OBJECT + align(REFERENCE) + align(ARRAY) + (5 * SizeConstants.SIZEOF_INT) + (3 * SizeConstants.SIZEOF_BOOLEAN)
         + SizeConstants.SIZEOF_LONG);
-    
+
     INTEGER = align(OBJECT + SizeConstants.SIZEOF_INT);
-    
+
     MAP_ENTRY = align(OBJECT + 5 * REFERENCE + SizeConstants.SIZEOF_BOOLEAN);
-    
+
     TREEMAP = align(OBJECT + (2 * SizeConstants.SIZEOF_INT) + align(7 * REFERENCE));
-    
+
     STRING = align(OBJECT + ARRAY + REFERENCE + 3 * SizeConstants.SIZEOF_INT);
-    
+
     CONCURRENT_HASHMAP = align((2 * SizeConstants.SIZEOF_INT) + ARRAY + (6 * REFERENCE) + OBJECT);
-    
+
     CONCURRENT_HASHMAP_ENTRY = align(REFERENCE + OBJECT + (3 * REFERENCE) + (2 * SizeConstants.SIZEOF_INT));
-    
+
     CONCURRENT_HASHMAP_SEGMENT = align(REFERENCE + OBJECT + (3 * SizeConstants.SIZEOF_INT) + SizeConstants.SIZEOF_FLOAT + ARRAY);
-    
+
     CONCURRENT_SKIPLISTMAP = align(SizeConstants.SIZEOF_INT + OBJECT + (8 * REFERENCE));
-    
+
     CONCURRENT_SKIPLISTMAP_ENTRY = align(align(OBJECT + (3 * REFERENCE)) + /* one node per entry */
     align((OBJECT + (3 * REFERENCE)) / 2)); /* one index per two entries */
-    
+
     REENTRANT_LOCK = align(OBJECT + (3 * REFERENCE));
-    
+
     ATOMIC_LONG = align(OBJECT + SizeConstants.SIZEOF_LONG);
-    
+
     ATOMIC_INTEGER = align(OBJECT + SizeConstants.SIZEOF_INT);
-    
+
     ATOMIC_BOOLEAN = align(OBJECT + SizeConstants.SIZEOF_BOOLEAN);
-    
+
     COPYONWRITE_ARRAYSET = align(OBJECT + REFERENCE);
-    
+
     COPYONWRITE_ARRAYLIST = align(OBJECT + (2 * REFERENCE) + ARRAY);
   }
-  
+
   /**
    * The estimate of the size of a class instance depends on whether the JVM uses 32 or 64 bit addresses, that is it depends on the size of an object reference.
    * It is a linear function of the size of a reference, e.g. 24 + 5*r where r is the size of a reference (usually 4 or 8 bytes).
-   * 
+   *
    * This method returns the coefficients of the linear function, e.g. {24, 5} in the above example.
-   * 
+   *
    * @param cl
    *          A class whose instance size is to be estimated
    * @return an array of 3 integers. The first integer is the size of the primitives, the second the number of arrays and the third the number of references.
@@ -169,7 +169,7 @@ public class ClassSize {
     int arrays = 0;
     // The number of references that a new object takes
     int references = nrOfRefsPerObj;
-    
+
     for (; null != cl; cl = cl.getSuperclass()) {
       Field[] field = cl.getDeclaredFields();
       if (null != field) {
@@ -183,7 +183,7 @@ public class ClassSize {
               references++;
             } else {// Is simple primitive
               String name = fieldClass.getName();
-              
+
               if (name.equals("int") || name.equals("I"))
                 primitives += SizeConstants.SIZEOF_INT;
               else if (name.equals("long") || name.equals("J"))
@@ -213,18 +213,18 @@ public class ClassSize {
     }
     return new int[] {primitives, arrays, references};
   }
-  
+
   /**
    * Estimate the static space taken up by a class instance given the coefficients returned by getSizeCoefficients.
-   * 
+   *
    * @param coeff
    *          the coefficients
-   * 
+   *
    * @return the size estimate, in bytes
    */
   private static long estimateBaseFromCoefficients(int[] coeff, boolean debug) {
     long size = coeff[0] + align(coeff[1] * ARRAY) + coeff[2] * REFERENCE;
-    
+
     // Round up to a multiple of 8
     size = align(size);
     if (debug) {
@@ -236,21 +236,21 @@ public class ClassSize {
     }
     return size;
   }
-  
+
   /**
    * Estimate the static space taken up by the fields of a class. This includes the space taken up by by references (the pointer) but not by the referenced
    * object. So the estimated size of an array field does not depend on the size of the array. Similarly the size of an object (reference) field does not depend
    * on the object.
-   * 
+   *
    * @return the size estimate in bytes.
    */
   public static long estimateBase(Class<?> cl, boolean debug) {
     return estimateBaseFromCoefficients(getSizeCoefficients(cl, debug), debug);
   }
-  
+
   /**
    * Aligns a number to 8.
-   * 
+   *
    * @param num
    *          number to align to 8
    * @return smallest number >= input that is a multiple of 8
@@ -258,10 +258,10 @@ public class ClassSize {
   public static int align(int num) {
     return (int) (align((long) num));
   }
-  
+
   /**
    * Aligns a number to 8.
-   * 
+   *
    * @param num
    *          number to align to 8
    * @return smallest number >= input that is a multiple of 8
@@ -271,5 +271,5 @@ public class ClassSize {
     // stored and sent together
     return ((num + 7) >> 3) << 3;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/HeapSize.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/HeapSize.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/HeapSize.java
index ca2402e..04db3cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/HeapSize.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/HeapSize.java
@@ -27,10 +27,10 @@ package org.apache.accumulo.core.file.blockfile.cache;
  * An Object's size is determined by the non-static data members in it, as well as the fixed {@link Object} overhead.
  * <p>
  * For example:
- * 
+ *
  * <pre>
  * public class SampleObject implements HeapSize {
- *   
+ *
  *   int[] numbers;
  *   int x;
  * }
@@ -41,5 +41,5 @@ public interface HeapSize {
    * @return Approximate 'exclusive deep size' of implementing object. Includes count of payload and hosting object sizings.
    */
   long heapSize();
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/LruBlockCache.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/LruBlockCache.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/LruBlockCache.java
index 83f363e..6cab164 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/LruBlockCache.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/LruBlockCache.java
@@ -36,111 +36,111 @@ import org.apache.commons.logging.LogFactory;
  * A block cache implementation that is memory-aware using {@link HeapSize}, memory-bound using an LRU eviction algorithm, and concurrent: backed by a
  * {@link ConcurrentHashMap} and with a non-blocking eviction thread giving constant-time {@link #cacheBlock} and {@link #getBlock} operations.
  * <p>
- * 
+ *
  * Contains three levels of block priority to allow for scan-resistance and in-memory families. A block is added with an inMemory flag if necessary, otherwise a
  * block becomes a single access priority. Once a blocked is accessed again, it changes to multiple access. This is used to prevent scans from thrashing the
  * cache, adding a least-frequently-used element to the eviction algorithm.
  * <p>
- * 
+ *
  * Each priority is given its own chunk of the total cache to ensure fairness during eviction. Each priority will retain close to its maximum size, however, if
  * any priority is not using its entire chunk the others are able to grow beyond their chunk size.
  * <p>
- * 
+ *
  * Instantiated at a minimum with the total size and average block size. All sizes are in bytes. The block size is not especially important as this cache is
  * fully dynamic in its sizing of blocks. It is only used for pre-allocating data structures and in initial heap estimation of the map.
  * <p>
- * 
+ *
  * The detailed constructor defines the sizes for the three priorities (they should total to the maximum size defined). It also sets the levels that trigger and
  * control the eviction thread.
  * <p>
- * 
+ *
  * The acceptable size is the cache size level which triggers the eviction process to start. It evicts enough blocks to get the size below the minimum size
  * specified.
  * <p>
- * 
+ *
  * Eviction happens in a separate thread and involves a single full-scan of the map. It determines how many bytes must be freed to reach the minimum size, and
  * then while scanning determines the fewest least-recently-used blocks necessary from each of the three priorities (would be 3 times bytes to free). It then
  * uses the priority chunk sizes to evict fairly according to the relative sizes and usage.
  */
 public class LruBlockCache implements BlockCache, HeapSize {
-  
+
   static final Log LOG = LogFactory.getLog(LruBlockCache.class);
-  
+
   /** Default Configuration Parameters */
-  
+
   /** Backing Concurrent Map Configuration */
   static final float DEFAULT_LOAD_FACTOR = 0.75f;
   static final int DEFAULT_CONCURRENCY_LEVEL = 16;
-  
+
   /** Eviction thresholds */
   static final float DEFAULT_MIN_FACTOR = 0.75f;
   static final float DEFAULT_ACCEPTABLE_FACTOR = 0.85f;
-  
+
   /** Priority buckets */
   static final float DEFAULT_SINGLE_FACTOR = 0.25f;
   static final float DEFAULT_MULTI_FACTOR = 0.50f;
   static final float DEFAULT_MEMORY_FACTOR = 0.25f;
-  
+
   /** Statistics thread */
   static final int statThreadPeriod = 60;
-  
+
   /** Concurrent map (the cache) */
   private final ConcurrentHashMap<String,CachedBlock> map;
-  
+
   /** Eviction lock (locked when eviction in process) */
   private final ReentrantLock evictionLock = new ReentrantLock(true);
-  
+
   /** Volatile boolean to track if we are in an eviction process or not */
   private volatile boolean evictionInProgress = false;
-  
+
   /** Eviction thread */
   private final EvictionThread evictionThread;
-  
+
   /** Statistics thread schedule pool (for heavy debugging, could remove) */
   private final ScheduledExecutorService scheduleThreadPool = Executors.newScheduledThreadPool(1, new NamingThreadFactory("LRUBlockCacheStats"));
-  
+
   /** Current size of cache */
   private final AtomicLong size;
-  
+
   /** Current number of cached elements */
   private final AtomicLong elements;
-  
+
   /** Cache access count (sequential ID) */
   private final AtomicLong count;
-  
+
   /** Cache statistics */
   private final CacheStats stats;
-  
+
   /** Maximum allowable size of cache (block put if size > max, evict) */
   private long maxSize;
-  
+
   /** Approximate block size */
   private long blockSize;
-  
+
   /** Acceptable size of cache (no evictions if size < acceptable) */
   private float acceptableFactor;
-  
+
   /** Minimum threshold of cache (when evicting, evict until size < min) */
   private float minFactor;
-  
+
   /** Single access bucket size */
   private float singleFactor;
-  
+
   /** Multiple access bucket size */
   private float multiFactor;
-  
+
   /** In-memory bucket size */
   private float memoryFactor;
-  
+
   /** Overhead of the structure itself */
   private long overhead;
-  
+
   /**
    * Default constructor. Specify maximum size and expected average block size (approximation is fine).
-   * 
+   *
    * <p>
    * All other factors will be calculated based on defaults specified in this class.
-   * 
+   *
    * @param maxSize
    *          maximum size of cache, in bytes
    * @param blockSize
@@ -149,7 +149,7 @@ public class LruBlockCache implements BlockCache, HeapSize {
   public LruBlockCache(long maxSize, long blockSize) {
     this(maxSize, blockSize, true);
   }
-  
+
   /**
    * Constructor used for testing. Allows disabling of the eviction thread.
    */
@@ -157,10 +157,10 @@ public class LruBlockCache implements BlockCache, HeapSize {
     this(maxSize, blockSize, evictionThread, (int) Math.ceil(1.2 * maxSize / blockSize), DEFAULT_LOAD_FACTOR, DEFAULT_CONCURRENCY_LEVEL, DEFAULT_MIN_FACTOR,
         DEFAULT_ACCEPTABLE_FACTOR, DEFAULT_SINGLE_FACTOR, DEFAULT_MULTI_FACTOR, DEFAULT_MEMORY_FACTOR);
   }
-  
+
   /**
    * Configurable constructor. Use this constructor if not using defaults.
-   * 
+   *
    * @param maxSize
    *          maximum size of this cache, in bytes
    * @param blockSize
@@ -208,7 +208,7 @@ public class LruBlockCache implements BlockCache, HeapSize {
     this.elements = new AtomicLong(0);
     this.overhead = calculateOverhead(maxSize, blockSize, mapConcurrencyLevel);
     this.size = new AtomicLong(this.overhead);
-    
+
     if (evictionThread) {
       this.evictionThread = new EvictionThread(this);
       this.evictionThread.start();
@@ -224,22 +224,22 @@ public class LruBlockCache implements BlockCache, HeapSize {
     }
     this.scheduleThreadPool.scheduleAtFixedRate(new StatisticsThread(this), statThreadPeriod, statThreadPeriod, TimeUnit.SECONDS);
   }
-  
+
   public void setMaxSize(long maxSize) {
     this.maxSize = maxSize;
     if (this.size.get() > acceptableSize() && !evictionInProgress) {
       runEviction();
     }
   }
-  
+
   // BlockCache implementation
-  
+
   /**
    * Cache the block with the specified name and buffer.
    * <p>
    * It is assumed this will NEVER be called on an already cached block. If that is done, it is assumed that you are reinserting the same exact block due to a
    * race condition and will update the buffer but not modify the size of the cache.
-   * 
+   *
    * @param blockName
    *          block name
    * @param buf
@@ -252,7 +252,7 @@ public class LruBlockCache implements BlockCache, HeapSize {
     if (cb != null) {
       stats.duplicateReads();
       cb.access(count.incrementAndGet());
-      
+
     } else {
       cb = new CachedBlock(blockName, buf, count.incrementAndGet(), inMemory);
       long newSize = size.addAndGet(cb.heapSize());
@@ -262,16 +262,16 @@ public class LruBlockCache implements BlockCache, HeapSize {
         runEviction();
       }
     }
-    
+
     return cb;
   }
-  
+
   /**
    * Cache the block with the specified name and buffer.
    * <p>
    * It is assumed this will NEVER be called on an already cached block. If that is done, it is assumed that you are reinserting the same exact block due to a
    * race condition and will update the buffer but not modify the size of the cache.
-   * 
+   *
    * @param blockName
    *          block name
    * @param buf
@@ -280,15 +280,15 @@ public class LruBlockCache implements BlockCache, HeapSize {
   public CacheEntry cacheBlock(String blockName, byte buf[]) {
     return cacheBlock(blockName, buf, false);
   }
-  
+
   /**
    * Get the buffer of the block with the specified name.
-   * 
+   *
    * @param blockName
    *          block name
    * @return buffer of specified block name, or null if not in cache
    */
-  
+
   public CachedBlock getBlock(String blockName) {
     CachedBlock cb = map.get(blockName);
     if (cb == null) {
@@ -299,7 +299,7 @@ public class LruBlockCache implements BlockCache, HeapSize {
     cb.access(count.incrementAndGet());
     return cb;
   }
-  
+
   protected long evictBlock(CachedBlock block) {
     map.remove(block.getName());
     size.addAndGet(-1 * block.heapSize());
@@ -307,7 +307,7 @@ public class LruBlockCache implements BlockCache, HeapSize {
     stats.evicted();
     return block.heapSize();
   }
-  
+
   /**
    * Multi-threaded call to run the eviction process.
    */
@@ -318,31 +318,31 @@ public class LruBlockCache implements BlockCache, HeapSize {
       evictionThread.evict();
     }
   }
-  
+
   /**
    * Eviction method.
    */
   void evict() {
-    
+
     // Ensure only one eviction at a time
     if (!evictionLock.tryLock())
       return;
-    
+
     try {
       evictionInProgress = true;
-      
+
       long bytesToFree = size.get() - minSize();
-      
+
       LOG.debug("Block cache LRU eviction started.  Attempting to free " + bytesToFree + " bytes");
-      
+
       if (bytesToFree <= 0)
         return;
-      
+
       // Instantiate priority buckets
       BlockBucket bucketSingle = new BlockBucket(bytesToFree, blockSize, singleSize());
       BlockBucket bucketMulti = new BlockBucket(bytesToFree, blockSize, multiSize());
       BlockBucket bucketMemory = new BlockBucket(bytesToFree, blockSize, memorySize());
-      
+
       // Scan entire map putting into appropriate buckets
       for (CachedBlock cachedBlock : map.values()) {
         switch (cachedBlock.getPriority()) {
@@ -360,16 +360,16 @@ public class LruBlockCache implements BlockCache, HeapSize {
           }
         }
       }
-      
+
       PriorityQueue<BlockBucket> bucketQueue = new PriorityQueue<BlockBucket>(3);
-      
+
       bucketQueue.add(bucketSingle);
       bucketQueue.add(bucketMulti);
       bucketQueue.add(bucketMemory);
-      
+
       int remainingBuckets = 3;
       long bytesFreed = 0;
-      
+
       BlockBucket bucket;
       while ((bucket = bucketQueue.poll()) != null) {
         long overflow = bucket.overflow();
@@ -379,43 +379,43 @@ public class LruBlockCache implements BlockCache, HeapSize {
         }
         remainingBuckets--;
       }
-      
+
       float singleMB = ((float) bucketSingle.totalSize()) / ((float) (1024 * 1024));
       float multiMB = ((float) bucketMulti.totalSize()) / ((float) (1024 * 1024));
       float memoryMB = ((float) bucketMemory.totalSize()) / ((float) (1024 * 1024));
-      
+
       LOG.debug("Block cache LRU eviction completed. " + "Freed " + bytesFreed + " bytes.  " + "Priority Sizes: " + "Single=" + singleMB + "MB ("
           + bucketSingle.totalSize() + "), " + "Multi=" + multiMB + "MB (" + bucketMulti.totalSize() + ")," + "Memory=" + memoryMB + "MB ("
           + bucketMemory.totalSize() + ")");
-      
+
     } finally {
       stats.evict();
       evictionInProgress = false;
       evictionLock.unlock();
     }
   }
-  
+
   /**
    * Used to group blocks into priority buckets. There will be a BlockBucket for each priority (single, multi, memory). Once bucketed, the eviction algorithm
    * takes the appropriate number of elements out of each according to configuration parameters and their relatives sizes.
    */
   private class BlockBucket implements Comparable<BlockBucket> {
-    
+
     private CachedBlockQueue queue;
     private long totalSize = 0;
     private long bucketSize;
-    
+
     public BlockBucket(long bytesToFree, long blockSize, long bucketSize) {
       this.bucketSize = bucketSize;
       queue = new CachedBlockQueue(bytesToFree, blockSize);
       totalSize = 0;
     }
-    
+
     public void add(CachedBlock block) {
       totalSize += block.heapSize();
       queue.add(block);
     }
-    
+
     public long free(long toFree) {
       CachedBlock[] blocks = queue.get();
       long freedBytes = 0;
@@ -427,99 +427,99 @@ public class LruBlockCache implements BlockCache, HeapSize {
       }
       return freedBytes;
     }
-    
+
     public long overflow() {
       return totalSize - bucketSize;
     }
-    
+
     public long totalSize() {
       return totalSize;
     }
-    
+
     @Override
     public int compareTo(BlockBucket that) {
       if (this.overflow() == that.overflow())
         return 0;
       return this.overflow() > that.overflow() ? 1 : -1;
     }
-    
+
     @Override
     public boolean equals(Object that) {
       if (that instanceof BlockBucket)
-        return compareTo((BlockBucket)that) == 0;
+        return compareTo((BlockBucket) that) == 0;
       return false;
     }
   }
-  
+
   /**
    * Get the maximum size of this cache.
-   * 
+   *
    * @return max size in bytes
    */
   public long getMaxSize() {
     return this.maxSize;
   }
-  
+
   /**
    * Get the current size of this cache.
-   * 
+   *
    * @return current size in bytes
    */
   public long getCurrentSize() {
     return this.size.get();
   }
-  
+
   /**
    * Get the current size of this cache.
-   * 
+   *
    * @return current size in bytes
    */
   public long getFreeSize() {
     return getMaxSize() - getCurrentSize();
   }
-  
+
   /**
    * Get the size of this cache (number of cached blocks)
-   * 
+   *
    * @return number of cached blocks
    */
   public long size() {
     return this.elements.get();
   }
-  
+
   /**
    * Get the number of eviction runs that have occurred
    */
   public long getEvictionCount() {
     return this.stats.getEvictionCount();
   }
-  
+
   /**
    * Get the number of blocks that have been evicted during the lifetime of this cache.
    */
   public long getEvictedCount() {
     return this.stats.getEvictedCount();
   }
-  
+
   /*
    * Eviction thread. Sits in waiting state until an eviction is triggered when the cache size grows above the acceptable level.<p>
-   * 
+   *
    * Thread is triggered into action by {@link LruBlockCache#runEviction()}
    */
   private static class EvictionThread extends Thread {
     private WeakReference<LruBlockCache> cache;
     private boolean running = false;
-    
+
     public EvictionThread(LruBlockCache cache) {
       super("LruBlockCache.EvictionThread");
       setDaemon(true);
       this.cache = new WeakReference<LruBlockCache>(cache);
     }
-    
+
     public synchronized boolean running() {
       return running;
     }
-    
+
     @Override
     public void run() {
       while (true) {
@@ -535,32 +535,32 @@ public class LruBlockCache implements BlockCache, HeapSize {
         cache.evict();
       }
     }
-    
+
     public void evict() {
       synchronized (this) {
         this.notify();
       }
     }
   }
-  
+
   /*
    * Statistics thread. Periodically prints the cache statistics to the log.
    */
   private static class StatisticsThread extends Thread {
     LruBlockCache lru;
-    
+
     public StatisticsThread(LruBlockCache lru) {
       super("LruBlockCache.StatisticsThread");
       setDaemon(true);
       this.lru = lru;
     }
-    
+
     @Override
     public void run() {
       lru.logStats();
     }
   }
-  
+
   public void logStats() {
     // Log size
     long totalSize = heapSize();
@@ -574,17 +574,17 @@ public class LruBlockCache implements BlockCache, HeapSize {
         + "Hit Ratio=" + stats.getHitRatio() * 100 + "%, " + "Miss Ratio=" + stats.getMissRatio() * 100 + "%, " + "Evicted/Run=" + stats.evictedPerEviction()
         + ", " + "Duplicate Reads=" + stats.getDuplicateReads());
   }
-  
+
   /**
    * Get counter statistics for this cache.
-   * 
+   *
    * <p>
    * Includes: total accesses, hits, misses, evicted blocks, and runs of the eviction processes.
    */
   public CacheStats getStats() {
     return this.stats;
   }
-  
+
   public static class CacheStats {
     private final AtomicLong accessCount = new AtomicLong(0);
     private final AtomicLong hitCount = new AtomicLong(0);
@@ -592,101 +592,101 @@ public class LruBlockCache implements BlockCache, HeapSize {
     private final AtomicLong evictionCount = new AtomicLong(0);
     private final AtomicLong evictedCount = new AtomicLong(0);
     private final AtomicLong duplicateReads = new AtomicLong(0);
-    
+
     public void miss() {
       missCount.incrementAndGet();
       accessCount.incrementAndGet();
     }
-    
+
     public void hit() {
       hitCount.incrementAndGet();
       accessCount.incrementAndGet();
     }
-    
+
     public void evict() {
       evictionCount.incrementAndGet();
     }
-    
+
     public void duplicateReads() {
       duplicateReads.incrementAndGet();
     }
-    
+
     public void evicted() {
       evictedCount.incrementAndGet();
     }
-    
+
     public long getRequestCount() {
       return accessCount.get();
     }
-    
+
     public long getMissCount() {
       return missCount.get();
     }
-    
+
     public long getHitCount() {
       return hitCount.get();
     }
-    
+
     public long getEvictionCount() {
       return evictionCount.get();
     }
-    
+
     public long getDuplicateReads() {
       return duplicateReads.get();
     }
-    
+
     public long getEvictedCount() {
       return evictedCount.get();
     }
-    
+
     public double getHitRatio() {
       return ((float) getHitCount() / (float) getRequestCount());
     }
-    
+
     public double getMissRatio() {
       return ((float) getMissCount() / (float) getRequestCount());
     }
-    
+
     public double evictedPerEviction() {
       return (float) ((float) getEvictedCount() / (float) getEvictionCount());
     }
   }
-  
+
   public final static long CACHE_FIXED_OVERHEAD = ClassSize.align((3 * SizeConstants.SIZEOF_LONG) + (8 * ClassSize.REFERENCE)
       + (5 * SizeConstants.SIZEOF_FLOAT) + SizeConstants.SIZEOF_BOOLEAN + ClassSize.OBJECT);
-  
+
   // HeapSize implementation
   public long heapSize() {
     return getCurrentSize();
   }
-  
+
   public static long calculateOverhead(long maxSize, long blockSize, int concurrency) {
     return CACHE_FIXED_OVERHEAD + ClassSize.CONCURRENT_HASHMAP + ((int) Math.ceil(maxSize * 1.2 / blockSize) * ClassSize.CONCURRENT_HASHMAP_ENTRY)
         + (concurrency * ClassSize.CONCURRENT_HASHMAP_SEGMENT);
   }
-  
+
   // Simple calculators of sizes given factors and maxSize
-  
+
   private long acceptableSize() {
     return (long) Math.floor(this.maxSize * this.acceptableFactor);
   }
-  
+
   private long minSize() {
     return (long) Math.floor(this.maxSize * this.minFactor);
   }
-  
+
   private long singleSize() {
     return (long) Math.floor(this.maxSize * this.singleFactor * this.minFactor);
   }
-  
+
   private long multiSize() {
     return (long) Math.floor(this.maxSize * this.multiFactor * this.minFactor);
   }
-  
+
   private long memorySize() {
     return (long) Math.floor(this.maxSize * this.memoryFactor * this.minFactor);
   }
-  
+
   public void shutdown() {
     this.scheduleThreadPool.shutdown();
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/SizeConstants.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/SizeConstants.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/SizeConstants.java
index f3e483f..e5004cd 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/SizeConstants.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/cache/SizeConstants.java
@@ -17,42 +17,42 @@
 package org.apache.accumulo.core.file.blockfile.cache;
 
 public class SizeConstants {
-  
+
   public static final int SIZEOF_BOOLEAN = Byte.SIZE / Byte.SIZE;
-  
+
   /**
    * Size of byte in bytes
    */
   public static final int SIZEOF_BYTE = SIZEOF_BOOLEAN;
-  
+
   /**
    * Size of char in bytes
    */
   public static final int SIZEOF_CHAR = Character.SIZE / Byte.SIZE;
-  
+
   /**
    * Size of double in bytes
    */
   public static final int SIZEOF_DOUBLE = Double.SIZE / Byte.SIZE;
-  
+
   /**
    * Size of float in bytes
    */
   public static final int SIZEOF_FLOAT = Float.SIZE / Byte.SIZE;
-  
+
   /**
    * Size of int in bytes
    */
   public static final int SIZEOF_INT = Integer.SIZE / Byte.SIZE;
-  
+
   /**
    * Size of long in bytes
    */
   public static final int SIZEOF_LONG = Long.SIZE / Byte.SIZE;
-  
+
   /**
    * Size of short in bytes
    */
   public static final int SIZEOF_SHORT = Short.SIZE / Byte.SIZE;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/blockfile/impl/CachableBlockFile.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/blockfile/impl/CachableBlockFile.java b/core/src/main/java/org/apache/accumulo/core/file/blockfile/impl/CachableBlockFile.java
index 095a218..4d65c9f 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/blockfile/impl/CachableBlockFile.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/blockfile/impl/CachableBlockFile.java
@@ -41,105 +41,105 @@ import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
 /***
- * 
+ *
  * This is a wrapper class for BCFile that includes a cache for independent caches for datablocks and metadatablocks
  */
 
 public class CachableBlockFile {
-  
+
   private CachableBlockFile() {};
-  
+
   private static final Logger log = Logger.getLogger(CachableBlockFile.class);
-  
+
   public static class Writer implements BlockFileWriter {
     private BCFile.Writer _bc;
     private BlockWrite _bw;
     private FSDataOutputStream fsout = null;
-    
+
     public Writer(FileSystem fs, Path fName, String compressAlgor, Configuration conf, AccumuloConfiguration accumuloConfiguration) throws IOException {
       this.fsout = fs.create(fName);
       init(fsout, compressAlgor, conf, accumuloConfiguration);
     }
-    
+
     public Writer(FSDataOutputStream fsout, String compressAlgor, Configuration conf, AccumuloConfiguration accumuloConfiguration) throws IOException {
       this.fsout = fsout;
       init(fsout, compressAlgor, conf, accumuloConfiguration);
     }
-    
+
     private void init(FSDataOutputStream fsout, String compressAlgor, Configuration conf, AccumuloConfiguration accumuloConfiguration) throws IOException {
       _bc = new BCFile.Writer(fsout, compressAlgor, conf, false, accumuloConfiguration);
     }
-    
+
     public ABlockWriter prepareMetaBlock(String name) throws IOException {
       _bw = new BlockWrite(_bc.prepareMetaBlock(name));
       return _bw;
     }
-    
+
     public ABlockWriter prepareMetaBlock(String name, String compressionName) throws IOException {
       _bw = new BlockWrite(_bc.prepareMetaBlock(name, compressionName));
       return _bw;
     }
-    
+
     public ABlockWriter prepareDataBlock() throws IOException {
       _bw = new BlockWrite(_bc.prepareDataBlock());
       return _bw;
     }
-    
+
     public void close() throws IOException {
-      
+
       _bw.close();
       _bc.close();
-      
+
       if (this.fsout != null) {
         this.fsout.close();
       }
-      
+
     }
-    
+
   }
-  
+
   public static class BlockWrite extends DataOutputStream implements ABlockWriter {
     BlockAppender _ba;
-    
+
     public BlockWrite(BlockAppender ba) {
       super(ba);
       this._ba = ba;
     };
-    
+
     @Override
     public long getCompressedSize() throws IOException {
       return _ba.getCompressedSize();
     }
-    
+
     @Override
     public long getRawSize() throws IOException {
       return _ba.getRawSize();
     }
-    
+
     @Override
     public void close() throws IOException {
-      
+
       _ba.close();
     }
-    
+
     @Override
     public DataOutputStream getStream() throws IOException {
-      
+
       return this;
     }
-    
+
     @Override
     public long getStartPos() throws IOException {
       return _ba.getStartPos();
-    }    
-    
+    }
+
   }
-  
+
   /**
-   * 
-   * 
+   *
+   *
    * Class wraps the BCFile reader.
-   * 
+   *
    */
   public static class Reader implements BlockFileReader {
     private BCFile.Reader _bc;
@@ -151,83 +151,84 @@ public class CachableBlockFile {
     private Configuration conf;
     private boolean closed = false;
     private AccumuloConfiguration accumuloConfiguration = null;
-    
+
     private interface BlockLoader {
       BlockReader get() throws IOException;
-      
+
       String getInfo();
     }
-    
+
     private class OffsetBlockLoader implements BlockLoader {
-      
+
       private int blockIndex;
-      
+
       OffsetBlockLoader(int blockIndex) {
         this.blockIndex = blockIndex;
       }
-      
+
       @Override
       public BlockReader get() throws IOException {
         return getBCFile(accumuloConfiguration).getDataBlock(blockIndex);
       }
-      
+
       @Override
       public String getInfo() {
         return "" + blockIndex;
       }
-      
+
     }
-    
+
     private class RawBlockLoader implements BlockLoader {
-      
+
       private long offset;
       private long compressedSize;
       private long rawSize;
-      
+
       RawBlockLoader(long offset, long compressedSize, long rawSize) {
         this.offset = offset;
         this.compressedSize = compressedSize;
         this.rawSize = rawSize;
       }
-      
+
       @Override
       public BlockReader get() throws IOException {
         return getBCFile(accumuloConfiguration).getDataBlock(offset, compressedSize, rawSize);
       }
-      
+
       @Override
       public String getInfo() {
         return "" + offset + "," + compressedSize + "," + rawSize;
       }
     }
-    
+
     private class MetaBlockLoader implements BlockLoader {
-      
+
       private String name;
       private AccumuloConfiguration accumuloConfiguration;
-      
+
       MetaBlockLoader(String name, AccumuloConfiguration accumuloConfiguration) {
         this.name = name;
         this.accumuloConfiguration = accumuloConfiguration;
       }
-      
+
       @Override
       public BlockReader get() throws IOException {
         return getBCFile(accumuloConfiguration).getMetaBlock(name);
       }
-      
+
       @Override
       public String getInfo() {
         return name;
       }
     }
-    
-    public Reader(FileSystem fs, Path dataFile, Configuration conf, BlockCache data, BlockCache index, AccumuloConfiguration accumuloConfiguration) throws IOException {
-      
+
+    public Reader(FileSystem fs, Path dataFile, Configuration conf, BlockCache data, BlockCache index, AccumuloConfiguration accumuloConfiguration)
+        throws IOException {
+
       /*
        * Grab path create input stream grab len create file
        */
-      
+
       fileName = dataFile.toString();
       this._dCache = data;
       this._iCache = index;
@@ -235,8 +236,9 @@ public class CachableBlockFile {
       this.conf = conf;
       this.accumuloConfiguration = accumuloConfiguration;
     }
-    
-    public Reader(FSDataInputStream fsin, long len, Configuration conf, BlockCache data, BlockCache index, AccumuloConfiguration accumuloConfiguration) throws IOException {
+
+    public Reader(FSDataInputStream fsin, long len, Configuration conf, BlockCache data, BlockCache index, AccumuloConfiguration accumuloConfiguration)
+        throws IOException {
       this._dCache = data;
       this._iCache = index;
       init(fsin, len, conf, accumuloConfiguration);
@@ -246,50 +248,50 @@ public class CachableBlockFile {
       // this.fin = fsin;
       init(fsin, len, conf, accumuloConfiguration);
     }
-    
+
     private void init(FSDataInputStream fsin, long len, Configuration conf, AccumuloConfiguration accumuloConfiguration) throws IOException {
       this._bc = new BCFile.Reader(this, fsin, len, conf, accumuloConfiguration);
     }
-    
+
     private synchronized BCFile.Reader getBCFile(AccumuloConfiguration accumuloConfiguration) throws IOException {
       if (closed)
         throw new IllegalStateException("File " + fileName + " is closed");
-      
+
       if (_bc == null) {
         // lazily open file if needed
         Path path = new Path(fileName);
         fin = fs.open(path);
         init(fin, fs.getFileStatus(path).getLen(), conf, accumuloConfiguration);
       }
-      
+
       return _bc;
     }
-    
+
     public BlockRead getCachedMetaBlock(String blockName) throws IOException {
       String _lookup = fileName + "M" + blockName;
-      
+
       if (_iCache != null) {
         CacheEntry cacheEntry = _iCache.getBlock(_lookup);
-        
+
         if (cacheEntry != null) {
           return new CachedBlockRead(cacheEntry, cacheEntry.getBuffer());
         }
-        
+
       }
-      
+
       return null;
     }
-    
+
     public BlockRead cacheMetaBlock(String blockName, BlockReader _currBlock) throws IOException {
       String _lookup = fileName + "M" + blockName;
       return cacheBlock(_lookup, _iCache, _currBlock, blockName);
     }
-    
+
     public void cacheMetaBlock(String blockName, byte[] b) {
-      
+
       if (_iCache == null)
         return;
-      
+
       String _lookup = fileName + "M" + blockName;
       try {
         _iCache.cacheBlock(_lookup, b);
@@ -297,42 +299,42 @@ public class CachableBlockFile {
         log.warn("Already cached block: " + _lookup, e);
       }
     }
-    
+
     private BlockRead getBlock(String _lookup, BlockCache cache, BlockLoader loader) throws IOException {
-      
+
       BlockReader _currBlock;
-      
+
       if (cache != null) {
         CacheEntry cb = null;
         cb = cache.getBlock(_lookup);
-        
+
         if (cb != null) {
           return new CachedBlockRead(cb, cb.getBuffer());
         }
-        
+
       }
       /**
        * grab the currBlock at this point the block is still in the data stream
-       * 
+       *
        */
       _currBlock = loader.get();
-      
+
       /**
        * If the block is bigger than the cache just return the stream
        */
       return cacheBlock(_lookup, cache, _currBlock, loader.getInfo());
-      
+
     }
-    
+
     private BlockRead cacheBlock(String _lookup, BlockCache cache, BlockReader _currBlock, String block) throws IOException {
-      
+
       if ((cache == null) || (_currBlock.getRawSize() > cache.getMaxSize())) {
         return new BlockRead(_currBlock, _currBlock.getRawSize());
       } else {
-        
+
         /**
          * Try to fully read block for meta data if error try to close file
-         * 
+         *
          */
         byte b[] = null;
         try {
@@ -344,26 +346,26 @@ public class CachableBlockFile {
         } finally {
           _currBlock.close();
         }
-        
+
         CacheEntry ce = null;
         try {
           ce = cache.cacheBlock(_lookup, b);
         } catch (Exception e) {
           log.warn("Already cached block: " + _lookup, e);
         }
-        
+
         if (ce == null)
           return new BlockRead(new DataInputStream(new ByteArrayInputStream(b)), b.length);
         else
           return new CachedBlockRead(ce, ce.getBuffer());
-        
+
       }
     }
-    
+
     /**
      * It is intended that once the BlockRead object is returned to the caller, that the caller will read the entire block and then call close on the BlockRead
      * class.
-     * 
+     *
      * NOTE: In the case of multi-read threads: This method can do redundant work where an entry is read from disk and other threads check the cache before it
      * has been inserted.
      */
@@ -371,102 +373,102 @@ public class CachableBlockFile {
       String _lookup = this.fileName + "M" + blockName;
       return getBlock(_lookup, _iCache, new MetaBlockLoader(blockName, accumuloConfiguration));
     }
-    
+
     @Override
     public ABlockReader getMetaBlock(long offset, long compressedSize, long rawSize) throws IOException {
       String _lookup = this.fileName + "R" + offset;
       return getBlock(_lookup, _iCache, new RawBlockLoader(offset, compressedSize, rawSize));
     }
-    
+
     /**
      * It is intended that once the BlockRead object is returned to the caller, that the caller will read the entire block and then call close on the BlockRead
      * class.
-     * 
+     *
      * NOTE: In the case of multi-read threads: This method can do redundant work where an entry is read from disk and other threads check the cache before it
      * has been inserted.
      */
-    
+
     public BlockRead getDataBlock(int blockIndex) throws IOException {
       String _lookup = this.fileName + "O" + blockIndex;
       return getBlock(_lookup, _dCache, new OffsetBlockLoader(blockIndex));
-      
+
     }
-    
+
     @Override
     public ABlockReader getDataBlock(long offset, long compressedSize, long rawSize) throws IOException {
       String _lookup = this.fileName + "R" + offset;
       return getBlock(_lookup, _dCache, new RawBlockLoader(offset, compressedSize, rawSize));
     }
-    
+
     public synchronized void close() throws IOException {
       if (closed)
         return;
-      
+
       closed = true;
-      
+
       if (_bc != null)
         _bc.close();
-      
+
       if (fin != null) {
         fin.close();
       }
     }
-    
+
   }
-  
+
   static class SeekableByteArrayInputStream extends ByteArrayInputStream {
-    
+
     public SeekableByteArrayInputStream(byte[] buf) {
       super(buf);
     }
-    
+
     public SeekableByteArrayInputStream(byte buf[], int offset, int length) {
       super(buf, offset, length);
       throw new UnsupportedOperationException("Seek code assumes offset is zero"); // do not need this constructor, documenting that seek will not work
-                                                                                  // unless offset it kept track of
+                                                                                   // unless offset it kept track of
     }
-    
+
     public void seek(int position) {
       if (pos < 0 || pos >= buf.length)
         throw new IllegalArgumentException("pos = " + pos + " buf.lenght = " + buf.length);
       this.pos = position;
     }
-    
+
     public int getPosition() {
       return this.pos;
     }
-    
+
   }
 
   public static class CachedBlockRead extends BlockRead {
     private SeekableByteArrayInputStream seekableInput;
     private final CacheEntry cb;
-    
+
     public CachedBlockRead(CacheEntry cb, byte buf[]) {
       this(new SeekableByteArrayInputStream(buf), buf.length, cb);
     }
-    
+
     private CachedBlockRead(SeekableByteArrayInputStream seekableInput, long size, CacheEntry cb) {
-        super(seekableInput, size);
-        this.seekableInput = seekableInput;
-        this.cb = cb;
-      }
+      super(seekableInput, size);
+      this.seekableInput = seekableInput;
+      this.cb = cb;
+    }
 
     @Override
     public void seek(int position) {
       seekableInput.seek(position);
     }
-    
+
     @Override
     public int getPosition() {
       return seekableInput.getPosition();
     }
-    
+
     @Override
     public boolean isIndexable() {
       return true;
     }
-    
+
     @Override
     public <T> T getIndex(Class<T> clazz) {
       T bi = null;
@@ -475,7 +477,7 @@ public class CachableBlockFile {
         SoftReference<T> softRef = (SoftReference<T>) cb.getIndex();
         if (softRef != null)
           bi = softRef.get();
-        
+
         if (bi == null) {
           try {
             bi = clazz.newInstance();
@@ -485,33 +487,33 @@ public class CachableBlockFile {
           cb.setIndex(new SoftReference<T>(bi));
         }
       }
-      
+
       return bi;
     }
   }
 
   /**
-   * 
+   *
    * Class provides functionality to read one block from the underlying BCFile Since We are caching blocks in the Reader class as bytearrays, this class will
    * wrap a DataInputStream(ByteArrayStream(cachedBlock)).
-   * 
-   * 
+   *
+   *
    */
   public static class BlockRead extends DataInputStream implements ABlockReader {
     private long size;
-    
+
     public BlockRead(InputStream in, long size) {
       super(in);
       this.size = size;
     }
-    
+
     /**
      * Size is the size of the bytearray that was read form the cache
      */
     public long getRawSize() {
       return size;
     }
-    
+
     /**
      * It is intended that the caller of this method will close the stream we also only intend that this be called once per BlockRead. This method is provide
      * for methods up stream that expect to receive a DataInputStream object.
@@ -520,26 +522,26 @@ public class CachableBlockFile {
     public DataInputStream getStream() throws IOException {
       return this;
     }
-    
+
     @Override
     public boolean isIndexable() {
       return false;
     }
-    
+
     @Override
     public void seek(int position) {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public int getPosition() {
       throw new UnsupportedOperationException();
     }
-    
+
     @Override
     public <T> T getIndex(Class<T> clazz) {
       throw new UnsupportedOperationException();
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnFamilyFunctor.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnFamilyFunctor.java b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnFamilyFunctor.java
index e200b44..3660291 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnFamilyFunctor.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnFamilyFunctor.java
@@ -22,23 +22,23 @@ import org.apache.accumulo.core.data.Range;
 import org.apache.hadoop.util.bloom.Key;
 
 public class ColumnFamilyFunctor implements KeyFunctor {
-  
+
   public static final PartialKey kDepth = PartialKey.ROW_COLFAM;
-  
+
   @Override
   public Key transform(org.apache.accumulo.core.data.Key acuKey) {
-    
+
     byte keyData[];
-    
+
     ByteSequence row = acuKey.getRowData();
     ByteSequence cf = acuKey.getColumnFamilyData();
     keyData = new byte[row.length() + cf.length()];
     System.arraycopy(row.getBackingArray(), row.offset(), keyData, 0, row.length());
     System.arraycopy(cf.getBackingArray(), cf.offset(), keyData, row.length(), cf.length());
-    
+
     return new Key(keyData, 1.0);
   }
-  
+
   @Override
   public Key transform(Range range) {
     if (RowFunctor.isRangeInBloomFilter(range, PartialKey.ROW_COLFAM)) {
@@ -46,5 +46,5 @@ public class ColumnFamilyFunctor implements KeyFunctor {
     }
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnQualifierFunctor.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnQualifierFunctor.java b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnQualifierFunctor.java
index 7456486..d759a8a 100644
--- a/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnQualifierFunctor.java
+++ b/core/src/main/java/org/apache/accumulo/core/file/keyfunctor/ColumnQualifierFunctor.java
@@ -22,11 +22,11 @@ import org.apache.accumulo.core.data.Range;
 import org.apache.hadoop.util.bloom.Key;
 
 public class ColumnQualifierFunctor implements KeyFunctor {
-  
+
   @Override
   public org.apache.hadoop.util.bloom.Key transform(org.apache.accumulo.core.data.Key acuKey) {
     byte keyData[];
-    
+
     ByteSequence row = acuKey.getRowData();
     ByteSequence cf = acuKey.getColumnFamilyData();
     ByteSequence cq = acuKey.getColumnQualifierData();
@@ -34,10 +34,10 @@ public class ColumnQualifierFunctor implements KeyFunctor {
     System.arraycopy(row.getBackingArray(), row.offset(), keyData, 0, row.length());
     System.arraycopy(cf.getBackingArray(), cf.offset(), keyData, row.length(), cf.length());
     System.arraycopy(cq.getBackingArray(), cq.offset(), keyData, row.length() + cf.length(), cq.length());
-    
+
     return new org.apache.hadoop.util.bloom.Key(keyData, 1.0);
   }
-  
+
   @Override
   public Key transform(Range range) {
     if (RowFunctor.isRangeInBloomFilter(range, PartialKey.ROW_COLFAM_COLQUAL)) {
@@ -45,5 +45,5 @@ public class ColumnQualifierFunctor implements KeyFunctor {
     }
     return null;
   }
-  
+
 }


[07/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/TestIngest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/TestIngest.java b/test/src/main/java/org/apache/accumulo/test/TestIngest.java
index 47033f3..76bedf2 100644
--- a/test/src/main/java/org/apache/accumulo/test/TestIngest.java
+++ b/test/src/main/java/org/apache/accumulo/test/TestIngest.java
@@ -57,61 +57,62 @@ import org.apache.log4j.Logger;
 
 import com.beust.jcommander.Parameter;
 
-
 public class TestIngest {
   public static final Authorizations AUTHS = new Authorizations("L1", "L2", "G1", "GROUP2");
-  
+
   public static class Opts extends ClientOnDefaultTable {
-    
-    @Parameter(names="--createTable")
+
+    @Parameter(names = "--createTable")
     public boolean createTable = false;
-    
-    @Parameter(names="--splits", description="the number of splits to use when creating the table")
+
+    @Parameter(names = "--splits", description = "the number of splits to use when creating the table")
     public int numsplits = 1;
-    
-    @Parameter(names="--start", description="the starting row number")
+
+    @Parameter(names = "--start", description = "the starting row number")
     public int startRow = 0;
-    
-    @Parameter(names="--rows", description="the number of rows to ingest")
+
+    @Parameter(names = "--rows", description = "the number of rows to ingest")
     public int rows = 100000;
-    
-    @Parameter(names="--cols", description="the number of columns to ingest per row")
+
+    @Parameter(names = "--cols", description = "the number of columns to ingest per row")
     public int cols = 1;
-    
-    @Parameter(names="--random", description="insert random rows and use the given number to seed the psuedo-random number generator")
+
+    @Parameter(names = "--random", description = "insert random rows and use the given number to seed the psuedo-random number generator")
     public Integer random = null;
-    
-    @Parameter(names="--size", description="the size of the value to ingest")
+
+    @Parameter(names = "--size", description = "the size of the value to ingest")
     public int dataSize = 1000;
-    
-    @Parameter(names="--delete", description="delete values instead of inserting them")
+
+    @Parameter(names = "--delete", description = "delete values instead of inserting them")
     public boolean delete = false;
-    
-    @Parameter(names={"-ts", "--timestamp"}, description="timestamp to use for all values")
+
+    @Parameter(names = {"-ts", "--timestamp"}, description = "timestamp to use for all values")
     public long timestamp = -1;
-    
-    @Parameter(names="--rfile", description="generate data into a file that can be imported")
+
+    @Parameter(names = "--rfile", description = "generate data into a file that can be imported")
     public String outputFile = null;
-    
-    @Parameter(names="--stride", description="the difference between successive row ids")
+
+    @Parameter(names = "--stride", description = "the difference between successive row ids")
     public int stride;
 
-    @Parameter(names={"-cf","--columnFamily"}, description="place columns in this column family")
+    @Parameter(names = {"-cf", "--columnFamily"}, description = "place columns in this column family")
     public String columnFamily = "colf";
 
-    @Parameter(names={"-cv","--columnVisibility"}, description="place columns in this column family", converter=VisibilityConverter.class)
+    @Parameter(names = {"-cv", "--columnVisibility"}, description = "place columns in this column family", converter = VisibilityConverter.class)
     public ColumnVisibility columnVisibility = new ColumnVisibility();
-    
-    public Opts() { super("test_ingest"); }
+
+    public Opts() {
+      super("test_ingest");
+    }
   }
-  
+
   @SuppressWarnings("unused")
   private static final Logger log = Logger.getLogger(TestIngest.class);
-  
+
   public static void createTable(Connector conn, Opts args) throws AccumuloException, AccumuloSecurityException, TableExistsException {
     if (args.createTable) {
       TreeSet<Text> splits = getSplitPoints(args.startRow, args.startRow + args.rows, args.numsplits);
-      
+
       if (!conn.tableOperations().exists(args.getTableName()))
         conn.tableOperations().create(args.getTableName());
       try {
@@ -122,27 +123,27 @@ public class TestIngest {
       }
     }
   }
-  
+
   public static TreeSet<Text> getSplitPoints(long start, long end, long numsplits) {
     long splitSize = (end - start) / numsplits;
-    
+
     long pos = start + splitSize;
-    
+
     TreeSet<Text> splits = new TreeSet<Text>();
-    
+
     while (pos < end) {
       splits.add(new Text(String.format("row_%010d", pos)));
       pos += splitSize;
     }
     return splits;
   }
-  
+
   public static byte[][] generateValues(int dataSize) {
-    
+
     byte[][] bytevals = new byte[10][];
-    
+
     byte[] letters = {'1', '2', '3', '4', '5', '6', '7', '8', '9', '0'};
-    
+
     for (int i = 0; i < 10; i++) {
       bytevals[i] = new byte[dataSize];
       for (int j = 0; j < dataSize; j++)
@@ -150,46 +151,46 @@ public class TestIngest {
     }
     return bytevals;
   }
-  
+
   private static byte ROW_PREFIX[] = "row_".getBytes(UTF_8);
   private static byte COL_PREFIX[] = "col_".getBytes(UTF_8);
-  
+
   public static Text generateRow(int rowid, int startRow) {
     return new Text(FastFormat.toZeroPaddedString(rowid + startRow, 10, 10, ROW_PREFIX));
   }
-  
+
   public static byte[] genRandomValue(Random random, byte dest[], int seed, int row, int col) {
     random.setSeed((row ^ seed) ^ col);
     random.nextBytes(dest);
     toPrintableChars(dest);
-    
+
     return dest;
   }
-  
+
   public static void toPrintableChars(byte[] dest) {
     // transform to printable chars
     for (int i = 0; i < dest.length; i++) {
       dest[i] = (byte) (((0xff & dest[i]) % 92) + ' ');
     }
   }
-  
+
   public static void main(String[] args) throws Exception {
-    
+
     Opts opts = new Opts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(TestIngest.class.getName(), args, bwOpts);
 
     String name = TestIngest.class.getSimpleName();
     DistributedTrace.enable(name);
-    
+
     try {
       opts.startTracing(name);
-      
+
       if (opts.debug)
         Logger.getLogger(TabletServerBatchWriter.class.getName()).setLevel(Level.TRACE);
-      
+
       // test batch update
-      
+
       ingest(opts.getConnector(), opts, bwOpts);
     } catch (Exception e) {
       throw new RuntimeException(e);
@@ -199,34 +200,33 @@ public class TestIngest {
     }
   }
 
-  public static void ingest(Connector connector, Opts opts, BatchWriterOpts bwOpts) throws IOException, AccumuloException, AccumuloSecurityException, TableNotFoundException,
-      MutationsRejectedException, TableExistsException {
+  public static void ingest(Connector connector, Opts opts, BatchWriterOpts bwOpts) throws IOException, AccumuloException, AccumuloSecurityException,
+      TableNotFoundException, MutationsRejectedException, TableExistsException {
     long stopTime;
-    
+
     byte[][] bytevals = generateValues(opts.dataSize);
-    
+
     byte randomValue[] = new byte[opts.dataSize];
     Random random = new Random();
-    
+
     long bytesWritten = 0;
 
     createTable(connector, opts);
-    
+
     BatchWriter bw = null;
     FileSKVWriter writer = null;
-    
+
     if (opts.outputFile != null) {
       Configuration conf = CachedConfiguration.getInstance();
       FileSystem fs = FileSystem.get(conf);
-      writer = FileOperations.getInstance().openWriter(opts.outputFile + "." + RFile.EXTENSION, fs, conf,
-          AccumuloConfiguration.getDefaultConfiguration());
+      writer = FileOperations.getInstance().openWriter(opts.outputFile + "." + RFile.EXTENSION, fs, conf, AccumuloConfiguration.getDefaultConfiguration());
       writer.startDefaultLocalityGroup();
     } else {
       bw = connector.createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig());
       connector.securityOperations().changeUserAuthorizations(opts.principal, AUTHS);
     }
     Text labBA = new Text(opts.columnVisibility.getExpression());
-    
+
     long startTime = System.currentTimeMillis();
     for (int i = 0; i < opts.rows; i++) {
       int rowid;
@@ -235,13 +235,13 @@ public class TestIngest {
       } else {
         rowid = i;
       }
-      
+
       Text row = generateRow(rowid, opts.startRow);
       Mutation m = new Mutation(row);
       for (int j = 0; j < opts.cols; j++) {
         Text colf = new Text(opts.columnFamily);
         Text colq = new Text(FastFormat.toZeroPaddedString(j, 7, 10, COL_PREFIX));
-        
+
         if (writer != null) {
           Key key = new Key(row, colf, colq, labBA);
           if (opts.timestamp >= 0) {
@@ -249,15 +249,15 @@ public class TestIngest {
           } else {
             key.setTimestamp(startTime);
           }
-          
+
           if (opts.delete) {
             key.setDeleted(true);
           } else {
             key.setDeleted(false);
           }
-          
+
           bytesWritten += key.getSize();
-          
+
           if (opts.delete) {
             writer.append(key, new Value(new byte[0]));
           } else {
@@ -267,16 +267,16 @@ public class TestIngest {
             } else {
               value = bytevals[j % bytevals.length];
             }
-            
+
             Value v = new Value(value);
             writer.append(key, v);
             bytesWritten += v.getSize();
           }
-          
+
         } else {
           Key key = new Key(row, colf, colq, labBA);
           bytesWritten += key.getSize();
-          
+
           if (opts.delete) {
             if (opts.timestamp >= 0)
               m.putDelete(colf, colq, opts.columnVisibility, opts.timestamp);
@@ -290,22 +290,22 @@ public class TestIngest {
               value = bytevals[j % bytevals.length];
             }
             bytesWritten += value.length;
-            
+
             if (opts.timestamp >= 0) {
               m.put(colf, colq, opts.columnVisibility, opts.timestamp, new Value(value, true));
             } else {
               m.put(colf, colq, opts.columnVisibility, new Value(value, true));
-              
+
             }
           }
         }
-        
+
       }
       if (bw != null)
         bw.addMutation(m);
-      
+
     }
-    
+
     if (writer != null) {
       writer.close();
     } else if (bw != null) {
@@ -317,22 +317,22 @@ public class TestIngest {
             System.err.println("ERROR : Not authorized to write to : " + entry.getKey() + " due to " + entry.getValue());
           }
         }
-        
+
         if (e.getConstraintViolationSummaries().size() > 0) {
           for (ConstraintViolationSummary cvs : e.getConstraintViolationSummaries()) {
             System.err.println("ERROR : Constraint violates : " + cvs);
           }
         }
-        
+
         throw e;
       }
     }
-    
+
     stopTime = System.currentTimeMillis();
-    
+
     int totalValues = opts.rows * opts.cols;
     double elapsed = (stopTime - startTime) / 1000.0;
-    
+
     System.out.printf("%,12d records written | %,8d records/sec | %,12d bytes written | %,8d bytes/sec | %6.3f secs   %n", totalValues,
         (int) (totalValues / elapsed), bytesWritten, (int) (bytesWritten / elapsed), elapsed);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/TestRandomDeletes.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/TestRandomDeletes.java b/test/src/main/java/org/apache/accumulo/test/TestRandomDeletes.java
index 1b553f4..ba5874d 100644
--- a/test/src/main/java/org/apache/accumulo/test/TestRandomDeletes.java
+++ b/test/src/main/java/org/apache/accumulo/test/TestRandomDeletes.java
@@ -40,30 +40,30 @@ import org.apache.log4j.Logger;
 public class TestRandomDeletes {
   private static final Logger log = Logger.getLogger(TestRandomDeletes.class);
   private static Authorizations auths = new Authorizations("L1", "L2", "G1", "GROUP2");
-  
+
   static private class RowColumn implements Comparable<RowColumn> {
     Text row;
     Column column;
     long timestamp;
-    
+
     public RowColumn(Text row, Column column, long timestamp) {
       this.row = row;
       this.column = column;
       this.timestamp = timestamp;
     }
-    
+
     public int compareTo(RowColumn other) {
       int result = row.compareTo(other.row);
       if (result != 0)
         return result;
       return column.compareTo(other.column);
     }
-    
+
     public String toString() {
       return row.toString() + ":" + column.toString();
     }
   }
-  
+
   private static TreeSet<RowColumn> scanAll(ClientOnDefaultTable opts, ScannerOpts scanOpts, String tableName) throws Exception {
     TreeSet<RowColumn> result = new TreeSet<RowColumn>();
     Connector conn = opts.getConnector();
@@ -77,26 +77,28 @@ public class TestRandomDeletes {
     }
     return result;
   }
-  
-  private static long scrambleDeleteHalfAndCheck(ClientOnDefaultTable opts, ScannerOpts scanOpts, BatchWriterOpts bwOpts, String tableName, Set<RowColumn> rows) throws Exception {
+
+  private static long scrambleDeleteHalfAndCheck(ClientOnDefaultTable opts, ScannerOpts scanOpts, BatchWriterOpts bwOpts, String tableName, Set<RowColumn> rows)
+      throws Exception {
     int result = 0;
     ArrayList<RowColumn> entries = new ArrayList<RowColumn>(rows);
     java.util.Collections.shuffle(entries);
-    
+
     Connector connector = opts.getConnector();
     BatchWriter mutations = connector.createBatchWriter(tableName, bwOpts.getBatchWriterConfig());
-    
+
     for (int i = 0; i < (entries.size() + 1) / 2; i++) {
       RowColumn rc = entries.get(i);
       Mutation m = new Mutation(rc.row);
-      m.putDelete(new Text(rc.column.columnFamily), new Text(rc.column.columnQualifier), new ColumnVisibility(rc.column.getColumnVisibility()), rc.timestamp + 1);
+      m.putDelete(new Text(rc.column.columnFamily), new Text(rc.column.columnQualifier), new ColumnVisibility(rc.column.getColumnVisibility()),
+          rc.timestamp + 1);
       mutations.addMutation(m);
       rows.remove(rc);
       result++;
     }
-    
+
     mutations.close();
-    
+
     Set<RowColumn> current = scanAll(opts, scanOpts, tableName);
     current.removeAll(rows);
     if (current.size() > 0) {
@@ -104,25 +106,24 @@ public class TestRandomDeletes {
     }
     return result;
   }
-  
+
   static public void main(String[] args) {
-    
+
     ClientOnDefaultTable opts = new ClientOnDefaultTable("test_ingest");
     ScannerOpts scanOpts = new ScannerOpts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(TestRandomDeletes.class.getName(), args, scanOpts, bwOpts);
-    
+
     log.info("starting random delete test");
 
-    
     try {
       long deleted = 0;
-      
+
       String tableName = opts.getTableName();
-      
+
       TreeSet<RowColumn> doomed = scanAll(opts, scanOpts, tableName);
       log.info("Got " + doomed.size() + " rows");
-      
+
       long startTime = System.currentTimeMillis();
       while (true) {
         long half = scrambleDeleteHalfAndCheck(opts, scanOpts, bwOpts, tableName, doomed);
@@ -131,7 +132,7 @@ public class TestRandomDeletes {
           break;
       }
       long stopTime = System.currentTimeMillis();
-      
+
       long elapsed = (stopTime - startTime) / 1000;
       log.info("deleted " + deleted + " values in " + elapsed + " seconds");
     } catch (Exception e) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/VerifyIngest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/VerifyIngest.java b/test/src/main/java/org/apache/accumulo/test/VerifyIngest.java
index 02eead2..8717c26 100644
--- a/test/src/main/java/org/apache/accumulo/test/VerifyIngest.java
+++ b/test/src/main/java/org/apache/accumulo/test/VerifyIngest.java
@@ -34,7 +34,6 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.trace.DistributedTrace;
 import org.apache.accumulo.core.trace.Trace;
-
 import org.apache.hadoop.io.Text;
 import org.apache.log4j.Logger;
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/WrongTabletTest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/WrongTabletTest.java b/test/src/main/java/org/apache/accumulo/test/WrongTabletTest.java
index e4abed3..31e7b06 100644
--- a/test/src/main/java/org/apache/accumulo/test/WrongTabletTest.java
+++ b/test/src/main/java/org/apache/accumulo/test/WrongTabletTest.java
@@ -58,8 +58,8 @@ public class WrongTabletTest {
 
       Mutation mutation = new Mutation(new Text("row_0003750001"));
       mutation.putDelete(new Text("colf"), new Text("colq"));
-      client.update(Tracer.traceInfo(), context.rpcCreds(), new KeyExtent(new Text("!!"), null,
-          new Text("row_0003750000")).toThrift(), mutation.toThrift(), TDurability.DEFAULT);
+      client.update(Tracer.traceInfo(), context.rpcCreds(), new KeyExtent(new Text("!!"), null, new Text("row_0003750000")).toThrift(), mutation.toThrift(),
+          TDurability.DEFAULT);
     } catch (Exception e) {
       throw new RuntimeException(e);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousBatchWalker.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousBatchWalker.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousBatchWalker.java
index 1a74962..a2687bb 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousBatchWalker.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousBatchWalker.java
@@ -43,81 +43,81 @@ import com.beust.jcommander.validators.PositiveInteger;
 public class ContinuousBatchWalker {
 
   static class Opts extends ContinuousWalk.Opts {
-    @Parameter(names="--numToScan", description="Number rows to scan between sleeps", required=true, validateWith=PositiveInteger.class)
+    @Parameter(names = "--numToScan", description = "Number rows to scan between sleeps", required = true, validateWith = PositiveInteger.class)
     long numToScan = 0;
   }
 
   public static void main(String[] args) throws Exception {
-    
+
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     BatchScannerOpts bsOpts = new BatchScannerOpts();
     opts.parseArgs(ContinuousBatchWalker.class.getName(), args, scanOpts, bsOpts);
-    
+
     Random r = new Random();
     Authorizations auths = opts.randomAuths.getAuths(r);
 
     Connector conn = opts.getConnector();
     Scanner scanner = ContinuousUtil.createScanner(conn, opts.getTableName(), auths);
     scanner.setBatchSize(scanOpts.scanBatchSize);
-    
+
     BatchScanner bs = conn.createBatchScanner(opts.getTableName(), auths, bsOpts.scanThreads);
     bs.setTimeout(bsOpts.scanTimeout, TimeUnit.MILLISECONDS);
 
     while (true) {
       Set<Text> batch = getBatch(scanner, opts.min, opts.max, scanOpts.scanBatchSize, r);
       List<Range> ranges = new ArrayList<Range>(batch.size());
-      
+
       for (Text row : batch) {
         ranges.add(new Range(row));
       }
-      
+
       runBatchScan(scanOpts.scanBatchSize, bs, batch, ranges);
-      
+
       UtilWaitThread.sleep(opts.sleepTime);
     }
-    
+
   }
-  
+
   /*
    * private static void runSequentialScan(Scanner scanner, List<Range> ranges) { Set<Text> srowsSeen = new HashSet<Text>(); long st1 =
    * System.currentTimeMillis(); int scount = 0; for (Range range : ranges) { scanner.setRange(range);
-   * 
+   *
    * for (Entry<Key,Value> entry : scanner) { srowsSeen.add(entry.getKey().getRow()); scount++; } }
-   * 
-   * 
+   *
+   *
    * long st2 = System.currentTimeMillis(); System.out.println("SRQ "+(st2 - st1)+" "+srowsSeen.size() +" "+scount); }
    */
-  
+
   private static void runBatchScan(int batchSize, BatchScanner bs, Set<Text> batch, List<Range> ranges) {
     bs.setRanges(ranges);
-    
+
     Set<Text> rowsSeen = new HashSet<Text>();
-    
+
     int count = 0;
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     for (Entry<Key,Value> entry : bs) {
       ContinuousWalk.validate(entry.getKey(), entry.getValue());
-      
+
       rowsSeen.add(entry.getKey().getRow());
-      
+
       addRow(batchSize, entry.getValue());
-      
+
       count++;
     }
     bs.close();
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     if (!rowsSeen.equals(batch)) {
       HashSet<Text> copy1 = new HashSet<Text>(rowsSeen);
       HashSet<Text> copy2 = new HashSet<Text>(batch);
-      
+
       copy1.removeAll(batch);
       copy2.removeAll(rowsSeen);
-      
+
       System.out.printf("DIF %d %d %d%n", t1, copy1.size(), copy2.size());
       System.err.printf("DIF %d %d %d%n", t1, copy1.size(), copy2.size());
       System.err.println("Extra seen : " + copy1);
@@ -125,12 +125,12 @@ public class ContinuousBatchWalker {
     } else {
       System.out.printf("BRQ %d %d %d %d %d%n", t1, (t2 - t1), rowsSeen.size(), count, (int) (rowsSeen.size() / ((t2 - t1) / 1000.0)));
     }
-    
+
   }
-  
+
   private static void addRow(int batchSize, Value v) {
     byte[] val = v.get();
-    
+
     int offset = ContinuousWalk.getPrevRowOffset(val);
     if (offset > 1) {
       Text prevRow = new Text();
@@ -140,19 +140,19 @@ public class ContinuousBatchWalker {
       }
     }
   }
-  
+
   private static HashSet<Text> rowsToQuery = new HashSet<Text>();
-  
+
   private static Set<Text> getBatch(Scanner scanner, long min, long max, int batchSize, Random r) {
-    
+
     while (rowsToQuery.size() < batchSize) {
       byte[] scanStart = ContinuousIngest.genRow(min, max, r);
       scanner.setRange(new Range(new Text(scanStart), null));
-      
+
       int count = 0;
-      
+
       long t1 = System.currentTimeMillis();
-      
+
       Iterator<Entry<Key,Value>> iter = scanner.iterator();
       while (iter.hasNext() && rowsToQuery.size() < 3 * batchSize) {
         Entry<Key,Value> entry = iter.next();
@@ -160,24 +160,24 @@ public class ContinuousBatchWalker {
         addRow(batchSize, entry.getValue());
         count++;
       }
-      
+
       long t2 = System.currentTimeMillis();
-      
+
       System.out.println("FSB " + t1 + " " + (t2 - t1) + " " + count);
-      
+
       UtilWaitThread.sleep(100);
     }
-    
+
     HashSet<Text> ret = new HashSet<Text>();
-    
+
     Iterator<Text> iter = rowsToQuery.iterator();
-    
+
     for (int i = 0; i < batchSize; i++) {
       ret.add(iter.next());
       iter.remove();
     }
-    
+
     return ret;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousIngest.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousIngest.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousIngest.java
index f54b8db..dba6ac9 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousIngest.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousIngest.java
@@ -54,9 +54,8 @@ import org.apache.log4j.PatternLayout;
 import com.beust.jcommander.IStringConverter;
 import com.beust.jcommander.Parameter;
 
-
 public class ContinuousIngest {
-  
+
   static public class BaseOpts extends MapReduceClientOnDefaultTable {
     public class DebugConverter implements IStringConverter<String> {
       @Override
@@ -72,64 +71,66 @@ public class ContinuousIngest {
         return debugLog;
       }
     }
-    
-    @Parameter(names="--min", description="lowest random row number to use")
+
+    @Parameter(names = "--min", description = "lowest random row number to use")
     long min = 0;
-    
-    @Parameter(names="--max", description="maximum random row number to use")
+
+    @Parameter(names = "--max", description = "maximum random row number to use")
     long max = Long.MAX_VALUE;
-    
-    @Parameter(names="--debugLog", description="file to write debugging output", converter=DebugConverter.class)
+
+    @Parameter(names = "--debugLog", description = "file to write debugging output", converter = DebugConverter.class)
     String debugLog = null;
 
-    BaseOpts() { super("ci"); }
+    BaseOpts() {
+      super("ci");
+    }
   }
-  
+
   public static class ShortConverter implements IStringConverter<Short> {
     @Override
     public Short convert(String value) {
       return Short.valueOf(value);
     }
   }
-  
+
   static public class Opts extends BaseOpts {
-    @Parameter(names="--num", description="the number of entries to ingest")
+    @Parameter(names = "--num", description = "the number of entries to ingest")
     long num = Long.MAX_VALUE;
-    
-    @Parameter(names="--maxColF", description="maximum column family value to use", converter=ShortConverter.class)
+
+    @Parameter(names = "--maxColF", description = "maximum column family value to use", converter = ShortConverter.class)
     short maxColF = Short.MAX_VALUE;
-    
-    @Parameter(names="--maxColQ", description="maximum column qualifier value to use", converter=ShortConverter.class)
+
+    @Parameter(names = "--maxColQ", description = "maximum column qualifier value to use", converter = ShortConverter.class)
     short maxColQ = Short.MAX_VALUE;
- 
-    @Parameter(names="--addCheckSum", description="turn on checksums")
+
+    @Parameter(names = "--addCheckSum", description = "turn on checksums")
     boolean checksum = false;
-    
-    @Parameter(names="--visibilities", description="read the visibilities to ingest with from a file")
+
+    @Parameter(names = "--visibilities", description = "read the visibilities to ingest with from a file")
     String visFile = null;
   }
-  
+
   private static final byte[] EMPTY_BYTES = new byte[0];
-  
+
   private static List<ColumnVisibility> visibilities;
-  
+
   private static void initVisibilities(Opts opts) throws Exception {
     if (opts.visFile == null) {
       visibilities = Collections.singletonList(new ColumnVisibility());
       return;
     }
-    
+
     visibilities = new ArrayList<ColumnVisibility>();
-    
+
     FileSystem fs = FileSystem.get(new Configuration());
     BufferedReader in = new BufferedReader(new InputStreamReader(fs.open(new Path(opts.visFile)), UTF_8));
-    
+
     String line;
-    
+
     while ((line = in.readLine()) != null) {
       visibilities.add(new ColumnVisibility(line));
     }
-    
+
     in.close();
   }
 
@@ -138,35 +139,35 @@ public class ContinuousIngest {
   }
 
   public static void main(String[] args) throws Exception {
-    
+
     Opts opts = new Opts();
     BatchWriterOpts bwOpts = new BatchWriterOpts();
     opts.parseArgs(ContinuousIngest.class.getName(), args, bwOpts);
-    
+
     initVisibilities(opts);
 
     if (opts.min < 0 || opts.max < 0 || opts.max <= opts.min) {
       throw new IllegalArgumentException("bad min and max");
     }
     Connector conn = opts.getConnector();
-    
+
     if (!conn.tableOperations().exists(opts.getTableName())) {
       throw new TableNotFoundException(null, opts.getTableName(), "Consult the README and create the table before starting ingest.");
     }
 
     BatchWriter bw = conn.createBatchWriter(opts.getTableName(), bwOpts.getBatchWriterConfig());
     bw = Trace.wrapAll(bw, new CountSampler(1024));
-    
+
     Random r = new Random();
-    
+
     byte[] ingestInstanceId = UUID.randomUUID().toString().getBytes(UTF_8);
-    
+
     System.out.printf("UUID %d %s%n", System.currentTimeMillis(), new String(ingestInstanceId, UTF_8));
-    
+
     long count = 0;
     final int flushInterval = 1000000;
     final int maxDepth = 25;
-    
+
     // always want to point back to flushed data. This way the previous item should
     // always exist in accumulo when verifying data. To do this make insert N point
     // back to the row from insert (N - flushInterval). The array below is used to keep
@@ -175,9 +176,9 @@ public class ContinuousIngest {
     long firstRows[] = new long[flushInterval];
     int firstColFams[] = new int[flushInterval];
     int firstColQuals[] = new int[flushInterval];
-    
+
     long lastFlushTime = System.currentTimeMillis();
-    
+
     out: while (true) {
       // generate first set of nodes
       ColumnVisibility cv = getVisibility(r);
@@ -186,22 +187,22 @@ public class ContinuousIngest {
         long rowLong = genLong(opts.min, opts.max, r);
         prevRows[index] = rowLong;
         firstRows[index] = rowLong;
-        
+
         int cf = r.nextInt(opts.maxColF);
         int cq = r.nextInt(opts.maxColQ);
-        
+
         firstColFams[index] = cf;
         firstColQuals[index] = cq;
-        
+
         Mutation m = genMutation(rowLong, cf, cq, cv, ingestInstanceId, count, null, r, opts.checksum);
         count++;
         bw.addMutation(m);
       }
-      
+
       lastFlushTime = flush(bw, count, flushInterval, lastFlushTime);
       if (count >= opts.num)
         break out;
-      
+
       // generate subsequent sets of nodes that link to previous set of nodes
       for (int depth = 1; depth < maxDepth; depth++) {
         for (int index = 0; index < flushInterval; index++) {
@@ -212,12 +213,12 @@ public class ContinuousIngest {
           count++;
           bw.addMutation(m);
         }
-        
+
         lastFlushTime = flush(bw, count, flushInterval, lastFlushTime);
         if (count >= opts.num)
           break out;
       }
-      
+
       // create one big linked list, this makes all of the first inserts
       // point to something
       for (int index = 0; index < flushInterval - 1; index++) {
@@ -230,7 +231,7 @@ public class ContinuousIngest {
       if (count >= opts.num)
         break out;
     }
-    
+
     bw.close();
     opts.stopTracing();
   }
@@ -243,17 +244,17 @@ public class ContinuousIngest {
     lastFlushTime = t2;
     return lastFlushTime;
   }
-  
+
   public static Mutation genMutation(long rowLong, int cfInt, int cqInt, ColumnVisibility cv, byte[] ingestInstanceId, long count, byte[] prevRow, Random r,
       boolean checksum) {
     // Adler32 is supposed to be faster, but according to wikipedia is not good for small data.... so used CRC32 instead
     CRC32 cksum = null;
-    
+
     byte[] rowString = genRow(rowLong);
-    
+
     byte[] cfString = FastFormat.toZeroPaddedString(cfInt, 4, 16, EMPTY_BYTES);
     byte[] cqString = FastFormat.toZeroPaddedString(cqInt, 4, 16, EMPTY_BYTES);
-    
+
     if (checksum) {
       cksum = new CRC32();
       cksum.update(rowString);
@@ -261,25 +262,25 @@ public class ContinuousIngest {
       cksum.update(cqString);
       cksum.update(cv.getExpression());
     }
-    
+
     Mutation m = new Mutation(new Text(rowString));
-    
+
     m.put(new Text(cfString), new Text(cqString), cv, createValue(ingestInstanceId, count, prevRow, cksum));
     return m;
   }
-  
+
   public static final long genLong(long min, long max, Random r) {
     return ((r.nextLong() & 0x7fffffffffffffffl) % (max - min)) + min;
   }
-  
+
   static final byte[] genRow(long min, long max, Random r) {
     return genRow(genLong(min, max, r));
   }
-  
+
   static final byte[] genRow(long rowLong) {
     return FastFormat.toZeroPaddedString(rowLong, 16, 16, EMPTY_BYTES);
   }
-  
+
   private static Value createValue(byte[] ingestInstanceId, long count, byte[] prevRow, Checksum cksum) {
     int dataLen = ingestInstanceId.length + 16 + (prevRow == null ? 0 : prevRow.length) + 3;
     if (cksum != null)
@@ -297,17 +298,17 @@ public class ContinuousIngest {
       System.arraycopy(prevRow, 0, val, index, prevRow.length);
       index += prevRow.length;
     }
-    
+
     val[index++] = ':';
-    
+
     if (cksum != null) {
       cksum.update(val, 0, index);
       cksum.getValue();
       FastFormat.toZeroPaddedString(val, index, cksum.getValue(), 8, 16, EMPTY_BYTES);
     }
-    
+
     // System.out.println("val "+new String(val));
-    
+
     return new Value(val);
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousMoru.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousMoru.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousMoru.java
index 797413f..89ff515 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousMoru.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousMoru.java
@@ -50,7 +50,7 @@ import com.beust.jcommander.validators.PositiveInteger;
 /**
  * A map only job that reads a table created by continuous ingest and creates doubly linked list. This map reduce job tests the ability of a map only job to
  * read and write to accumulo at the same time. This map reduce job mutates the table in such a way that it should not create any undefined nodes.
- * 
+ *
  */
 public class ContinuousMoru extends Configured implements Tool {
   private static final String PREFIX = ContinuousMoru.class.getSimpleName() + ".";
@@ -59,49 +59,49 @@ public class ContinuousMoru extends Configured implements Tool {
   private static final String MAX = PREFIX + "MAX";
   private static final String MIN = PREFIX + "MIN";
   private static final String CI_ID = PREFIX + "CI_ID";
-  
+
   static enum Counts {
     SELF_READ;
   }
-  
+
   public static class CMapper extends Mapper<Key,Value,Text,Mutation> {
-    
+
     private short max_cf;
     private short max_cq;
     private Random random;
     private String ingestInstanceId;
     private byte[] iiId;
     private long count;
-    
+
     private static final ColumnVisibility EMPTY_VIS = new ColumnVisibility();
-    
+
     @Override
     public void setup(Context context) throws IOException, InterruptedException {
       int max_cf = context.getConfiguration().getInt(MAX_CF, -1);
       int max_cq = context.getConfiguration().getInt(MAX_CQ, -1);
-      
+
       if (max_cf > Short.MAX_VALUE || max_cq > Short.MAX_VALUE)
         throw new IllegalArgumentException();
-      
+
       this.max_cf = (short) max_cf;
       this.max_cq = (short) max_cq;
-      
+
       random = new Random();
       ingestInstanceId = context.getConfiguration().get(CI_ID);
       iiId = ingestInstanceId.getBytes(UTF_8);
-      
+
       count = 0;
     }
-    
+
     @Override
     public void map(Key key, Value data, Context context) throws IOException, InterruptedException {
-      
+
       ContinuousWalk.validate(key, data);
-      
+
       if (WritableComparator.compareBytes(iiId, 0, iiId.length, data.get(), 0, iiId.length) != 0) {
         // only rewrite data not written by this M/R job
         byte[] val = data.get();
-        
+
         int offset = ContinuousWalk.getPrevRowOffset(val);
         if (offset > 0) {
           long rowLong = Long.parseLong(new String(val, offset, 16, UTF_8), 16);
@@ -109,24 +109,24 @@ public class ContinuousMoru extends Configured implements Tool {
               .toArray(), random, true);
           context.write(null, m);
         }
-        
+
       } else {
         ContinuousVerify.increment(context.getCounter(Counts.SELF_READ));
       }
     }
   }
-  
+
   static class Opts extends BaseOpts {
-    @Parameter(names = "--maxColF", description = "maximum column family value to use", converter=ShortConverter.class)
+    @Parameter(names = "--maxColF", description = "maximum column family value to use", converter = ShortConverter.class)
     short maxColF = Short.MAX_VALUE;
-    
-    @Parameter(names = "--maxColQ", description = "maximum column qualifier value to use", converter=ShortConverter.class)
+
+    @Parameter(names = "--maxColQ", description = "maximum column qualifier value to use", converter = ShortConverter.class)
     short maxColQ = Short.MAX_VALUE;
-    
+
     @Parameter(names = "--maxMappers", description = "the maximum number of mappers to use", required = true, validateWith = PositiveInteger.class)
     int maxMaps = 0;
   }
-  
+
   @Override
   public int run(String[] args) throws IOException, InterruptedException, ClassNotFoundException, AccumuloSecurityException {
     Opts opts = new Opts();
@@ -136,10 +136,10 @@ public class ContinuousMoru extends Configured implements Tool {
     @SuppressWarnings("deprecation")
     Job job = new Job(getConf(), this.getClass().getSimpleName() + "_" + System.currentTimeMillis());
     job.setJarByClass(this.getClass());
-    
+
     job.setInputFormatClass(AccumuloInputFormat.class);
     opts.setAccumuloConfigs(job);
-    
+
     // set up ranges
     try {
       Set<Range> ranges = opts.getConnector().tableOperations().splitRangeByTablets(opts.getTableName(), new Range(), opts.maxMaps);
@@ -148,28 +148,28 @@ public class ContinuousMoru extends Configured implements Tool {
     } catch (Exception e) {
       throw new IOException(e);
     }
-    
+
     job.setMapperClass(CMapper.class);
-    
+
     job.setNumReduceTasks(0);
-    
+
     job.setOutputFormatClass(AccumuloOutputFormat.class);
     AccumuloOutputFormat.setBatchWriterOptions(job, bwOpts.getBatchWriterConfig());
-    
+
     Configuration conf = job.getConfiguration();
     conf.setLong(MIN, opts.min);
     conf.setLong(MAX, opts.max);
     conf.setInt(MAX_CF, opts.maxColF);
     conf.setInt(MAX_CQ, opts.maxColQ);
     conf.set(CI_ID, UUID.randomUUID().toString());
-    
+
     job.waitForCompletion(true);
     opts.stopTracing();
     return job.isSuccessful() ? 0 : 1;
   }
-  
+
   /**
-   * 
+   *
    * @param args
    *          instanceName zookeepers username password table columns outputpath
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousQuery.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousQuery.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousQuery.java
index dcc3e49..73048f6 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousQuery.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousQuery.java
@@ -33,28 +33,28 @@ import org.apache.hadoop.io.Text;
 import com.beust.jcommander.Parameter;
 
 public class ContinuousQuery {
-  
+
   public static class Opts extends BaseOpts {
-    @Parameter(names="--sleep", description="the time to wait between queries", converter=TimeConverter.class)
+    @Parameter(names = "--sleep", description = "the time to wait between queries", converter = TimeConverter.class)
     long sleepTime = 100;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(ContinuousQuery.class.getName(), args, scanOpts);
-    
+
     Connector conn = opts.getConnector();
     Scanner scanner = ContinuousUtil.createScanner(conn, opts.getTableName(), opts.auths);
     scanner.setBatchSize(scanOpts.scanBatchSize);
-    
+
     Random r = new Random();
-    
+
     while (true) {
       byte[] row = ContinuousIngest.genRow(opts.min, opts.max, r);
-      
+
       int count = 0;
-      
+
       long t1 = System.currentTimeMillis();
       scanner.setRange(new Range(new Text(row)));
       for (Entry<Key,Value> entry : scanner) {
@@ -62,9 +62,9 @@ public class ContinuousQuery {
         count++;
       }
       long t2 = System.currentTimeMillis();
-      
+
       System.out.printf("SRQ %d %s %d %d%n", t1, new String(row, UTF_8), (t2 - t1), count);
-      
+
       if (opts.sleepTime > 0)
         Thread.sleep(opts.sleepTime);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousScanner.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousScanner.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousScanner.java
index 60154df..f68377a 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousScanner.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousScanner.java
@@ -36,50 +36,50 @@ import com.beust.jcommander.Parameter;
 import com.beust.jcommander.validators.PositiveInteger;
 
 public class ContinuousScanner {
-  
+
   static class Opts extends ContinuousWalk.Opts {
-    @Parameter(names="--numToScan", description="Number rows to scan between sleeps", required=true, validateWith=PositiveInteger.class)
+    @Parameter(names = "--numToScan", description = "Number rows to scan between sleeps", required = true, validateWith = PositiveInteger.class)
     long numToScan = 0;
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(ContinuousScanner.class.getName(), args, scanOpts);
-    
+
     Random r = new Random();
 
     long distance = 1000000000000l;
-    
+
     Connector conn = opts.getConnector();
     Authorizations auths = opts.randomAuths.getAuths(r);
     Scanner scanner = ContinuousUtil.createScanner(conn, opts.getTableName(), auths);
     scanner.setBatchSize(scanOpts.scanBatchSize);
-    
+
     double delta = Math.min(.05, .05 / (opts.numToScan / 1000.0));
-    
+
     while (true) {
       long startRow = ContinuousIngest.genLong(opts.min, opts.max - distance, r);
       byte[] scanStart = ContinuousIngest.genRow(startRow);
       byte[] scanStop = ContinuousIngest.genRow(startRow + distance);
-      
+
       scanner.setRange(new Range(new Text(scanStart), new Text(scanStop)));
-      
+
       int count = 0;
       Iterator<Entry<Key,Value>> iter = scanner.iterator();
-      
+
       long t1 = System.currentTimeMillis();
-      
+
       while (iter.hasNext()) {
         Entry<Key,Value> entry = iter.next();
         ContinuousWalk.validate(entry.getKey(), entry.getValue());
         count++;
       }
-      
+
       long t2 = System.currentTimeMillis();
-      
+
       // System.out.println("P1 " +count +" "+((1-delta) * numToScan)+" "+((1+delta) * numToScan)+" "+numToScan);
-      
+
       if (count < (1 - delta) * opts.numToScan || count > (1 + delta) * opts.numToScan) {
         if (count == 0) {
           distance = distance * 10;
@@ -91,15 +91,15 @@ public class ContinuousScanner {
           ratio = ratio - (ratio - 1.0) * (2.0 / 3.0);
           distance = (long) (ratio * distance);
         }
-        
+
         // System.out.println("P2 "+delta +" "+numToScan+" "+distance+"  "+((double)numToScan/count ));
       }
-      
+
       System.out.printf("SCN %d %s %d %d%n", t1, new String(scanStart, UTF_8), (t2 - t1), count);
-      
+
       if (opts.sleepTime > 0)
         UtilWaitThread.sleep(opts.sleepTime);
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousStatsCollector.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousStatsCollector.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousStatsCollector.java
index 1e3b636..7c2f93b 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousStatsCollector.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousStatsCollector.java
@@ -58,15 +58,15 @@ import org.apache.hadoop.mapred.JobClient;
 import org.apache.log4j.Logger;
 
 public class ContinuousStatsCollector {
-  
+
   private static final Logger log = Logger.getLogger(ContinuousStatsCollector.class);
-  
+
   static class StatsCollectionTask extends TimerTask {
-    
+
     private final String tableId;
     private final Opts opts;
     private final int scanBatchSize;
-    
+
     public StatsCollectionTask(Opts opts, int scanBatchSize) {
       this.opts = opts;
       this.scanBatchSize = scanBatchSize;
@@ -76,7 +76,7 @@ public class ContinuousStatsCollector {
               + " ACCUMULO_DU ACCUMULO_DIRS ACCUMULO_FILES TABLE_DU TABLE_DIRS TABLE_FILES"
               + " MAP_TASK MAX_MAP_TASK REDUCE_TASK MAX_REDUCE_TASK TASK_TRACKERS BLACK_LISTED MIN_FILES/TABLET MAX_FILES/TABLET AVG_FILES/TABLET STDDEV_FILES/TABLET");
     }
-    
+
     @Override
     public void run() {
       try {
@@ -84,37 +84,37 @@ public class ContinuousStatsCollector {
         String fsStats = getFSStats();
         String mrStats = getMRStats();
         String tabletStats = getTabletStats();
-        
+
         System.out.println(System.currentTimeMillis() + " " + acuStats + " " + fsStats + " " + mrStats + " " + tabletStats);
       } catch (Exception e) {
-        log.error(System.currentTimeMillis()+" - Failed to collect stats", e);
+        log.error(System.currentTimeMillis() + " - Failed to collect stats", e);
       }
     }
-    
+
     private String getTabletStats() throws Exception {
-      
+
       Connector conn = opts.getConnector();
       Scanner scanner = conn.createScanner(MetadataTable.NAME, opts.auths);
       scanner.setBatchSize(scanBatchSize);
       scanner.fetchColumnFamily(DataFileColumnFamily.NAME);
       scanner.addScanIterator(new IteratorSetting(1000, "cfc", ColumnFamilyCounter.class.getName()));
       scanner.setRange(new KeyExtent(new Text(tableId), null, null).toMetadataRange());
-      
+
       Stat s = new Stat();
-      
+
       int count = 0;
       for (Entry<Key,Value> entry : scanner) {
         count++;
         s.addStat(Long.parseLong(entry.getValue().toString()));
       }
-      
+
       if (count > 0)
         return String.format("%d %d %.3f %.3f", s.getMin(), s.getMax(), s.getAverage(), s.getStdDev());
       else
         return "0 0 0 0";
-      
+
     }
-    
+
     private String getFSStats() throws Exception {
       VolumeManager fs = VolumeManagerImpl.get();
       long length1 = 0, dcount1 = 0, fcount1 = 0;
@@ -129,22 +129,22 @@ public class ContinuousStatsCollector {
         dcount2 += contentSummary.getDirectoryCount();
         fcount2 += contentSummary.getFileCount();
       }
-      
+
       return "" + length1 + " " + dcount1 + " " + fcount1 + " " + length2 + " " + dcount2 + " " + fcount2;
     }
-    
+
     private String getACUStats() throws Exception {
-      
+
       MasterClientService.Iface client = null;
       try {
         ClientContext context = new ClientContext(opts.getInstance(), new Credentials(opts.principal, opts.getToken()), new ServerConfigurationFactory(
             opts.getInstance()).getConfiguration());
         client = MasterClient.getConnectionWithRetry(context);
         MasterMonitorInfo stats = client.getMasterStats(Tracer.traceInfo(), context.rpcCreds());
-        
+
         TableInfo all = new TableInfo();
         Map<String,TableInfo> tableSummaries = new HashMap<String,TableInfo>();
-        
+
         for (TabletServerStatus server : stats.tServerInfo) {
           for (Entry<String,TableInfo> info : server.tableMap.entrySet()) {
             TableInfo tableSummary = tableSummaries.get(info.getKey());
@@ -156,42 +156,42 @@ public class ContinuousStatsCollector {
             TableInfoUtil.add(all, info.getValue());
           }
         }
-        
+
         TableInfo ti = tableSummaries.get(tableId);
-        
+
         return "" + stats.tServerInfo.size() + " " + all.recs + " " + (long) all.ingestRate + " " + (long) all.queryRate + " " + ti.recs + " "
             + ti.recsInMemory + " " + (long) ti.ingestRate + " " + (long) ti.queryRate + " " + ti.tablets + " " + ti.onlineTablets;
-        
+
       } finally {
         if (client != null)
           MasterClient.close(client);
       }
-      
+
     }
-    
+
   }
-  
+
   private static String getMRStats() throws Exception {
     Configuration conf = CachedConfiguration.getInstance();
     // No alternatives for hadoop 20
     JobClient jc = new JobClient(new org.apache.hadoop.mapred.JobConf(conf));
-    
+
     ClusterStatus cs = jc.getClusterStatus(false);
-    
+
     return "" + cs.getMapTasks() + " " + cs.getMaxMapTasks() + " " + cs.getReduceTasks() + " " + cs.getMaxReduceTasks() + " " + cs.getTaskTrackers() + " "
         + cs.getBlacklistedTrackers();
-    
+
   }
-  
+
   static class Opts extends ClientOnRequiredTable {}
-  
+
   public static void main(String[] args) {
     Opts opts = new Opts();
     ScannerOpts scanOpts = new ScannerOpts();
     opts.parseArgs(ContinuousStatsCollector.class.getName(), args, scanOpts);
     Timer jtimer = new Timer();
-    
+
     jtimer.schedule(new StatsCollectionTask(opts, scanOpts.scanBatchSize), 0, 30000);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousVerify.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousVerify.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousVerify.java
index 049f9b8..461d226 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousVerify.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousVerify.java
@@ -234,7 +234,7 @@ public class ContinuousVerify extends Configured implements Tool {
   }
 
   /**
-   * 
+   *
    * @param args
    *          instanceName zookeepers username password table columns outputpath
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousWalk.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousWalk.java b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousWalk.java
index 9253093..60f8ec2 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousWalk.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/ContinuousWalk.java
@@ -45,7 +45,7 @@ import com.beust.jcommander.IStringConverter;
 import com.beust.jcommander.Parameter;
 
 public class ContinuousWalk {
-  
+
   static public class Opts extends ContinuousQuery.Opts {
     class RandomAuthsConverter implements IStringConverter<RandomAuths> {
       @Override
@@ -57,35 +57,35 @@ public class ContinuousWalk {
         }
       }
     }
-    
+
     @Parameter(names = "--authsFile", description = "read the authorities to use from a file")
     RandomAuths randomAuths = new RandomAuths();
   }
-  
+
   static class BadChecksumException extends RuntimeException {
     private static final long serialVersionUID = 1L;
-    
+
     public BadChecksumException(String msg) {
       super(msg);
     }
-    
+
   }
-  
+
   static class RandomAuths {
     private List<Authorizations> auths;
-    
+
     RandomAuths() {
       auths = Collections.singletonList(Authorizations.EMPTY);
     }
-    
+
     RandomAuths(String file) throws IOException {
       if (file == null) {
         auths = Collections.singletonList(Authorizations.EMPTY);
         return;
       }
-      
+
       auths = new ArrayList<Authorizations>();
-      
+
       FileSystem fs = FileSystem.get(new Configuration());
       BufferedReader in = new BufferedReader(new InputStreamReader(fs.open(new Path(file)), UTF_8));
       try {
@@ -97,30 +97,30 @@ public class ContinuousWalk {
         in.close();
       }
     }
-    
+
     Authorizations getAuths(Random r) {
       return auths.get(r.nextInt(auths.size()));
     }
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(ContinuousWalk.class.getName(), args);
-    
+
     Connector conn = opts.getConnector();
-    
+
     Random r = new Random();
-    
+
     ArrayList<Value> values = new ArrayList<Value>();
-    
+
     while (true) {
       Scanner scanner = ContinuousUtil.createScanner(conn, opts.getTableName(), opts.randomAuths.getAuths(r));
       String row = findAStartRow(opts.min, opts.max, scanner, r);
-      
+
       while (row != null) {
-        
+
         values.clear();
-        
+
         long t1 = System.currentTimeMillis();
         Span span = Trace.on("walk");
         try {
@@ -133,9 +133,9 @@ public class ContinuousWalk {
           span.stop();
         }
         long t2 = System.currentTimeMillis();
-        
+
         System.out.printf("SRQ %d %s %d %d%n", t1, row, (t2 - t1), values.size());
-        
+
         if (values.size() > 0) {
           row = getPrevRow(values.get(r.nextInt(values.size())));
         } else {
@@ -143,27 +143,27 @@ public class ContinuousWalk {
           System.err.printf("MIS %d %s%n", t1, row);
           row = null;
         }
-        
+
         if (opts.sleepTime > 0)
           Thread.sleep(opts.sleepTime);
       }
-      
+
       if (opts.sleepTime > 0)
         Thread.sleep(opts.sleepTime);
     }
   }
-  
+
   private static String findAStartRow(long min, long max, Scanner scanner, Random r) {
-    
+
     byte[] scanStart = ContinuousIngest.genRow(min, max, r);
     scanner.setRange(new Range(new Text(scanStart), null));
     scanner.setBatchSize(100);
-    
+
     int count = 0;
     String pr = null;
-    
+
     long t1 = System.currentTimeMillis();
-    
+
     for (Entry<Key,Value> entry : scanner) {
       validate(entry.getKey(), entry.getValue());
       pr = getPrevRow(entry.getValue());
@@ -171,66 +171,66 @@ public class ContinuousWalk {
       if (pr != null)
         break;
     }
-    
+
     long t2 = System.currentTimeMillis();
-    
+
     System.out.printf("FSR %d %s %d %d%n", t1, new String(scanStart, UTF_8), (t2 - t1), count);
-    
+
     return pr;
   }
-  
+
   static int getPrevRowOffset(byte val[]) {
     if (val.length == 0)
       throw new IllegalArgumentException();
     if (val[53] != ':')
       throw new IllegalArgumentException(new String(val, UTF_8));
-    
+
     // prev row starts at 54
     if (val[54] != ':') {
       if (val[54 + 16] != ':')
         throw new IllegalArgumentException(new String(val, UTF_8));
       return 54;
     }
-    
+
     return -1;
   }
-  
+
   static String getPrevRow(Value value) {
-    
+
     byte[] val = value.get();
     int offset = getPrevRowOffset(val);
     if (offset > 0) {
       return new String(val, offset, 16, UTF_8);
     }
-    
+
     return null;
   }
-  
+
   static int getChecksumOffset(byte val[]) {
     if (val[val.length - 1] != ':') {
       if (val[val.length - 9] != ':')
         throw new IllegalArgumentException(new String(val, UTF_8));
       return val.length - 8;
     }
-    
+
     return -1;
   }
-  
+
   static void validate(Key key, Value value) throws BadChecksumException {
     int ckOff = getChecksumOffset(value.get());
     if (ckOff < 0)
       return;
-    
+
     long storedCksum = Long.parseLong(new String(value.get(), ckOff, 8, UTF_8), 16);
-    
+
     CRC32 cksum = new CRC32();
-    
+
     cksum.update(key.getRowData().toArray());
     cksum.update(key.getColumnFamilyData().toArray());
     cksum.update(key.getColumnQualifierData().toArray());
     cksum.update(key.getColumnVisibilityData().toArray());
     cksum.update(value.get(), 0, ckOff);
-    
+
     if (cksum.getValue() != storedCksum) {
       throw new BadChecksumException("Checksum invalid " + key + " " + value);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/GenSplits.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/GenSplits.java b/test/src/main/java/org/apache/accumulo/test/continuous/GenSplits.java
index 1320ed5..ba39f1c 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/GenSplits.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/GenSplits.java
@@ -23,27 +23,27 @@ import com.beust.jcommander.Parameter;
 import com.beust.jcommander.ParameterException;
 
 /**
- * 
+ *
  */
 public class GenSplits {
-  
+
   static class Opts {
     @Parameter(names = "--min", description = "minimum row")
     long minRow = 0;
-    
+
     @Parameter(names = "--max", description = "maximum row")
     long maxRow = Long.MAX_VALUE;
-    
+
     @Parameter(description = "<num tablets>")
     List<String> args = null;
   }
 
   public static void main(String[] args) {
-    
+
     Opts opts = new Opts();
     JCommander jcommander = new JCommander(opts);
     jcommander.setProgramName(GenSplits.class.getSimpleName());
-    
+
     try {
       jcommander.parse(args);
     } catch (ParameterException pe) {
@@ -56,14 +56,14 @@ public class GenSplits {
       jcommander.usage();
       System.exit(-1);
     }
-    
+
     int numTablets = Integer.parseInt(opts.args.get(0));
-    
+
     if (numTablets < 1) {
       System.err.println("ERROR: numTablets < 1");
       System.exit(-1);
     }
-    
+
     if (opts.minRow >= opts.maxRow) {
       System.err.println("ERROR: min >= max");
       System.exit(-1);
@@ -73,13 +73,13 @@ public class GenSplits {
     long distance = ((opts.maxRow - opts.minRow) / numTablets) + 1;
     long split = distance;
     for (int i = 0; i < numSplits; i++) {
-      
+
       String s = String.format("%016x", split + opts.minRow);
-      
+
       while (s.charAt(s.length() - 1) == '0') {
         s = s.substring(0, s.length() - 1);
       }
-      
+
       System.out.println(s);
       split += distance;
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/Histogram.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/Histogram.java b/test/src/main/java/org/apache/accumulo/test/continuous/Histogram.java
index b3aae46..6362afd 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/Histogram.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/Histogram.java
@@ -34,15 +34,15 @@ import java.util.TreeSet;
 
 class HistData<T> implements Comparable<HistData<T>>, Serializable {
   private static final long serialVersionUID = 1L;
-  
+
   T bin;
   long count;
-  
+
   HistData(T bin) {
     this.bin = bin;
     count = 0;
   }
-  
+
   @SuppressWarnings("unchecked")
   public int compareTo(HistData<T> o) {
     return ((Comparable<T>) bin).compareTo(o.bin);
@@ -50,55 +50,55 @@ class HistData<T> implements Comparable<HistData<T>>, Serializable {
 }
 
 public class Histogram<T> implements Serializable {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   protected long sum;
   protected HashMap<T,HistData<T>> counts;
-  
+
   public Histogram() {
     sum = 0;
     counts = new HashMap<T,HistData<T>>();
   }
-  
+
   public void addPoint(T x) {
     addPoint(x, 1);
   }
-  
+
   public void addPoint(T x, long y) {
-    
+
     HistData<T> hd = counts.get(x);
     if (hd == null) {
       hd = new HistData<T>(x);
       counts.put(x, hd);
     }
-    
+
     hd.count += y;
     sum += y;
   }
-  
+
   public long getCount(T x) {
     HistData<T> hd = counts.get(x);
     if (hd == null)
       return 0;
     return hd.count;
   }
-  
+
   public double getPercentage(T x) {
     if (getSum() == 0) {
       return 0;
     }
     return (double) getCount(x) / (double) getSum() * 100.0;
   }
-  
+
   public long getSum() {
     return sum;
   }
-  
+
   public List<T> getKeysInCountSortedOrder() {
-    
+
     ArrayList<HistData<T>> sortedCounts = new ArrayList<HistData<T>>(counts.values());
-    
+
     Collections.sort(sortedCounts, new Comparator<HistData<T>>() {
       public int compare(HistData<T> o1, HistData<T> o2) {
         if (o1.count < o2.count)
@@ -108,60 +108,60 @@ public class Histogram<T> implements Serializable {
         return 0;
       }
     });
-    
+
     ArrayList<T> sortedKeys = new ArrayList<T>();
-    
+
     for (Iterator<HistData<T>> iter = sortedCounts.iterator(); iter.hasNext();) {
       HistData<T> hd = iter.next();
       sortedKeys.add(hd.bin);
     }
-    
+
     return sortedKeys;
   }
-  
+
   public void print(StringBuilder out) {
     TreeSet<HistData<T>> sortedCounts = new TreeSet<HistData<T>>(counts.values());
-    
+
     int maxValueLen = 0;
-    
+
     for (Iterator<HistData<T>> iter = sortedCounts.iterator(); iter.hasNext();) {
       HistData<T> hd = iter.next();
       if (("" + hd.bin).length() > maxValueLen) {
         maxValueLen = ("" + hd.bin).length();
       }
     }
-    
+
     double psum = 0;
-    
+
     for (Iterator<HistData<T>> iter = sortedCounts.iterator(); iter.hasNext();) {
       HistData<T> hd = iter.next();
-      
+
       psum += getPercentage(hd.bin);
-      
+
       out.append(String.format(" %" + (maxValueLen + 1) + "s %,16d %6.2f%s %6.2f%s%n", hd.bin + "", hd.count, getPercentage(hd.bin), "%", psum, "%"));
     }
     out.append(String.format("%n %" + (maxValueLen + 1) + "s %,16d %n", "TOTAL", sum));
   }
-  
+
   public void save(String file) throws IOException {
-    
+
     FileOutputStream fos = new FileOutputStream(file);
     BufferedOutputStream bos = new BufferedOutputStream(fos);
     PrintStream ps = new PrintStream(bos, false, UTF_8.name());
-    
+
     TreeSet<HistData<T>> sortedCounts = new TreeSet<HistData<T>>(counts.values());
     for (Iterator<HistData<T>> iter = sortedCounts.iterator(); iter.hasNext();) {
       HistData<T> hd = iter.next();
       ps.println(" " + hd.bin + " " + hd.count);
     }
-    
+
     ps.close();
   }
-  
+
   public Set<T> getKeys() {
     return counts.keySet();
   }
-  
+
   public void clear() {
     counts.clear();
     sum = 0;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/PrintScanTimeHistogram.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/PrintScanTimeHistogram.java b/test/src/main/java/org/apache/accumulo/test/continuous/PrintScanTimeHistogram.java
index cab3126..d77f427 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/PrintScanTimeHistogram.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/PrintScanTimeHistogram.java
@@ -23,38 +23,39 @@ import java.io.FileNotFoundException;
 import java.io.IOException;
 import java.io.InputStream;
 import java.io.InputStreamReader;
+
 import org.apache.log4j.Logger;
 
 public class PrintScanTimeHistogram {
-  
+
   private static final Logger log = Logger.getLogger(PrintScanTimeHistogram.class);
 
   public static void main(String[] args) throws Exception {
     Histogram<String> srqHist = new Histogram<String>();
     Histogram<String> fsrHist = new Histogram<String>();
-    
+
     processFile(System.in, srqHist, fsrHist);
-    
+
     StringBuilder report = new StringBuilder();
     report.append(String.format("%n *** Single row queries histogram *** %n"));
     srqHist.print(report);
     log.info(report);
-    
+
     report = new StringBuilder();
     report.append(String.format("%n *** Find start rows histogram *** %n"));
     fsrHist.print(report);
     log.info(report);
   }
-  
+
   private static void processFile(InputStream ins, Histogram<String> srqHist, Histogram<String> fsrHist) throws FileNotFoundException, IOException {
     String line;
     BufferedReader in = new BufferedReader(new InputStreamReader(ins, UTF_8));
-    
+
     while ((line = in.readLine()) != null) {
-      
+
       try {
         String[] tokens = line.split(" ");
-        
+
         String type = tokens[0];
         if (type.equals("SRQ")) {
           long delta = Long.parseLong(tokens[3]);
@@ -66,16 +67,16 @@ public class PrintScanTimeHistogram {
           fsrHist.addPoint(point);
         }
       } catch (Exception e) {
-        log.error("Failed to process line '"+line+"'.", e);
+        log.error("Failed to process line '" + line + "'.", e);
       }
     }
-    
+
     in.close();
   }
-  
+
   private static String generateHistPoint(long delta) {
     String point;
-    
+
     if (delta / 1000.0 < .1) {
       point = String.format("%07.2f", delta / 1000.0);
       if (point.equals("0000.10"))
@@ -89,5 +90,5 @@ public class PrintScanTimeHistogram {
     }
     return point;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/TimeBinner.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/TimeBinner.java b/test/src/main/java/org/apache/accumulo/test/continuous/TimeBinner.java
index 0824948..186e8d0 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/TimeBinner.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/TimeBinner.java
@@ -33,16 +33,16 @@ import org.apache.accumulo.core.cli.Help;
 import com.beust.jcommander.Parameter;
 
 public class TimeBinner {
-  
+
   enum Operation {
     AVG, SUM, MIN, MAX, COUNT, CUMULATIVE, AMM, // avg,min,max
     AMM_HACK1 // special case
   }
-  
+
   private static class DoubleWrapper {
     double d;
   }
-  
+
   private static DoubleWrapper get(long l, HashMap<Long,DoubleWrapper> m, double init) {
     DoubleWrapper dw = m.get(l);
     if (dw == null) {
@@ -52,49 +52,49 @@ public class TimeBinner {
     }
     return dw;
   }
-  
+
   static class Opts extends Help {
-    @Parameter(names="--period", description="period", converter=TimeConverter.class, required=true)
+    @Parameter(names = "--period", description = "period", converter = TimeConverter.class, required = true)
     long period = 0;
-    @Parameter(names="--timeColumn", description="time column", required=true)
+    @Parameter(names = "--timeColumn", description = "time column", required = true)
     int timeColumn = 0;
-    @Parameter(names="--dataColumn", description="data column", required=true)
+    @Parameter(names = "--dataColumn", description = "data column", required = true)
     int dataColumn = 0;
-    @Parameter(names="--operation", description="one of: AVG, SUM, MIN, MAX, COUNT", required=true)
+    @Parameter(names = "--operation", description = "one of: AVG, SUM, MIN, MAX, COUNT", required = true)
     String operation;
-    @Parameter(names="--dateFormat", description="a SimpleDataFormat string that describes the data format")
+    @Parameter(names = "--dateFormat", description = "a SimpleDataFormat string that describes the data format")
     String dateFormat = "MM/dd/yy-HH:mm:ss";
   }
-  
+
   public static void main(String[] args) throws Exception {
     Opts opts = new Opts();
     opts.parseArgs(TimeBinner.class.getName(), args);
-    
+
     Operation operation = Operation.valueOf(opts.operation);
     SimpleDateFormat sdf = new SimpleDateFormat(opts.dateFormat);
-    
+
     BufferedReader in = new BufferedReader(new InputStreamReader(System.in, UTF_8));
-    
+
     String line = null;
-    
+
     HashMap<Long,DoubleWrapper> aggregation1 = new HashMap<Long,DoubleWrapper>();
     HashMap<Long,DoubleWrapper> aggregation2 = new HashMap<Long,DoubleWrapper>();
     HashMap<Long,DoubleWrapper> aggregation3 = new HashMap<Long,DoubleWrapper>();
     HashMap<Long,DoubleWrapper> aggregation4 = new HashMap<Long,DoubleWrapper>();
-    
+
     while ((line = in.readLine()) != null) {
-      
+
       try {
         String tokens[] = line.split("\\s+");
-        
+
         long time = (long) Double.parseDouble(tokens[opts.timeColumn]);
         double data = Double.parseDouble(tokens[opts.dataColumn]);
-        
+
         time = (time / opts.period) * opts.period;
-        
+
         double data_min = data;
         double data_max = data;
-        
+
         switch (operation) {
           case AMM_HACK1: {
             data_min = Double.parseDouble(tokens[opts.dataColumn - 2]);
@@ -105,20 +105,20 @@ public class TimeBinner {
             DoubleWrapper mindw = get(time, aggregation3, Double.POSITIVE_INFINITY);
             if (data < mindw.d)
               mindw.d = data_min;
-            
+
             DoubleWrapper maxdw = get(time, aggregation4, Double.NEGATIVE_INFINITY);
             if (data > maxdw.d)
               maxdw.d = data_max;
-            
+
             // fall through to AVG
           }
           case AVG: {
             DoubleWrapper sumdw = get(time, aggregation1, 0);
             DoubleWrapper countdw = get(time, aggregation2, 0);
-            
+
             sumdw.d += data;
             countdw.d++;
-            
+
             break;
           }
           case MAX: {
@@ -147,20 +147,20 @@ public class TimeBinner {
             break;
           }
         }
-        
+
       } catch (Exception e) {
         System.err.println("Failed to process line : " + line + "  " + e.getMessage());
       }
     }
-    
+
     TreeMap<Long,DoubleWrapper> sorted = new TreeMap<Long,DoubleWrapper>(aggregation1);
-    
+
     Set<Entry<Long,DoubleWrapper>> es = sorted.entrySet();
-    
+
     double cumulative = 0;
     for (Entry<Long,DoubleWrapper> entry : es) {
       String value;
-      
+
       switch (operation) {
         case AMM_HACK1:
         case AMM: {
@@ -181,9 +181,9 @@ public class TimeBinner {
         default:
           value = "" + entry.getValue().d;
       }
-      
+
       System.out.println(sdf.format(new Date(entry.getKey())) + " " + value);
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/continuous/UndefinedAnalyzer.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/continuous/UndefinedAnalyzer.java b/test/src/main/java/org/apache/accumulo/test/continuous/UndefinedAnalyzer.java
index dffd6c3..7d2c65b 100644
--- a/test/src/main/java/org/apache/accumulo/test/continuous/UndefinedAnalyzer.java
+++ b/test/src/main/java/org/apache/accumulo/test/continuous/UndefinedAnalyzer.java
@@ -51,7 +51,7 @@ import com.beust.jcommander.Parameter;
 /**
  * BUGS This code does not handle the fact that these files could include log events from previous months. It therefore it assumes all dates are in the current
  * month. One solution might be to skip log files that haven't been touched in the last month, but that doesn't prevent newer files that have old dates in them.
- * 
+ *
  */
 public class UndefinedAnalyzer {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/BadCombiner.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/BadCombiner.java b/test/src/main/java/org/apache/accumulo/test/functional/BadCombiner.java
index b7ee6fc..c589137 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/BadCombiner.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/BadCombiner.java
@@ -23,10 +23,10 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.Combiner;
 
 public class BadCombiner extends Combiner {
-  
+
   @Override
   public Value reduce(Key key, Iterator<Value> iter) {
     throw new IllegalStateException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/BadIterator.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/BadIterator.java b/test/src/main/java/org/apache/accumulo/test/functional/BadIterator.java
index 1c62720..5d13d1d 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/BadIterator.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/BadIterator.java
@@ -23,17 +23,17 @@ import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 import org.apache.accumulo.core.iterators.WrappingIterator;
 
 public class BadIterator extends WrappingIterator {
-  
+
   @Override
   public Key getTopKey() {
     throw new NullPointerException();
   }
-  
+
   @Override
   public boolean hasTop() {
     throw new NullPointerException();
   }
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/CacheTestClean.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/CacheTestClean.java b/test/src/main/java/org/apache/accumulo/test/functional/CacheTestClean.java
index d112b5b..62afb32 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/CacheTestClean.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/CacheTestClean.java
@@ -24,17 +24,17 @@ import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.apache.commons.io.FileUtils;
 
 public class CacheTestClean {
-  
+
   public static void main(String[] args) throws Exception {
     String rootDir = args[0];
     File reportDir = new File(args[1]);
-    
+
     IZooReaderWriter zoo = ZooReaderWriter.getInstance();
-    
+
     if (zoo.exists(rootDir)) {
       zoo.recursiveDelete(rootDir, NodeMissingPolicy.FAIL);
     }
-    
+
     if (reportDir.exists()) {
       FileUtils.deleteDirectory(reportDir);
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/CacheTestReader.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/CacheTestReader.java b/test/src/main/java/org/apache/accumulo/test/functional/CacheTestReader.java
index fdc704d..82eef6c 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/CacheTestReader.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/CacheTestReader.java
@@ -35,48 +35,48 @@ public class CacheTestReader {
     String reportDir = args[1];
     String keepers = args[2];
     int numData = CacheTestWriter.NUM_DATA;
-    
+
     File myfile = new File(reportDir + "/" + UUID.randomUUID());
     myfile.deleteOnExit();
-    
+
     ZooCache zc = new ZooCache(keepers, 30000);
-    
+
     while (true) {
       if (myfile.exists())
         myfile.delete();
-      
+
       if (zc.get(rootDir + "/die") != null) {
         return;
       }
-      
+
       Map<String,String> readData = new TreeMap<String,String>();
-      
+
       for (int i = 0; i < numData; i++) {
         byte[] v = zc.get(rootDir + "/data" + i);
         if (v != null)
           readData.put(rootDir + "/data" + i, new String(v, UTF_8));
       }
-      
+
       byte[] v = zc.get(rootDir + "/dataS");
       if (v != null)
         readData.put(rootDir + "/dataS", new String(v, UTF_8));
-      
+
       List<String> children = zc.getChildren(rootDir + "/dir");
       if (children != null)
         for (String child : children) {
           readData.put(rootDir + "/dir/" + child, "");
         }
-      
+
       FileOutputStream fos = new FileOutputStream(myfile);
       ObjectOutputStream oos = new ObjectOutputStream(fos);
-      
+
       oos.writeObject(readData);
-      
+
       fos.close();
       oos.close();
-      
+
       UtilWaitThread.sleep(20);
     }
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/CacheTestWriter.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/CacheTestWriter.java b/test/src/main/java/org/apache/accumulo/test/functional/CacheTestWriter.java
index e1be8e6..3a3baf0 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/CacheTestWriter.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/CacheTestWriter.java
@@ -35,40 +35,40 @@ import org.apache.accumulo.fate.zookeeper.ZooUtil.NodeMissingPolicy;
 import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 
 public class CacheTestWriter {
-  
+
   static final int NUM_DATA = 3;
-  
+
   public static void main(String[] args) throws Exception {
     IZooReaderWriter zk = ZooReaderWriter.getInstance();
-    
+
     String rootDir = args[0];
     File reportDir = new File(args[1]);
     int numReaders = Integer.parseInt(args[2]);
     int numVerifications = Integer.parseInt(args[3]);
     int numData = NUM_DATA;
-    
+
     boolean dataSExists = false;
     int count = 0;
-    
+
     zk.putPersistentData(rootDir, new byte[0], NodeExistsPolicy.FAIL);
     for (int i = 0; i < numData; i++) {
       zk.putPersistentData(rootDir + "/data" + i, new byte[0], NodeExistsPolicy.FAIL);
     }
-    
+
     zk.putPersistentData(rootDir + "/dir", new byte[0], NodeExistsPolicy.FAIL);
-    
+
     ArrayList<String> children = new ArrayList<String>();
-    
+
     Random r = new Random();
-    
+
     while (count++ < numVerifications) {
-      
+
       Map<String,String> expectedData = null;
       // change children in dir
-      
+
       for (int u = 0; u < r.nextInt(4) + 1; u++) {
         expectedData = new TreeMap<String,String>();
-        
+
         if (r.nextFloat() < .5) {
           String child = UUID.randomUUID().toString();
           zk.putPersistentData(rootDir + "/dir/" + child, new byte[0], NodeExistsPolicy.SKIP);
@@ -78,32 +78,32 @@ public class CacheTestWriter {
           String child = children.remove(index);
           zk.recursiveDelete(rootDir + "/dir/" + child, NodeMissingPolicy.FAIL);
         }
-        
+
         for (String child : children) {
           expectedData.put(rootDir + "/dir/" + child, "");
         }
-        
+
         // change values
         for (int i = 0; i < numData; i++) {
           byte data[] = Long.toString(r.nextLong(), 16).getBytes(UTF_8);
           zk.putPersistentData(rootDir + "/data" + i, data, NodeExistsPolicy.OVERWRITE);
           expectedData.put(rootDir + "/data" + i, new String(data, UTF_8));
         }
-        
+
         // test a data node that does not always exists...
         if (r.nextFloat() < .5) {
-          
+
           byte data[] = Long.toString(r.nextLong(), 16).getBytes(UTF_8);
-          
+
           if (!dataSExists) {
             zk.putPersistentData(rootDir + "/dataS", data, NodeExistsPolicy.SKIP);
             dataSExists = true;
           } else {
             zk.putPersistentData(rootDir + "/dataS", data, NodeExistsPolicy.OVERWRITE);
           }
-          
+
           expectedData.put(rootDir + "/dataS", new String(data, UTF_8));
-          
+
         } else {
           if (dataSExists) {
             zk.recursiveDelete(rootDir + "/dataS", NodeMissingPolicy.FAIL);
@@ -111,34 +111,34 @@ public class CacheTestWriter {
           }
         }
       }
-      
+
       // change children in dir and change values
-      
+
       System.out.println("expectedData " + expectedData);
-      
+
       // wait for all readers to see changes
       while (true) {
-        
+
         File[] files = reportDir.listFiles();
-        
+
         System.out.println("files.length " + files.length);
-        
+
         if (files.length == numReaders) {
           boolean ok = true;
-          
+
           for (int i = 0; i < files.length; i++) {
             try {
               FileInputStream fis = new FileInputStream(files[i]);
               ObjectInputStream ois = new ObjectInputStream(fis);
-              
+
               @SuppressWarnings("unchecked")
               Map<String,String> readerMap = (Map<String,String>) ois.readObject();
-              
+
               fis.close();
               ois.close();
-              
+
               System.out.println("read " + readerMap);
-              
+
               if (!readerMap.equals(expectedData)) {
                 System.out.println("maps not equals");
                 ok = false;
@@ -148,16 +148,16 @@ public class CacheTestWriter {
               ok = false;
             }
           }
-          
+
           if (ok)
             break;
         }
-        
+
         UtilWaitThread.sleep(5);
       }
     }
-    
+
     zk.putPersistentData(rootDir + "/die", new byte[0], NodeExistsPolicy.FAIL);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/DropModIter.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/DropModIter.java b/test/src/main/java/org/apache/accumulo/test/functional/DropModIter.java
index 20fe856..32a178d 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/DropModIter.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/DropModIter.java
@@ -26,26 +26,26 @@ import org.apache.accumulo.core.iterators.SkippingIterator;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 public class DropModIter extends SkippingIterator {
-  
+
   private int mod;
   private int drop;
-  
+
   @Override
   public SortedKeyValueIterator<Key,Value> deepCopy(IteratorEnvironment env) {
     throw new UnsupportedOperationException();
   }
-  
+
   @Override
   public void init(SortedKeyValueIterator<Key,Value> source, Map<String,String> options, IteratorEnvironment env) throws IOException {
     super.init(source, options, env);
     this.mod = Integer.parseInt(options.get("mod"));
     this.drop = Integer.parseInt(options.get("drop"));
   }
-  
+
   protected void consume() throws IOException {
     while (getSource().hasTop() && Integer.parseInt(getSource().getTopKey().getRow().toString()) % mod == drop) {
       getSource().next();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/SlowConstraint.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/SlowConstraint.java b/test/src/main/java/org/apache/accumulo/test/functional/SlowConstraint.java
index 8e29955..187da35 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/SlowConstraint.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/SlowConstraint.java
@@ -23,19 +23,19 @@ import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.util.UtilWaitThread;
 
 /**
- * 
+ *
  */
 public class SlowConstraint implements Constraint {
-  
+
   @Override
   public String getViolationDescription(short violationCode) {
     return null;
   }
-  
+
   @Override
   public List<Short> check(Environment env, Mutation mutation) {
     UtilWaitThread.sleep(20000);
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/SlowIterator.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/SlowIterator.java b/test/src/main/java/org/apache/accumulo/test/functional/SlowIterator.java
index cb29688..f84a4d9 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/SlowIterator.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/SlowIterator.java
@@ -37,9 +37,9 @@ public class SlowIterator extends WrappingIterator {
 
   private long sleepTime = 0;
   private long seekSleepTime = 0;
-  
+
   public static void setSleepTime(IteratorSetting is, long millis) {
-    is.addOption(SLEEP_TIME, Long.toString(millis));  
+    is.addOption(SLEEP_TIME, Long.toString(millis));
   }
 
   public static void setSeekSleepTime(IteratorSetting is, long t) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/test/src/main/java/org/apache/accumulo/test/functional/ZombieTServer.java
----------------------------------------------------------------------
diff --git a/test/src/main/java/org/apache/accumulo/test/functional/ZombieTServer.java b/test/src/main/java/org/apache/accumulo/test/functional/ZombieTServer.java
index eb84533..3bb44ff 100644
--- a/test/src/main/java/org/apache/accumulo/test/functional/ZombieTServer.java
+++ b/test/src/main/java/org/apache/accumulo/test/functional/ZombieTServer.java
@@ -102,8 +102,8 @@ public class ZombieTServer {
     TransactionWatcher watcher = new TransactionWatcher();
     final ThriftClientHandler tch = new ThriftClientHandler(context, watcher);
     Processor<Iface> processor = new Processor<Iface>(tch);
-    ServerAddress serverPort = TServerUtils.startTServer(context.getConfiguration(), HostAndPort.fromParts("0.0.0.0", port), processor, "ZombieTServer", "walking dead", 2, 1, 1000,
-        10 * 1024 * 1024, null, -1);
+    ServerAddress serverPort = TServerUtils.startTServer(context.getConfiguration(), HostAndPort.fromParts("0.0.0.0", port), processor, "ZombieTServer",
+        "walking dead", 2, 1, 1000, 10 * 1024 * 1024, null, -1);
 
     String addressString = serverPort.address.toString();
     String zPath = ZooUtil.getRoot(context.getInstance()) + Constants.ZTSERVERS + "/" + addressString;


[40/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/Mutation.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Mutation.java b/core/src/main/java/org/apache/accumulo/core/data/Mutation.java
index 0751ba4..6f7d433 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Mutation.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Mutation.java
@@ -40,15 +40,15 @@ import com.google.common.base.Preconditions;
  * <p>
  * Mutation represents an action that manipulates a row in a table. A mutation holds a list of column/value pairs that represent an atomic set of modifications
  * to make to a row.
- * 
+ *
  * <p>
  * Convenience methods which takes columns and value as CharSequence (String implements CharSequence) are provided. CharSequence is converted to UTF-8 by
  * constructing a new Text object.
- * 
+ *
  * <p>
  * When always passing in the same data as a CharSequence/String, it's probably more efficient to call the Text put methods. This way the data is only encoded
  * once and only one Text object is created.
- * 
+ *
  * <p>
  * All of the put methods append data to the mutation; they do not overwrite anything that was previously put. The mutation holds a list of all columns/values
  * that were put into it.
@@ -61,62 +61,64 @@ import com.google.common.base.Preconditions;
  * </p>
  */
 public class Mutation implements Writable {
-  
+
   /**
-   * Internally, this class keeps most mutation data in a byte buffer. If a cell
-   * value put into a mutation exceeds this size, then it is stored in a
-   * separate buffer, and a reference to it is inserted into the main buffer.
+   * Internally, this class keeps most mutation data in a byte buffer. If a cell value put into a mutation exceeds this size, then it is stored in a separate
+   * buffer, and a reference to it is inserted into the main buffer.
    */
   static final int VALUE_SIZE_COPY_CUTOFF = 1 << 15;
-  
+
   /**
-   * Formats available for serializing Mutations. The formats are described
-   * in a <a href="doc-files/mutation-serialization.html">separate document</a>.
+   * Formats available for serializing Mutations. The formats are described in a <a href="doc-files/mutation-serialization.html">separate document</a>.
    */
   public static enum SERIALIZED_FORMAT {
-     VERSION1,
-     VERSION2
+    VERSION1, VERSION2
   };
-  
+
   private boolean useOldDeserialize = false;
   private byte[] row;
   private byte[] data;
   private int entries;
   private List<byte[]> values;
-  
+
   private UnsynchronizedBuffer.Writer buffer;
-  
+
   private List<ColumnUpdate> updates;
 
   private static final Set<String> EMPTY = Collections.emptySet();
   private Set<String> replicationSources = EMPTY;
-  
+
   private static final byte[] EMPTY_BYTES = new byte[0];
-  
+
   private void serialize() {
     if (buffer != null) {
       data = buffer.toArray();
       buffer = null;
     }
   }
-  
+
   /**
    * Creates a new mutation. A defensive copy is made.
    *
-   * @param row row ID
+   * @param row
+   *          row ID
    * @since 1.5.0
    */
   public Mutation(byte[] row) {
     this(row, 0, row.length);
   }
-  
+
   /**
    * Creates a new mutation. A defensive copy is made.
    *
-   * @param row byte array containing row ID
-   * @param start starting index of row ID in byte array
-   * @param length length of row ID in byte array
-   * @throws IndexOutOfBoundsException if start or length is invalid
+   * @param row
+   *          byte array containing row ID
+   * @param start
+   *          starting index of row ID in byte array
+   * @param length
+   *          length of row ID in byte array
+   * @throws IndexOutOfBoundsException
+   *           if start or length is invalid
    * @since 1.5.0
    */
   public Mutation(byte[] row, int start, int length) {
@@ -124,34 +126,37 @@ public class Mutation implements Writable {
     System.arraycopy(row, start, this.row, 0, length);
     buffer = new UnsynchronizedBuffer.Writer();
   }
-  
+
   /**
    * Creates a new mutation. A defensive copy is made.
    *
-   * @param row row ID
+   * @param row
+   *          row ID
    */
   public Mutation(Text row) {
     this(row.getBytes(), 0, row.getLength());
   }
-  
+
   /**
    * Creates a new mutation.
    *
-   * @param row row ID
+   * @param row
+   *          row ID
    */
   public Mutation(CharSequence row) {
     this(new Text(row.toString()));
   }
-  
+
   /**
    * Creates a new mutation.
    */
   public Mutation() {}
-  
+
   /**
    * Creates a new mutation from a Thrift mutation.
    *
-   * @param tmutation Thrift mutation
+   * @param tmutation
+   *          Thrift mutation
    */
   public Mutation(TMutation tmutation) {
     this.row = ByteBufferUtil.toBytes(tmutation.row);
@@ -162,7 +167,7 @@ public class Mutation implements Writable {
     if (tmutation.isSetSources()) {
       this.replicationSources = new HashSet<>(tmutation.sources);
     }
-    
+
     if (this.row == null) {
       throw new IllegalArgumentException("null row");
     }
@@ -170,11 +175,12 @@ public class Mutation implements Writable {
       throw new IllegalArgumentException("null serialized data");
     }
   }
-  
+
   /**
    * Creates a new mutation by copying another.
    *
-   * @param m mutation to copy
+   * @param m
+   *          mutation to copy
    */
   public Mutation(Mutation m) {
     m.serialize();
@@ -184,7 +190,7 @@ public class Mutation implements Writable {
     this.values = m.values;
     this.replicationSources = m.replicationSources;
   }
-  
+
   /**
    * Gets the row ID for this mutation. Not a defensive copy.
    *
@@ -193,28 +199,28 @@ public class Mutation implements Writable {
   public byte[] getRow() {
     return row;
   }
-  
+
   private void put(byte b[]) {
     put(b, b.length);
   }
-  
+
   private void put(byte b[], int length) {
     buffer.writeVLong(length);
     buffer.add(b, 0, length);
   }
-  
+
   private void put(boolean b) {
     buffer.add(b);
   }
-  
+
   private void put(int i) {
     buffer.writeVLong(i);
   }
-  
+
   private void put(long l) {
     buffer.writeVLong(l);
   }
-  
+
   private void put(byte[] cf, byte[] cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
     put(cf, cf.length, cq, cq.length, cv, hasts, ts, deleted, val, val.length);
   }
@@ -225,7 +231,7 @@ public class Mutation implements Writable {
   private void put(Text cf, Text cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
     put(cf.getBytes(), cf.getLength(), cq.getBytes(), cq.getLength(), cv, hasts, ts, deleted, val, val.length);
   }
-  
+
   private void put(byte[] cf, int cfLength, byte[] cq, int cqLength, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val, int valLength) {
     if (buffer == null) {
       throw new IllegalStateException("Can not add to mutation after serializing it");
@@ -238,7 +244,7 @@ public class Mutation implements Writable {
       put(ts);
     }
     put(deleted);
-    
+
     if (valLength < VALUE_SIZE_COPY_CUTOFF) {
       put(val, valLength);
     } else {
@@ -250,14 +256,14 @@ public class Mutation implements Writable {
       values.add(copy);
       put(-1 * values.size());
     }
-    
+
     entries++;
   }
-  
+
   private void put(CharSequence cf, CharSequence cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
     put(new Text(cf.toString()), new Text(cq.toString()), cv, hasts, ts, deleted, val);
   }
-  
+
   private void put(Text cf, Text cq, byte[] cv, boolean hasts, long ts, boolean deleted, Text val) {
     put(cf.getBytes(), cf.getLength(), cq.getBytes(), cq.getLength(), cv, hasts, ts, deleted, val.getBytes(), val.getLength());
   }
@@ -267,358 +273,429 @@ public class Mutation implements Writable {
   }
 
   /**
-   * Puts a modification in this mutation. Column visibility is empty;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param value
+   *          cell value
    */
   public void put(Text columnFamily, Text columnQualifier, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value.get());
   }
-  
+
   /**
-   * Puts a modification in this mutation. Timestamp is not set. All parameters
-   * are defensively copied.
+   * Puts a modification in this mutation. Timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param value
+   *          cell value
    */
   public void put(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value.get());
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty. All
-   * appropriate parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    */
   public void put(Text columnFamily, Text columnQualifier, long timestamp, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value.get());
   }
-  
+
   /**
-   * Puts a modification in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a modification in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    */
   public void put(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility, long timestamp, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value.get());
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Matches empty column visibility;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a deletion in this mutation. Matches empty column visibility; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
    */
   public void putDelete(Text columnFamily, Text columnQualifier) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Timestamp is not set. All parameters are
-   * defensively copied.
+   * Puts a deletion in this mutation. Timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
    */
   public void putDelete(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Matches empty column visibility. All
-   * appropriate parameters are defensively copied.
+   * Puts a deletion in this mutation. Matches empty column visibility. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
    */
   public void putDelete(Text columnFamily, Text columnQualifier, long timestamp) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a deletion in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
    */
   public void putDelete(Text columnFamily, Text columnQualifier, ColumnVisibility columnVisibility, long timestamp) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value.get());
   }
-  
+
   /**
-   * Puts a modification in this mutation. Timestamp is not set. All parameters
-   * are defensively copied.
+   * Puts a modification in this mutation. Timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value.get());
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty. All
-   * appropriate parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, long timestamp, Value value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value.get());
   }
-  
+
   /**
-   * Puts a modification in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a modification in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, long timestamp, Value value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value.get());
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Matches empty column visibility;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a deletion in this mutation. Matches empty column visibility; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
    */
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Timestamp is not set. All appropriate
-   * parameters are defensively copied.
+   * Puts a deletion in this mutation. Timestamp is not set. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
    */
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Matches empty column visibility. All
-   * appropriate parameters are defensively copied.
+   * Puts a deletion in this mutation. Matches empty column visibility. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
    */
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier, long timestamp) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a deletion in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
    */
   public void putDelete(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, long timestamp) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, CharSequence value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Timestamp is not set. All parameters
-   * are defensively copied.
+   * Puts a modification in this mutation. Timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, CharSequence value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty. All
-   * appropriate parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, long timestamp, CharSequence value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a modification in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    */
   public void put(CharSequence columnFamily, CharSequence columnQualifier, ColumnVisibility columnVisibility, long timestamp, CharSequence value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param value
+   *          cell value
    * @since 1.5.0
    */
   public void put(byte[] columnFamily, byte[] columnQualifier, byte[] value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Timestamp is not set. All parameters
-   * are defensively copied.
+   * Puts a modification in this mutation. Timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param value
+   *          cell value
    * @since 1.5.0
    */
   public void put(byte[] columnFamily, byte[] columnQualifier, ColumnVisibility columnVisibility, byte[] value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. Column visibility is empty. All
-   * appropriate parameters are defensively copied.
+   * Puts a modification in this mutation. Column visibility is empty. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    * @since 1.5.0
    */
   public void put(byte[] columnFamily, byte[] columnQualifier, long timestamp, byte[] value) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, false, value);
   }
-  
+
   /**
-   * Puts a modification in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a modification in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
-   * @param value cell value
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
+   * @param value
+   *          cell value
    * @since 1.5.0
    */
   public void put(byte[] columnFamily, byte[] columnQualifier, ColumnVisibility columnVisibility, long timestamp, byte[] value) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), true, timestamp, false, value);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Matches empty column visibility;
-   * timestamp is not set. All parameters are defensively copied.
+   * Puts a deletion in this mutation. Matches empty column visibility; timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
    * @since 1.5.0
    */
   public void putDelete(byte[] columnFamily, byte[] columnQualifier) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, false, 0l, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Timestamp is not set. All parameters are
-   * defensively copied.
+   * Puts a deletion in this mutation. Timestamp is not set. All parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
    * @since 1.5.0
    */
   public void putDelete(byte[] columnFamily, byte[] columnQualifier, ColumnVisibility columnVisibility) {
     put(columnFamily, columnQualifier, columnVisibility.getExpression(), false, 0l, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. Matches empty column visibility. All
-   * appropriate parameters are defensively copied.
+   * Puts a deletion in this mutation. Matches empty column visibility. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param timestamp timestamp
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param timestamp
+   *          timestamp
    * @since 1.5.0
    */
   public void putDelete(byte[] columnFamily, byte[] columnQualifier, long timestamp) {
     put(columnFamily, columnQualifier, EMPTY_BYTES, true, timestamp, true, EMPTY_BYTES);
   }
-  
+
   /**
-   * Puts a deletion in this mutation. All appropriate parameters are
-   * defensively copied.
+   * Puts a deletion in this mutation. All appropriate parameters are defensively copied.
    *
-   * @param columnFamily column family
-   * @param columnQualifier column qualifier
-   * @param columnVisibility column visibility
-   * @param timestamp timestamp
+   * @param columnFamily
+   *          column family
+   * @param columnQualifier
+   *          column qualifier
+   * @param columnVisibility
+   *          column visibility
+   * @param timestamp
+   *          timestamp
    * @since 1.5.0
    */
   public void putDelete(byte[] columnFamily, byte[] columnQualifier, ColumnVisibility columnVisibility, long timestamp) {
@@ -629,50 +706,49 @@ public class Mutation implements Writable {
     int len = in.readInt();
     if (len == 0)
       return EMPTY_BYTES;
-    
+
     byte bytes[] = new byte[len];
     in.readBytes(bytes);
     return bytes;
   }
-  
+
   private byte[] readBytes(UnsynchronizedBuffer.Reader in) {
-    int len = (int)in.readVLong();
+    int len = (int) in.readVLong();
     if (len == 0)
       return EMPTY_BYTES;
-    
+
     byte bytes[] = new byte[len];
     in.readBytes(bytes);
     return bytes;
   }
-  
+
   /**
-   * Gets the modifications and deletions in this mutation. After calling
-   * this method, further modifications to this mutation are ignored. Changes
-   * made to the returned updates do not affect this mutation.
+   * Gets the modifications and deletions in this mutation. After calling this method, further modifications to this mutation are ignored. Changes made to the
+   * returned updates do not affect this mutation.
    *
    * @return list of modifications and deletions
    */
   public List<ColumnUpdate> getUpdates() {
     serialize();
-    
+
     UnsynchronizedBuffer.Reader in = new UnsynchronizedBuffer.Reader(data);
-    
+
     if (updates == null) {
       if (entries == 1) {
         updates = Collections.singletonList(deserializeColumnUpdate(in));
       } else {
         ColumnUpdate[] tmpUpdates = new ColumnUpdate[entries];
-        
+
         for (int i = 0; i < entries; i++)
           tmpUpdates[i] = deserializeColumnUpdate(in);
-        
+
         updates = Arrays.asList(tmpUpdates);
       }
     }
-    
+
     return updates;
   }
-  
+
   protected ColumnUpdate newColumnUpdate(byte[] cf, byte[] cq, byte[] cv, boolean hasts, long ts, boolean deleted, byte[] val) {
     return new ColumnUpdate(cf, cq, cv, hasts, ts, deleted, val);
   }
@@ -686,10 +762,10 @@ public class Mutation implements Writable {
     if (hasts)
       ts = in.readVLong();
     boolean deleted = in.readBoolean();
-    
+
     byte[] val;
-    int valLen = (int)in.readVLong();
-    
+    int valLen = (int) in.readVLong();
+
     if (valLen < 0) {
       val = values.get((-1 * valLen) - 1);
     } else if (valLen == 0) {
@@ -698,12 +774,12 @@ public class Mutation implements Writable {
       val = new byte[valLen];
       in.readBytes(val);
     }
-    
+
     return newColumnUpdate(cf, cq, cv, hasts, ts, deleted, val);
   }
-  
+
   private int cachedValLens = -1;
-  
+
   /**
    * Gets the byte length of all large values stored in this mutation.
    *
@@ -713,19 +789,19 @@ public class Mutation implements Writable {
   long getValueLengths() {
     if (values == null)
       return 0;
-    
+
     if (cachedValLens == -1) {
       int tmpCVL = 0;
       for (byte[] val : values)
         tmpCVL += val.length;
-      
+
       cachedValLens = tmpCVL;
     }
-    
+
     return cachedValLens;
-    
+
   }
-  
+
   /**
    * Gets the total number of bytes in this mutation.
    *
@@ -735,17 +811,16 @@ public class Mutation implements Writable {
     serialize();
     return row.length + data.length + getValueLengths();
   }
-  
+
   /**
-   * Gets an estimate of the amount of memory used by this mutation. The
-   * estimate includes data sizes and object overhead.
+   * Gets an estimate of the amount of memory used by this mutation. The estimate includes data sizes and object overhead.
    *
    * @return memory usage estimate
    */
   public long estimatedMemoryUsed() {
     return numBytes() + 238;
   }
-  
+
   /**
    * Gets the number of modifications / deletions in this mutation.
    *
@@ -757,9 +832,9 @@ public class Mutation implements Writable {
 
   /**
    * Add a new element to the set of peers which this Mutation originated from
-   * 
+   *
    * @param peer
-   *         the peer to add
+   *          the peer to add
    * @since 1.7.0
    */
   public void addReplicationSource(String peer) {
@@ -772,7 +847,7 @@ public class Mutation implements Writable {
 
   /**
    * Set the replication peers which this Mutation originated from
-   * 
+   *
    * @param sources
    *          Set of peer names which have processed this update
    * @since 1.7.0
@@ -784,6 +859,7 @@ public class Mutation implements Writable {
 
   /**
    * Return the replication sources for this Mutation
+   *
    * @return An unmodifiable view of the replication sources
    */
   public Set<String> getReplicationSources() {
@@ -792,10 +868,10 @@ public class Mutation implements Writable {
     }
     return Collections.unmodifiableSet(replicationSources);
   }
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
-    
+
     // Clear out cached column updates and value lengths so
     // that we recalculate them based on the (potentially) new
     // data we are about to read in.
@@ -803,7 +879,7 @@ public class Mutation implements Writable {
     cachedValLens = -1;
     buffer = null;
     useOldDeserialize = false;
-    
+
     byte first = in.readByte();
     if ((first & 0x80) != 0x80) {
       oldReadFields(first, in);
@@ -818,7 +894,7 @@ public class Mutation implements Writable {
     data = new byte[len];
     in.readFully(data);
     entries = WritableUtils.readVInt(in);
-    
+
     boolean valuesPresent = (first & 0x01) == 0x01;
     if (!valuesPresent) {
       values = null;
@@ -841,24 +917,23 @@ public class Mutation implements Writable {
       }
     }
   }
-  
+
   protected void droppingOldTimestamp(long ts) {}
 
   private void oldReadFields(byte first, DataInput in) throws IOException {
 
-    byte b = (byte)in.readByte();
-    byte c = (byte)in.readByte();
-    byte d = (byte)in.readByte();
-    
-    int len = (((first & 0xff) << 24) | ((b & 0xff) << 16) |
-        ((c & 0xff) << 8) | (d & 0xff));
+    byte b = (byte) in.readByte();
+    byte c = (byte) in.readByte();
+    byte d = (byte) in.readByte();
+
+    int len = (((first & 0xff) << 24) | ((b & 0xff) << 16) | ((c & 0xff) << 8) | (d & 0xff));
     row = new byte[len];
     in.readFully(row);
     len = in.readInt();
     byte[] localData = new byte[len];
     in.readFully(localData);
     int localEntries = in.readInt();
-    
+
     List<byte[]> localValues;
     boolean valuesPresent = in.readBoolean();
     if (!valuesPresent) {
@@ -873,7 +948,7 @@ public class Mutation implements Writable {
         localValues.add(val);
       }
     }
-    
+
     // convert data to new format
     UnsynchronizedBuffer.Reader din = new UnsynchronizedBuffer.Reader(localData);
     buffer = new UnsynchronizedBuffer.Writer();
@@ -884,10 +959,10 @@ public class Mutation implements Writable {
       boolean hasts = din.readBoolean();
       long ts = din.readLong();
       boolean deleted = din.readBoolean();
-      
+
       byte[] val;
       int valLen = din.readInt();
-      
+
       if (valLen < 0) {
         val = localValues.get((-1 * valLen) - 1);
       } else if (valLen == 0) {
@@ -896,7 +971,7 @@ public class Mutation implements Writable {
         val = new byte[valLen];
         din.readBytes(val);
       }
-      
+
       put(cf, cq, cv, hasts, ts, deleted, val);
       if (!hasts)
         droppingOldTimestamp(ts);
@@ -905,24 +980,24 @@ public class Mutation implements Writable {
     serialize();
 
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     serialize();
-    byte hasValues = (values == null) ? 0 : (byte)1;
+    byte hasValues = (values == null) ? 0 : (byte) 1;
     if (!replicationSources.isEmpty()) {
       // Use 2nd least-significant bit for whether or not we have replication sources
       hasValues = (byte) (0x02 | hasValues);
     }
-    out.write((byte)(0x80 | hasValues));
-    
+    out.write((byte) (0x80 | hasValues));
+
     WritableUtils.writeVInt(out, row.length);
     out.write(row);
 
     WritableUtils.writeVInt(out, data.length);
     out.write(data);
     WritableUtils.writeVInt(out, entries);
-    
+
     if (0x01 == (0x01 & hasValues)) {
       WritableUtils.writeVInt(out, values.size());
       for (int i = 0; i < values.size(); i++) {
@@ -938,7 +1013,7 @@ public class Mutation implements Writable {
       }
     }
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o == this) {
@@ -949,18 +1024,18 @@ public class Mutation implements Writable {
     }
     return false;
   }
-  
+
   @Override
   public int hashCode() {
     return toThrift().hashCode();
   }
+
   /**
-   * Checks if this mutation equals another. Two mutations are equal if they
-   * target the same row and have the same modifications and deletions, in order.
-   * This method may be removed in a  future API revision in favor of
-   * {@link #equals(Object)}. See ACCUMULO-1627 for more information.
+   * Checks if this mutation equals another. Two mutations are equal if they target the same row and have the same modifications and deletions, in order. This
+   * method may be removed in a future API revision in favor of {@link #equals(Object)}. See ACCUMULO-1627 for more information.
    *
-   * @param m mutation to compare
+   * @param m
+   *          mutation to compare
    * @return true if this mutation equals the other, false otherwise
    */
   public boolean equals(Mutation m) {
@@ -978,21 +1053,21 @@ public class Mutation implements Writable {
 
       if (values == null && m.values == null)
         return true;
-      
+
       if (values != null && m.values != null && values.size() == m.values.size()) {
         for (int i = 0; i < values.size(); i++) {
           if (!Arrays.equals(values.get(i), m.values.get(i)))
             return false;
         }
-        
+
         return true;
       }
-      
+
     }
-    
+
     return false;
   }
-  
+
   /**
    * Converts this mutation to Thrift.
    *
@@ -1006,7 +1081,7 @@ public class Mutation implements Writable {
     }
     return tmutation;
   }
-  
+
   /**
    * Gets the serialization format used to (de)serialize this mutation.
    *

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/PartialKey.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/PartialKey.java b/core/src/main/java/org/apache/accumulo/core/data/PartialKey.java
index 5636de0..f4289d2 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/PartialKey.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/PartialKey.java
@@ -21,22 +21,23 @@ package org.apache.accumulo.core.data;
  */
 public enum PartialKey {
   ROW(1), ROW_COLFAM(2), ROW_COLFAM_COLQUAL(3), ROW_COLFAM_COLQUAL_COLVIS(4), ROW_COLFAM_COLQUAL_COLVIS_TIME(5),
-  //everything with delete flag
-  ROW_COLFAM_COLQUAL_COLVIS_TIME_DEL(6) 
-  ;
-  
+  // everything with delete flag
+  ROW_COLFAM_COLQUAL_COLVIS_TIME_DEL(6);
+
   int depth;
-  
+
   private PartialKey(int depth) {
     this.depth = depth;
   }
-  
+
   /**
    * Get a partial key specification by depth of the specification.
    *
-   * @param depth depth of scope (i.e., number of fields included)
+   * @param depth
+   *          depth of scope (i.e., number of fields included)
    * @return partial key
-   * @throws IllegalArgumentException if no partial key has the given depth
+   * @throws IllegalArgumentException
+   *           if no partial key has the given depth
    */
   public static PartialKey getByDepth(int depth) {
     for (PartialKey d : PartialKey.values())
@@ -44,7 +45,7 @@ public enum PartialKey {
         return d;
     throw new IllegalArgumentException("Invalid legacy depth " + depth);
   }
-  
+
   /**
    * Gets the depth of this partial key.
    *

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/Range.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Range.java b/core/src/main/java/org/apache/accumulo/core/data/Range.java
index 56e823e..05bb9f3 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Range.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Range.java
@@ -31,111 +31,137 @@ import org.apache.hadoop.io.WritableComparable;
 
 /**
  * This class is used to specify a range of Accumulo keys.
- * 
+ *
  * @see Key
  */
 public class Range implements WritableComparable<Range> {
-  
+
   private Key start;
   private Key stop;
   private boolean startKeyInclusive;
   private boolean stopKeyInclusive;
   private boolean infiniteStartKey;
   private boolean infiniteStopKey;
-  
+
   /**
    * Creates a range that goes from negative to positive infinity
    */
   public Range() {
     this((Key) null, true, (Key) null, true);
   }
-  
+
   /**
    * Creates a range from startKey inclusive to endKey inclusive.
-   * 
-   * @param startKey starting key; set to null for negative infinity
-   * @param endKey ending key; set to null for positive infinity
-   * @throws IllegalArgumentException if end key is before start key
+   *
+   * @param startKey
+   *          starting key; set to null for negative infinity
+   * @param endKey
+   *          ending key; set to null for positive infinity
+   * @throws IllegalArgumentException
+   *           if end key is before start key
    */
   public Range(Key startKey, Key endKey) {
     this(startKey, true, endKey, true);
   }
-  
+
   /**
    * Creates a range that covers an entire row.
-   * 
-   * @param row row to cover; set to null to cover all rows
+   *
+   * @param row
+   *          row to cover; set to null to cover all rows
    */
   public Range(CharSequence row) {
     this(row, true, row, true);
   }
-  
+
   /**
    * Creates a range that covers an entire row.
-   * 
-   * @param row row to cover; set to null to cover all rows
+   *
+   * @param row
+   *          row to cover; set to null to cover all rows
    */
   public Range(Text row) {
     this(row, true, row, true);
   }
-  
+
   /**
    * Creates a range from startRow inclusive to endRow inclusive.
-   * 
-   * @param startRow starting row; set to null for negative infinity
-   * @param endRow ending row; set to null for positive infinity
-   * @throws IllegalArgumentException if end row is before start row
+   *
+   * @param startRow
+   *          starting row; set to null for negative infinity
+   * @param endRow
+   *          ending row; set to null for positive infinity
+   * @throws IllegalArgumentException
+   *           if end row is before start row
    */
   public Range(Text startRow, Text endRow) {
     this(startRow, true, endRow, true);
   }
-  
+
   /**
    * Creates a range from startRow inclusive to endRow inclusive.
-   * 
-   * @param startRow starting row; set to null for negative infinity
-   * @param endRow ending row; set to null for positive infinity
-   * @throws IllegalArgumentException if end row is before start row
+   *
+   * @param startRow
+   *          starting row; set to null for negative infinity
+   * @param endRow
+   *          ending row; set to null for positive infinity
+   * @throws IllegalArgumentException
+   *           if end row is before start row
    */
   public Range(CharSequence startRow, CharSequence endRow) {
     this(startRow, true, endRow, true);
   }
-  
+
   /**
    * Creates a range from startRow to endRow.
-   * 
-   * @param startRow starting row; set to null for negative infinity
-   * @param startRowInclusive true to include start row, false to skip
-   * @param endRow ending row; set to null for positive infinity
-   * @param endRowInclusive true to include start row, false to skip
-   * @throws IllegalArgumentException if end row is before start row
+   *
+   * @param startRow
+   *          starting row; set to null for negative infinity
+   * @param startRowInclusive
+   *          true to include start row, false to skip
+   * @param endRow
+   *          ending row; set to null for positive infinity
+   * @param endRowInclusive
+   *          true to include start row, false to skip
+   * @throws IllegalArgumentException
+   *           if end row is before start row
    */
   public Range(Text startRow, boolean startRowInclusive, Text endRow, boolean endRowInclusive) {
     this((startRow == null ? null : (startRowInclusive ? new Key(startRow) : new Key(startRow).followingKey(PartialKey.ROW))), true, (endRow == null ? null
         : (endRowInclusive ? new Key(endRow).followingKey(PartialKey.ROW) : new Key(endRow))), false);
   }
-  
+
   /**
    * Creates a range from startRow to endRow.
-   * 
-   * @param startRow starting row; set to null for negative infinity
-   * @param startRowInclusive true to include start row, false to skip
-   * @param endRow ending row; set to null for positive infinity
-   * @param endRowInclusive true to include start row, false to skip
-   * @throws IllegalArgumentException if end row is before start row
+   *
+   * @param startRow
+   *          starting row; set to null for negative infinity
+   * @param startRowInclusive
+   *          true to include start row, false to skip
+   * @param endRow
+   *          ending row; set to null for positive infinity
+   * @param endRowInclusive
+   *          true to include start row, false to skip
+   * @throws IllegalArgumentException
+   *           if end row is before start row
    */
   public Range(CharSequence startRow, boolean startRowInclusive, CharSequence endRow, boolean endRowInclusive) {
     this(startRow == null ? null : new Text(startRow.toString()), startRowInclusive, endRow == null ? null : new Text(endRow.toString()), endRowInclusive);
   }
-  
+
   /**
    * Creates a range from startKey to endKey.
-   * 
-   * @param startKey starting key; set to null for negative infinity
-   * @param startKeyInclusive true to include start key, false to skip
-   * @param endKey ending key; set to null for positive infinity
-   * @param endKeyInclusive true to include start key, false to skip
-   * @throws IllegalArgumentException if end key is before start key
+   *
+   * @param startKey
+   *          starting key; set to null for negative infinity
+   * @param startKeyInclusive
+   *          true to include start key, false to skip
+   * @param endKey
+   *          ending key; set to null for positive infinity
+   * @param endKeyInclusive
+   *          true to include start key, false to skip
+   * @throws IllegalArgumentException
+   *           if end key is before start key
    */
   public Range(Key startKey, boolean startKeyInclusive, Key endKey, boolean endKeyInclusive) {
     this.start = startKey;
@@ -144,21 +170,22 @@ public class Range implements WritableComparable<Range> {
     this.stop = endKey;
     this.stopKeyInclusive = endKeyInclusive;
     this.infiniteStopKey = stop == null;
-    
+
     if (!infiniteStartKey && !infiniteStopKey && beforeStartKey(endKey)) {
       throw new IllegalArgumentException("Start key must be less than end key in range (" + startKey + ", " + endKey + ")");
     }
   }
-  
+
   /**
    * Copies a range.
    *
-   * @param range range to copy
+   * @param range
+   *          range to copy
    */
   public Range(Range range) {
     this(range.start, range.startKeyInclusive, range.infiniteStartKey, range.stop, range.stopKeyInclusive, range.infiniteStopKey);
   }
-  
+
   /**
    * Creates a range from start to stop.
    *
@@ -174,8 +201,8 @@ public class Range implements WritableComparable<Range> {
    *          true if start key is negative infinity (null)
    * @param infiniteStopKey
    *          true if stop key is positive infinity (null)
-   * @throws IllegalArgumentException if stop is before start, or infiniteStartKey is true but start is not null, or infiniteStopKey is true but stop is not
-   *          null
+   * @throws IllegalArgumentException
+   *           if stop is before start, or infiniteStartKey is true but start is not null, or infiniteStopKey is true but stop is not null
    */
   public Range(Key start, Key stop, boolean startKeyInclusive, boolean stopKeyInclusive, boolean infiniteStartKey, boolean infiniteStopKey) {
     this(start, startKeyInclusive, infiniteStartKey, stop, stopKeyInclusive, infiniteStopKey);
@@ -185,9 +212,8 @@ public class Range implements WritableComparable<Range> {
   }
 
   /**
-   * Creates a range from start to stop. Unlike the public six-argument method,
-   * this one does not assure that stop is after start, which helps performance
-   * in cases where that assurance is already in place.
+   * Creates a range from start to stop. Unlike the public six-argument method, this one does not assure that stop is after start, which helps performance in
+   * cases where that assurance is already in place.
    *
    * @param start
    *          set this to null when negative infinity is needed
@@ -201,15 +227,16 @@ public class Range implements WritableComparable<Range> {
    *          determines if the range includes the end key
    * @param infiniteStopKey
    *          true if stop key is positive infinity (null)
-   * @throws IllegalArgumentException if infiniteStartKey is true but start is not null, or infiniteStopKey is true but stop is not null
+   * @throws IllegalArgumentException
+   *           if infiniteStartKey is true but start is not null, or infiniteStopKey is true but stop is not null
    */
   protected Range(Key start, boolean startKeyInclusive, boolean infiniteStartKey, Key stop, boolean stopKeyInclusive, boolean infiniteStopKey) {
     if (infiniteStartKey && start != null)
       throw new IllegalArgumentException();
-    
+
     if (infiniteStopKey && stop != null)
       throw new IllegalArgumentException();
-    
+
     this.start = start;
     this.stop = stop;
     this.startKeyInclusive = startKeyInclusive;
@@ -217,20 +244,21 @@ public class Range implements WritableComparable<Range> {
     this.infiniteStartKey = infiniteStartKey;
     this.infiniteStopKey = infiniteStopKey;
   }
-  
+
   /**
    * Creates a range from a Thrift range.
    *
-   * @param trange Thrift range
+   * @param trange
+   *          Thrift range
    */
   public Range(TRange trange) {
-    this(trange.start == null ? null : new Key(trange.start), trange.startKeyInclusive, trange.infiniteStartKey,
-        trange.stop == null ? null : new Key(trange.stop), trange.stopKeyInclusive, trange.infiniteStopKey);
+    this(trange.start == null ? null : new Key(trange.start), trange.startKeyInclusive, trange.infiniteStartKey, trange.stop == null ? null : new Key(
+        trange.stop), trange.stopKeyInclusive, trange.infiniteStopKey);
     if (!infiniteStartKey && !infiniteStopKey && beforeStartKey(stop)) {
       throw new IllegalArgumentException("Start key must be less than end key in range (" + start + ", " + stop + ")");
     }
   }
-  
+
   /**
    * Gets the start key, or null if the start is negative infinity.
    *
@@ -242,23 +270,24 @@ public class Range implements WritableComparable<Range> {
     }
     return start;
   }
-  
+
   /**
    * Determines if the given key is before the start key of this range.
    *
-   * @param key key to check
+   * @param key
+   *          key to check
    * @return true if the given key is before the range, otherwise false
    */
   public boolean beforeStartKey(Key key) {
     if (infiniteStartKey) {
       return false;
     }
-    
+
     if (startKeyInclusive)
       return key.compareTo(start) < 0;
     return key.compareTo(start) <= 0;
   }
-  
+
   /**
    * Gets the ending key, or null if the end is positive infinity.
    *
@@ -270,60 +299,63 @@ public class Range implements WritableComparable<Range> {
     }
     return stop;
   }
-  
+
   /**
    * Determines if the given key is after the ending key of this range.
    *
-   * @param key key to check
+   * @param key
+   *          key to check
    * @return true if the given key is after the range, otherwise false
    */
   public boolean afterEndKey(Key key) {
     if (infiniteStopKey)
       return false;
-    
+
     if (stopKeyInclusive)
       return stop.compareTo(key) < 0;
     return stop.compareTo(key) <= 0;
   }
-  
+
   @Override
   public int hashCode() {
     int startHash = infiniteStartKey ? 0 : start.hashCode() + (startKeyInclusive ? 1 : 0);
     int stopHash = infiniteStopKey ? 0 : stop.hashCode() + (stopKeyInclusive ? 1 : 0);
-    
+
     return startHash + stopHash;
   }
-  
+
   @Override
   public boolean equals(Object o) {
     if (o instanceof Range)
       return equals((Range) o);
     return false;
   }
-  
+
   /**
    * Determines if this range equals another.
    *
-   * @param otherRange range to compare
+   * @param otherRange
+   *          range to compare
    * @return true if ranges are equals, false otherwise
    * @see #compareTo(Range)
    */
   public boolean equals(Range otherRange) {
-    
+
     return compareTo(otherRange) == 0;
   }
-  
+
   /**
    * Compares this range to another range. Compares in order: start key, inclusiveness of start key, end key, inclusiveness of end key. Infinite keys sort
    * first, and non-infinite keys are compared with {@link Key#compareTo(Key)}. Inclusive sorts before non-inclusive.
    *
-   * @param o range to compare
+   * @param o
+   *          range to compare
    * @return comparison result
    */
   @Override
   public int compareTo(Range o) {
     int comp;
-    
+
     if (infiniteStartKey)
       if (o.infiniteStartKey)
         comp = 0;
@@ -338,9 +370,9 @@ public class Range implements WritableComparable<Range> {
           comp = -1;
         else if (!startKeyInclusive && o.startKeyInclusive)
           comp = 1;
-      
+
     }
-    
+
     if (comp == 0)
       if (infiniteStopKey)
         if (o.infiniteStopKey)
@@ -357,60 +389,62 @@ public class Range implements WritableComparable<Range> {
           else if (!stopKeyInclusive && o.stopKeyInclusive)
             comp = -1;
       }
-    
+
     return comp;
   }
-  
+
   /**
    * Determines if the given key falls within this range.
    *
-   * @param key key to consider
+   * @param key
+   *          key to consider
    * @return true if the given key falls within the range, false otherwise
    */
   public boolean contains(Key key) {
     return !beforeStartKey(key) && !afterEndKey(key);
   }
-  
+
   /**
    * Merges overlapping and adjacent ranges. For example given the following input:
-   * 
+   *
    * <pre>
    * [a,c], (c, d], (g,m), (j,t]
    * </pre>
-   * 
+   *
    * the following ranges would be returned:
-   * 
+   *
    * <pre>
    * [a,d], (g,t]
    * </pre>
-   * 
-   * @param ranges to merge
+   *
+   * @param ranges
+   *          to merge
    * @return list of merged ranges
    */
   public static List<Range> mergeOverlapping(Collection<Range> ranges) {
     if (ranges.size() == 0)
       return Collections.emptyList();
-    
+
     List<Range> ral = new ArrayList<Range>(ranges);
     Collections.sort(ral);
-    
+
     ArrayList<Range> ret = new ArrayList<Range>(ranges.size());
-    
+
     Range currentRange = ral.get(0);
     boolean currentStartKeyInclusive = ral.get(0).startKeyInclusive;
-    
+
     for (int i = 1; i < ral.size(); i++) {
       // because of inclusive switch, equal keys may not be seen
-      
+
       if (currentRange.infiniteStopKey) {
         // this range has the minimal start key and
         // an infinite end key so it will contain all
         // other ranges
         break;
       }
-      
+
       Range range = ral.get(i);
-      
+
       boolean startKeysEqual;
       if (range.infiniteStartKey) {
         // previous start key must be infinite because it is sorted
@@ -423,11 +457,11 @@ public class Range implements WritableComparable<Range> {
       } else {
         startKeysEqual = false;
       }
-      
+
       if (startKeysEqual || currentRange.contains(range.start)
           || (!currentRange.stopKeyInclusive && range.startKeyInclusive && range.start.equals(currentRange.stop))) {
         int cmp;
-        
+
         if (range.infiniteStopKey || (cmp = range.stop.compareTo(currentRange.stop)) > 0 || (cmp == 0 && range.stopKeyInclusive)) {
           currentRange = new Range(currentRange.getStartKey(), currentStartKeyInclusive, range.getEndKey(), range.stopKeyInclusive);
         }/* else currentRange contains ral.get(i) */
@@ -437,50 +471,54 @@ public class Range implements WritableComparable<Range> {
         currentStartKeyInclusive = range.startKeyInclusive;
       }
     }
-    
+
     ret.add(currentRange);
-    
+
     return ret;
   }
-  
+
   /**
    * Creates a range which represents the intersection of this range and the passed in range. The following example will print true.
-   * 
+   *
    * <pre>
    * Range range1 = new Range(&quot;a&quot;, &quot;f&quot;);
    * Range range2 = new Range(&quot;c&quot;, &quot;n&quot;);
    * Range range3 = range1.clip(range2);
    * System.out.println(range3.equals(new Range(&quot;c&quot;, &quot;f&quot;)));
    * </pre>
-   * 
-   * @param range range to clip to
+   *
+   * @param range
+   *          range to clip to
    * @return the intersection of this range and the given range
-   * @throws IllegalArgumentException if ranges does not overlap
+   * @throws IllegalArgumentException
+   *           if ranges does not overlap
    */
   public Range clip(Range range) {
     return clip(range, false);
   }
-  
+
   /**
-   * Creates a range which represents the intersection of this range and the passed in range. Unlike {@link #clip(Range)},
-   * this method can optionally return null if the ranges do not overlap, instead of throwing an exception. The returnNullIfDisjoint parameter controls this
-   * behavior.
-   * 
-   * @param range range to clip to
-   * @param returnNullIfDisjoint true to return null if ranges are disjoint, false to throw an exception
+   * Creates a range which represents the intersection of this range and the passed in range. Unlike {@link #clip(Range)}, this method can optionally return
+   * null if the ranges do not overlap, instead of throwing an exception. The returnNullIfDisjoint parameter controls this behavior.
+   *
+   * @param range
+   *          range to clip to
+   * @param returnNullIfDisjoint
+   *          true to return null if ranges are disjoint, false to throw an exception
    * @return the intersection of this range and the given range, or null if ranges do not overlap and returnNullIfDisjoint is true
-   * @throws IllegalArgumentException if ranges does not overlap and returnNullIfDisjoint is false
+   * @throws IllegalArgumentException
+   *           if ranges does not overlap and returnNullIfDisjoint is false
    * @see Range#clip(Range)
    */
-  
+
   public Range clip(Range range, boolean returnNullIfDisjoint) {
-    
+
     Key sk = range.getStartKey();
     boolean ski = range.isStartKeyInclusive();
-    
+
     Key ek = range.getEndKey();
     boolean eki = range.isEndKeyInclusive();
-    
+
     if (range.getStartKey() == null) {
       if (getStartKey() != null) {
         sk = getStartKey();
@@ -495,7 +533,7 @@ public class Range implements WritableComparable<Range> {
       sk = getStartKey();
       ski = isStartKeyInclusive();
     }
-    
+
     if (range.getEndKey() == null) {
       if (getEndKey() != null) {
         ek = getEndKey();
@@ -510,67 +548,70 @@ public class Range implements WritableComparable<Range> {
       ek = getEndKey();
       eki = isEndKeyInclusive();
     }
-    
+
     return new Range(sk, ski, ek, eki);
   }
-  
+
   /**
    * Creates a new range that is bounded by the columns passed in. The start key in the returned range will have a column >= to the minimum column. The end key
    * in the returned range will have a column <= the max column.
-   * 
-   * @param min minimum column
-   * @param max maximum column
+   *
+   * @param min
+   *          minimum column
+   * @param max
+   *          maximum column
    * @return a column bounded range
-   * @throws IllegalArgumentException if the minimum column compares greater than the maximum column
+   * @throws IllegalArgumentException
+   *           if the minimum column compares greater than the maximum column
    */
   public Range bound(Column min, Column max) {
-    
+
     if (min.compareTo(max) > 0) {
       throw new IllegalArgumentException("min column > max column " + min + " " + max);
     }
-    
+
     Key sk = getStartKey();
     boolean ski = isStartKeyInclusive();
-    
+
     if (sk != null) {
-      
+
       ByteSequence cf = sk.getColumnFamilyData();
       ByteSequence cq = sk.getColumnQualifierData();
-      
+
       ByteSequence mincf = new ArrayByteSequence(min.columnFamily);
       ByteSequence mincq;
-      
+
       if (min.columnQualifier != null)
         mincq = new ArrayByteSequence(min.columnQualifier);
       else
         mincq = new ArrayByteSequence(new byte[0]);
-      
+
       int cmp = cf.compareTo(mincf);
-      
+
       if (cmp < 0 || (cmp == 0 && cq.compareTo(mincq) < 0)) {
         ski = true;
         sk = new Key(sk.getRowData().toArray(), mincf.toArray(), mincq.toArray(), new byte[0], Long.MAX_VALUE, true);
       }
     }
-    
+
     Key ek = getEndKey();
     boolean eki = isEndKeyInclusive();
-    
+
     if (ek != null) {
       ByteSequence row = ek.getRowData();
       ByteSequence cf = ek.getColumnFamilyData();
       ByteSequence cq = ek.getColumnQualifierData();
       ByteSequence cv = ek.getColumnVisibilityData();
-      
+
       ByteSequence maxcf = new ArrayByteSequence(max.columnFamily);
       ByteSequence maxcq = null;
       if (max.columnQualifier != null)
         maxcq = new ArrayByteSequence(max.columnQualifier);
-      
+
       boolean set = false;
-      
+
       int comp = cf.compareTo(maxcf);
-      
+
       if (comp > 0) {
         set = true;
       } else if (comp == 0 && maxcq != null && cq.compareTo(maxcq) > 0) {
@@ -580,7 +621,7 @@ public class Range implements WritableComparable<Range> {
         row = row.subSequence(0, row.length() - 1);
         set = true;
       }
-      
+
       if (set) {
         eki = false;
         if (maxcq == null)
@@ -589,16 +630,16 @@ public class Range implements WritableComparable<Range> {
           ek = new Key(row.toArray(), maxcf.toArray(), maxcq.toArray(), new byte[0], 0, false).followingKey(PartialKey.ROW_COLFAM_COLQUAL);
       }
     }
-    
+
     return new Range(sk, ski, ek, eki);
   }
-  
+
   @Override
   public String toString() {
     return ((startKeyInclusive && start != null) ? "[" : "(") + (start == null ? "-inf" : start) + "," + (stop == null ? "+inf" : stop)
         + ((stopKeyInclusive && stop != null) ? "]" : ")");
   }
-  
+
   @Override
   public void readFields(DataInput in) throws IOException {
     infiniteStartKey = in.readBoolean();
@@ -609,14 +650,14 @@ public class Range implements WritableComparable<Range> {
     } else {
       start = null;
     }
-    
+
     if (!infiniteStopKey) {
       stop = new Key();
       stop.readFields(in);
     } else {
       stop = null;
     }
-    
+
     startKeyInclusive = in.readBoolean();
     stopKeyInclusive = in.readBoolean();
 
@@ -624,7 +665,7 @@ public class Range implements WritableComparable<Range> {
       throw new InvalidObjectException("Start key must be less than end key in range (" + start + ", " + stop + ")");
     }
   }
-  
+
   @Override
   public void write(DataOutput out) throws IOException {
     out.writeBoolean(infiniteStartKey);
@@ -636,7 +677,7 @@ public class Range implements WritableComparable<Range> {
     out.writeBoolean(startKeyInclusive);
     out.writeBoolean(stopKeyInclusive);
   }
-  
+
   /**
    * Gets whether the start key of this range is inclusive.
    *
@@ -645,7 +686,7 @@ public class Range implements WritableComparable<Range> {
   public boolean isStartKeyInclusive() {
     return startKeyInclusive;
   }
-  
+
   /**
    * Gets whether the end key of this range is inclusive.
    *
@@ -654,7 +695,7 @@ public class Range implements WritableComparable<Range> {
   public boolean isEndKeyInclusive() {
     return stopKeyInclusive;
   }
-  
+
   /**
    * Converts this range to Thrift.
    *
@@ -664,7 +705,7 @@ public class Range implements WritableComparable<Range> {
     return new TRange(start == null ? null : start.toThrift(), stop == null ? null : stop.toThrift(), startKeyInclusive, stopKeyInclusive, infiniteStartKey,
         infiniteStopKey);
   }
-  
+
   /**
    * Gets whether the start key is negative infinity.
    *
@@ -673,7 +714,7 @@ public class Range implements WritableComparable<Range> {
   public boolean isInfiniteStartKey() {
     return infiniteStartKey;
   }
-  
+
   /**
    * Gets whether the end key is positive infinity.
    *
@@ -682,238 +723,289 @@ public class Range implements WritableComparable<Range> {
   public boolean isInfiniteStopKey() {
     return infiniteStopKey;
   }
-  
+
   /**
-   * Creates a range that covers an exact row. Returns the same Range as
-   * {@link #Range(Text)}.
-   * 
-   * @param row row to cover; set to null to cover all rows
+   * Creates a range that covers an exact row. Returns the same Range as {@link #Range(Text)}.
+   *
+   * @param row
+   *          row to cover; set to null to cover all rows
    */
   public static Range exact(Text row) {
     return new Range(row);
   }
-  
+
   /**
    * Creates a range that covers an exact row and column family.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
    */
   public static Range exact(Text row, Text cf) {
     Key startKey = new Key(row, cf);
     return new Range(startKey, true, startKey.followingKey(PartialKey.ROW_COLFAM), false);
   }
-  
+
   /**
    * Creates a range that covers an exact row, column family, and column qualifier.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
    */
   public static Range exact(Text row, Text cf, Text cq) {
     Key startKey = new Key(row, cf, cq);
     return new Range(startKey, true, startKey.followingKey(PartialKey.ROW_COLFAM_COLQUAL), false);
   }
-  
+
   /**
    * Creates a range that covers an exact row, column family, column qualifier, and column visibility.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
-   * @param cv column visibility to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
+   * @param cv
+   *          column visibility to cover
    */
   public static Range exact(Text row, Text cf, Text cq, Text cv) {
     Key startKey = new Key(row, cf, cq, cv);
     return new Range(startKey, true, startKey.followingKey(PartialKey.ROW_COLFAM_COLQUAL_COLVIS), false);
   }
-  
+
   /**
    * Creates a range that covers an exact row, column family, column qualifier, column visibility, and timestamp.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
-   * @param cv column visibility to cover
-   * @param ts timestamp to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
+   * @param cv
+   *          column visibility to cover
+   * @param ts
+   *          timestamp to cover
    */
   public static Range exact(Text row, Text cf, Text cq, Text cv, long ts) {
     Key startKey = new Key(row, cf, cq, cv, ts);
     return new Range(startKey, true, startKey.followingKey(PartialKey.ROW_COLFAM_COLQUAL_COLVIS_TIME), false);
   }
-  
+
   /**
    * Returns a Text that sorts just after all Texts beginning with a prefix.
-   * 
-   * @param prefix to follow
-   * @return prefix that immediately follows the given prefix when sorted, or
-   * null if no prefix can follow (i.e., the string is all 0xff bytes)
+   *
+   * @param prefix
+   *          to follow
+   * @return prefix that immediately follows the given prefix when sorted, or null if no prefix can follow (i.e., the string is all 0xff bytes)
    */
   public static Text followingPrefix(Text prefix) {
     byte[] prefixBytes = prefix.getBytes();
-    
+
     // find the last byte in the array that is not 0xff
     int changeIndex = prefix.getLength() - 1;
     while (changeIndex >= 0 && prefixBytes[changeIndex] == (byte) 0xff)
       changeIndex--;
     if (changeIndex < 0)
       return null;
-    
+
     // copy prefix bytes into new array
     byte[] newBytes = new byte[changeIndex + 1];
     System.arraycopy(prefixBytes, 0, newBytes, 0, changeIndex + 1);
-    
+
     // increment the selected byte
     newBytes[changeIndex]++;
     return new Text(newBytes);
   }
-  
+
   /**
    * Returns a Range that covers all rows beginning with a prefix.
-   * 
-   * @param rowPrefix prefix of rows to cover
+   *
+   * @param rowPrefix
+   *          prefix of rows to cover
    */
   public static Range prefix(Text rowPrefix) {
     Text fp = followingPrefix(rowPrefix);
     return new Range(new Key(rowPrefix), true, fp == null ? null : new Key(fp), false);
   }
-  
+
   /**
    * Returns a Range that covers all column families beginning with a prefix within a given row.
-   * 
-   * @param row row to cover
-   * @param cfPrefix prefix of column families to cover
+   *
+   * @param row
+   *          row to cover
+   * @param cfPrefix
+   *          prefix of column families to cover
    */
   public static Range prefix(Text row, Text cfPrefix) {
     Text fp = followingPrefix(cfPrefix);
     return new Range(new Key(row, cfPrefix), true, fp == null ? new Key(row).followingKey(PartialKey.ROW) : new Key(row, fp), false);
   }
-  
+
   /**
    * Returns a Range that covers all column qualifiers beginning with a prefix within a given row and column family.
-   * 
-   * @param row row to cover
-   * @param cf column family to cover
-   * @param cqPrefix prefix of column qualifiers to cover
+   *
+   * @param row
+   *          row to cover
+   * @param cf
+   *          column family to cover
+   * @param cqPrefix
+   *          prefix of column qualifiers to cover
    */
   public static Range prefix(Text row, Text cf, Text cqPrefix) {
     Text fp = followingPrefix(cqPrefix);
     return new Range(new Key(row, cf, cqPrefix), true, fp == null ? new Key(row, cf).followingKey(PartialKey.ROW_COLFAM) : new Key(row, cf, fp), false);
   }
-  
+
   /**
    * Returns a Range that covers all column visibilities beginning with a prefix within a given row, column family, and column qualifier.
-   * 
-   * @param row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
-   * @param cvPrefix prefix of column visibilities to cover
+   *
+   * @param row
+   *          row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
+   * @param cvPrefix
+   *          prefix of column visibilities to cover
    */
   public static Range prefix(Text row, Text cf, Text cq, Text cvPrefix) {
     Text fp = followingPrefix(cvPrefix);
     return new Range(new Key(row, cf, cq, cvPrefix), true, fp == null ? new Key(row, cf, cq).followingKey(PartialKey.ROW_COLFAM_COLQUAL) : new Key(row, cf, cq,
         fp), false);
   }
-  
+
   /**
    * Creates a range that covers an exact row.
-   * 
-   * @param row row to cover; set to null to cover all rows
+   *
+   * @param row
+   *          row to cover; set to null to cover all rows
    * @see #exact(Text)
    */
   public static Range exact(CharSequence row) {
     return Range.exact(new Text(row.toString()));
   }
-  
+
   /**
    * Creates a range that covers an exact row and column family.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
    * @see #exact(Text, Text)
    */
   public static Range exact(CharSequence row, CharSequence cf) {
     return Range.exact(new Text(row.toString()), new Text(cf.toString()));
   }
-  
+
   /**
    * Creates a range that covers an exact row, column family, and column qualifier.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
    * @see #exact(Text, Text, Text)
    */
   public static Range exact(CharSequence row, CharSequence cf, CharSequence cq) {
     return Range.exact(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()));
   }
-  
+
   /**
    * Creates a range that covers an exact row, column family, column qualifier, and column visibility.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
-   * @param cv column visibility to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
+   * @param cv
+   *          column visibility to cover
    * @see #exact(Text, Text, Text, Text)
    */
   public static Range exact(CharSequence row, CharSequence cf, CharSequence cq, CharSequence cv) {
     return Range.exact(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()), new Text(cv.toString()));
   }
-  
+
   /**
    * Creates a range that covers an exact row, column family, column qualifier, column visibility, and timestamp.
-   * 
-   * @param row row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
-   * @param cv column visibility to cover
-   * @param ts timestamp to cover
+   *
+   * @param row
+   *          row row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
+   * @param cv
+   *          column visibility to cover
+   * @param ts
+   *          timestamp to cover
    * @see #exact(Text, Text, Text, Text, long)
    */
   public static Range exact(CharSequence row, CharSequence cf, CharSequence cq, CharSequence cv, long ts) {
     return Range.exact(new Text(row.toString()), new Text(cf.toString()), new Text(cq.toString()), new Text(cv.toString()), ts);
   }
+
   /**
    * Returns a Range that covers all rows beginning with a prefix.
-   * 
-   * @param rowPrefix prefix of rows to cover
+   *
+   * @param rowPrefix
+   *          prefix of rows to cover
    * @see #prefix(Text)
    */
   public static Range prefix(CharSequence rowPrefix) {
     return Range.prefix(new Text(rowPrefix.toString()));
   }
-  
+
   /**
    * Returns a Range that covers all column families beginning with a prefix within a given row.
-   * 
-   * @param row row to cover
-   * @param cfPrefix prefix of column families to cover
+   *
+   * @param row
+   *          row to cover
+   * @param cfPrefix
+   *          prefix of column families to cover
    * @see #prefix(Text, Text)
    */
   public static Range prefix(CharSequence row, CharSequence cfPrefix) {
     return Range.prefix(new Text(row.toString()), new Text(cfPrefix.toString()));
   }
-  
+
   /**
    * Returns a Range that covers all column qualifiers beginning with a prefix within a given row and column family.
-   * 
-   * @param row row to cover
-   * @param cf column family to cover
-   * @param cqPrefix prefix of column qualifiers to cover
+   *
+   * @param row
+   *          row to cover
+   * @param cf
+   *          column family to cover
+   * @param cqPrefix
+   *          prefix of column qualifiers to cover
    * @see #prefix(Text, Text, Text)
    */
   public static Range prefix(CharSequence row, CharSequence cf, CharSequence cqPrefix) {
     return Range.prefix(new Text(row.toString()), new Text(cf.toString()), new Text(cqPrefix.toString()));
   }
+
   /**
    * Returns a Range that covers all column visibilities beginning with a prefix within a given row, column family, and column qualifier.
-   * 
-   * @param row row to cover
-   * @param cf column family to cover
-   * @param cq column qualifier to cover
-   * @param cvPrefix prefix of column visibilities to cover
+   *
+   * @param row
+   *          row to cover
+   * @param cf
+   *          column family to cover
+   * @param cq
+   *          column qualifier to cover
+   * @param cvPrefix
+   *          prefix of column visibilities to cover
    * @see #prefix(Text, Text, Text, Text)
    */
   public static Range prefix(CharSequence row, CharSequence cf, CharSequence cq, CharSequence cvPrefix) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/data/Value.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/data/Value.java b/core/src/main/java/org/apache/accumulo/core/data/Value.java
index 48eae02..e77e8a7 100644
--- a/core/src/main/java/org/apache/accumulo/core/data/Value.java
+++ b/core/src/main/java/org/apache/accumulo/core/data/Value.java
@@ -51,7 +51,7 @@ public class Value implements WritableComparable<Object> {
   /**
    * Creates a Value using a byte array as the initial value. The given byte array is used directly as the backing array, so later changes made to the array
    * reflect into the new Value.
-   * 
+   *
    * @param bytes
    *          May not be null
    */
@@ -61,7 +61,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Creates a Value using the bytes in a buffer as the initial value. Makes a defensive copy.
-   * 
+   *
    * @param bytes
    *          May not be null
    */
@@ -72,7 +72,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * @deprecated A copy of the bytes in the buffer is always made. Use {@link #Value(ByteBuffer)} instead.
-   * 
+   *
    * @param bytes
    *          bytes of value (may not be null)
    * @param copy
@@ -86,7 +86,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Creates a Value using a byte array as the initial value.
-   * 
+   *
    * @param bytes
    *          may not be null
    * @param copy
@@ -105,7 +105,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Creates a new Value based on another.
-   * 
+   *
    * @param ibw
    *          may not be null.
    */
@@ -115,7 +115,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Creates a Value based on a range in a byte array. A copy of the bytes is always made.
-   * 
+   *
    * @param newData
    *          source of copy, may not be null
    * @param offset
@@ -133,7 +133,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Gets the byte data of this value.
-   * 
+   *
    * @return the underlying byte array directly.
    */
   public byte[] get() {
@@ -143,7 +143,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Sets the byte data of this value. The given byte array is used directly as the backing array, so later changes made to the array reflect into this Value.
-   * 
+   *
    * @param b
    *          may not be null
    */
@@ -154,7 +154,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Sets the byte data of this value. The given byte array is copied.
-   * 
+   *
    * @param b
    *          may not be null
    */
@@ -166,7 +166,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Gets the size of this value.
-   * 
+   *
    * @return size in bytes
    */
   public int getSize() {
@@ -195,7 +195,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Define the sort order of the BytesWritable.
-   * 
+   *
    * @param right_obj
    *          The other bytes writable
    * @return Positive if left is bigger than right, 0 if they are equal, and negative if left is smaller than right.
@@ -207,7 +207,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Compares the bytes in this object to the specified byte array
-   * 
+   *
    * @return Positive if left is bigger than right, 0 if they are equal, and negative if left is smaller than right.
    */
   public int compareTo(final byte[] that) {
@@ -254,7 +254,7 @@ public class Value implements WritableComparable<Object> {
 
   /**
    * Converts a list of byte arrays to a two-dimensional array.
-   * 
+   *
    * @param array
    *          list of byte arrays
    * @return two-dimensional byte array containing one given byte array per row


[15/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/ExportTable.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/ExportTable.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/ExportTable.java
index f309879..4661dea8 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/ExportTable.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/ExportTable.java
@@ -64,9 +64,9 @@ import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 
 class ExportInfo implements Serializable {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   public String tableName;
   public String tableID;
   public String exportDir;
@@ -74,14 +74,14 @@ class ExportInfo implements Serializable {
 }
 
 class WriteExportFiles extends MasterRepo {
-  
+
   private static final long serialVersionUID = 1L;
   private final ExportInfo tableInfo;
-  
+
   WriteExportFiles(ExportInfo tableInfo) {
     this.tableInfo = tableInfo;
   }
-  
+
   private void checkOffline(Connector conn) throws Exception {
     if (Tables.getTableState(conn.getInstance(), tableInfo.tableID) != TableState.OFFLINE) {
       Tables.clearCache(conn.getInstance());
@@ -91,43 +91,43 @@ class WriteExportFiles extends MasterRepo {
       }
     }
   }
-  
+
   @Override
   public long isReady(long tid, Master master) throws Exception {
-    
+
     long reserved = Utils.reserveNamespace(tableInfo.namespaceID, tid, false, true, TableOperation.EXPORT)
         + Utils.reserveTable(tableInfo.tableID, tid, false, true, TableOperation.EXPORT);
     if (reserved > 0)
       return reserved;
-    
+
     Connector conn = master.getConnector();
-    
+
     checkOffline(conn);
-    
+
     Scanner metaScanner = conn.createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     metaScanner.setRange(new KeyExtent(new Text(tableInfo.tableID), null, null).toMetadataRange());
-    
+
     // scan for locations
     metaScanner.fetchColumnFamily(TabletsSection.CurrentLocationColumnFamily.NAME);
     metaScanner.fetchColumnFamily(TabletsSection.FutureLocationColumnFamily.NAME);
-    
+
     if (metaScanner.iterator().hasNext()) {
       return 500;
     }
-    
+
     // use the same range to check for walogs that we used to check for hosted (or future hosted) tablets
     // this is done as a separate scan after we check for locations, because walogs are okay only if there is no location
     metaScanner.clearColumns();
     metaScanner.fetchColumnFamily(LogColumnFamily.NAME);
-    
+
     if (metaScanner.iterator().hasNext()) {
       throw new ThriftTableOperationException(tableInfo.tableID, tableInfo.tableName, TableOperation.EXPORT, TableOperationExceptionType.OTHER,
           "Write ahead logs found for table");
     }
-    
+
     return 0;
   }
-  
+
   @Override
   public Repo<Master> call(long tid, Master master) throws Exception {
     try {
@@ -141,25 +141,25 @@ class WriteExportFiles extends MasterRepo {
     Utils.unreserveHdfsDirectory(new Path(tableInfo.exportDir).toString(), tid);
     return null;
   }
-  
+
   @Override
   public void undo(long tid, Master env) throws Exception {
     Utils.unreserveNamespace(tableInfo.namespaceID, tid, false);
     Utils.unreserveTable(tableInfo.tableID, tid, false);
   }
-  
+
   public static void exportTable(VolumeManager fs, AccumuloServerContext context, String tableName, String tableID, String exportDir) throws Exception {
-    
+
     fs.mkdirs(new Path(exportDir));
     Path exportMetaFilePath = fs.getVolumeByPath(new Path(exportDir)).getFileSystem().makeQualified(new Path(exportDir, Constants.EXPORT_FILE));
-    
+
     FSDataOutputStream fileOut = fs.create(exportMetaFilePath, false);
     ZipOutputStream zipOut = new ZipOutputStream(fileOut);
     BufferedOutputStream bufOut = new BufferedOutputStream(zipOut);
     DataOutputStream dataOut = new DataOutputStream(bufOut);
-    
+
     try {
-      
+
       zipOut.putNextEntry(new ZipEntry(Constants.EXPORT_INFO_FILE));
       OutputStreamWriter osw = new OutputStreamWriter(dataOut, UTF_8);
       osw.append(ExportTable.EXPORT_VERSION_PROP + ":" + ExportTable.VERSION + "\n");
@@ -170,72 +170,72 @@ class WriteExportFiles extends MasterRepo {
       osw.append("srcTableID:" + tableID + "\n");
       osw.append(ExportTable.DATA_VERSION_PROP + ":" + ServerConstants.DATA_VERSION + "\n");
       osw.append("srcCodeVersion:" + Constants.VERSION + "\n");
-      
+
       osw.flush();
       dataOut.flush();
-      
+
       exportConfig(context, tableID, zipOut, dataOut);
       dataOut.flush();
-      
+
       Map<String,String> uniqueFiles = exportMetadata(fs, context, tableID, zipOut, dataOut);
-      
+
       dataOut.close();
       dataOut = null;
-      
+
       createDistcpFile(fs, exportDir, exportMetaFilePath, uniqueFiles);
-      
+
     } finally {
       if (dataOut != null)
         dataOut.close();
     }
   }
-  
+
   private static void createDistcpFile(VolumeManager fs, String exportDir, Path exportMetaFilePath, Map<String,String> uniqueFiles) throws IOException {
     BufferedWriter distcpOut = new BufferedWriter(new OutputStreamWriter(fs.create(new Path(exportDir, "distcp.txt"), false), UTF_8));
-    
+
     try {
       for (String file : uniqueFiles.values()) {
         distcpOut.append(file);
         distcpOut.newLine();
       }
-      
+
       distcpOut.append(exportMetaFilePath.toString());
       distcpOut.newLine();
-      
+
       distcpOut.close();
       distcpOut = null;
-      
+
     } finally {
       if (distcpOut != null)
         distcpOut.close();
     }
   }
-  
+
   private static Map<String,String> exportMetadata(VolumeManager fs, AccumuloServerContext context, String tableID, ZipOutputStream zipOut,
       DataOutputStream dataOut) throws IOException, TableNotFoundException, AccumuloException, AccumuloSecurityException {
     zipOut.putNextEntry(new ZipEntry(Constants.EXPORT_METADATA_FILE));
-    
+
     Map<String,String> uniqueFiles = new HashMap<String,String>();
-    
+
     Scanner metaScanner = context.getConnector().createScanner(MetadataTable.NAME, Authorizations.EMPTY);
     metaScanner.fetchColumnFamily(DataFileColumnFamily.NAME);
     TabletsSection.TabletColumnFamily.PREV_ROW_COLUMN.fetch(metaScanner);
     TabletsSection.ServerColumnFamily.TIME_COLUMN.fetch(metaScanner);
     metaScanner.setRange(new KeyExtent(new Text(tableID), null, null).toMetadataRange());
-    
+
     for (Entry<Key,Value> entry : metaScanner) {
       entry.getKey().write(dataOut);
       entry.getValue().write(dataOut);
-      
+
       if (entry.getKey().getColumnFamily().equals(DataFileColumnFamily.NAME)) {
         String path = fs.getFullPath(entry.getKey()).toString();
         String tokens[] = path.split("/");
         if (tokens.length < 1) {
           throw new RuntimeException("Illegal path " + path);
         }
-        
+
         String filename = tokens[tokens.length - 1];
-        
+
         String existingPath = uniqueFiles.get(filename);
         if (existingPath == null) {
           uniqueFiles.put(filename, path);
@@ -243,30 +243,30 @@ class WriteExportFiles extends MasterRepo {
           // make sure file names are unique, should only apply for tables with file names generated by Accumulo 1.3 and earlier
           throw new IOException("Cannot export table with nonunique file names " + filename + ". Major compact table.");
         }
-        
+
       }
     }
     return uniqueFiles;
   }
-  
+
   private static void exportConfig(AccumuloServerContext context, String tableID, ZipOutputStream zipOut, DataOutputStream dataOut) throws AccumuloException,
       AccumuloSecurityException, TableNotFoundException, IOException {
     Connector conn = context.getConnector();
-    
+
     DefaultConfiguration defaultConfig = AccumuloConfiguration.getDefaultConfiguration();
     Map<String,String> siteConfig = conn.instanceOperations().getSiteConfiguration();
     Map<String,String> systemConfig = conn.instanceOperations().getSystemConfiguration();
-    
+
     TableConfiguration tableConfig = context.getServerConfigurationFactory().getTableConfiguration(tableID);
-    
+
     OutputStreamWriter osw = new OutputStreamWriter(dataOut, UTF_8);
-    
+
     // only put props that are different than defaults and higher level configurations
     zipOut.putNextEntry(new ZipEntry(Constants.EXPORT_TABLE_CONFIG_FILE));
     for (Entry<String,String> prop : tableConfig) {
       if (prop.getKey().startsWith(Property.TABLE_PREFIX.getKey())) {
         Property key = Property.getPropertyByKey(prop.getKey());
-        
+
         if (key == null || !defaultConfig.get(key).equals(prop.getValue())) {
           if (!prop.getValue().equals(siteConfig.get(prop.getKey())) && !prop.getValue().equals(systemConfig.get(prop.getKey()))) {
             osw.append(prop.getKey() + "=" + prop.getValue() + "\n");
@@ -274,16 +274,16 @@ class WriteExportFiles extends MasterRepo {
         }
       }
     }
-    
+
     osw.flush();
   }
 }
 
 public class ExportTable extends MasterRepo {
   private static final long serialVersionUID = 1L;
-  
+
   private final ExportInfo tableInfo;
-  
+
   public ExportTable(String tableName, String tableId, String exportDir) {
     tableInfo = new ExportInfo();
     tableInfo.tableName = tableName;
@@ -291,25 +291,25 @@ public class ExportTable extends MasterRepo {
     tableInfo.tableID = tableId;
     tableInfo.namespaceID = Tables.getNamespaceId(HdfsZooInstance.getInstance(), tableId);
   }
-  
+
   @Override
   public long isReady(long tid, Master environment) throws Exception {
     return Utils.reserveHdfsDirectory(new Path(tableInfo.exportDir).toString(), tid);
   }
-  
+
   @Override
   public Repo<Master> call(long tid, Master env) throws Exception {
     return new WriteExportFiles(tableInfo);
   }
-  
+
   @Override
   public void undo(long tid, Master env) throws Exception {
     Utils.unreserveHdfsDirectory(new Path(tableInfo.exportDir).toString(), tid);
   }
-  
+
   public static final int VERSION = 1;
-  
+
   public static final String DATA_VERSION_PROP = "srcDataVersion";
   public static final String EXPORT_VERSION_PROP = "exportVersion";
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/MasterRepo.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/MasterRepo.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/MasterRepo.java
index dfd287c..8f65fac 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/MasterRepo.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/MasterRepo.java
@@ -21,29 +21,29 @@ import org.apache.accumulo.master.Master;
 import org.apache.log4j.Logger;
 
 public abstract class MasterRepo implements Repo<Master> {
-  
+
   private static final long serialVersionUID = 1L;
   protected static final Logger log = Logger.getLogger(MasterRepo.class);
-  
+
   @Override
   public long isReady(long tid, Master environment) throws Exception {
     return 0;
   }
-  
+
   @Override
   public void undo(long tid, Master environment) throws Exception {}
-  
+
   @Override
   public String getDescription() {
     return this.getClass().getSimpleName();
   }
-  
+
   @Override
   public String getReturn() {
     return null;
   }
-  
+
   @Override
   abstract public Repo<Master> call(long tid, Master environment) throws Exception;
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/TableRangeOp.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/TableRangeOp.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/TableRangeOp.java
index ccb5d69..68c3e72 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/TableRangeOp.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/TableRangeOp.java
@@ -32,14 +32,14 @@ import org.apache.hadoop.io.Text;
 
 /**
  * Merge makes things hard.
- * 
+ *
  * Typically, a client will read the list of tablets, and begin an operation on that tablet at the location listed in the metadata table. When a tablet splits,
  * the information read from the metadata table doesn't match reality, so the operation fails, and must be retried. But the operation will take place either on
  * the parent, or at a later time on the children. It won't take place on just half of the tablet.
- * 
+ *
  * However, when a merge occurs, the operation may have succeeded on one section of the merged area, and not on the others, when the merge occurs. There is no
  * way to retry the request at a later time on an unmodified tablet.
- * 
+ *
  * The code below uses read-write lock to prevent some operations while a merge is taking place. Normal operations, like bulk imports, will grab the read lock
  * and prevent merges (writes) while they run. Merge operations will lock out some operations while they run.
  */
@@ -87,8 +87,7 @@ public class TableRangeOp extends MasterRepo {
   @Override
   public long isReady(long tid, Master environment) throws Exception {
     String namespaceId = Tables.getNamespaceId(environment.getInstance(), tableId);
-    return Utils.reserveNamespace(namespaceId, tid, false, true, TableOperation.MERGE)
-        + Utils.reserveTable(tableId, tid, true, true, TableOperation.MERGE);
+    return Utils.reserveNamespace(namespaceId, tid, false, true, TableOperation.MERGE) + Utils.reserveTable(tableId, tid, true, true, TableOperation.MERGE);
   }
 
   public TableRangeOp(MergeInfo.Operation op, String tableId, Text startRow, Text endRow) throws ThriftTableOperationException {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/TraceRepo.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/TraceRepo.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/TraceRepo.java
index 2571030..43f27bd 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/TraceRepo.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/TraceRepo.java
@@ -23,23 +23,23 @@ import org.apache.accumulo.core.trace.thrift.TInfo;
 import org.apache.accumulo.fate.Repo;
 
 /**
- * 
+ *
  */
 public class TraceRepo<T> implements Repo<T> {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   long traceId;
   long parentId;
   Repo<T> repo;
-  
+
   public TraceRepo(Repo<T> repo) {
     this.repo = repo;
     TInfo tinfo = Tracer.traceInfo();
     traceId = tinfo.traceId;
     parentId = tinfo.parentId;
   }
-  
+
   @Override
   public long isReady(long tid, T environment) throws Exception {
     Span span = Trace.trace(new TInfo(traceId, parentId), repo.getDescription());
@@ -49,7 +49,7 @@ public class TraceRepo<T> implements Repo<T> {
       span.stop();
     }
   }
-  
+
   @Override
   public Repo<T> call(long tid, T environment) throws Exception {
     Span span = Trace.trace(new TInfo(traceId, parentId), repo.getDescription());
@@ -62,7 +62,7 @@ public class TraceRepo<T> implements Repo<T> {
       span.stop();
     }
   }
-  
+
   @Override
   public void undo(long tid, T environment) throws Exception {
     Span span = Trace.trace(new TInfo(traceId, parentId), repo.getDescription());
@@ -72,15 +72,15 @@ public class TraceRepo<T> implements Repo<T> {
       span.stop();
     }
   }
-  
+
   @Override
   public String getDescription() {
     return repo.getDescription();
   }
-  
+
   @Override
   public String getReturn() {
     return repo.getReturn();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tableOps/Utils.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tableOps/Utils.java b/server/master/src/main/java/org/apache/accumulo/master/tableOps/Utils.java
index 64947fb..0f6025a 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tableOps/Utils.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tableOps/Utils.java
@@ -118,8 +118,7 @@ public class Utils {
   public static long reserveHdfsDirectory(String directory, long tid) throws KeeperException, InterruptedException {
     Instance instance = HdfsZooInstance.getInstance();
 
-    String resvPath = ZooUtil.getRoot(instance) + Constants.ZHDFS_RESERVATIONS + "/"
-        + Base64.encodeBase64String(directory.getBytes(UTF_8));
+    String resvPath = ZooUtil.getRoot(instance) + Constants.ZHDFS_RESERVATIONS + "/" + Base64.encodeBase64String(directory.getBytes(UTF_8));
 
     IZooReaderWriter zk = ZooReaderWriter.getInstance();
 
@@ -131,8 +130,7 @@ public class Utils {
 
   public static void unreserveHdfsDirectory(String directory, long tid) throws KeeperException, InterruptedException {
     Instance instance = HdfsZooInstance.getInstance();
-    String resvPath = ZooUtil.getRoot(instance) + Constants.ZHDFS_RESERVATIONS + "/"
-        + Base64.encodeBase64String(directory.getBytes(UTF_8));
+    String resvPath = ZooUtil.getRoot(instance) + Constants.ZHDFS_RESERVATIONS + "/" + Base64.encodeBase64String(directory.getBytes(UTF_8));
     ZooReservation.release(ZooReaderWriter.getInstance(), resvPath, String.format("%016x", tid));
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/main/java/org/apache/accumulo/master/tserverOps/ShutdownTServer.java
----------------------------------------------------------------------
diff --git a/server/master/src/main/java/org/apache/accumulo/master/tserverOps/ShutdownTServer.java b/server/master/src/main/java/org/apache/accumulo/master/tserverOps/ShutdownTServer.java
index 53c7c6f..6cce8ee 100644
--- a/server/master/src/main/java/org/apache/accumulo/master/tserverOps/ShutdownTServer.java
+++ b/server/master/src/main/java/org/apache/accumulo/master/tserverOps/ShutdownTServer.java
@@ -24,8 +24,8 @@ import org.apache.accumulo.core.zookeeper.ZooUtil;
 import org.apache.accumulo.fate.Repo;
 import org.apache.accumulo.fate.zookeeper.IZooReaderWriter;
 import org.apache.accumulo.fate.zookeeper.ZooUtil.NodeExistsPolicy;
-import org.apache.accumulo.master.Master;
 import org.apache.accumulo.master.EventCoordinator.Listener;
+import org.apache.accumulo.master.Master;
 import org.apache.accumulo.master.tableOps.MasterRepo;
 import org.apache.accumulo.server.master.LiveTServerSet.TServerConnection;
 import org.apache.accumulo.server.master.state.TServerInstance;
@@ -35,22 +35,22 @@ import org.apache.log4j.Logger;
 import org.apache.thrift.transport.TTransportException;
 
 public class ShutdownTServer extends MasterRepo {
-  
+
   private static final long serialVersionUID = 1L;
   private static final Logger log = Logger.getLogger(ShutdownTServer.class);
   private TServerInstance server;
   private boolean force;
-  
+
   public ShutdownTServer(TServerInstance server, boolean force) {
     this.server = server;
     this.force = force;
   }
-  
+
   @Override
   public long isReady(long tid, Master environment) throws Exception {
     return 0;
   }
-  
+
   @Override
   public Repo<Master> call(long tid, Master master) throws Exception {
     // suppress assignment of tablets to the server
@@ -62,7 +62,7 @@ public class ShutdownTServer extends MasterRepo {
       zoo.putPersistentData(path, "forced down".getBytes(UTF_8), NodeExistsPolicy.OVERWRITE);
       return null;
     }
-    
+
     // TODO move this to isReady() and drop while loop? - ACCUMULO-1259
     Listener listener = master.getEventCoordinator().getListener();
     master.shutdownTServer(server);
@@ -85,10 +85,10 @@ public class ShutdownTServer extends MasterRepo {
       }
       listener.waitForEvents(1000);
     }
-    
+
     return null;
   }
-  
+
   @Override
   public void undo(long tid, Master m) throws Exception {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/DefaultMapTest.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/DefaultMapTest.java b/server/master/src/test/java/org/apache/accumulo/master/DefaultMapTest.java
index 3389aa3..c0e9a4a 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/DefaultMapTest.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/DefaultMapTest.java
@@ -16,13 +16,14 @@
  */
 package org.apache.accumulo.master;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+
 import org.apache.accumulo.server.util.DefaultMap;
 import org.junit.Test;
 
-import static org.junit.Assert.*;
-
 public class DefaultMapTest {
-  
+
   @Test
   public void testDefaultMap() {
     DefaultMap<String,String> map = new DefaultMap<String,String>("");
@@ -32,5 +33,5 @@ public class DefaultMapTest {
     assertEquals(empty, "");
     assertTrue(empty == map.get("otherKey"));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/TestMergeState.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/TestMergeState.java b/server/master/src/test/java/org/apache/accumulo/master/TestMergeState.java
index 5c46ddc..d5951cc 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/TestMergeState.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/TestMergeState.java
@@ -54,7 +54,7 @@ import org.junit.Test;
 import com.google.common.net.HostAndPort;
 
 /**
- * 
+ *
  */
 public class TestMergeState {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/replication/DistributedWorkQueueWorkAssignerHelperTest.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/replication/DistributedWorkQueueWorkAssignerHelperTest.java b/server/master/src/test/java/org/apache/accumulo/master/replication/DistributedWorkQueueWorkAssignerHelperTest.java
index feee6d5..789a634 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/replication/DistributedWorkQueueWorkAssignerHelperTest.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/replication/DistributedWorkQueueWorkAssignerHelperTest.java
@@ -27,7 +27,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class DistributedWorkQueueWorkAssignerHelperTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/replication/MasterReplicationCoordinatorTest.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/replication/MasterReplicationCoordinatorTest.java b/server/master/src/test/java/org/apache/accumulo/master/replication/MasterReplicationCoordinatorTest.java
index a2ea329..36e71f8 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/replication/MasterReplicationCoordinatorTest.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/replication/MasterReplicationCoordinatorTest.java
@@ -30,7 +30,7 @@ import org.junit.Test;
 import com.google.common.net.HostAndPort;
 
 /**
- * 
+ *
  */
 public class MasterReplicationCoordinatorTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/replication/StatusMakerTest.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/replication/StatusMakerTest.java b/server/master/src/test/java/org/apache/accumulo/master/replication/StatusMakerTest.java
index 6a8fa4d..7ee3a89 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/replication/StatusMakerTest.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/replication/StatusMakerTest.java
@@ -54,7 +54,7 @@ import com.google.common.collect.Iterables;
 import com.google.common.collect.Sets;
 
 /**
- * 
+ *
  */
 public class StatusMakerTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/state/MergeInfoTest.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/state/MergeInfoTest.java b/server/master/src/test/java/org/apache/accumulo/master/state/MergeInfoTest.java
index d7fc619..8ac3e42 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/state/MergeInfoTest.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/state/MergeInfoTest.java
@@ -26,7 +26,7 @@ import org.junit.Assert;
 import org.junit.Test;
 
 public class MergeInfoTest {
-  
+
   MergeInfo readWrite(MergeInfo info) throws Exception {
     DataOutputBuffer buffer = new DataOutputBuffer();
     info.write(buffer);
@@ -39,11 +39,11 @@ public class MergeInfoTest {
     Assert.assertEquals(info.getOperation(), info2.getOperation());
     return info2;
   }
-  
+
   KeyExtent ke(String tableId, String endRow, String prevEndRow) {
     return new KeyExtent(new Text(tableId), endRow == null ? null : new Text(endRow), prevEndRow == null ? null : new Text(prevEndRow));
   }
-  
+
   @Test
   public void testWritable() throws Exception {
     MergeInfo info;
@@ -54,7 +54,7 @@ public class MergeInfoTest {
     Assert.assertTrue(info.isDelete());
     info.setState(MergeState.COMPLETE);
   }
-  
+
   @Test
   public void testNeedsToBeChopped() throws Exception {
     MergeInfo info = new MergeInfo(ke("x", "b", "a"), MergeInfo.Operation.DELETE);
@@ -72,5 +72,5 @@ public class MergeInfoTest {
     Assert.assertFalse(info.needsToBeChopped(ke("x", "c", "bb")));
     Assert.assertTrue(info.needsToBeChopped(ke("x", "b", "a")));
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/master/src/test/java/org/apache/accumulo/master/state/RootTabletStateStoreTest.java
----------------------------------------------------------------------
diff --git a/server/master/src/test/java/org/apache/accumulo/master/state/RootTabletStateStoreTest.java b/server/master/src/test/java/org/apache/accumulo/master/state/RootTabletStateStoreTest.java
index 3479d35..abceae4 100644
--- a/server/master/src/test/java/org/apache/accumulo/master/state/RootTabletStateStoreTest.java
+++ b/server/master/src/test/java/org/apache/accumulo/master/state/RootTabletStateStoreTest.java
@@ -34,8 +34,8 @@ import org.apache.accumulo.server.master.state.DistributedStore;
 import org.apache.accumulo.server.master.state.DistributedStoreException;
 import org.apache.accumulo.server.master.state.TServerInstance;
 import org.apache.accumulo.server.master.state.TabletLocationState;
-import org.apache.accumulo.server.master.state.ZooTabletStateStore;
 import org.apache.accumulo.server.master.state.TabletLocationState.BadLocationStateException;
+import org.apache.accumulo.server.master.state.ZooTabletStateStore;
 import org.apache.hadoop.io.Text;
 import org.junit.Assert;
 import org.junit.Test;
@@ -43,16 +43,16 @@ import org.junit.Test;
 import com.google.common.net.HostAndPort;
 
 public class RootTabletStateStoreTest {
-  
+
   static class Node {
     Node(String name) {
       this.name = name;
     }
-    
+
     List<Node> children = new ArrayList<Node>();
     String name;
     byte[] value = new byte[] {};
-    
+
     Node find(String name) {
       for (Node node : children)
         if (node.name.equals(name))
@@ -60,11 +60,11 @@ public class RootTabletStateStoreTest {
       return null;
     }
   };
-  
+
   static class FakeZooStore implements DistributedStore {
-    
+
     Node root = new Node("/");
-    
+
     private Node recurse(Node root, String[] path, int depth) {
       if (depth == path.length)
         return root;
@@ -73,12 +73,12 @@ public class RootTabletStateStoreTest {
         return null;
       return recurse(child, path, depth + 1);
     }
-    
+
     private Node navigate(String path) {
       path = path.replaceAll("/$", "");
       return recurse(root, path.split("/"), 1);
     }
-    
+
     @Override
     public List<String> getChildren(String path) throws DistributedStoreException {
       Node node = navigate(path);
@@ -89,17 +89,17 @@ public class RootTabletStateStoreTest {
         children.add(child.name);
       return children;
     }
-    
+
     @Override
     public void put(String path, byte[] bs) throws DistributedStoreException {
       create(path).value = bs;
     }
-    
+
     private Node create(String path) {
       String[] parts = path.split("/");
       return recurseCreate(root, parts, 1);
     }
-    
+
     private Node recurseCreate(Node root, String[] path, int index) {
       if (path.length == index)
         return root;
@@ -110,7 +110,7 @@ public class RootTabletStateStoreTest {
       }
       return recurseCreate(node, path, index + 1);
     }
-    
+
     @Override
     public void remove(String path) throws DistributedStoreException {
       String[] parts = path.split("/");
@@ -122,7 +122,7 @@ public class RootTabletStateStoreTest {
       if (child != null)
         parent.children.remove(child);
     }
-    
+
     @Override
     public byte[] get(String path) throws DistributedStoreException {
       Node node = navigate(path);
@@ -131,7 +131,7 @@ public class RootTabletStateStoreTest {
       return null;
     }
   }
-  
+
   @Test
   public void testFakeZoo() throws DistributedStoreException {
     DistributedStore store = new FakeZooStore();
@@ -149,7 +149,7 @@ public class RootTabletStateStoreTest {
     children = store.getChildren("/a/b");
     assertEquals(new HashSet<String>(children), new HashSet<String>(Arrays.asList("b")));
   }
-  
+
   @Test
   public void testRootTabletStateStore() throws DistributedStoreException {
     ZooTabletStateStore tstore = new ZooTabletStateStore(new FakeZooStore());
@@ -190,18 +190,18 @@ public class RootTabletStateStoreTest {
       count++;
     }
     assertEquals(count, 1);
-    
+
     KeyExtent notRoot = new KeyExtent(new Text("0"), null, null);
     try {
       tstore.setLocations(Collections.singletonList(new Assignment(notRoot, server)));
       Assert.fail("should not get here");
     } catch (IllegalArgumentException ex) {}
-    
+
     try {
       tstore.setFutureLocations(Collections.singletonList(new Assignment(notRoot, server)));
       Assert.fail("should not get here");
     } catch (IllegalArgumentException ex) {}
-    
+
     TabletLocationState broken = null;
     try {
       broken = new TabletLocationState(notRoot, server, null, null, null, false);
@@ -213,7 +213,7 @@ public class RootTabletStateStoreTest {
       Assert.fail("should not get here");
     } catch (IllegalArgumentException ex) {}
   }
-  
+
   // @Test
   // public void testMetaDataStore() { } // see functional test
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/Monitor.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/Monitor.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/Monitor.java
index 218bcaf..1a2904c 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/Monitor.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/Monitor.java
@@ -523,6 +523,7 @@ public class Monitor {
     public final long scanCount;
     public final Long oldestScan;
     public final long fetched;
+
     ScanStats(List<ActiveScan> active) {
       this.scanCount = active.size();
       long oldest = -1;
@@ -535,9 +536,10 @@ public class Monitor {
   }
 
   static final Map<HostAndPort,ScanStats> allScans = new HashMap<HostAndPort,ScanStats>();
-  public static Map<HostAndPort, ScanStats> getScans() {
+
+  public static Map<HostAndPort,ScanStats> getScans() {
     synchronized (allScans) {
-      return new TreeMap<HostAndPort, ScanStats>(allScans);
+      return new TreeMap<HostAndPort,ScanStats>(allScans);
     }
   }
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/ZooKeeperStatus.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/ZooKeeperStatus.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/ZooKeeperStatus.java
index 907d1f1..89e879e 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/ZooKeeperStatus.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/ZooKeeperStatus.java
@@ -33,22 +33,22 @@ import org.apache.thrift.transport.TTransportException;
 import com.google.common.net.HostAndPort;
 
 public class ZooKeeperStatus implements Runnable {
-  
+
   private static final Logger log = Logger.getLogger(ZooKeeperStatus.class);
-  
+
   private volatile boolean stop = false;
-  
+
   public static class ZooKeeperState implements Comparable<ZooKeeperState> {
     public final String keeper;
     public final String mode;
     public final int clients;
-    
+
     public ZooKeeperState(String keeper, String mode, int clients) {
       this.keeper = keeper;
       this.mode = mode;
       this.clients = clients;
     }
-    
+
     @Override
     public int compareTo(ZooKeeperState other) {
       if (this == other) {
@@ -68,29 +68,29 @@ public class ZooKeeperStatus implements Runnable {
       }
     }
   }
-  
+
   private static SortedSet<ZooKeeperState> status = new TreeSet<ZooKeeperState>();
-  
+
   public static Collection<ZooKeeperState> getZooKeeperStatus() {
     return status;
   }
-  
+
   public void stop() {
     this.stop = true;
   }
-  
+
   @Override
   public void run() {
-    
+
     while (!stop) {
-      
+
       TreeSet<ZooKeeperState> update = new TreeSet<ZooKeeperState>();
-      
+
       String zookeepers[] = SiteConfiguration.getInstance().get(Property.INSTANCE_ZK_HOST).split(",");
       for (String keeper : zookeepers) {
         int clients = 0;
         String mode = "unknown";
-        
+
         String[] parts = keeper.split(":");
         TTransport transport = null;
         try {
@@ -99,7 +99,7 @@ public class ZooKeeperStatus implements Runnable {
             addr = HostAndPort.fromParts(parts[0], Integer.parseInt(parts[1]));
           else
             addr = HostAndPort.fromParts(parts[0], 2181);
-          
+
           transport = TTimeoutTransport.create(addr, 10 * 1000l);
           transport.write("stat\n".getBytes(UTF_8), 0, 5);
           StringBuilder response = new StringBuilder();
@@ -137,5 +137,5 @@ public class ZooKeeperStatus implements Runnable {
       UtilWaitThread.sleep(5 * 1000);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/GcStatusServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/GcStatusServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/GcStatusServlet.java
index 0f0db9e..926eedc 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/GcStatusServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/GcStatusServlet.java
@@ -29,18 +29,18 @@ import org.apache.accumulo.monitor.util.celltypes.DurationType;
 import org.apache.accumulo.monitor.util.celltypes.NumberType;
 
 public class GcStatusServlet extends BasicServlet {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "Garbage Collector Status";
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     GCStatus status = Monitor.getGcStatus();
-    
+
     if (status != null) {
       Table gcActivity = new Table("gcActivity", "Collection&nbsp;Activity");
       gcActivity.addSortableColumn("Activity");
@@ -50,7 +50,7 @@ public class GcStatusServlet extends BasicServlet {
       gcActivity.addSortableColumn("In&nbsp;Use", new NumberType<Long>(), null);
       gcActivity.addSortableColumn("Errors", new NumberType<Long>(0l, 1l), null);
       gcActivity.addSortableColumn("Duration", new DurationType(), null);
-      
+
       if (status.last.finished > 0)
         gcActivity.addRow("File&nbsp;Collection,&nbsp;Last&nbsp;Cycle", status.last.finished, status.last.candidates, status.last.deleted, status.last.inUse,
             status.last.errors, status.last.finished - status.last.started);
@@ -68,5 +68,5 @@ public class GcStatusServlet extends BasicServlet {
       banner(sb, "error", "Collector is Unavailable");
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/JSONServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/JSONServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/JSONServlet.java
index 09dbb6e..224e1a1 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/JSONServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/JSONServlet.java
@@ -36,19 +36,19 @@ import com.google.gson.Gson;
 
 public class JSONServlet extends BasicServlet {
   private static final long serialVersionUID = 1L;
-  
+
   private Gson gson = new Gson();
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "JSON Report";
   }
-  
+
   @Override
   protected void pageStart(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     resp.setContentType("application/json");
   }
-  
+
   private static Map<String,Object> addServer(String ip, String hostname, double osload, double ingest, double query, double ingestMB, double queryMB,
       int scans, double scansessions, long holdtime) {
     Map<String,Object> map = new HashMap<String,Object>();
@@ -64,41 +64,41 @@ public class JSONServlet extends BasicServlet {
     map.put("holdtime", holdtime);
     return map;
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     if (Monitor.getMmi() == null || Monitor.getMmi().tableMap == null) {
       return;
     }
-    
+
     Map<String,Object> results = new HashMap<String,Object>();
     List<Map<String,Object>> servers = new ArrayList<Map<String,Object>>();
-    
+
     for (TabletServerStatus status : Monitor.getMmi().tServerInfo) {
       TableInfo summary = TableInfoUtil.summarizeTableStats(status);
       servers.add(addServer(status.name, TServerLinkType.displayName(status.name), status.osLoad, summary.ingestRate, summary.queryRate,
           summary.ingestByteRate / 1000000.0, summary.queryByteRate / 1000000.0, summary.scans.running + summary.scans.queued, Monitor.getLookupRate(),
           status.holdTime));
     }
-    
+
     for (Entry<String,Byte> entry : Monitor.getMmi().badTServers.entrySet()) {
       Map<String,Object> badServer = new HashMap<String,Object>();
       badServer.put("ip", entry.getKey());
       badServer.put("bad", true);
       servers.add(badServer);
     }
-    
+
     for (DeadServer dead : Monitor.getMmi().deadTabletServers) {
       Map<String,Object> deadServer = new HashMap<String,Object>();
       deadServer.put("ip", dead.server);
       deadServer.put("dead", true);
       servers.add(deadServer);
     }
-    
+
     results.put("servers", servers);
     sb.append(gson.toJson(results));
   }
-  
+
   @Override
   protected void pageEnd(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/LogServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/LogServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/LogServlet.java
index 7c2172a..77f14de 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/LogServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/LogServlet.java
@@ -33,14 +33,14 @@ import org.apache.log4j.Level;
 import org.apache.log4j.spi.LoggingEvent;
 
 public class LogServlet extends BasicServlet {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "Recent Logs";
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     AccumuloConfiguration conf = Monitor.getContext().getConfiguration();
@@ -79,7 +79,7 @@ public class LogServlet extends BasicServlet {
           default:
             text.append(c);
         }
-        
+
       }
       StringBuilder builder = new StringBuilder(text.toString());
       if (ev.getThrowableStrRep() != null)
@@ -95,13 +95,13 @@ public class LogServlet extends BasicServlet {
     if (!clear)
       sb.append("<div class='center'><a href='/op?action=clearLog&redir=").append(currentPage(req)).append("'>Clear&nbsp;All&nbsp;Events</a></div>\n");
   }
-  
+
   private static class LogLevelType extends StringType<Level> {
     @Override
     public String alignment() {
       return "center";
     }
-    
+
     @Override
     public String format(Object obj) {
       if (obj == null)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/MasterServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/MasterServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/MasterServlet.java
index 2f7086f..64b8648 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/MasterServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/MasterServlet.java
@@ -27,8 +27,6 @@ import java.util.Map;
 import javax.servlet.http.HttpServletRequest;
 import javax.servlet.http.HttpServletResponse;
 
-import org.apache.accumulo.monitor.util.celltypes.PreciseNumberType;
-
 import org.apache.accumulo.core.client.impl.Tables;
 import org.apache.accumulo.core.conf.Property;
 import org.apache.accumulo.core.master.thrift.DeadServer;
@@ -42,6 +40,7 @@ import org.apache.accumulo.monitor.util.Table;
 import org.apache.accumulo.monitor.util.TableRow;
 import org.apache.accumulo.monitor.util.celltypes.DurationType;
 import org.apache.accumulo.monitor.util.celltypes.NumberType;
+import org.apache.accumulo.monitor.util.celltypes.PreciseNumberType;
 import org.apache.accumulo.monitor.util.celltypes.ProgressChartType;
 import org.apache.accumulo.monitor.util.celltypes.StringType;
 import org.apache.accumulo.server.monitor.DedupedLogEvent;
@@ -51,26 +50,26 @@ import org.apache.log4j.Level;
 import com.google.common.base.Joiner;
 
 public class MasterServlet extends BasicServlet {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     List<String> masters = Monitor.getContext().getInstance().getMasterLocations();
     return "Master Server" + (masters.size() == 0 ? "" : ":" + AddressUtil.parseAddress(masters.get(0), false).getHostText());
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse response, StringBuilder sb) throws IOException {
     Map<String,String> tidToNameMap = Tables.getIdToNameMap(Monitor.getContext().getInstance());
-    
+
     doLogEventBanner(sb);
     TablesServlet.doProblemsBanner(sb);
     doMasterStatus(req, sb);
     doRecoveryList(req, sb);
     TablesServlet.doTableList(req, sb, tidToNameMap);
   }
-  
+
   private void doLogEventBanner(StringBuilder sb) {
     if (LogService.getInstance().getEvents().size() > 0) {
       int error = 0, warning = 0, total = 0;
@@ -90,9 +89,9 @@ public class MasterServlet extends BasicServlet {
           : "s", warning, warning == 1 ? "" : "s", total));
     }
   }
-  
+
   private void doMasterStatus(HttpServletRequest req, StringBuilder sb) throws IOException {
-    
+
     if (Monitor.getMmi() != null) {
       String gcStatus = "Waiting";
       if (Monitor.getGcStatus() != null) {
@@ -121,7 +120,7 @@ public class MasterServlet extends BasicServlet {
       if (Monitor.getMmi().serversShuttingDown != null && Monitor.getMmi().serversShuttingDown.size() > 0 && Monitor.getMmi().state == MasterState.NORMAL) {
         sb.append("<span class='warning'>Servers being stopped: " + Joiner.on(", ").join(Monitor.getMmi().serversShuttingDown) + "</span>\n");
       }
-      
+
       int guessHighLoad = ManagementFactory.getOperatingSystemMXBean().getAvailableProcessors();
       List<String> slaves = new ArrayList<String>();
       for (TabletServerStatus up : Monitor.getMmi().tServerInfo) {
@@ -131,7 +130,7 @@ public class MasterServlet extends BasicServlet {
         slaves.add(down.server);
       }
       List<String> masters = Monitor.getContext().getInstance().getMasterLocations();
-      
+
       Table masterStatus = new Table("masterStatus", "Master&nbsp;Status");
       masterStatus.addSortableColumn("Master", new StringType<String>(), "The hostname of the master server");
       masterStatus.addSortableColumn("#&nbsp;Online<br />Tablet&nbsp;Servers", new PreciseNumberType((int) (slaves.size() * 0.8 + 1.0), slaves.size(),
@@ -165,11 +164,11 @@ public class MasterServlet extends BasicServlet {
       row.add(ManagementFactory.getOperatingSystemMXBean().getSystemLoadAverage());
       masterStatus.addRow(row);
       masterStatus.generate(req, sb);
-      
+
     } else
       banner(sb, "error", "Master Server Not Running");
   }
-  
+
   private void doRecoveryList(HttpServletRequest req, StringBuilder sb) {
     MasterMonitorInfo mmi = Monitor.getMmi();
     if (mmi != null) {
@@ -197,5 +196,5 @@ public class MasterServlet extends BasicServlet {
         recoveryTable.generate(req, sb);
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ProblemServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ProblemServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ProblemServlet.java
index 4f7da48..1b3e944 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ProblemServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ProblemServlet.java
@@ -38,31 +38,31 @@ import org.apache.accumulo.server.problems.ProblemReports;
 import org.apache.accumulo.server.problems.ProblemType;
 
 public class ProblemServlet extends BasicServlet {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "Per-Table Problem Report";
   }
-  
+
   @Override
   protected void pageBody(final HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     Map<String,String> tidToNameMap = Tables.getIdToNameMap(Monitor.getContext().getInstance());
     doProblemSummary(req, sb, tidToNameMap);
     doProblemDetails(req, sb, req.getParameter("table"), tidToNameMap);
   }
-  
+
   private static void doProblemSummary(final HttpServletRequest req, StringBuilder sb, final Map<String,String> tidToNameMap) {
     if (Monitor.getProblemSummary().isEmpty() && Monitor.getProblemException() == null)
       return;
-    
+
     Table problemSummary = new Table("problemSummary", "Problem&nbsp;Summary", "error");
     problemSummary.addSortableColumn("Table", new TableProblemLinkType(tidToNameMap), null);
     for (ProblemType type : ProblemType.values())
       problemSummary.addSortableColumn(type.name(), new NumberType<Integer>(), null);
     problemSummary.addUnsortableColumn("Operations", new ClearTableProblemsLinkType(req, tidToNameMap), null);
-    
+
     if (Monitor.getProblemException() != null) {
       StringBuilder cell = new StringBuilder();
       cell.append("<b>Failed to obtain problem reports</b> : " + Monitor.getProblemException().getMessage());
@@ -87,12 +87,12 @@ public class ProblemServlet extends BasicServlet {
     }
     problemSummary.generate(req, sb);
   }
-  
+
   private static void doProblemDetails(final HttpServletRequest req, StringBuilder sb, String tableId, Map<String,String> tidToNameMap) {
-    
+
     if (Monitor.getProblemException() != null)
       return;
-    
+
     ArrayList<ProblemReport> problemReports = new ArrayList<ProblemReport>();
     Iterator<ProblemReport> iter = tableId == null ? ProblemReports.getInstance(Monitor.getContext()).iterator() : ProblemReports.getInstance(
         Monitor.getContext()).iterator(tableId);
@@ -109,7 +109,7 @@ public class ProblemServlet extends BasicServlet {
     problemTable.addSortableColumn("Exception");
     problemTable.addUnsortableColumn("Operations", new ClearProblemLinkType(req), null);
     for (ProblemReport pr : problemReports) {
-      
+
       TableRow row = problemTable.prepareRow();
       row.add(pr.getTableName());
       row.add(pr.getProblemType().name());
@@ -122,14 +122,14 @@ public class ProblemServlet extends BasicServlet {
     }
     problemTable.generate(req, sb);
   }
-  
+
   private static class TableProblemLinkType extends StringType<String> {
     private Map<String,String> tidToNameMap;
-    
+
     public TableProblemLinkType(Map<String,String> tidToNameMap) {
       this.tidToNameMap = tidToNameMap;
     }
-    
+
     @Override
     public String format(Object obj) {
       if (obj == null)
@@ -138,21 +138,21 @@ public class ProblemServlet extends BasicServlet {
       return String.format("<a href='/problems?table=%s'>%s</a>", encode(table), encode((Tables.getPrintableTableNameFromId(tidToNameMap, table))));
     }
   }
-  
+
   private static class ClearTableProblemsLinkType extends StringType<String> {
     private HttpServletRequest req;
     private Map<String,String> tidToNameMap;
-    
+
     public ClearTableProblemsLinkType(HttpServletRequest req, Map<String,String> tidToNameMap) {
       this.req = req;
       this.tidToNameMap = tidToNameMap;
     }
-    
+
     @Override
     public String alignment() {
       return "right";
     }
-    
+
     @Override
     public String format(Object obj) {
       if (obj == null)
@@ -162,19 +162,19 @@ public class ProblemServlet extends BasicServlet {
           Tables.getPrintableTableNameFromId(tidToNameMap, table));
     }
   }
-  
+
   private static class ClearProblemLinkType extends CellType<ProblemReport> {
     private HttpServletRequest req;
-    
+
     public ClearProblemLinkType(HttpServletRequest req) {
       this.req = req;
     }
-    
+
     @Override
     public String alignment() {
       return "right";
     }
-    
+
     @Override
     public String format(Object obj) {
       if (obj == null)
@@ -183,11 +183,11 @@ public class ProblemServlet extends BasicServlet {
       return String.format("<a href='/op?table=%s&action=clearProblem&redir=%s&resource=%s&ptype=%s'>clear this problem</a>", encode(p.getTableName()),
           currentPage(req), encode(p.getResource()), encode(p.getProblemType().name()));
     }
-    
+
     @Override
     public int compare(ProblemReport o1, ProblemReport o2) {
       return 0;
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ShellServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ShellServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ShellServlet.java
index 6b2b31f..55145f9 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ShellServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/ShellServlet.java
@@ -49,7 +49,7 @@ public class ShellServlet extends BasicServlet {
   protected String getTitle(HttpServletRequest req) {
     return "Shell";
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse response, StringBuilder sb) throws IOException {
     HttpSession session = req.getSession(true);
@@ -216,27 +216,27 @@ public class ShellServlet extends BasicServlet {
     return "<div id='login'><form method=POST action='" + requestURI + "'>"
         + "<table><tr><td>Mock:&nbsp</td><td><input type='checkbox' name='mock' value='mock'></td></tr>"
         + "<tr><td>Username:&nbsp;</td><td><input type='text' name='user'></td></tr>"
-        + "<tr><td>Password:&nbsp;</td><td><input type='password' name='pass'></td><td>"
-        + "<input type='hidden' name='" + CSRF_KEY + "' value='" + csrfToken + "'/><input type='submit' value='Enter'></td></tr></table></form></div>";
+        + "<tr><td>Password:&nbsp;</td><td><input type='password' name='pass'></td><td>" + "<input type='hidden' name='" + CSRF_KEY + "' value='" + csrfToken
+        + "'/><input type='submit' value='Enter'></td></tr></table></form></div>";
   }
-  
+
   private static class StringBuilderOutputStream extends OutputStream {
     StringBuilder sb = new StringBuilder();
-    
+
     @Override
     public void write(int b) throws IOException {
       sb.append((char) (0xff & b));
     }
-    
+
     public String get() {
       return sb.toString();
     }
-    
+
     public void clear() {
       sb.setLength(0);
     }
   }
-  
+
   private static class ShellExecutionThread extends InputStream implements Runnable {
     private Shell shell;
     StringBuilderOutputStream output;
@@ -244,7 +244,7 @@ public class ShellServlet extends BasicServlet {
     private int cmdIndex;
     private boolean done;
     private boolean readWait;
-    
+
     private ShellExecutionThread(String username, String password, String mock) throws IOException {
       this.done = false;
       this.cmd = null;
@@ -261,7 +261,7 @@ public class ShellServlet extends BasicServlet {
         throw new IOException("shell config error");
       }
     }
-    
+
     @Override
     public synchronized int read() throws IOException {
       if (cmd == null) {
@@ -287,7 +287,7 @@ public class ShellServlet extends BasicServlet {
       }
       return c;
     }
-    
+
     @Override
     public synchronized void run() {
       Thread.currentThread().setName("shell thread");
@@ -308,7 +308,7 @@ public class ShellServlet extends BasicServlet {
       done = true;
       this.notifyAll();
     }
-    
+
     public synchronized void addInputString(String s) {
       if (done)
         throw new IllegalStateException("adding string to exited shell");
@@ -319,7 +319,7 @@ public class ShellServlet extends BasicServlet {
       }
       this.notifyAll();
     }
-    
+
     public synchronized void waitUntilReady() {
       while (cmd != null) {
         try {
@@ -327,29 +327,29 @@ public class ShellServlet extends BasicServlet {
         } catch (InterruptedException e) {}
       }
     }
-    
+
     public synchronized String getOutput() {
       String s = output.get();
       output.clear();
       return s;
     }
-    
+
     public String getPrompt() {
       return shell.getDefaultPrompt();
     }
-    
+
     public void printInfo() throws IOException {
       shell.printInfo();
     }
-    
+
     public boolean isMasking() {
       return shell.isMasking();
     }
-    
+
     public synchronized boolean isWaitingForInput() {
       return readWait;
     }
-    
+
     public boolean isDone() {
       return done;
     }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TServersServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TServersServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TServersServlet.java
index e7ab43d..47a2ca0 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TServersServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TServersServlet.java
@@ -57,30 +57,30 @@ import org.apache.accumulo.server.util.TableInfoUtil;
 import com.google.common.net.HostAndPort;
 
 public class TServersServlet extends BasicServlet {
-  
+
   private static final long serialVersionUID = 1L;
   private static final TabletServerStatus NO_STATUS = new TabletServerStatus();
-  
+
   static class SecondType extends NumberType<Double> {
-    
+
     @Override
     public String format(Object obj) {
       if (obj == null)
         return "&mdash;";
       return Duration.format((long) (1000.0 * (Double) obj));
     }
-    
+
   }
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "Tablet Server Status";
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse response, StringBuilder sb) throws Exception {
     String tserverAddress = req.getParameter("s");
-    
+
     // Check to make sure tserver is a known address
     boolean tserverExists = false;
     if (tserverAddress != null && tserverAddress.isEmpty() == false) {
@@ -91,23 +91,23 @@ public class TServersServlet extends BasicServlet {
         }
       }
     }
-    
+
     if (tserverAddress == null || tserverAddress.isEmpty() || tserverExists == false) {
       doBadTserverList(req, sb);
-      
+
       doDeadTserverList(req, sb);
-      
+
       ArrayList<TabletServerStatus> tservers = new ArrayList<TabletServerStatus>();
       if (Monitor.getMmi() != null)
         tservers.addAll(Monitor.getMmi().tServerInfo);
-      
+
       Table tServerList = new Table("tservers", "Tablet&nbsp;Servers");
       tServerList.setSubCaption("Click on the <span style='color: #0000ff;'>server address</span> to view detailed performance statistics for that server.");
-      
+
       doTserverList(req, sb, tservers, null, tServerList);
       return;
     }
-    
+
     double totalElapsedForAll = 0;
     double splitStdDev = 0;
     double minorStdDev = 0;
@@ -119,7 +119,7 @@ public class TServersServlet extends BasicServlet {
     double currentMinorStdDev = 0;
     double currentMajorStdDev = 0;
     TabletStats total = new TabletStats(null, new ActionStats(), new ActionStats(), new ActionStats(), 0, 0, 0, 0);
-    
+
     HostAndPort address = HostAndPort.fromString(tserverAddress);
     TabletStats historical = new TabletStats(null, new ActionStats(), new ActionStats(), new ActionStats(), 0, 0, 0, 0);
     List<TabletStats> tsStats = new ArrayList<TabletStats>();
@@ -139,7 +139,7 @@ public class TServersServlet extends BasicServlet {
       log.error(e, e);
       return;
     }
-    
+
     Table perTabletResults = new Table("perTabletResults", "Detailed&nbsp;Current&nbsp;Operations");
     perTabletResults.setSubCaption("Per-tablet&nbsp;Details");
     perTabletResults.addSortableColumn("Table", new TableLinkType(), null);
@@ -153,7 +153,7 @@ public class TServersServlet extends BasicServlet {
     perTabletResults.addSortableColumn("Major&nbsp;Avg", new SecondType(), null);
     perTabletResults.addSortableColumn("Major&nbsp;Std&nbsp;Dev", new SecondType(), null);
     perTabletResults.addSortableColumn("Major&nbsp;Avg&nbsp;e/s", new NumberType<Double>(), null);
-    
+
     for (TabletStats info : tsStats) {
       if (info.extent == null) {
         historical = info;
@@ -162,7 +162,7 @@ public class TServersServlet extends BasicServlet {
       total.numEntries += info.numEntries;
       ActionStatsUpdator.update(total.minors, info.minors);
       ActionStatsUpdator.update(total.majors, info.majors);
-      
+
       KeyExtent extent = new KeyExtent(info.extent);
       String tableId = extent.getTableId().toString();
       MessageDigest digester = MessageDigest.getInstance("MD5");
@@ -171,7 +171,7 @@ public class TServersServlet extends BasicServlet {
       }
       String obscuredExtent = Base64.encodeBase64String(digester.digest());
       String displayExtent = String.format("<code>[%s]</code>", obscuredExtent);
-      
+
       TableRow row = perTabletResults.prepareRow();
       row.add(tableId);
       row.add(displayExtent);
@@ -186,7 +186,7 @@ public class TServersServlet extends BasicServlet {
       row.add(info.majors.elapsed != 0 ? info.majors.count / info.majors.elapsed : null);
       perTabletResults.addRow(row);
     }
-    
+
     // Calculate current averages oldServer adding in historical data
     if (total.minors.num != 0)
       currentMinorAvg = (long) (total.minors.elapsed / total.minors.num);
@@ -196,25 +196,25 @@ public class TServersServlet extends BasicServlet {
       currentMajorAvg = total.majors.elapsed / total.majors.num;
     if (total.majors.elapsed != 0 && total.majors.num != 0 && total.majors.elapsed > total.majors.num)
       currentMajorStdDev = stddev(total.majors.elapsed, total.majors.num, total.majors.sumDev);
-    
+
     // After these += operations, these variables are now total for current
     // tablets and historical tablets
     ActionStatsUpdator.update(total.minors, historical.minors);
     ActionStatsUpdator.update(total.majors, historical.majors);
     totalElapsedForAll += total.majors.elapsed + historical.splits.elapsed + total.minors.elapsed;
-    
+
     minorStdDev = stddev(total.minors.elapsed, total.minors.num, total.minors.sumDev);
     minorQueueStdDev = stddev(total.minors.queueTime, total.minors.num, total.minors.queueSumDev);
     majorStdDev = stddev(total.majors.elapsed, total.majors.num, total.majors.sumDev);
     majorQueueStdDev = stddev(total.majors.queueTime, total.majors.num, total.majors.queueSumDev);
     splitStdDev = stddev(historical.splits.num, historical.splits.elapsed, historical.splits.sumDev);
-    
+
     doDetailTable(req, sb, address, tsStats.size(), total, historical);
     doAllTimeTable(req, sb, total, historical, majorQueueStdDev, minorQueueStdDev, totalElapsedForAll, splitStdDev, majorStdDev, minorStdDev);
     doCurrentTabletOps(req, sb, currentMinorAvg, currentMinorStdDev, currentMajorAvg, currentMajorStdDev);
     perTabletResults.generate(req, sb);
   }
-  
+
   private void doCurrentTabletOps(HttpServletRequest req, StringBuilder sb, double currentMinorAvg, double currentMinorStdDev, double currentMajorAvg,
       double currentMajorStdDev) {
     Table currentTabletOps = new Table("currentTabletOps", "Current&nbsp;Tablet&nbsp;Operation&nbsp;Results");
@@ -225,10 +225,10 @@ public class TServersServlet extends BasicServlet {
     currentTabletOps.addRow(currentMinorAvg, currentMinorStdDev, currentMajorAvg, currentMajorStdDev);
     currentTabletOps.generate(req, sb);
   }
-  
+
   private void doAllTimeTable(HttpServletRequest req, StringBuilder sb, TabletStats total, TabletStats historical, double majorQueueStdDev,
       double minorQueueStdDev, double totalElapsedForAll, double splitStdDev, double majorStdDev, double minorStdDev) {
-    
+
     Table opHistoryDetails = new Table("opHistoryDetails", "All-Time&nbsp;Tablet&nbsp;Operation&nbsp;Results");
     opHistoryDetails.addSortableColumn("Operation");
     opHistoryDetails.addSortableColumn("Success", new NumberType<Integer>(), null);
@@ -238,7 +238,7 @@ public class TServersServlet extends BasicServlet {
     opHistoryDetails.addSortableColumn("Average<br />Time", new SecondType(), null);
     opHistoryDetails.addSortableColumn("Std.&nbsp;Dev.<br />Time", new SecondType(), null);
     opHistoryDetails.addSortableColumn("Percentage&nbsp;Time&nbsp;Spent", new ProgressChartType(totalElapsedForAll), null);
-    
+
     opHistoryDetails.addRow("Split", historical.splits.num, historical.splits.fail, null, null,
         historical.splits.num != 0 ? (historical.splits.elapsed / historical.splits.num) : null, splitStdDev, historical.splits.elapsed);
     opHistoryDetails.addRow("Major&nbsp;Compaction", total.majors.num, total.majors.fail, total.majors.num != 0 ? (total.majors.queueTime / total.majors.num)
@@ -247,7 +247,7 @@ public class TServersServlet extends BasicServlet {
         : null, minorQueueStdDev, total.minors.num != 0 ? (total.minors.elapsed / total.minors.num) : null, minorStdDev, total.minors.elapsed);
     opHistoryDetails.generate(req, sb);
   }
-  
+
   private void doDetailTable(HttpServletRequest req, StringBuilder sb, HostAndPort address, int numTablets, TabletStats total, TabletStats historical) {
     Table detailTable = new Table("tServerDetail", "Details");
     detailTable.setSubCaption(address.getHostText() + ":" + address.getPort());
@@ -259,7 +259,7 @@ public class TServersServlet extends BasicServlet {
     detailTable.addRow(numTablets, total.numEntries, total.minors.status, total.majors.status, historical.splits.status);
     detailTable.generate(req, sb);
   }
-  
+
   /*
    * omg there's so much undocumented stuff going on here. First, sumDev is a partial standard deviation computation. It is the (clue 1) sum of the squares of
    * (clue 2) seconds of elapsed time.
@@ -271,7 +271,7 @@ public class TServersServlet extends BasicServlet {
     }
     return 0;
   }
-  
+
   private void doBadTserverList(HttpServletRequest req, StringBuilder sb) {
     if (Monitor.getMmi() != null && !Monitor.getMmi().badTServers.isEmpty()) {
       Table badTServerList = new Table("badtservers", "Non-Functioning&nbsp;Tablet&nbsp;Servers", "error");
@@ -283,7 +283,7 @@ public class TServersServlet extends BasicServlet {
       badTServerList.generate(req, sb);
     }
   }
-  
+
   private void doDeadTserverList(HttpServletRequest req, StringBuilder sb) {
     MasterMonitorInfo mmi = Monitor.getMmi();
     if (mmi != null) {
@@ -293,7 +293,7 @@ public class TServersServlet extends BasicServlet {
       doDeadServerTable(req, sb, deadTServerList, obit);
     }
   }
-  
+
   public static void doDeadServerTable(HttpServletRequest req, StringBuilder sb, Table deadTServerList, List<DeadServer> obit) {
     if (obit != null && !obit.isEmpty()) {
       deadTServerList.addSortableColumn("Server");
@@ -306,11 +306,11 @@ public class TServersServlet extends BasicServlet {
       deadTServerList.generate(req, sb);
     }
   }
-  
+
   static void doTserverList(HttpServletRequest req, StringBuilder sb, List<TabletServerStatus> tservers, String tableId, Table tServerList) {
     int guessHighLoad = ManagementFactory.getOperatingSystemMXBean().getAvailableProcessors();
     long now = System.currentTimeMillis();
-    
+
     double avgLastContact = 0.;
     for (TabletServerStatus status : tservers) {
       avgLastContact += (now - status.lastContact);
@@ -336,7 +336,7 @@ public class TServersServlet extends BasicServlet {
     tServerList.addSortableColumn("Data Cache<br />Hit Rate", new PercentageType(), "The recent data cache hit rate.");
     tServerList.addSortableColumn("OS&nbsp;Load", new NumberType<Double>(0., guessHighLoad * 1., 0., guessHighLoad * 3.),
         "The Unix one minute load average. The average number of processes in the run queue over a one minute interval.");
-    
+
     log.debug("tableId: " + tableId);
     for (TabletServerStatus status : tservers) {
       if (status == null)
@@ -366,5 +366,5 @@ public class TServersServlet extends BasicServlet {
     }
     tServerList.generate(req, sb);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TablesServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TablesServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TablesServlet.java
index e5914f9..3700874 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TablesServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/TablesServlet.java
@@ -50,35 +50,35 @@ import org.apache.accumulo.server.util.TableInfoUtil;
 import org.apache.hadoop.io.Text;
 
 public class TablesServlet extends BasicServlet {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "Table Status";
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse response, StringBuilder sb) throws Exception {
     Map<String,String> tidToNameMap = Tables.getIdToNameMap(Monitor.getContext().getInstance());
     String tableId = req.getParameter("t");
-    
+
     doProblemsBanner(sb);
-    
+
     if (tableId == null || tableId.isEmpty() || tidToNameMap.containsKey(tableId) == false) {
       doTableList(req, sb, tidToNameMap);
       return;
     }
-    
+
     doTableDetails(req, sb, tidToNameMap, tableId);
   }
-  
+
   static void doProblemsBanner(StringBuilder sb) {
     int numProblems = Monitor.getProblemSummary().entrySet().size();
     if (numProblems > 0)
       banner(sb, "error", String.format("<a href='/problems'>Table Problems: %d Total</a>", numProblems));
   }
-  
+
   static void doTableList(HttpServletRequest req, StringBuilder sb, Map<String,String> tidToNameMap) {
     Table tableList = new Table("tableList", "Table&nbsp;List");
     tableList.addSortableColumn("Table&nbsp;Name", new TableLinkType(), null);
@@ -107,14 +107,14 @@ public class TablesServlet extends BasicServlet {
             + "Major Compactions are performed as a consequence of new files created from Minor Compactions and Bulk Load operations.  "
             + "They reduce the number of files used during queries.");
     SortedMap<String,TableInfo> tableStats = new TreeMap<String,TableInfo>();
-    
+
     if (Monitor.getMmi() != null && Monitor.getMmi().tableMap != null)
       for (Entry<String,TableInfo> te : Monitor.getMmi().tableMap.entrySet())
         tableStats.put(Tables.getPrintableTableNameFromId(tidToNameMap, te.getKey()), te.getValue());
-    
+
     Map<String,Double> compactingByTable = TableInfoUtil.summarizeTableStats(Monitor.getMmi());
     TableManager tableManager = TableManager.getInstance();
-    
+
     for (Entry<String,String> tableName_tableId : Tables.getNameToIdMap(Monitor.getContext().getInstance()).entrySet()) {
       String tableName = tableName_tableId.getKey();
       String tableId = tableName_tableId.getValue();
@@ -138,10 +138,10 @@ public class TablesServlet extends BasicServlet {
       row.add(tableInfo);
       tableList.addRow(row);
     }
-    
+
     tableList.generate(req, sb);
   }
-  
+
   private void doTableDetails(HttpServletRequest req, StringBuilder sb, Map<String,String> tidToNameMap, String tableId) {
     String displayName = Tables.getPrintableTableNameFromId(tidToNameMap, tableId);
     Instance instance = Monitor.getContext().getInstance();
@@ -152,7 +152,7 @@ public class TablesServlet extends BasicServlet {
       String systemTableName = MetadataTable.ID.equals(tableId) ? RootTable.NAME : MetadataTable.NAME;
       MetaDataTableScanner scanner = new MetaDataTableScanner(Monitor.getContext(), new Range(KeyExtent.getMetadataEntry(new Text(tableId), new Text()),
           KeyExtent.getMetadataEntry(new Text(tableId), null)), systemTableName);
-      
+
       while (scanner.hasNext()) {
         TabletLocationState state = scanner.next();
         if (state.current != null) {
@@ -165,9 +165,9 @@ public class TablesServlet extends BasicServlet {
       }
       scanner.close();
     }
-    
+
     log.debug("Locs: " + locs);
-    
+
     List<TabletServerStatus> tservers = new ArrayList<TabletServerStatus>();
     if (Monitor.getMmi() != null) {
       for (TabletServerStatus tss : Monitor.getMmi().tServerInfo) {
@@ -180,7 +180,7 @@ public class TablesServlet extends BasicServlet {
         }
       }
     }
-    
+
     Table tableDetails = new Table("participatingTServers", "Participating&nbsp;Tablet&nbsp;Servers");
     tableDetails.setSubCaption(displayName);
     TServersServlet.doTserverList(req, sb, tservers, tableId, tableDetails);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/VisServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/VisServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/VisServlet.java
index af1688e..eedf598 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/VisServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/VisServlet.java
@@ -29,9 +29,9 @@ import org.apache.accumulo.monitor.Monitor;
 
 public class VisServlet extends BasicServlet {
   private static final int concurrentScans = Monitor.getContext().getConfiguration().getCount(Property.TSERV_READ_AHEAD_MAXCONCURRENT);
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   public enum StatType {
     osload(ManagementFactory.getOperatingSystemMXBean().getAvailableProcessors(), true, 100, "OS Load"),
     ingest(1000, true, 1, "Ingest Entries"),
@@ -43,13 +43,13 @@ public class VisServlet extends BasicServlet {
     holdtime(60000, false, 1, "Hold Time"),
     allavg(1, false, 100, "Overall Avg", true),
     allmax(1, false, 100, "Overall Max", true);
-    
+
     private int max;
     private boolean adjustMax;
     private float significance;
     private String description;
     private boolean derived;
-    
+
     /**
      * @param max
      *          initial estimate of largest possible value for this stat
@@ -63,7 +63,7 @@ public class VisServlet extends BasicServlet {
     private StatType(int max, boolean adjustMax, float significance, String description) {
       this(max, adjustMax, significance, description, false);
     }
-    
+
     private StatType(int max, boolean adjustMax, float significance, String description, boolean derived) {
       this.max = max;
       this.adjustMax = adjustMax;
@@ -71,27 +71,27 @@ public class VisServlet extends BasicServlet {
       this.description = description;
       this.derived = derived;
     }
-    
+
     public int getMax() {
       return max;
     }
-    
+
     public boolean getAdjustMax() {
       return adjustMax;
     }
-    
+
     public float getSignificance() {
       return significance;
     }
-    
+
     public String getDescription() {
       return description;
     }
-    
+
     public boolean isDerived() {
       return derived;
     }
-    
+
     public static int numDerived() {
       int count = 0;
       for (StatType st : StatType.values())
@@ -100,7 +100,7 @@ public class VisServlet extends BasicServlet {
       return count;
     }
   }
-  
+
   public static class VisualizationConfig {
     boolean useCircles = true;
     StatType motion = StatType.allmax;
@@ -108,38 +108,38 @@ public class VisServlet extends BasicServlet {
     int spacing = 40;
     String url;
   }
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "Server Activity";
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse response, StringBuilder sb) throws IOException {
     StringBuffer urlsb = req.getRequestURL();
     urlsb.setLength(urlsb.lastIndexOf("/") + 1);
     VisualizationConfig cfg = new VisualizationConfig();
     cfg.url = urlsb.toString();
-    
+
     String s = req.getParameter("shape");
     if (s != null && (s.equals("square") || s.equals("squares"))) {
       cfg.useCircles = false;
     }
-    
+
     s = req.getParameter("motion");
     if (s != null) {
       try {
         cfg.motion = StatType.valueOf(s);
       } catch (Exception e) {}
     }
-    
+
     s = req.getParameter("color");
     if (s != null) {
       try {
         cfg.color = StatType.valueOf(s);
       } catch (Exception e) {}
     }
-    
+
     String size = req.getParameter("size");
     if (size != null) {
       if (size.equals("10"))
@@ -149,20 +149,20 @@ public class VisServlet extends BasicServlet {
       else if (size.equals("80"))
         cfg.spacing = 80;
     }
-    
+
     ArrayList<TabletServerStatus> tservers = new ArrayList<TabletServerStatus>();
     if (Monitor.getMmi() != null)
       tservers.addAll(Monitor.getMmi().tServerInfo);
-    
+
     if (tservers.size() == 0)
       return;
-    
+
     int width = (int) Math.ceil(Math.sqrt(tservers.size())) * cfg.spacing;
-    int height = (int) Math.ceil(tservers.size() / (double)width) * cfg.spacing;
+    int height = (int) Math.ceil(tservers.size() / (double) width) * cfg.spacing;
     doSettings(sb, cfg, width < 640 ? 640 : width, height < 640 ? 640 : height);
     doScript(sb, cfg, tservers);
   }
-  
+
   private void doSettings(StringBuilder sb, VisualizationConfig cfg, int width, int height) {
     sb.append("<div class='left'>\n");
     sb.append("<div id='parameters' class='nowrap'>\n");
@@ -193,13 +193,13 @@ public class VisServlet extends BasicServlet {
     sb.append("</div>\n");
     sb.append("</div>\n\n");
   }
-  
+
   private void addOptions(StringBuilder sb, StatType selectedStatType) {
     for (StatType st : StatType.values()) {
       sb.append("<option").append(st.equals(selectedStatType) ? " selected='true'>" : ">").append(st.getDescription()).append("</option>");
     }
   }
-  
+
   private void doScript(StringBuilder sb, VisualizationConfig cfg, ArrayList<TabletServerStatus> tservers) {
     // initialization of some javascript variables
     sb.append("<script type='text/javascript'>\n");
@@ -230,7 +230,7 @@ public class VisServlet extends BasicServlet {
     sb.append("}; // values will be converted by floor(this*value)/this\n");
     sb.append("var numNormalStats = ").append(StatType.values().length - StatType.numDerived()).append(";\n");
     sb.append("</script>\n");
-    
+
     sb.append("<script src='web/vis.js' type='text/javascript'></script>");
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/XMLServlet.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/XMLServlet.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/XMLServlet.java
index 3b115f9..1662069 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/XMLServlet.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/XMLServlet.java
@@ -38,40 +38,40 @@ import org.apache.accumulo.server.util.TableInfoUtil;
 
 public class XMLServlet extends BasicServlet {
   private static final long serialVersionUID = 1L;
-  
+
   @Override
   protected String getTitle(HttpServletRequest req) {
     return "XML Report";
   }
-  
+
   @Override
   protected void pageStart(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     resp.setContentType("text/xml;charset=" + UTF_8.name());
     sb.append("<?xml version=\"1.0\" encoding=\"" + UTF_8.name() + "\"?>\n");
     sb.append("<stats>\n");
   }
-  
+
   @Override
   protected void pageBody(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     double totalIngest = 0.;
     double totalQuery = 0.;
     double disk = 0.0;
     long totalEntries = 0L;
-    
+
     sb.append("\n<servers>\n");
     if (Monitor.getMmi() == null || Monitor.getMmi().tableMap == null) {
       sb.append("</servers>\n");
       return;
     }
     SortedMap<String,TableInfo> tableStats = new TreeMap<String,TableInfo>(Monitor.getMmi().tableMap);
-    
+
     for (TabletServerStatus status : Monitor.getMmi().tServerInfo) {
-      
+
       sb.append("\n<server id='").append(status.name).append("'>\n");
       sb.append("<hostname>").append(TServerLinkType.displayName(status.name)).append("</hostname>");
       sb.append("<lastContact>").append(System.currentTimeMillis() - status.lastContact).append("</lastContact>\n");
       sb.append("<osload>").append(status.osLoad).append("</osload>\n");
-      
+
       TableInfo summary = TableInfoUtil.summarizeTableStats(status);
       sb.append("<compactions>\n");
       sb.append("<major>").append("<running>").append(summary.majors.running).append("</running>").append("<queued>").append(summary.majors.queued)
@@ -79,9 +79,9 @@ public class XMLServlet extends BasicServlet {
       sb.append("<minor>").append("<running>").append(summary.minors.running).append("</running>").append("<queued>").append(summary.minors.queued)
           .append("</queued>").append("</minor>\n");
       sb.append("</compactions>\n");
-      
+
       sb.append("<tablets>").append(summary.tablets).append("</tablets>\n");
-      
+
       sb.append("<ingest>").append(summary.ingestRate).append("</ingest>\n");
       sb.append("<query>").append(summary.queryRate).append("</query>\n");
       sb.append("<ingestMB>").append(summary.ingestByteRate / 1000000.0).append("</ingestMB>\n");
@@ -95,41 +95,41 @@ public class XMLServlet extends BasicServlet {
       sb.append("</server>\n");
     }
     sb.append("\n</servers>\n");
-    
+
     sb.append("\n<masterGoalState>" + Monitor.getMmi().goalState + "</masterGoalState>\n");
     sb.append("\n<masterState>" + Monitor.getMmi().state + "</masterState>\n");
-    
+
     sb.append("\n<badTabletServers>\n");
     for (Entry<String,Byte> entry : Monitor.getMmi().badTServers.entrySet()) {
       sb.append(String.format("<badTabletServer id='%s' status='%s'/>\n", entry.getKey(), TabletServerState.getStateById(entry.getValue())));
     }
     sb.append("\n</badTabletServers>\n");
-    
+
     sb.append("\n<tabletServersShuttingDown>\n");
     for (String server : Monitor.getMmi().serversShuttingDown) {
       sb.append(String.format("<server id='%s'/>\n", server));
     }
     sb.append("\n</tabletServersShuttingDown>\n");
-    
+
     sb.append(String.format("\n<unassignedTablets>%d</unassignedTablets>\n", Monitor.getMmi().unassignedTablets));
-    
+
     sb.append("\n<deadTabletServers>\n");
     for (DeadServer dead : Monitor.getMmi().deadTabletServers) {
       sb.append(String.format("<deadTabletServer id='%s' lastChange='%d' status='%s'/>\n", dead.server, dead.lastStatus, dead.status));
     }
     sb.append("\n</deadTabletServers>\n");
-    
+
     sb.append("\n<deadLoggers>\n");
     for (DeadServer dead : Monitor.getMmi().deadTabletServers) {
       sb.append(String.format("<deadLogger id='%s' lastChange='%d' status='%s'/>\n", dead.server, dead.lastStatus, dead.status));
     }
     sb.append("\n</deadLoggers>\n");
-    
+
     sb.append("\n<tables>\n");
     Instance instance = Monitor.getContext().getInstance();
     for (Entry<String,TableInfo> entry : tableStats.entrySet()) {
       TableInfo tableInfo = entry.getValue();
-      
+
       sb.append("\n<table>\n");
       String tableId = entry.getKey();
       String tableName = "unknown";
@@ -163,7 +163,7 @@ public class XMLServlet extends BasicServlet {
       sb.append("</table>\n");
     }
     sb.append("\n</tables>\n");
-    
+
     sb.append("\n<totals>\n");
     sb.append("<ingestrate>").append(totalIngest).append("</ingestrate>\n");
     sb.append("<queryrate>").append(totalQuery).append("</queryrate>\n");
@@ -171,7 +171,7 @@ public class XMLServlet extends BasicServlet {
     sb.append("<numentries>").append(totalEntries).append("</numentries>\n");
     sb.append("</totals>\n");
   }
-  
+
   @Override
   protected void pageEnd(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) {
     sb.append("\n</stats>\n");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ListType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ListType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ListType.java
index 5ab3b2a..84322b5 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ListType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ListType.java
@@ -33,17 +33,17 @@ import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.hadoop.io.Text;
 
 public class ListType extends Basic {
-  
+
   private static final long serialVersionUID = 1L;
-  
+
   String getType(HttpServletRequest req) {
     return getStringParameter(req, "type", "<Unknown>");
   }
-  
+
   int getMinutes(HttpServletRequest req) {
     return getIntParameter(req, "minutes", Summary.DEFAULT_MINUTES);
   }
-  
+
   @Override
   public void pageBody(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) throws Exception {
     String type = getType(req);
@@ -68,7 +68,7 @@ public class ListType extends Basic {
     }
     trace.generate(req, sb);
   }
-  
+
   @Override
   public String getTitle(HttpServletRequest req) {
     return "Traces for " + getType(req) + " for the last " + getMinutes(req) + " minutes";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullKeyValueIterator.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullKeyValueIterator.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullKeyValueIterator.java
index 26cfb07..1ec7bfc 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullKeyValueIterator.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullKeyValueIterator.java
@@ -23,17 +23,17 @@ import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
 
 public class NullKeyValueIterator implements Iterator<Entry<Key,Value>> {
-  
+
   @Override
   public boolean hasNext() {
     return false;
   }
-  
+
   @Override
   public Entry<Key,Value> next() {
     return null;
   }
-  
+
   @Override
   public void remove() {}
 }


[14/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

Posted by ct...@apache.org.
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullScanner.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullScanner.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullScanner.java
index bf35557..750cdcd 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullScanner.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/NullScanner.java
@@ -28,79 +28,79 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.hadoop.io.Text;
 
 public class NullScanner implements Scanner {
-  
+
   @Override
   public void addScanIterator(IteratorSetting cfg) {}
-  
+
   @Override
   public void updateScanIteratorOption(String iteratorName, String key, String value) {}
-  
+
   @Override
   public void fetchColumnFamily(Text col) {}
-  
+
   @Override
   public void fetchColumn(Text colFam, Text colQual) {}
-  
+
   @Override
   public void clearColumns() {}
-  
+
   @Override
   public void clearScanIterators() {}
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {}
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
     return 0;
   }
-  
+
   @Override
   public void setRange(Range range) {}
-  
+
   @Override
   public Range getRange() {
     return null;
   }
-  
+
   @Override
   public void setBatchSize(int size) {
-    
+
   }
-  
+
   @Override
   public int getBatchSize() {
     return 0;
   }
-  
+
   @Override
   public void enableIsolation() {
-    
+
   }
-  
+
   @Override
   public void disableIsolation() {
-    
+
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     return new NullKeyValueIterator();
   }
-  
+
   @Override
   public void removeScanIterator(String iteratorName) {}
-  
+
   @Override
   public void setTimeout(long timeOut, TimeUnit timeUnit) {}
-  
+
   @Override
   public long getTimeout(TimeUnit timeUnit) {
     return 0;
   }
-  
+
   @Override
   public void close() {}
 
@@ -111,6 +111,6 @@ public class NullScanner implements Scanner {
 
   @Override
   public void setReadaheadThreshold(long batches) {
-    
+
   }
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTrace.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTrace.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTrace.java
index 067f184..214c12d 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTrace.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTrace.java
@@ -41,15 +41,15 @@ import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.hadoop.io.Text;
 
 public class ShowTrace extends Basic {
-  
+
   private static final long serialVersionUID = 1L;
   private static final String checkboxIdSuffix = "_checkbox";
   private static final String pageLoadFunctionName = "pageload";
-  
+
   String getTraceId(HttpServletRequest req) {
     return getStringParameter(req, "id", null);
   }
-  
+
   @Override
   public String getTitle(HttpServletRequest req) {
     String id = getTraceId(req);
@@ -57,7 +57,7 @@ public class ShowTrace extends Basic {
       return "No trace id specified";
     return "Trace ID " + id;
   }
-  
+
   @Override
   public void pageBody(HttpServletRequest req, HttpServletResponse resp, final StringBuilder sb) throws Exception {
     String id = getTraceId(req);
@@ -91,7 +91,7 @@ public class ShowTrace extends Basic {
     sb.append("    elt.style.display='none';\n ");
     sb.append(" }\n");
     sb.append("}\n");
-    
+
     sb.append("function ").append(pageLoadFunctionName).append("() {\n");
     sb.append("  var checkboxes = document.getElementsByTagName('input');\n");
     sb.append("  for (var i = 0; i < checkboxes.length; i++) {\n");
@@ -102,13 +102,13 @@ public class ShowTrace extends Basic {
     sb.append("    }\n");
     sb.append("  }\n");
     sb.append("}\n");
-    
+
     sb.append("</script>\n");
     sb.append("<div>");
     sb.append("<table><caption>");
     sb.append(String.format("<span class='table-caption'>Trace %s started at<br>%s</span></caption>", id, dateString(start)));
     sb.append("<tr><th>Time</th><th>Start</th><th>Service@Location</th><th>Name</th><th>Addl Data</th></tr>");
-    
+
     final long finalStart = start;
     Set<Long> visited = tree.visit(new SpanTreeVisitor() {
       @Override
@@ -135,7 +135,7 @@ public class ShowTrace extends Basic {
           sb.append("  <table class='indent,noborder'>\n");
           if (hasData) {
             sb.append("  <tr><th>Key</th><th>Value</th></tr>\n");
-            for (Entry<ByteBuffer, ByteBuffer> entry : node.data.entrySet()) {
+            for (Entry<ByteBuffer,ByteBuffer> entry : node.data.entrySet()) {
               String key = new String(entry.getKey().array(), entry.getKey().arrayOffset(), entry.getKey().limit(), UTF_8);
               String value = new String(entry.getValue().array(), entry.getValue().arrayOffset(), entry.getValue().limit(), UTF_8);
               sb.append("  <tr><td>" + BasicServlet.sanitize(key) + "</td>");
@@ -167,7 +167,7 @@ public class ShowTrace extends Basic {
     sb.append("</table>\n");
     sb.append("</div>\n");
   }
-  
+
   @Override
   protected String getBodyAttributes() {
     return " onload=\"" + pageLoadFunctionName + "()\" ";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTraceLinkType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTraceLinkType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTraceLinkType.java
index d9de107..f7a0e1d 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTraceLinkType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/ShowTraceLinkType.java
@@ -23,7 +23,7 @@ import org.apache.accumulo.tracer.TraceFormatter;
 import org.apache.accumulo.tracer.thrift.RemoteSpan;
 
 /**
- * 
+ *
  */
 public class ShowTraceLinkType extends StringType<RemoteSpan> {
   @Override
@@ -33,7 +33,7 @@ public class ShowTraceLinkType extends StringType<RemoteSpan> {
     RemoteSpan span = (RemoteSpan) obj;
     return String.format("<a href='/trace/show?id=%s'>%s</a>", Long.toHexString(span.traceId), TraceFormatter.formatDate(new Date(span.start)));
   }
-  
+
   @Override
   public int compare(RemoteSpan o1, RemoteSpan o2) {
     if (o1 == null && o2 == null)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/Summary.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/Summary.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/Summary.java
index 8cc451f..75541be 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/Summary.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/servlets/trace/Summary.java
@@ -37,26 +37,26 @@ import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.hadoop.io.Text;
 
 public class Summary extends Basic {
-  
+
   private static final long serialVersionUID = 1L;
   public static final int DEFAULT_MINUTES = 10;
-  
+
   int getMinutes(HttpServletRequest req) {
     return getIntParameter(req, "minutes", DEFAULT_MINUTES);
   }
-  
+
   @Override
   public String getTitle(HttpServletRequest req) {
     return "Traces for the last " + getMinutes(req) + " minutes";
   }
-  
+
   static private class Stats {
     int count;
     long min = Long.MAX_VALUE;
     long max = Long.MIN_VALUE;
     long total = 0l;
     long histogram[] = new long[] {0, 0, 0, 0, 0, 0};
-    
+
     void addSpan(RemoteSpan span) {
       count++;
       long ms = span.stop - span.start;
@@ -70,20 +70,20 @@ public class Summary extends Basic {
       }
       histogram[index]++;
     }
-    
+
     long average() {
       return total / count;
     }
   }
-  
+
   private static class ShowTypeLink extends StringType<String> {
-    
+
     int minutes;
-    
+
     public ShowTypeLink(int minutes) {
       this.minutes = minutes;
     }
-    
+
     @Override
     public String format(Object obj) {
       if (obj == null)
@@ -93,7 +93,7 @@ public class Summary extends Basic {
       return String.format("<a href='/trace/listType?type=%s&minutes=%d'>%s</a>", encodedType, minutes, type);
     }
   }
-  
+
   static private class HistogramType extends StringType<Stats> {
     @Override
     public String format(Object obj) {
@@ -110,7 +110,7 @@ public class Summary extends Basic {
       sb.append("</tr></table>");
       return sb.toString();
     }
-    
+
     @Override
     public int compare(Stats o1, Stats o2) {
       for (int i = 0; i < o1.histogram.length; i++) {
@@ -140,11 +140,11 @@ public class Summary extends Basic {
 
     return new Range(new Text("start:" + startHexTime), new Text("start:" + endHexTime));
   }
-  
+
   @Override
   public void pageBody(HttpServletRequest req, HttpServletResponse resp, StringBuilder sb) throws Exception {
     int minutes = getMinutes(req);
-    
+
     Scanner scanner = getScanner(sb);
     if (scanner == null) {
       return;
@@ -171,7 +171,7 @@ public class Summary extends Basic {
             "Histogram",
             new HistogramType(),
             "Counts of spans of different duration. Columns start at milliseconds, and each column is ten times longer: tens of milliseconds, seconds, tens of seconds, etc.");
-    
+
     for (Entry<String,Stats> entry : summary.entrySet()) {
       Stats stat = entry.getValue();
       trace.addRow(entry.getKey(), stat.count, stat.min, stat.max, stat.average(), stat);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableColumn.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableColumn.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableColumn.java
index a7330ac..1e00927 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableColumn.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableColumn.java
@@ -23,25 +23,25 @@ public class TableColumn<T> {
   private String title;
   private CellType<T> type;
   private String legend;
-  
+
   public TableColumn(String title, CellType<T> type, String legend) {
     this.title = title;
     this.type = type != null ? type : new StringType<T>();
     this.legend = legend;
   }
-  
+
   public void setTitle(String title) {
     this.title = title;
   }
-  
+
   public String getTitle() {
     return title;
   }
-  
+
   public String getLegend() {
     return legend;
   }
-  
+
   public CellType<T> getCellType() {
     return type;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableRow.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableRow.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableRow.java
index 5de0863..c1ce1c5 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableRow.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/TableRow.java
@@ -22,43 +22,43 @@ import java.util.Comparator;
 public class TableRow {
   private int size;
   private ArrayList<Object> row;
-  
+
   TableRow(int size) {
     this.size = size;
     this.row = new ArrayList<Object>(size);
   }
-  
+
   public boolean add(Object obj) {
     if (row.size() == size)
       throw new IllegalStateException("Row is full.");
     return row.add(obj);
   }
-  
+
   Object get(int index) {
     return row.get(index);
   }
-  
+
   int size() {
     return row.size();
   }
-  
+
   Object set(int i, Object value) {
     return row.set(i, value);
   }
-  
+
   public static <T> Comparator<TableRow> getComparator(int index, Comparator<T> comp) {
     return new TableRowComparator<T>(index, comp);
   }
-  
+
   private static class TableRowComparator<T> implements Comparator<TableRow> {
     private int index;
     private Comparator<T> comp;
-    
+
     public TableRowComparator(int index, Comparator<T> comp) {
       this.index = index;
       this.comp = comp;
     }
-    
+
     @SuppressWarnings("unchecked")
     @Override
     public int compare(TableRow o1, TableRow o2) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CellType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CellType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CellType.java
index 23071a8..71aa317 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CellType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CellType.java
@@ -20,15 +20,15 @@ import java.util.Comparator;
 
 public abstract class CellType<T> implements Comparator<T> {
   private boolean sortable = true;
-  
+
   abstract public String alignment();
-  
+
   abstract public String format(Object obj);
-  
+
   public final void setSortable(boolean sortable) {
     this.sortable = sortable;
   }
-  
+
   public final boolean isSortable() {
     return sortable;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CompactionsType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CompactionsType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CompactionsType.java
index 704aa42..e649b84 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CompactionsType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/CompactionsType.java
@@ -20,13 +20,13 @@ import org.apache.accumulo.core.master.thrift.Compacting;
 import org.apache.accumulo.core.master.thrift.TableInfo;
 
 public class CompactionsType extends CellType<TableInfo> {
-  
+
   private String fieldName;
-  
+
   public CompactionsType(String which) {
     this.fieldName = which;
   }
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
@@ -41,7 +41,7 @@ public class CompactionsType extends CellType<TableInfo> {
       c = new Compacting();
     return String.format("%s&nbsp;(%,d)", NumberType.commas(c.running, c.queued == 0 ? 0 : 1, summary.onlineTablets), c.queued);
   }
-  
+
   @Override
   public int compare(TableInfo o1, TableInfo o2) {
     if (o1 == null)
@@ -63,10 +63,10 @@ public class CompactionsType extends CellType<TableInfo> {
       return 1;
     return c1.running + c1.queued - c2.running - c2.queued;
   }
-  
+
   @Override
   public String alignment() {
     return "right";
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DateTimeType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DateTimeType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DateTimeType.java
index 8ff3b60..5a8ca58 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DateTimeType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DateTimeType.java
@@ -25,17 +25,17 @@ public class DateTimeType extends CellType<Long> {
   private SimpleDateFormat simple;
   private int dateFormat;
   private int timeFormat;
-  
+
   public DateTimeType(int dateFormat, int timeFormat) {
     this.dateFormat = dateFormat;
     this.timeFormat = timeFormat;
     this.simple = null;
   }
-  
+
   public DateTimeType(SimpleDateFormat fmt) {
     simple = fmt;
   }
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
@@ -47,7 +47,7 @@ public class DateTimeType extends CellType<Long> {
       return simple.format(new Date(millis)).replace(" ", "&nbsp;");
     return DateFormat.getDateTimeInstance(dateFormat, timeFormat, Locale.getDefault()).format(new Date(millis)).replace(" ", "&nbsp;");
   }
-  
+
   @Override
   public int compare(Long o1, Long o2) {
     if (o1 == null && o2 == null)
@@ -57,10 +57,10 @@ public class DateTimeType extends CellType<Long> {
     else
       return o1.compareTo(o2);
   }
-  
+
   @Override
   public String alignment() {
     return "right";
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DurationType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DurationType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DurationType.java
index 7f3472e..65dd214 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DurationType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/DurationType.java
@@ -21,16 +21,16 @@ import org.apache.accumulo.core.util.Duration;
 public class DurationType extends NumberType<Long> {
   private Long errMin;
   private Long errMax;
-  
+
   public DurationType() {
     this(null, null);
   }
-  
+
   public DurationType(Long errMin, Long errMax) {
     this.errMin = errMin;
     this.errMax = errMax;
   }
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
@@ -40,12 +40,12 @@ public class DurationType extends NumberType<Long> {
       return seconds(millis, errMin, errMax);
     return Duration.format(millis);
   }
-  
+
   private static String seconds(long secs, long errMin, long errMax) {
     String numbers = Duration.format(secs);
     if (secs < errMin || secs > errMax)
       return "<span class='error'>" + numbers + "</span>";
     return numbers;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/NumberType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/NumberType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/NumberType.java
index b285727..f00ad67 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/NumberType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/NumberType.java
@@ -19,24 +19,24 @@ package org.apache.accumulo.monitor.util.celltypes;
 import static org.apache.accumulo.core.util.NumUtil.bigNumberForQuantity;
 
 public class NumberType<T extends Number> extends CellType<T> {
-  
+
   protected final T warnMin, warnMax, errMin, errMax;
-  
+
   public NumberType(T warnMin, T warnMax, T errMin, T errMax) {
     this.warnMin = warnMin;
     this.warnMax = warnMax;
     this.errMin = errMin;
     this.errMax = errMax;
   }
-  
+
   public NumberType(T errMin, T errMax) {
     this(null, null, errMin, errMax);
   }
-  
+
   public NumberType() {
     this(null, null);
   }
-  
+
   @SuppressWarnings("unchecked")
   @Override
   public String format(Object obj) {
@@ -62,7 +62,7 @@ public class NumberType<T extends Number> extends CellType<T> {
     }
     return s;
   }
-  
+
   @Override
   public int compare(T o1, T o2) {
     if (o1 == null && o2 == null)
@@ -74,27 +74,27 @@ public class NumberType<T extends Number> extends CellType<T> {
     else
       return Double.valueOf(o1.doubleValue()).compareTo(o2.doubleValue());
   }
-  
+
   public static String commas(long i) {
     return bigNumberForQuantity(i);
   }
-  
+
   public static String commas(long i, long errMin, long errMax) {
     if (i < errMin || i > errMax)
       return String.format("<span class='error'>%s</span>", bigNumberForQuantity(i));
     return bigNumberForQuantity(i);
   }
-  
+
   public static String commas(double i) {
     return bigNumberForQuantity((long) i);
   }
-  
+
   public static String commas(double d, double errMin, double errMax) {
     if (d < errMin || d > errMax)
       return String.format("<span class='error'>%s</span>", bigNumberForQuantity(d));
     return bigNumberForQuantity(d);
   }
-  
+
   public static String commas(long i, long warnMin, long warnMax, long errMin, long errMax) {
     if (i < errMin || i > errMax)
       return String.format("<span class='error'>%s</span>", bigNumberForQuantity(i));
@@ -102,7 +102,7 @@ public class NumberType<T extends Number> extends CellType<T> {
       return String.format("<span class='warning'>%s</span>", bigNumberForQuantity(i));
     return bigNumberForQuantity(i);
   }
-  
+
   public static String commas(double d, double warnMin, double warnMax, double errMin, double errMax) {
     if (d < errMin || d > errMax)
       return String.format("<span class='error'>%s</span>", bigNumberForQuantity(d));
@@ -110,7 +110,7 @@ public class NumberType<T extends Number> extends CellType<T> {
       return String.format("<span class='warning'>%s</span>", bigNumberForQuantity(d));
     return bigNumberForQuantity(d);
   }
-  
+
   @Override
   public String alignment() {
     return "right";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PercentageType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PercentageType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PercentageType.java
index 2efc65f..462e33d 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PercentageType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PercentageType.java
@@ -17,24 +17,24 @@
 package org.apache.accumulo.monitor.util.celltypes;
 
 public class PercentageType extends CellType<Double> {
-  
+
   @Override
   public int compare(Double o1, Double o2) {
     return o1.compareTo(o2);
   }
-  
+
   @Override
   public String alignment() {
     return "right";
   }
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
       return "-";
-    
+
     return String.format("%.0f%s", 100 * (Double) obj, "%");
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberType.java
index c038d89..775e8dd 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberType.java
@@ -16,7 +16,6 @@
  */
 package org.apache.accumulo.monitor.util.celltypes;
 
-
 public class PreciseNumberType extends NumberType<Integer> {
 
   public PreciseNumberType(int warnMin, int warnMax, int errMin, int errMax) {
@@ -27,7 +26,7 @@ public class PreciseNumberType extends NumberType<Integer> {
   public String format(Object obj) {
     if (obj == null)
       return "-";
-    int i = ((Number)obj).intValue();
+    int i = ((Number) obj).intValue();
     String display = String.format("%,d", obj);
     if ((errMin != null && i < errMin) || (errMax != null && i > errMax)) {
       return String.format("<span class='error'>%s</span>", display);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/ProgressChartType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/ProgressChartType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/ProgressChartType.java
index 20db871..20836fd 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/ProgressChartType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/ProgressChartType.java
@@ -17,17 +17,17 @@
 package org.apache.accumulo.monitor.util.celltypes;
 
 public class ProgressChartType extends NumberType<Double> {
-  
+
   private double max;
-  
+
   public ProgressChartType() {
     this(1.0);
   }
-  
+
   public ProgressChartType(Double total) {
     max = total == null ? 1.0 : total;
   }
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
@@ -35,13 +35,13 @@ public class ProgressChartType extends NumberType<Double> {
     Double num = (Double) obj;
     return getChart(num, max);
   }
-  
+
   public static String getChart(double num, double total) {
     StringBuilder result = new StringBuilder();
     double percent = 0;
     if (total != 0)
       percent = (num / total) * 100;
-    
+
     int width = 0;
     if (percent < 1)
       width = 0;
@@ -49,7 +49,7 @@ public class ProgressChartType extends NumberType<Double> {
       width = 100;
     else
       width = (int) percent;
-    
+
     result.append("<div class='progress-chart'>");
     result.append("<div style='width: ").append(width).append("%;'></div>");
     result.append("</div>&nbsp;");

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/StringType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/StringType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/StringType.java
index 8807423..79f8d5e 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/StringType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/StringType.java
@@ -21,7 +21,7 @@ public class StringType<T> extends CellType<T> {
   public String format(Object obj) {
     return obj == null ? "-" : obj.toString();
   }
-  
+
   @Override
   public int compare(T o1, T o2) {
     if (o1 == null && o2 == null)
@@ -32,7 +32,7 @@ public class StringType<T> extends CellType<T> {
       return 1;
     return o1.toString().compareTo(o2.toString());
   }
-  
+
   @Override
   public String alignment() {
     return "left";

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TServerLinkType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TServerLinkType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TServerLinkType.java
index 26d623a..be3f527 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TServerLinkType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TServerLinkType.java
@@ -19,7 +19,7 @@ package org.apache.accumulo.monitor.util.celltypes;
 import org.apache.accumulo.core.master.thrift.TabletServerStatus;
 
 public class TServerLinkType extends CellType<TabletServerStatus> {
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
@@ -27,25 +27,25 @@ public class TServerLinkType extends CellType<TabletServerStatus> {
     TabletServerStatus status = (TabletServerStatus) obj;
     return String.format("<a href='/tservers?s=%s'>%s</a>", status.name, displayName(status));
   }
-  
+
   public static String displayName(TabletServerStatus status) {
     return displayName(status == null ? null : status.name);
   }
-  
+
   public static String displayName(String address) {
     if (address == null)
       return "--Unknown--";
     return address;
   }
-  
+
   @Override
   public int compare(TabletServerStatus o1, TabletServerStatus o2) {
     return displayName(o1).compareTo(displayName(o2));
   }
-  
+
   @Override
   public String alignment() {
     return "left";
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableLinkType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableLinkType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableLinkType.java
index 76041d4..0d176f2 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableLinkType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableLinkType.java
@@ -23,13 +23,13 @@ import org.apache.accumulo.monitor.servlets.BasicServlet;
 import org.apache.accumulo.server.client.HdfsZooInstance;
 
 public class TableLinkType extends CellType<String> {
-  
+
   private Map<String,String> tidToNameMap;
-  
+
   public TableLinkType() {
     tidToNameMap = Tables.getIdToNameMap(HdfsZooInstance.getInstance());
   }
-  
+
   @Override
   public String format(Object obj) {
     if (obj == null)
@@ -39,21 +39,21 @@ public class TableLinkType extends CellType<String> {
     // e.g. the root table's id of "+r" would not be interpreted properly
     return String.format("<a href='/tables?t=%s'>%s</a>", BasicServlet.encode(tableId), displayName(tableId));
   }
-  
+
   private String displayName(String tableId) {
     if (tableId == null)
       return "-";
     return Tables.getPrintableTableNameFromId(tidToNameMap, tableId);
   }
-  
+
   @Override
   public int compare(String o1, String o2) {
     return displayName(o1).compareTo(displayName(o2));
   }
-  
+
   @Override
   public String alignment() {
     return "left";
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableStateType.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableStateType.java b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableStateType.java
index 22fb498..724d220 100644
--- a/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableStateType.java
+++ b/server/monitor/src/main/java/org/apache/accumulo/monitor/util/celltypes/TableStateType.java
@@ -19,12 +19,12 @@ package org.apache.accumulo.monitor.util.celltypes;
 import org.apache.accumulo.core.master.state.tables.TableState;
 
 public class TableStateType extends CellType<TableState> {
-  
+
   @Override
   public String alignment() {
     return "center";
   }
-  
+
   @Override
   public String format(Object obj) {
     TableState state = obj == null ? TableState.UNKNOWN : (TableState) obj;
@@ -44,7 +44,7 @@ public class TableStateType extends CellType<TableState> {
     style = style != null ? " class='" + style + "'" : "";
     return String.format("<span%s>%s</span>", style, state);
   }
-  
+
   @Override
   public int compare(TableState o1, TableState o2) {
     if (o1 == null && o2 == null)
@@ -53,5 +53,5 @@ public class TableStateType extends CellType<TableState> {
       return -1;
     return o1.compareTo(o2);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/test/java/org/apache/accumulo/monitor/ShowTraceLinkTypeTest.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/test/java/org/apache/accumulo/monitor/ShowTraceLinkTypeTest.java b/server/monitor/src/test/java/org/apache/accumulo/monitor/ShowTraceLinkTypeTest.java
index ef4107b..b4840a7 100644
--- a/server/monitor/src/test/java/org/apache/accumulo/monitor/ShowTraceLinkTypeTest.java
+++ b/server/monitor/src/test/java/org/apache/accumulo/monitor/ShowTraceLinkTypeTest.java
@@ -16,19 +16,19 @@
  */
 package org.apache.accumulo.monitor;
 
+import java.nio.ByteBuffer;
 import java.util.ArrayList;
 import java.util.Collections;
 
-import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.accumulo.tracer.thrift.Annotation;
+import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.junit.Assert;
 import org.junit.Test;
 
-import java.nio.ByteBuffer;
-
 public class ShowTraceLinkTypeTest {
   private static RemoteSpan rs(long start, long stop, String description) {
-    return new RemoteSpan("sender", "svc", 0l, 0l, 0l, start, stop, description, Collections.<ByteBuffer, ByteBuffer>emptyMap(), Collections.<Annotation>emptyList());
+    return new RemoteSpan("sender", "svc", 0l, 0l, 0l, start, stop, description, Collections.<ByteBuffer,ByteBuffer> emptyMap(),
+        Collections.<Annotation> emptyList());
   }
 
   @Test

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/test/java/org/apache/accumulo/monitor/ZooKeeperStatusTest.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/test/java/org/apache/accumulo/monitor/ZooKeeperStatusTest.java b/server/monitor/src/test/java/org/apache/accumulo/monitor/ZooKeeperStatusTest.java
index a90396b..7f56931 100644
--- a/server/monitor/src/test/java/org/apache/accumulo/monitor/ZooKeeperStatusTest.java
+++ b/server/monitor/src/test/java/org/apache/accumulo/monitor/ZooKeeperStatusTest.java
@@ -26,31 +26,31 @@ import org.junit.Assert;
 import org.junit.Test;
 
 public class ZooKeeperStatusTest {
-  
+
   @Test
   public void zkHostSortingTest() {
     List<String> expectedHosts = Arrays.asList("rack1node1", "rack2node1", "rack4node1", "rack4node4");
-    
+
     // Add the states in a not correctly sorted order
     TreeSet<ZooKeeperState> states = new TreeSet<ZooKeeperState>();
     states.add(new ZooKeeperState("rack4node4", "leader", 10));
     states.add(new ZooKeeperState("rack4node1", "follower", 10));
     states.add(new ZooKeeperState("rack1node1", "follower", 10));
     states.add(new ZooKeeperState("rack2node1", "follower", 10));
-    
+
     List<String> actualHosts = new ArrayList<String>(4);
     for (ZooKeeperState state : states) {
       actualHosts.add(state.keeper);
     }
-    
+
     // Assert we have 4 of each
     Assert.assertEquals(expectedHosts.size(), actualHosts.size());
-    
+
     // Assert the ordering is correct
     for (int i = 0; i < expectedHosts.size(); i++) {
       Assert.assertEquals(expectedHosts.get(i), actualHosts.get(i));
     }
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/test/java/org/apache/accumulo/monitor/servlets/trace/SummaryTest.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/test/java/org/apache/accumulo/monitor/servlets/trace/SummaryTest.java b/server/monitor/src/test/java/org/apache/accumulo/monitor/servlets/trace/SummaryTest.java
index 23cdc0b..5c33d66 100644
--- a/server/monitor/src/test/java/org/apache/accumulo/monitor/servlets/trace/SummaryTest.java
+++ b/server/monitor/src/test/java/org/apache/accumulo/monitor/servlets/trace/SummaryTest.java
@@ -20,7 +20,7 @@ import org.junit.Before;
 import org.junit.Test;
 
 /**
- * 
+ *
  */
 public class SummaryTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/monitor/src/test/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberTypeTest.java
----------------------------------------------------------------------
diff --git a/server/monitor/src/test/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberTypeTest.java b/server/monitor/src/test/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberTypeTest.java
index 6e741ae..3bb11e6 100644
--- a/server/monitor/src/test/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberTypeTest.java
+++ b/server/monitor/src/test/java/org/apache/accumulo/monitor/util/celltypes/PreciseNumberTypeTest.java
@@ -16,8 +16,9 @@
  */
 package org.apache.accumulo.monitor.util.celltypes;
 
+import static org.junit.Assert.assertEquals;
+
 import org.junit.Test;
-import static org.junit.Assert.*;
 
 public class PreciseNumberTypeTest {
 

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/main/java/org/apache/accumulo/tracer/AsyncSpanReceiver.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/main/java/org/apache/accumulo/tracer/AsyncSpanReceiver.java b/server/tracer/src/main/java/org/apache/accumulo/tracer/AsyncSpanReceiver.java
index 7c54b55..55378ad 100644
--- a/server/tracer/src/main/java/org/apache/accumulo/tracer/AsyncSpanReceiver.java
+++ b/server/tracer/src/main/java/org/apache/accumulo/tracer/AsyncSpanReceiver.java
@@ -16,15 +16,6 @@
  */
 package org.apache.accumulo.tracer;
 
-import org.apache.accumulo.core.trace.DistributedTrace;
-import org.apache.accumulo.tracer.thrift.Annotation;
-import org.apache.accumulo.tracer.thrift.RemoteSpan;
-import org.apache.log4j.Logger;
-import org.htrace.HTraceConfiguration;
-import org.htrace.Span;
-import org.htrace.SpanReceiver;
-import org.htrace.TimelineAnnotation;
-
 import java.net.InetAddress;
 import java.net.UnknownHostException;
 import java.nio.ByteBuffer;
@@ -38,6 +29,15 @@ import java.util.Timer;
 import java.util.TimerTask;
 import java.util.concurrent.ConcurrentLinkedQueue;
 
+import org.apache.accumulo.core.trace.DistributedTrace;
+import org.apache.accumulo.tracer.thrift.Annotation;
+import org.apache.accumulo.tracer.thrift.RemoteSpan;
+import org.apache.log4j.Logger;
+import org.htrace.HTraceConfiguration;
+import org.htrace.Span;
+import org.htrace.SpanReceiver;
+import org.htrace.TimelineAnnotation;
+
 /**
  * Deliver Span information periodically to a destination.
  * <ul>
@@ -49,18 +49,18 @@ import java.util.concurrent.ConcurrentLinkedQueue;
 public abstract class AsyncSpanReceiver<SpanKey,Destination> implements SpanReceiver {
 
   private static final Logger log = Logger.getLogger(AsyncSpanReceiver.class);
-  
+
   private final Map<SpanKey,Destination> clients = new HashMap<SpanKey,Destination>();
-  
+
   protected String host = null;
   protected String service = null;
-  
+
   protected abstract Destination createDestination(SpanKey key) throws Exception;
-  
+
   protected abstract void send(Destination resource, RemoteSpan span) throws Exception;
-  
+
   protected abstract SpanKey getSpanKey(Map<ByteBuffer,ByteBuffer> data);
-  
+
   Timer timer = new Timer("SpanSender", true);
   protected final AbstractQueue<RemoteSpan> sendQueue = new ConcurrentLinkedQueue<RemoteSpan>();
 
@@ -78,7 +78,7 @@ public abstract class AsyncSpanReceiver<SpanKey,Destination> implements SpanRece
           log.warn("Exception sending spans to destination", ex);
         }
       }
-      
+
     }, millis, millis);
   }
 
@@ -113,11 +113,11 @@ public abstract class AsyncSpanReceiver<SpanKey,Destination> implements SpanRece
     }
   }
 
-  public static Map<ByteBuffer, ByteBuffer> convertToByteBuffers(Map<byte[], byte[]> bytesMap) {
+  public static Map<ByteBuffer,ByteBuffer> convertToByteBuffers(Map<byte[],byte[]> bytesMap) {
     if (bytesMap == null)
       return null;
-    Map<ByteBuffer, ByteBuffer> result = new HashMap<ByteBuffer, ByteBuffer>();
-    for (Entry<byte[], byte[]> bytes : bytesMap.entrySet()) {
+    Map<ByteBuffer,ByteBuffer> result = new HashMap<ByteBuffer,ByteBuffer>();
+    for (Entry<byte[],byte[]> bytes : bytesMap.entrySet()) {
       result.put(ByteBuffer.wrap(bytes.getKey()), ByteBuffer.wrap(bytes.getValue()));
     }
     return result;
@@ -135,12 +135,12 @@ public abstract class AsyncSpanReceiver<SpanKey,Destination> implements SpanRece
 
   @Override
   public void receiveSpan(Span s) {
-    Map<ByteBuffer, ByteBuffer> data = convertToByteBuffers(s.getKVAnnotations());
+    Map<ByteBuffer,ByteBuffer> data = convertToByteBuffers(s.getKVAnnotations());
     SpanKey dest = getSpanKey(data);
     if (dest != null) {
       List<Annotation> annotations = convertToAnnotations(s.getTimelineAnnotations());
-      sendQueue.add(new RemoteSpan(host, service==null ? s.getProcessId() : service, s.getTraceId(), s.getSpanId(), s.getParentId(),
-          s.getStartTimeMillis(), s.getStopTimeMillis(), s.getDescription(), data, annotations));
+      sendQueue.add(new RemoteSpan(host, service == null ? s.getProcessId() : service, s.getTraceId(), s.getSpanId(), s.getParentId(), s.getStartTimeMillis(),
+          s.getStopTimeMillis(), s.getDescription(), data, annotations));
     }
   }
 
@@ -170,5 +170,5 @@ public abstract class AsyncSpanReceiver<SpanKey,Destination> implements SpanRece
     }
     service = conf.get(DistributedTrace.TRACE_SERVICE_PROPERTY, service);
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/main/java/org/apache/accumulo/tracer/SendSpansViaThrift.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/main/java/org/apache/accumulo/tracer/SendSpansViaThrift.java b/server/tracer/src/main/java/org/apache/accumulo/tracer/SendSpansViaThrift.java
index 2190570..c585cc1 100644
--- a/server/tracer/src/main/java/org/apache/accumulo/tracer/SendSpansViaThrift.java
+++ b/server/tracer/src/main/java/org/apache/accumulo/tracer/SendSpansViaThrift.java
@@ -18,6 +18,11 @@ package org.apache.accumulo.tracer;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
 
+import java.net.InetSocketAddress;
+import java.net.Socket;
+import java.nio.ByteBuffer;
+import java.util.Map;
+
 import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.accumulo.tracer.thrift.SpanReceiver.Client;
 import org.apache.thrift.protocol.TBinaryProtocol;
@@ -25,18 +30,13 @@ import org.apache.thrift.protocol.TProtocol;
 import org.apache.thrift.transport.TSocket;
 import org.apache.thrift.transport.TTransport;
 
-import java.net.InetSocketAddress;
-import java.net.Socket;
-import java.nio.ByteBuffer;
-import java.util.Map;
-
 /**
  * Send Span data to a destination using thrift.
  */
 public class SendSpansViaThrift extends AsyncSpanReceiver<String,Client> {
-  
+
   private static final org.apache.log4j.Logger log = org.apache.log4j.Logger.getLogger(SendSpansViaThrift.class);
-  
+
   private static final String THRIFT = "thrift://";
 
   public SendSpansViaThrift() {
@@ -46,7 +46,7 @@ public class SendSpansViaThrift extends AsyncSpanReceiver<String,Client> {
   public SendSpansViaThrift(long millis) {
     super(millis);
   }
-  
+
   @Override
   protected Client createDestination(String destination) throws Exception {
     if (destination == null)
@@ -54,7 +54,7 @@ public class SendSpansViaThrift extends AsyncSpanReceiver<String,Client> {
     try {
       int portSeparatorIndex = destination.lastIndexOf(':');
       String host = destination.substring(0, portSeparatorIndex);
-      int port = Integer.parseInt(destination.substring(portSeparatorIndex+1));
+      int port = Integer.parseInt(destination.substring(portSeparatorIndex + 1));
       log.debug("Connecting to " + host + ":" + port);
       InetSocketAddress addr = new InetSocketAddress(host, port);
       Socket sock = new Socket();
@@ -67,7 +67,7 @@ public class SendSpansViaThrift extends AsyncSpanReceiver<String,Client> {
       return null;
     }
   }
-  
+
   @Override
   protected void send(Client client, RemoteSpan s) throws Exception {
     if (client != null) {
@@ -93,5 +93,5 @@ public class SendSpansViaThrift extends AsyncSpanReceiver<String,Client> {
     }
     return null;
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/main/java/org/apache/accumulo/tracer/SpanTree.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/main/java/org/apache/accumulo/tracer/SpanTree.java b/server/tracer/src/main/java/org/apache/accumulo/tracer/SpanTree.java
index 6337ee4..e8f8056 100644
--- a/server/tracer/src/main/java/org/apache/accumulo/tracer/SpanTree.java
+++ b/server/tracer/src/main/java/org/apache/accumulo/tracer/SpanTree.java
@@ -23,22 +23,22 @@ import java.util.List;
 import java.util.Map;
 import java.util.Set;
 
-import org.htrace.Span;
 import org.apache.accumulo.tracer.thrift.RemoteSpan;
+import org.htrace.Span;
 
 public class SpanTree {
   final Map<Long,List<Long>> parentChildren = new HashMap<Long,List<Long>>();
   public final Map<Long,RemoteSpan> nodes = new HashMap<Long,RemoteSpan>();
-  
+
   public SpanTree() {}
-  
+
   public void addNode(RemoteSpan span) {
     nodes.put(span.spanId, span);
     if (parentChildren.get(span.parentId) == null)
       parentChildren.put(span.parentId, new ArrayList<Long>());
     parentChildren.get(span.parentId).add(span.spanId);
   }
-  
+
   public Set<Long> visit(SpanTreeVisitor visitor) {
     Set<Long> visited = new HashSet<Long>();
     List<Long> root = parentChildren.get(Long.valueOf(Span.ROOT_SPAN_ID));
@@ -50,7 +50,7 @@ public class SpanTree {
     recurse(0, null, rootSpan, visitor, visited);
     return visited;
   }
-  
+
   private void recurse(int level, RemoteSpan parent, RemoteSpan node, SpanTreeVisitor visitor, Set<Long> visited) {
     // improbable case: duplicate spanId in a trace tree: prevent
     // infinite recursion

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceDump.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceDump.java b/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceDump.java
index 4468866..64fee7e 100644
--- a/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceDump.java
+++ b/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceDump.java
@@ -41,24 +41,26 @@ import org.htrace.Span;
 
 import com.beust.jcommander.Parameter;
 
-
 public class TraceDump {
   static final long DEFAULT_TIME_IN_MILLIS = 10 * 60 * 1000l;
 
   static class Opts extends ClientOnDefaultTable {
-    @Parameter(names={"-l", "--list"}, description="List recent traces")
+    @Parameter(names = {"-l", "--list"}, description = "List recent traces")
     boolean list = false;
-    @Parameter(names={"-s", "--start"}, description="The start time of traces to display")
+    @Parameter(names = {"-s", "--start"}, description = "The start time of traces to display")
     String start;
-    @Parameter(names={"-e", "--end"}, description="The end time of traces to display")
+    @Parameter(names = {"-e", "--end"}, description = "The end time of traces to display")
     String end;
-    @Parameter(names={"-d", "--dump"}, description="Dump the traces")
+    @Parameter(names = {"-d", "--dump"}, description = "Dump the traces")
     boolean dump = false;
-    @Parameter(names={"-i", "--instance"}, description="URL to point to accumulo.")
+    @Parameter(names = {"-i", "--instance"}, description = "URL to point to accumulo.")
     String instance;
-    @Parameter(description=" <trace id> { <trace id> ... }")
+    @Parameter(description = " <trace id> { <trace id> ... }")
     List<String> traceIds = new ArrayList<String>();
-    Opts() { super("trace");}
+
+    Opts() {
+      super("trace");
+    }
   }
 
   public static void main(String[] args) throws Exception {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceFormatter.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceFormatter.java b/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceFormatter.java
index 4829eb2..0ec93b4 100644
--- a/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceFormatter.java
+++ b/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceFormatter.java
@@ -24,22 +24,21 @@ import java.util.Date;
 import java.util.Iterator;
 import java.util.Map.Entry;
 
-import org.apache.accumulo.tracer.thrift.Annotation;
-import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.util.format.DefaultFormatter;
 import org.apache.accumulo.core.util.format.Formatter;
+import org.apache.accumulo.tracer.thrift.Annotation;
+import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.commons.lang.NotImplementedException;
 import org.apache.hadoop.io.Text;
 import org.apache.thrift.TException;
 import org.apache.thrift.protocol.TCompactProtocol;
 import org.apache.thrift.transport.TMemoryInputTransport;
 
-
 /**
  * A formatter than can be used in the shell to display trace information.
- * 
+ *
  */
 public class TraceFormatter implements Formatter {
   public static final String DATE_FORMAT = "yyyy/MM/dd HH:mm:ss.SSS";
@@ -50,16 +49,16 @@ public class TraceFormatter implements Formatter {
       return new SimpleDateFormat(DATE_FORMAT);
     }
   };
-  
+
   public static String formatDate(final Date date) {
     return formatter.get().format(date);
   }
-  
+
   private final static Text SPAN_CF = new Text("span");
-  
+
   private Iterator<Entry<Key,Value>> scanner;
   private boolean printTimeStamps;
-  
+
   public static RemoteSpan getRemoteSpan(Entry<Key,Value> entry) {
     TMemoryInputTransport transport = new TMemoryInputTransport(entry.getValue().get());
     TCompactProtocol protocol = new TCompactProtocol(transport);
@@ -71,12 +70,12 @@ public class TraceFormatter implements Formatter {
     }
     return span;
   }
-  
+
   @Override
   public boolean hasNext() {
     return scanner.hasNext();
   }
-  
+
   @Override
   public String next() {
     Entry<Key,Value> next = scanner.next();
@@ -93,7 +92,7 @@ public class TraceFormatter implements Formatter {
       result.append(String.format(" %12s:%s%n", "start", dateFormatter.format(span.start)));
       result.append(String.format(" %12s:%s%n", "ms", span.stop - span.start));
       if (span.data != null) {
-        for (Entry<ByteBuffer, ByteBuffer> entry : span.data.entrySet()) {
+        for (Entry<ByteBuffer,ByteBuffer> entry : span.data.entrySet()) {
           String key = new String(entry.getKey().array(), entry.getKey().arrayOffset(), entry.getKey().limit(), UTF_8);
           String value = new String(entry.getValue().array(), entry.getValue().arrayOffset(), entry.getValue().limit(), UTF_8);
           result.append(String.format(" %12s:%s%n", key, value));
@@ -104,7 +103,7 @@ public class TraceFormatter implements Formatter {
           result.append(String.format(" %12s:%s:%s%n", "annotation", annotation.getMsg(), dateFormatter.format(annotation.getTime())));
         }
       }
-      
+
       if (printTimeStamps) {
         result.append(String.format(" %-12s:%d%n", "timestamp", next.getKey().getTimestamp()));
       }
@@ -112,12 +111,12 @@ public class TraceFormatter implements Formatter {
     }
     return DefaultFormatter.formatEntry(next, printTimeStamps);
   }
-  
+
   @Override
   public void remove() {
     throw new NotImplementedException();
   }
-  
+
   @Override
   public void initialize(Iterable<Entry<Key,Value>> scanner, boolean printTimestamps) {
     this.scanner = scanner.iterator();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceServer.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceServer.java b/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceServer.java
index 7e33300..3063cdc 100644
--- a/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceServer.java
+++ b/server/tracer/src/main/java/org/apache/accumulo/tracer/TraceServer.java
@@ -54,7 +54,6 @@ import org.apache.accumulo.server.security.SecurityUtil;
 import org.apache.accumulo.server.util.time.SimpleTimer;
 import org.apache.accumulo.server.zookeeper.ZooReaderWriter;
 import org.apache.accumulo.start.classloader.vfs.AccumuloVFSClassLoader;
-import org.htrace.Span;
 import org.apache.accumulo.tracer.thrift.RemoteSpan;
 import org.apache.accumulo.tracer.thrift.SpanReceiver.Iface;
 import org.apache.accumulo.tracer.thrift.SpanReceiver.Processor;
@@ -73,6 +72,7 @@ import org.apache.zookeeper.WatchedEvent;
 import org.apache.zookeeper.Watcher;
 import org.apache.zookeeper.Watcher.Event.EventType;
 import org.apache.zookeeper.Watcher.Event.KeeperState;
+import org.htrace.Span;
 
 public class TraceServer implements Watcher {
 
@@ -184,8 +184,8 @@ public class TraceServer implements Watcher {
           at = new PasswordToken(conf.get(p).getBytes(UTF_8));
         } else {
           Properties props = new Properties();
-          AuthenticationToken token = AccumuloVFSClassLoader.getClassLoader().loadClass(conf.get(Property.TRACE_TOKEN_TYPE)).asSubclass(AuthenticationToken.class)
-              .newInstance();
+          AuthenticationToken token = AccumuloVFSClassLoader.getClassLoader().loadClass(conf.get(Property.TRACE_TOKEN_TYPE))
+              .asSubclass(AuthenticationToken.class).newInstance();
 
           int prefixLength = Property.TRACE_TOKEN_PROPERTY_PREFIX.getKey().length();
           for (Entry<String,String> entry : loginMap.entrySet()) {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tracer/src/test/java/org/apache/accumulo/tracer/TracerTest.java
----------------------------------------------------------------------
diff --git a/server/tracer/src/test/java/org/apache/accumulo/tracer/TracerTest.java b/server/tracer/src/test/java/org/apache/accumulo/tracer/TracerTest.java
index c1264bb..76ed910 100644
--- a/server/tracer/src/test/java/org/apache/accumulo/tracer/TracerTest.java
+++ b/server/tracer/src/test/java/org/apache/accumulo/tracer/TracerTest.java
@@ -33,8 +33,8 @@ import java.util.concurrent.Callable;
 
 import org.apache.accumulo.core.trace.Span;
 import org.apache.accumulo.core.trace.Trace;
-import org.apache.accumulo.core.trace.wrappers.TraceWrap;
 import org.apache.accumulo.core.trace.thrift.TInfo;
+import org.apache.accumulo.core.trace.wrappers.TraceWrap;
 import org.apache.accumulo.tracer.thrift.TestService;
 import org.apache.accumulo.tracer.thrift.TestService.Iface;
 import org.apache.accumulo.tracer.thrift.TestService.Processor;
@@ -52,7 +52,6 @@ import org.htrace.wrappers.TraceProxy;
 import org.junit.Before;
 import org.junit.Test;
 
-
 public class TracerTest {
   static class SpanStruct {
     public SpanStruct(long traceId, long spanId, long parentId, long start, long stop, String description, Map<byte[],byte[]> data) {
@@ -65,7 +64,7 @@ public class TracerTest {
       this.description = description;
       this.data = data;
     }
-    
+
     public long traceId;
     public long spanId;
     public long parentId;
@@ -73,17 +72,17 @@ public class TracerTest {
     public long stop;
     public String description;
     public Map<byte[],byte[]> data;
-    
+
     public long millis() {
       return stop - start;
     }
   }
-  
+
   static class TestReceiver implements SpanReceiver {
     public Map<Long,List<SpanStruct>> traces = new HashMap<Long,List<SpanStruct>>();
-    
+
     @Override
-    public void receiveSpan(org.htrace.Span s)  {
+    public void receiveSpan(org.htrace.Span s) {
       long traceId = s.getTraceId();
       SpanStruct span = new SpanStruct(traceId, s.getSpanId(), s.getParentId(), s.getStartTimeMillis(), s.getStopTimeMillis(), s.getDescription(),
           s.getKVAnnotations());
@@ -91,48 +90,46 @@ public class TracerTest {
         traces.put(traceId, new ArrayList<SpanStruct>());
       traces.get(traceId).add(span);
     }
-    
+
     @Override
-    public void configure(HTraceConfiguration conf) {
-    }
+    public void configure(HTraceConfiguration conf) {}
 
     @Override
-    public void close() throws IOException {
-    }
+    public void close() throws IOException {}
   }
-  
+
   @SuppressWarnings("deprecation")
   @Test
   public void testTrace() throws Exception {
     TestReceiver tracer = new TestReceiver();
     org.htrace.Trace.addReceiver(tracer);
-    
+
     assertFalse(Trace.isTracing());
     Trace.start("nop").stop();
     assertTrue(tracer.traces.size() == 0);
     assertFalse(Trace.isTracing());
-    
+
     Trace.on("nop").stop();
     assertTrue(tracer.traces.size() == 1);
     assertFalse(Trace.isTracing());
-    
+
     Span start = Trace.on("testing");
     assertEquals(Trace.currentTrace().getSpan(), start.getScope().getSpan());
     assertTrue(Trace.isTracing());
-    
+
     Span span = Trace.start("shortest trace ever");
     span.stop();
     long traceId = Trace.currentTraceId();
     assertNotNull(tracer.traces.get(traceId));
     assertTrue(tracer.traces.get(traceId).size() == 1);
     assertEquals("shortest trace ever", tracer.traces.get(traceId).get(0).description);
-    
+
     Span pause = Trace.start("pause");
     Thread.sleep(100);
     pause.stop();
     assertTrue(tracer.traces.get(traceId).size() == 2);
     assertTrue(tracer.traces.get(traceId).get(1).millis() >= 100);
-    
+
     Thread t = new Thread(Trace.wrap(new Runnable() {
       @Override
       public void run() {
@@ -141,13 +138,13 @@ public class TracerTest {
     }), "My Task");
     t.start();
     t.join();
-    
+
     assertTrue(tracer.traces.get(traceId).size() == 3);
     assertEquals("My Task", tracer.traces.get(traceId).get(2).description);
     Trace.off();
     assertFalse(Trace.isTracing());
   }
-  
+
   static class Service implements TestService.Iface {
     @Override
     public boolean checkTrace(TInfo t, String message) throws TException {
@@ -159,12 +156,12 @@ public class TracerTest {
       }
     }
   }
-  
+
   @Test
   public void testThrift() throws Exception {
     TestReceiver tracer = new TestReceiver();
     org.htrace.Trace.addReceiver(tracer);
-    
+
     ServerSocket socket = new ServerSocket(0);
     TServerSocket transport = new TServerSocket(socket);
     transport.listen();
@@ -181,17 +178,17 @@ public class TracerTest {
     TestService.Iface client = new TestService.Client(new TBinaryProtocol(clientTransport), new TBinaryProtocol(clientTransport));
     client = TraceWrap.client(client);
     assertFalse(client.checkTrace(null, "test"));
-    
+
     Span start = Trace.on("start");
     assertTrue(client.checkTrace(null, "my test"));
     start.stop();
-    
+
     assertNotNull(tracer.traces.get(start.traceId()));
     String traces[] = {"my test", "checkTrace", "client:checkTrace", "start"};
     assertTrue(tracer.traces.get(start.traceId()).size() == traces.length);
     for (int i = 0; i < traces.length; i++)
       assertEquals(traces[i], tracer.traces.get(start.traceId()).get(i).description);
-    
+
     tserver.stop();
     t.join(100);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/server/GarbageCollectionLogger.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/server/GarbageCollectionLogger.java b/server/tserver/src/main/java/org/apache/accumulo/server/GarbageCollectionLogger.java
index 8bb163d..7d09fe3 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/server/GarbageCollectionLogger.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/server/GarbageCollectionLogger.java
@@ -28,14 +28,13 @@ import org.apache.log4j.Logger;
 
 public class GarbageCollectionLogger {
   private static final Logger log = Logger.getLogger(GarbageCollectionLogger.class);
-  
+
   private final HashMap<String,Long> prevGcTime = new HashMap<String,Long>();
   private long lastMemorySize = 0;
   private long gcTimeIncreasedCount = 0;
   private static long lastMemoryCheckTime = 0;
-  
-  public GarbageCollectionLogger() {
-  }
+
+  public GarbageCollectionLogger() {}
 
   public synchronized void logGCInfo(AccumuloConfiguration conf) {
     final long now = System.currentTimeMillis();
@@ -99,7 +98,7 @@ public class GarbageCollectionLogger {
       final long diff = now - lastMemoryCheckTime;
       if (diff > keepAliveTimeout) {
         log.warn(String.format("GC pause checker not called in a timely fashion. Expected every %.1f seconds but was %.1f seconds since last check",
-                    keepAliveTimeout / 1000., diff / 1000.));
+            keepAliveTimeout / 1000., diff / 1000.));
       }
       lastMemoryCheckTime = now;
       return;
@@ -113,5 +112,4 @@ public class GarbageCollectionLogger {
     lastMemoryCheckTime = now;
   }
 
-
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/BulkFailedCopyProcessor.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/BulkFailedCopyProcessor.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/BulkFailedCopyProcessor.java
index 034becd..cfb5fb4 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/BulkFailedCopyProcessor.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/BulkFailedCopyProcessor.java
@@ -19,6 +19,7 @@ package org.apache.accumulo.tserver;
 import static java.nio.charset.StandardCharsets.UTF_8;
 
 import java.io.IOException;
+
 import org.apache.accumulo.core.conf.SiteConfiguration;
 import org.apache.accumulo.core.util.CachedConfiguration;
 import org.apache.accumulo.server.fs.VolumeManager;
@@ -54,7 +55,7 @@ public class BulkFailedCopyProcessor implements Processor {
       VolumeManager vm = VolumeManagerImpl.get(SiteConfiguration.getInstance());
       FileSystem origFs = vm.getVolumeByPath(orig).getFileSystem();
       FileSystem destFs = vm.getVolumeByPath(dest).getFileSystem();
-      
+
       FileUtil.copy(origFs, orig, destFs, tmp, false, true, CachedConfiguration.getInstance());
       destFs.rename(tmp, dest);
       log.debug("copied " + orig + " to " + dest);

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/CompactionQueue.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/CompactionQueue.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/CompactionQueue.java
index a0574d9..0cb04a7 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/CompactionQueue.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/CompactionQueue.java
@@ -27,122 +27,122 @@ import java.util.concurrent.TimeUnit;
 
 @SuppressWarnings({"rawtypes", "unchecked"})
 public class CompactionQueue extends AbstractQueue<Runnable> implements BlockingQueue<Runnable> {
-  
+
   private List<Comparable> task = new LinkedList<Comparable>();
-  
+
   @Override
   public synchronized Runnable poll() {
     if (task.size() == 0)
       return null;
-    
+
     Comparable min = Collections.min(task);
     task.remove(min);
     return (Runnable) min;
   }
-  
+
   @Override
   public synchronized Runnable peek() {
     if (task.size() == 0)
       return null;
-    
+
     Comparable min = Collections.min(task);
     return (Runnable) min;
   }
-  
+
   @Override
   public synchronized boolean offer(Runnable e) {
     task.add((Comparable) e);
     notify();
     return true;
   }
-  
+
   @Override
   public synchronized void put(Runnable e) throws InterruptedException {
     task.add((Comparable) e);
     notify();
   }
-  
+
   @Override
   public synchronized boolean offer(Runnable e, long timeout, TimeUnit unit) throws InterruptedException {
     task.add((Comparable) e);
     notify();
     return true;
   }
-  
+
   @Override
   public synchronized Runnable take() throws InterruptedException {
     while (task.size() == 0) {
       wait();
     }
-    
+
     return poll();
   }
-  
+
   @Override
   public synchronized Runnable poll(long timeout, TimeUnit unit) throws InterruptedException {
     if (task.size() == 0) {
       wait(unit.toMillis(timeout));
     }
-    
+
     if (task.size() == 0)
       return null;
-    
+
     return poll();
   }
-  
+
   @Override
   public synchronized int remainingCapacity() {
     return Integer.MAX_VALUE;
   }
-  
+
   @Override
   public synchronized int drainTo(Collection<? super Runnable> c) {
     return drainTo(c, task.size());
   }
-  
+
   @Override
   public synchronized int drainTo(Collection<? super Runnable> c, int maxElements) {
     Collections.sort(task);
-    
+
     int num = Math.min(task.size(), maxElements);
-    
+
     Iterator<Comparable> iter = task.iterator();
     for (int i = 0; i < num; i++) {
       c.add((Runnable) iter.next());
       iter.remove();
     }
-    
+
     return num;
   }
-  
+
   @Override
   public synchronized Iterator<Runnable> iterator() {
     Collections.sort(task);
-    
+
     final Iterator<Comparable> iter = task.iterator();
-    
+
     return new Iterator<Runnable>() {
-      
+
       @Override
       public boolean hasNext() {
         return iter.hasNext();
       }
-      
+
       @Override
       public Runnable next() {
         return (Runnable) iter.next();
       }
-      
+
       @Override
       public void remove() {
         iter.remove();
       }
     };
   }
-  
+
   @Override
   public synchronized int size() {
     return task.size();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/ConditionalMutationSet.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/ConditionalMutationSet.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/ConditionalMutationSet.java
index bbd7007..79a1176 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/ConditionalMutationSet.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/ConditionalMutationSet.java
@@ -29,14 +29,14 @@ import org.apache.accumulo.tserver.data.ServerConditionalMutation;
 import org.apache.hadoop.io.WritableComparator;
 
 /**
- * 
+ *
  */
 public class ConditionalMutationSet {
 
   interface DeferFilter {
     void defer(List<ServerConditionalMutation> scml, List<ServerConditionalMutation> okMutations, List<ServerConditionalMutation> deferred);
   }
-  
+
   static class DuplicateFilter implements DeferFilter {
     public void defer(List<ServerConditionalMutation> scml, List<ServerConditionalMutation> okMutations, List<ServerConditionalMutation> deferred) {
       okMutations.add(scml.get(0));
@@ -49,14 +49,14 @@ public class ConditionalMutationSet {
       }
     }
   }
-  
+
   static void defer(Map<KeyExtent,List<ServerConditionalMutation>> updates, Map<KeyExtent,List<ServerConditionalMutation>> deferredMutations, DeferFilter filter) {
     for (Entry<KeyExtent,List<ServerConditionalMutation>> entry : updates.entrySet()) {
       List<ServerConditionalMutation> scml = entry.getValue();
       List<ServerConditionalMutation> okMutations = new ArrayList<ServerConditionalMutation>(scml.size());
       List<ServerConditionalMutation> deferred = new ArrayList<ServerConditionalMutation>();
       filter.defer(scml, okMutations, deferred);
-      
+
       if (deferred.size() > 0) {
         scml.clear();
         scml.addAll(okMutations);
@@ -71,7 +71,7 @@ public class ConditionalMutationSet {
       }
     }
   }
-  
+
   static void deferDuplicatesRows(Map<KeyExtent,List<ServerConditionalMutation>> updates, Map<KeyExtent,List<ServerConditionalMutation>> deferred) {
     defer(updates, deferred, new DuplicateFilter());
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/FileManager.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/FileManager.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/FileManager.java
index 98ccd07..0a7de95 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/FileManager.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/FileManager.java
@@ -56,22 +56,22 @@ import org.apache.hadoop.fs.Path;
 import org.apache.log4j.Logger;
 
 public class FileManager {
-  
+
   private static final Logger log = Logger.getLogger(FileManager.class);
-  
+
   int maxOpen;
-  
+
   private static class OpenReader implements Comparable<OpenReader> {
     long releaseTime;
     FileSKVIterator reader;
     String fileName;
-    
+
     public OpenReader(String fileName, FileSKVIterator reader) {
       this.fileName = fileName;
       this.reader = reader;
       this.releaseTime = System.currentTimeMillis();
     }
-    
+
     @Override
     public int compareTo(OpenReader o) {
       if (releaseTime < o.releaseTime) {
@@ -82,7 +82,7 @@ public class FileManager {
         return 0;
       }
     }
-    
+
     @Override
     public boolean equals(Object obj) {
       if (obj instanceof OpenReader) {
@@ -90,40 +90,40 @@ public class FileManager {
       }
       return false;
     }
-    
+
     @Override
     public int hashCode() {
       return fileName.hashCode();
     }
   }
-  
+
   private Map<String,List<OpenReader>> openFiles;
   private HashMap<FileSKVIterator,String> reservedReaders;
-  
+
   private Semaphore filePermits;
-  
+
   private VolumeManager fs;
-  
+
   // the data cache and index cache are allocated in
   // TabletResourceManager and passed through the file opener to
   // CachableBlockFile which can handle the caches being
   // null if unallocated
   private BlockCache dataCache = null;
   private BlockCache indexCache = null;
-  
+
   private long maxIdleTime;
 
   private final AccumuloServerContext context;
 
   private class IdleFileCloser implements Runnable {
-    
+
     @Override
     public void run() {
-      
+
       long curTime = System.currentTimeMillis();
-      
+
       ArrayList<FileSKVIterator> filesToClose = new ArrayList<FileSKVIterator>();
-      
+
       // determine which files to close in a sync block, and then close the
       // files outside of the sync block
       synchronized (FileManager.this) {
@@ -131,31 +131,31 @@ public class FileManager {
         while (iter.hasNext()) {
           Entry<String,List<OpenReader>> entry = iter.next();
           List<OpenReader> ofl = entry.getValue();
-          
+
           for (Iterator<OpenReader> oflIter = ofl.iterator(); oflIter.hasNext();) {
             OpenReader openReader = oflIter.next();
-            
+
             if (curTime - openReader.releaseTime > maxIdleTime) {
-              
+
               filesToClose.add(openReader.reader);
               oflIter.remove();
             }
           }
-          
+
           if (ofl.size() == 0) {
             iter.remove();
           }
         }
       }
-      
+
       closeReaders(filesToClose);
-      
+
     }
-    
+
   }
-  
+
   /**
-   * 
+   *
    * @param dataCache
    *          : underlying file can and should be able to handle a null cache
    * @param indexCache
@@ -168,11 +168,11 @@ public class FileManager {
     this.context = context;
     this.dataCache = dataCache;
     this.indexCache = indexCache;
-    
+
     this.filePermits = new Semaphore(maxOpen, true);
     this.maxOpen = maxOpen;
     this.fs = fs;
-    
+
     this.openFiles = new HashMap<String,List<OpenReader>>();
     this.reservedReaders = new HashMap<FileSKVIterator,String>();
 
@@ -180,57 +180,57 @@ public class FileManager {
     SimpleTimer.getInstance(context.getConfiguration()).schedule(new IdleFileCloser(), maxIdleTime, maxIdleTime / 2);
 
   }
-  
+
   private static int countReaders(Map<String,List<OpenReader>> files) {
     int count = 0;
-    
+
     for (List<OpenReader> list : files.values()) {
       count += list.size();
     }
-    
+
     return count;
   }
-  
+
   private List<FileSKVIterator> takeLRUOpenFiles(int numToTake) {
-    
+
     ArrayList<OpenReader> openReaders = new ArrayList<OpenReader>();
-    
+
     for (Entry<String,List<OpenReader>> entry : openFiles.entrySet()) {
       openReaders.addAll(entry.getValue());
     }
-    
+
     Collections.sort(openReaders);
-    
+
     ArrayList<FileSKVIterator> ret = new ArrayList<FileSKVIterator>();
-    
+
     for (int i = 0; i < numToTake && i < openReaders.size(); i++) {
       OpenReader or = openReaders.get(i);
-      
+
       List<OpenReader> ofl = openFiles.get(or.fileName);
       if (!ofl.remove(or)) {
         throw new RuntimeException("Failed to remove open reader that should have been there");
       }
-      
+
       if (ofl.size() == 0) {
         openFiles.remove(or.fileName);
       }
-      
+
       ret.add(or.reader);
     }
-    
+
     return ret;
   }
-  
+
   private static <T> List<T> getFileList(String file, Map<String,List<T>> files) {
     List<T> ofl = files.get(file);
     if (ofl == null) {
       ofl = new ArrayList<T>();
       files.put(file, ofl);
     }
-    
+
     return ofl;
   }
-  
+
   private void closeReaders(List<FileSKVIterator> filesToClose) {
     for (FileSKVIterator reader : filesToClose) {
       try {
@@ -240,12 +240,12 @@ public class FileManager {
       }
     }
   }
-  
+
   private List<String> takeOpenFiles(Collection<String> files, List<FileSKVIterator> reservedFiles, Map<FileSKVIterator,String> readersReserved) {
     List<String> filesToOpen = new LinkedList<String>(files);
     for (Iterator<String> iterator = filesToOpen.iterator(); iterator.hasNext();) {
       String file = iterator.next();
-      
+
       List<OpenReader> ofl = openFiles.get(file);
       if (ofl != null && ofl.size() > 0) {
         OpenReader openReader = ofl.remove(ofl.size() - 1);
@@ -256,55 +256,55 @@ public class FileManager {
         }
         iterator.remove();
       }
-      
+
     }
     return filesToOpen;
   }
-  
+
   private synchronized String getReservedReadeFilename(FileSKVIterator reader) {
     return reservedReaders.get(reader);
   }
-  
+
   private List<FileSKVIterator> reserveReaders(KeyExtent tablet, Collection<String> files, boolean continueOnFailure) throws IOException {
-    
+
     if (!tablet.isMeta() && files.size() >= maxOpen) {
       throw new IllegalArgumentException("requested files exceeds max open");
     }
-    
+
     if (files.size() == 0) {
       return Collections.emptyList();
     }
-    
+
     List<String> filesToOpen = null;
     List<FileSKVIterator> filesToClose = Collections.emptyList();
     List<FileSKVIterator> reservedFiles = new ArrayList<FileSKVIterator>();
     Map<FileSKVIterator,String> readersReserved = new HashMap<FileSKVIterator,String>();
-    
+
     if (!tablet.isMeta()) {
       filePermits.acquireUninterruptibly(files.size());
     }
-    
+
     // now that the we are past the semaphore, we have the authority
     // to open files.size() files
-    
+
     // determine what work needs to be done in sync block
     // but do the work of opening and closing files outside
     // a synch block
     synchronized (this) {
-      
+
       filesToOpen = takeOpenFiles(files, reservedFiles, readersReserved);
-      
+
       int numOpen = countReaders(openFiles);
-      
+
       if (filesToOpen.size() + numOpen + reservedReaders.size() > maxOpen) {
         filesToClose = takeLRUOpenFiles((filesToOpen.size() + numOpen + reservedReaders.size()) - maxOpen);
       }
     }
-    
+
     // close files before opening files to ensure we stay under resource
     // limitations
     closeReaders(filesToClose);
-    
+
     // open any files that need to be opened
     for (String file : filesToOpen) {
       try {
@@ -330,36 +330,36 @@ public class FileManager {
         } else {
           // close whatever files were opened
           closeReaders(reservedFiles);
-          
+
           if (!tablet.isMeta()) {
             filePermits.release(files.size());
           }
-          
+
           log.error("Failed to open file " + file + " " + e.getMessage());
           throw new IOException("Failed to open " + file, e);
         }
       }
     }
-    
+
     synchronized (this) {
       // update set of reserved readers
       reservedReaders.putAll(readersReserved);
     }
-    
+
     return reservedFiles;
   }
-  
+
   private void releaseReaders(KeyExtent tablet, List<FileSKVIterator> readers, boolean sawIOException) {
     // put files in openFiles
-    
+
     synchronized (this) {
-      
+
       // check that readers were actually reserved ... want to make sure a thread does
       // not try to release readers they never reserved
       if (!reservedReaders.keySet().containsAll(readers)) {
         throw new IllegalArgumentException("Asked to release readers that were never reserved ");
       }
-      
+
       for (FileSKVIterator reader : readers) {
         try {
           reader.closeDeepCopies();
@@ -368,69 +368,69 @@ public class FileManager {
           sawIOException = true;
         }
       }
-      
+
       for (FileSKVIterator reader : readers) {
         String fileName = reservedReaders.remove(reader);
         if (!sawIOException)
           getFileList(fileName, openFiles).add(new OpenReader(fileName, reader));
       }
     }
-    
+
     if (sawIOException)
       closeReaders(readers);
-    
+
     // decrement the semaphore
     if (!tablet.isMeta()) {
       filePermits.release(readers.size());
     }
-    
+
   }
-  
+
   static class FileDataSource implements DataSource {
-    
+
     private SortedKeyValueIterator<Key,Value> iter;
     private ArrayList<FileDataSource> deepCopies;
     private boolean current = true;
     private IteratorEnvironment env;
     private String file;
     private AtomicBoolean iflag;
-    
+
     FileDataSource(String file, SortedKeyValueIterator<Key,Value> iter) {
       this.file = file;
       this.iter = iter;
       this.deepCopies = new ArrayList<FileManager.FileDataSource>();
     }
-    
+
     public FileDataSource(IteratorEnvironment env, SortedKeyValueIterator<Key,Value> deepCopy, ArrayList<FileDataSource> deepCopies) {
       this.iter = deepCopy;
       this.env = env;
       this.deepCopies = deepCopies;
       deepCopies.add(this);
     }
-    
+
     @Override
     public boolean isCurrent() {
       return current;
     }
-    
+
     @Override
     public DataSource getNewDataSource() {
       current = true;
       return this;
     }
-    
+
     @Override
     public DataSource getDeepCopyDataSource(IteratorEnvironment env) {
       return new FileDataSource(env, iter.deepCopy(env), deepCopies);
     }
-    
+
     @Override
     public SortedKeyValueIterator<Key,Value> iterator() throws IOException {
       if (iflag != null)
         ((InterruptibleIterator) this.iter).setInterruptFlag(iflag);
       return iter;
     }
-    
+
     void unsetIterator() {
       current = false;
       iter = null;
@@ -439,7 +439,7 @@ public class FileManager {
         fds.iter = null;
       }
     }
-    
+
     void setIterator(SortedKeyValueIterator<Key,Value> iter) {
       current = false;
       this.iter = iter;
@@ -457,16 +457,16 @@ public class FileManager {
     public void setInterruptFlag(AtomicBoolean flag) {
       this.iflag = flag;
     }
-    
+
   }
-  
+
   public class ScanFileManager {
-    
+
     private ArrayList<FileDataSource> dataSources;
     private ArrayList<FileSKVIterator> tabletReservedReaders;
     private KeyExtent tablet;
     private boolean continueOnFailure;
-    
+
     ScanFileManager(KeyExtent tablet) {
       tabletReservedReaders = new ArrayList<FileSKVIterator>();
       dataSources = new ArrayList<FileDataSource>();
@@ -478,35 +478,35 @@ public class FileManager {
         continueOnFailure = false;
       }
     }
-    
+
     private List<FileSKVIterator> openFileRefs(Collection<FileRef> files) throws TooManyFilesException, IOException {
       List<String> strings = new ArrayList<String>(files.size());
       for (FileRef ref : files)
         strings.add(ref.path().toString());
       return openFiles(strings);
     }
-    
+
     private List<FileSKVIterator> openFiles(Collection<String> files) throws TooManyFilesException, IOException {
       // one tablet can not open more than maxOpen files, otherwise it could get stuck
       // forever waiting on itself to release files
-      
+
       if (tabletReservedReaders.size() + files.size() >= maxOpen) {
         throw new TooManyFilesException("Request to open files would exceed max open files reservedReaders.size()=" + tabletReservedReaders.size()
             + " files.size()=" + files.size() + " maxOpen=" + maxOpen + " tablet = " + tablet);
       }
-      
+
       List<FileSKVIterator> newlyReservedReaders = reserveReaders(tablet, files, continueOnFailure);
-      
+
       tabletReservedReaders.addAll(newlyReservedReaders);
       return newlyReservedReaders;
     }
-    
+
     public synchronized List<InterruptibleIterator> openFiles(Map<FileRef,DataFileValue> files, boolean detachable) throws IOException {
-      
+
       List<FileSKVIterator> newlyReservedReaders = openFileRefs(files.keySet());
-      
+
       ArrayList<InterruptibleIterator> iters = new ArrayList<InterruptibleIterator>();
-      
+
       for (FileSKVIterator reader : newlyReservedReaders) {
         String filename = getReservedReadeFilename(reader);
         InterruptibleIterator iter;
@@ -522,30 +522,30 @@ public class FileManager {
         if (value.isTimeSet()) {
           iter = new TimeSettingIterator(iter, value.getTime());
         }
-        
+
         iters.add(iter);
       }
-      
+
       return iters;
     }
-    
+
     public synchronized void detach() {
-      
+
       releaseReaders(tablet, tabletReservedReaders, false);
       tabletReservedReaders.clear();
-      
+
       for (FileDataSource fds : dataSources)
         fds.unsetIterator();
     }
-    
+
     public synchronized void reattach() throws IOException {
       if (tabletReservedReaders.size() != 0)
         throw new IllegalStateException();
-      
+
       Collection<String> files = new ArrayList<String>();
       for (FileDataSource fds : dataSources)
         files.add(fds.file);
-      
+
       List<FileSKVIterator> newlyReservedReaders = openFiles(files);
       Map<String,List<FileSKVIterator>> map = new HashMap<String,List<FileSKVIterator>>();
       for (FileSKVIterator reader : newlyReservedReaders) {
@@ -555,27 +555,27 @@ public class FileManager {
           list = new LinkedList<FileSKVIterator>();
           map.put(fileName, list);
         }
-        
+
         list.add(reader);
       }
-      
+
       for (FileDataSource fds : dataSources) {
         FileSKVIterator reader = map.get(fds.file).remove(0);
         fds.setIterator(reader);
       }
     }
-    
+
     public synchronized void releaseOpenFiles(boolean sawIOException) {
       releaseReaders(tablet, tabletReservedReaders, sawIOException);
       tabletReservedReaders.clear();
       dataSources.clear();
     }
-    
+
     public synchronized int getNumOpenFiles() {
       return tabletReservedReaders.size();
     }
   }
-  
+
   public ScanFileManager newScanFileManager(KeyExtent tablet) {
     return new ScanFileManager(tablet);
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/server/tserver/src/main/java/org/apache/accumulo/tserver/HoldTimeoutException.java
----------------------------------------------------------------------
diff --git a/server/tserver/src/main/java/org/apache/accumulo/tserver/HoldTimeoutException.java b/server/tserver/src/main/java/org/apache/accumulo/tserver/HoldTimeoutException.java
index 2175851..1bd2c2c 100644
--- a/server/tserver/src/main/java/org/apache/accumulo/tserver/HoldTimeoutException.java
+++ b/server/tserver/src/main/java/org/apache/accumulo/tserver/HoldTimeoutException.java
@@ -18,7 +18,7 @@ package org.apache.accumulo.tserver;
 
 public class HoldTimeoutException extends RuntimeException {
   private static final long serialVersionUID = 1L;
-  
+
   public HoldTimeoutException(String why) {
     super(why);
   }