You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by ct...@apache.org on 2015/01/09 03:44:47 UTC

[43/66] [abbrv] accumulo git commit: ACCUMULO-3451 Format master branch (1.7.0-SNAPSHOT)

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
index 37caf15..77081bf 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/AccumuloRowInputFormat.java
@@ -36,16 +36,16 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
 /**
  * This class allows MapReduce jobs to use Accumulo as the source of data. This {@link InputFormat} provides row names as {@link Text} as keys, and a
  * corresponding {@link PeekingIterator} as a value, which in turn makes the {@link Key}/{@link Value} pairs for that row available to the Map function.
- * 
+ *
  * The user must specify the following via static configurator methods:
- * 
+ *
  * <ul>
  * <li>{@link AccumuloRowInputFormat#setConnectorInfo(Job, String, AuthenticationToken)}
  * <li>{@link AccumuloRowInputFormat#setInputTableName(Job, String)}
  * <li>{@link AccumuloRowInputFormat#setScanAuthorizations(Job, Authorizations)}
  * <li>{@link AccumuloRowInputFormat#setZooKeeperInstance(Job, ClientConfiguration)} OR {@link AccumuloRowInputFormat#setMockInstance(Job, String)}
  * </ul>
- * 
+ *
  * Other static methods are optional.
  */
 public class AccumuloRowInputFormat extends InputFormatBase<Text,PeekingIterator<Entry<Key,Value>>> {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
index e58e350..a60cb80 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputFormatBase.java
@@ -54,7 +54,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the table name from the configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return the table name
@@ -67,7 +67,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param tableName
@@ -80,7 +80,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Sets the input ranges to scan for the single input table associated with this job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param ranges
@@ -93,7 +93,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the ranges to scan over from a job.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return the ranges
@@ -106,7 +106,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Restricts the columns that will be mapped over for this job for the default input table.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param columnFamilyColumnQualifierPairs
@@ -120,7 +120,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets the columns to be mapped over from this job.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return a set of columns
@@ -133,7 +133,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Encode an iterator on the single input table for this job.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param cfg
@@ -146,7 +146,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Gets a list of the iterator settings (for iterators to apply to a scanner) from this configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return a list of iterators
@@ -160,10 +160,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -177,7 +177,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return false if the feature is disabled, true otherwise
@@ -190,10 +190,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Controls the use of the {@link IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -206,7 +206,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -220,10 +220,10 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
   /**
    * Controls the use of the {@link ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack to be constructed within the Map
    * task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -236,7 +236,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -252,26 +252,26 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param job
    *          the Hadoop job instance to be configured
    * @param enableFeature
@@ -284,7 +284,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return true if the feature is enabled, false otherwise
@@ -297,7 +297,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
   /**
    * Initializes an Accumulo {@link org.apache.accumulo.core.client.impl.TabletLocator} based on the configuration.
-   * 
+   *
    * @param context
    *          the Hadoop context for the configured job
    * @return an Accumulo tablet locator
@@ -315,7 +315,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
     /**
      * Apply the configured iterators from the configuration to the scanner for the specified table name
-     * 
+     *
      * @param context
      *          the Hadoop context for the configured job
      * @param scanner
@@ -329,7 +329,7 @@ public abstract class InputFormatBase<K,V> extends AbstractInputFormat<K,V> {
 
     /**
      * Apply the configured iterators from the configuration to the scanner.
-     * 
+     *
      * @param context
      *          the Hadoop context for the configured job
      * @param scanner

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
index fa3b7eb..03473f2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/InputTableConfig.java
@@ -48,7 +48,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Creates a batch scan config object out of a previously serialized batch scan config object.
-   * 
+   *
    * @param input
    *          the data input of the serialized batch scan config
    */
@@ -58,7 +58,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Sets the input ranges to scan for all tables associated with this job. This will be added to any per-table ranges that have been set using
-   * 
+   *
    * @param ranges
    *          the ranges that will be mapped over
    * @since 1.6.0
@@ -77,7 +77,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Restricts the columns that will be mapped over for this job for the default input table.
-   * 
+   *
    * @param columns
    *          a pair of {@link Text} objects corresponding to column family and column qualifier. If the column qualifier is null, the entire column family is
    *          selected. An empty set is the default and is equivalent to scanning the all columns.
@@ -97,7 +97,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Set iterators on to be used in the query.
-   * 
+   *
    * @param iterators
    *          the configurations for the iterators
    * @since 1.6.0
@@ -117,10 +117,10 @@ public class InputTableConfig implements Writable {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param autoAdjustRanges
    *          the feature is enabled if true, disabled otherwise
    * @see #setRanges(java.util.List)
@@ -133,7 +133,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @return false if the feature is disabled, true otherwise
    * @since 1.6.0
    * @see #setAutoAdjustRanges(boolean)
@@ -146,10 +146,10 @@ public class InputTableConfig implements Writable {
    * Controls the use of the {@link org.apache.accumulo.core.client.ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack
    * to be constructed within the Map task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be
    * available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param useLocalIterators
    *          the feature is enabled if true, disabled otherwise
    * @since 1.6.0
@@ -161,7 +161,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @return true if the feature is enabled, false otherwise
    * @since 1.6.0
    * @see #setUseLocalIterators(boolean)
@@ -175,26 +175,26 @@ public class InputTableConfig implements Writable {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath. The accumulo-site.xml may need to be on the mapper's classpath if HDFS or the Accumulo directory in HDFS are non-standard.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param offlineScan
    *          the feature is enabled if true, disabled otherwise
    * @since 1.6.0
@@ -206,7 +206,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @return true if the feature is enabled, false otherwise
    * @since 1.6.0
    * @see #setOfflineScan(boolean)
@@ -217,10 +217,10 @@ public class InputTableConfig implements Writable {
 
   /**
    * Controls the use of the {@link org.apache.accumulo.core.client.IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param useIsolatedScanners
    *          the feature is enabled if true, disabled otherwise
    * @since 1.6.0
@@ -232,7 +232,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @return true if the feature is enabled, false otherwise
    * @since 1.6.0
    * @see #setUseIsolatedScanners(boolean)
@@ -243,7 +243,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Writes the state for the current object out to the specified {@link DataOutput}
-   * 
+   *
    * @param dataOutput
    *          the output for which to write the object's state
    */
@@ -286,7 +286,7 @@ public class InputTableConfig implements Writable {
 
   /**
    * Reads the fields in the {@link DataInput} into the current object
-   * 
+   *
    * @param dataInput
    *          the input fields to read into the current object
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
index 29cf95d..fe27b01 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/RangeInputSplit.java
@@ -33,8 +33,8 @@ import org.apache.accumulo.core.client.ClientConfiguration;
 import org.apache.accumulo.core.client.Instance;
 import org.apache.accumulo.core.client.IteratorSetting;
 import org.apache.accumulo.core.client.ZooKeeperInstance;
-import org.apache.accumulo.core.client.mapreduce.lib.impl.InputConfigurator;
 import org.apache.accumulo.core.client.mapreduce.lib.impl.ConfiguratorBase.TokenSource;
+import org.apache.accumulo.core.client.mapreduce.lib.impl.InputConfigurator;
 import org.apache.accumulo.core.client.mock.MockInstance;
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken;
 import org.apache.accumulo.core.client.security.tokens.AuthenticationToken.AuthenticationTokenSerializer;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
index ae1d46f..b2b5150 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/ConfiguratorBase.java
@@ -22,6 +22,7 @@ import static java.nio.charset.StandardCharsets.UTF_8;
 import java.io.IOException;
 import java.net.URI;
 import java.net.URISyntaxException;
+
 import org.apache.accumulo.core.Constants;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
 import org.apache.accumulo.core.client.ClientConfiguration;
@@ -47,7 +48,7 @@ public class ConfiguratorBase {
 
   /**
    * Configuration keys for {@link Instance#getConnector(String, AuthenticationToken)}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum ConnectorInfo {
@@ -70,7 +71,7 @@ public class ConfiguratorBase {
 
   /**
    * Configuration keys for {@link Instance}, {@link ZooKeeperInstance}, and {@link MockInstance}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum InstanceOpts {
@@ -79,17 +80,16 @@ public class ConfiguratorBase {
 
   /**
    * Configuration keys for general configuration options.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum GeneralOpts {
-    LOG_LEVEL,
-    VISIBILITY_CACHE_SIZE
+    LOG_LEVEL, VISIBILITY_CACHE_SIZE
   }
 
   /**
    * Provides a configuration key for a given feature enum, prefixed by the implementingClass
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param e
@@ -102,23 +102,23 @@ public class ConfiguratorBase {
   }
 
   /**
-  * Provides a configuration key for a given feature enum.
-  * 
-  * @param e
-  *          the enum used to provide the unique part of the configuration key
-  * @return the configuration key
-  */
+   * Provides a configuration key for a given feature enum.
+   *
+   * @param e
+   *          the enum used to provide the unique part of the configuration key
+   * @return the configuration key
+   */
   protected static String enumToConfKey(Enum<?> e) {
-	  return  e.getDeclaringClass().getSimpleName() + "." +  StringUtils.camelize(e.name().toLowerCase());
+    return e.getDeclaringClass().getSimpleName() + "." + StringUtils.camelize(e.name().toLowerCase());
   }
 
   /**
    * Sets the connector information needed to communicate with Accumulo in this job.
-   * 
+   *
    * <p>
    * <b>WARNING:</b> The serialized token is stored in the configuration and shared with all MapReduce tasks. It is BASE64 encoded to provide a charset safe
    * conversion to a string, and is not intended to be secure.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -144,11 +144,11 @@ public class ConfiguratorBase {
 
   /**
    * Sets the connector information needed to communicate with Accumulo in this job.
-   * 
+   *
    * <p>
    * Pulls a token file into the Distributed Cache that contains the authentication token in an attempt to be more secure than storing the password in the
    * Configuration. Token file created with "bin/accumulo create-token".
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -179,7 +179,7 @@ public class ConfiguratorBase {
 
   /**
    * Determines if the connector info has already been set for this instance.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -194,7 +194,7 @@ public class ConfiguratorBase {
 
   /**
    * Gets the user name from the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -209,7 +209,7 @@ public class ConfiguratorBase {
 
   /**
    * Gets the authenticated token from either the specified token file or directly from the configuration, whichever was used when the job was configured.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -237,7 +237,7 @@ public class ConfiguratorBase {
 
   /**
    * Reads from the token file in distributed cache. Currently, the token file stores data separated by colons e.g. principal:token_class:token
-   * 
+   *
    * @param conf
    *          the Hadoop context for the configured job
    * @return path to the token file as a String
@@ -275,7 +275,7 @@ public class ConfiguratorBase {
 
   /**
    * Configures a {@link ZooKeeperInstance} for this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -296,7 +296,7 @@ public class ConfiguratorBase {
 
   /**
    * Configures a {@link MockInstance} for this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -317,7 +317,7 @@ public class ConfiguratorBase {
 
   /**
    * Initializes an Accumulo {@link Instance} based on the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -348,7 +348,7 @@ public class ConfiguratorBase {
 
   /**
    * Sets the log level for this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -365,7 +365,7 @@ public class ConfiguratorBase {
 
   /**
    * Gets the log level from this configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -380,7 +380,7 @@ public class ConfiguratorBase {
 
   /**
    * Sets the valid visibility count for this job.
-   * 
+   *
    * @param conf
    *          the Hadoop configuration object to configure
    * @param visibilityCacheSize
@@ -392,13 +392,13 @@ public class ConfiguratorBase {
 
   /**
    * Gets the valid visibility count for this job.
-   * 
+   *
    * @param conf
    *          the Hadoop configuration object to configure
    * @return the valid visibility count
    */
   public static int getVisibilityCacheSize(Configuration conf) {
-    return conf.getInt(enumToConfKey(GeneralOpts.VISIBILITY_CACHE_SIZE),Constants.DEFAULT_VISIBILITY_CACHE_SIZE);
+    return conf.getInt(enumToConfKey(GeneralOpts.VISIBILITY_CACHE_SIZE), Constants.DEFAULT_VISIBILITY_CACHE_SIZE);
   }
 
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
index ce84209..882c6d3 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/FileOutputConfigurator.java
@@ -31,7 +31,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for {@link AccumuloConfiguration}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum Opts {
@@ -41,7 +41,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * The supported Accumulo properties we set in this OutputFormat, that change the behavior of the RecordWriter.<br />
    * These properties correspond to the supported public static setter methods available to this class.
-   * 
+   *
    * @param property
    *          the Accumulo property to check
    * @since 1.6.0
@@ -61,7 +61,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Helper for transforming Accumulo configuration properties into something that can be stored safely inside the Hadoop Job configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -86,7 +86,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * This helper method provides an AccumuloConfiguration object constructed from the Accumulo defaults, and overridden with Accumulo properties that have been
    * stored in the Job's configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -104,7 +104,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the compression type to use for data blocks. Specifying a compression may require additional libraries to be available to your Job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -122,10 +122,10 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the size for data blocks within each file.<br />
    * Data blocks are a span of key/value pairs stored in the file that are compressed and indexed as a group.
-   * 
+   *
    * <p>
    * Making this value smaller may increase seek performance, but at the cost of increasing the size of the indexes (which can also affect seek performance).
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -140,7 +140,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the size for file blocks in the file system; file blocks are managed, and replicated, by the underlying file system.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -156,7 +156,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the size for index blocks within each file; smaller blocks means a deeper index hierarchy within the file, while larger blocks mean a more shallow
    * index hierarchy within the file. This can affect the performance of queries.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -171,7 +171,7 @@ public class FileOutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the file system replication factor for the resulting file, overriding the file system default.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
index af84bb4..5405ac0 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/InputConfigurator.java
@@ -81,7 +81,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for {@link Scanner}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum ScanOpts {
@@ -90,7 +90,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for various features.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum Features {
@@ -99,7 +99,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -115,7 +115,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the name of the input table, over which this job will scan.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -128,7 +128,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the {@link Authorizations} used to scan. Must be a subset of the user's authorization. Defaults to the empty set.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -144,7 +144,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the authorizations to set for the scans from the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -160,7 +160,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the input ranges to scan on all input tables for this job. If not set, the entire table will be scanned.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -189,7 +189,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the ranges to scan over from a job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -215,7 +215,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets a list of the iterator settings (for iterators to apply to a scanner) from this configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -249,7 +249,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Restricts the columns that will be mapped over for the single input table on this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -286,7 +286,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the columns to be mapped over from this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -326,7 +326,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Encode an iterator on the input for the single input table associated with this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -364,10 +364,10 @@ public class InputConfigurator extends ConfiguratorBase {
   /**
    * Controls the automatic adjustment of ranges for this job. This feature merges overlapping ranges, then splits them to align with tablet boundaries.
    * Disabling this feature will cause exactly one Map task to be created for each specified range. The default setting is enabled. *
-   * 
+   *
    * <p>
    * By default, this feature is <b>enabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -383,7 +383,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration has auto-adjust ranges enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -398,10 +398,10 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Controls the use of the {@link IsolatedScanner} in this job.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -416,7 +416,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration has isolation enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -432,10 +432,10 @@ public class InputConfigurator extends ConfiguratorBase {
   /**
    * Controls the use of the {@link ClientSideIteratorScanner} in this job. Enabling this feature will cause the iterator stack to be constructed within the Map
    * task, rather than within the Accumulo TServer. To use this feature, all classes needed for those iterators must be available on the classpath for the task.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -450,7 +450,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration uses local iterators.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -468,26 +468,26 @@ public class InputConfigurator extends ConfiguratorBase {
    * Enable reading offline tables. By default, this feature is disabled and only online tables are scanned. This will make the map reduce job directly read the
    * table's files. If the table is not offline, then the job will fail. If the table comes online during the map reduce job, it is likely that the job will
    * fail.
-   * 
+   *
    * <p>
    * To use this option, the map reduce user will need access to read the Accumulo directory in HDFS.
-   * 
+   *
    * <p>
    * Reading the offline table will create the scan time iterator stack in the map process. So any iterators that are configured for the table will need to be
    * on the mapper's classpath.
-   * 
+   *
    * <p>
    * One way to use this feature is to clone a table, take the clone offline, and use the clone as the input table for a map reduce job. If you plan to map
    * reduce over the data many times, it may be better to the compact the table, clone it, take it offline, and use the clone for all map reduce jobs. The
    * reason to do this is that compaction will reduce each tablet in the table to one file, and it is faster to read from one file.
-   * 
+   *
    * <p>
    * There are two possible advantages to reading a tables file directly out of HDFS. First, you may see better read performance. Second, it will support
    * speculative execution better. When reading an online table speculative execution can put more load on an already slow tablet server.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -502,7 +502,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether a configuration has the offline table scan feature enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -517,7 +517,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets configurations for multiple tables at a time.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -544,7 +544,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Returns all {@link InputTableConfig} objects associated with this job.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -577,7 +577,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Returns the {@link InputTableConfig} for the given table
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -594,7 +594,7 @@ public class InputConfigurator extends ConfiguratorBase {
 
   /**
    * Initializes an Accumulo {@link TabletLocator} based on the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -619,7 +619,7 @@ public class InputConfigurator extends ConfiguratorBase {
   // InputFormat doesn't have the equivalent of OutputFormat's checkOutputSpecs(JobContext job)
   /**
    * Check whether a configuration is fully configured to be used with an Accumulo {@link org.apache.hadoop.mapreduce.InputFormat}.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -674,7 +674,7 @@ public class InputConfigurator extends ConfiguratorBase {
   /**
    * Returns the {@link org.apache.accumulo.core.client.mapreduce.InputTableConfig} for the configuration based on the properties set using the single-table
    * input methods.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
index 13b67d5..55e980c 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/OutputConfigurator.java
@@ -23,6 +23,7 @@ import java.io.ByteArrayOutputStream;
 import java.io.DataInputStream;
 import java.io.DataOutputStream;
 import java.io.IOException;
+
 import org.apache.accumulo.core.client.BatchWriter;
 import org.apache.accumulo.core.client.BatchWriterConfig;
 import org.apache.hadoop.conf.Configuration;
@@ -34,7 +35,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for {@link BatchWriter}.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum WriteOpts {
@@ -43,7 +44,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Configuration keys for various features.
-   * 
+   *
    * @since 1.6.0
    */
   public static enum Features {
@@ -53,7 +54,7 @@ public class OutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the default table name to use if one emits a null in place of a table name for a given mutation. Table names can only be alpha-numeric and
    * underscores.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -69,7 +70,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the default table name from the configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -85,7 +86,7 @@ public class OutputConfigurator extends ConfiguratorBase {
   /**
    * Sets the configuration for for the job's {@link BatchWriter} instances. If not set, a new {@link BatchWriterConfig}, with sensible built-in defaults is
    * used. Setting the configuration multiple times overwrites any previous configuration.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -109,7 +110,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Gets the {@link BatchWriterConfig} settings.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -137,10 +138,10 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the directive to create new tables, as necessary. Table names can only be alpha-numeric and underscores.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -155,7 +156,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether tables are permitted to be created as needed.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -170,10 +171,10 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Sets the directive to use simulation mode for this job. In simulation mode, no output is produced. This is useful for testing.
-   * 
+   *
    * <p>
    * By default, this feature is <b>disabled</b>.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf
@@ -188,7 +189,7 @@ public class OutputConfigurator extends ConfiguratorBase {
 
   /**
    * Determines whether this feature is enabled.
-   * 
+   *
    * @param implementingClass
    *          the class whose name will be used as a prefix for the property configuration key
    * @param conf

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
index 243160d..34ea7d2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/impl/package-info.java
@@ -21,13 +21,13 @@
  * InputFormats/OutputFormats, so as not to clutter their API with methods that don't match the conventions for that framework. These classes may be useful to
  * input/output plugins for other frameworks, so they can reuse the same configuration options and/or serialize them into a
  * {@link org.apache.hadoop.conf.Configuration} instance in a standard way.
- * 
+ *
  * <p>
  * It is not expected these will change much (except when new features are added), but end users should not use these classes. They should use the static
  * configurators on the {@link org.apache.hadoop.mapreduce.InputFormat} or {@link org.apache.hadoop.mapreduce.OutputFormat} they are configuring, which in turn
  * may use these classes to implement their own static configurators. Once again, these classes are intended for internal use, but may be useful to developers
  * of plugins for other frameworks that read/write to Accumulo.
- * 
+ *
  * @since 1.6.0
  */
 package org.apache.accumulo.core.client.mapreduce.lib.impl;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
index c59841d..bd4857e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mapreduce/lib/partition/KeyRangePartitioner.java
@@ -28,29 +28,29 @@ import org.apache.hadoop.mapreduce.Partitioner;
  */
 public class KeyRangePartitioner extends Partitioner<Key,Writable> implements Configurable {
   private RangePartitioner rp = new RangePartitioner();
-  
+
   @Override
   public int getPartition(Key key, Writable value, int numPartitions) {
     return rp.getPartition(key.getRow(), value, numPartitions);
   }
-  
+
   @Override
   public Configuration getConf() {
     return rp.getConf();
   }
-  
+
   @Override
   public void setConf(Configuration conf) {
     rp.setConf(conf);
   }
-  
+
   /**
    * Sets the hdfs file name to use, containing a newline separated list of Base64 encoded split points that represent ranges for partitioning
    */
   public static void setSplitFile(Job job, String file) {
     RangePartitioner.setSplitFile(job, file);
   }
-  
+
   /**
    * Sets the number of random sub-bins per range
    */

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
index 840db41..d4d4004 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/IteratorAdapter.java
@@ -27,18 +27,18 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
 
 public class IteratorAdapter implements Iterator<Entry<Key,Value>> {
-  
+
   SortedKeyValueIterator<Key,Value> inner;
-  
+
   public IteratorAdapter(SortedKeyValueIterator<Key,Value> inner) {
     this.inner = inner;
   }
-  
+
   @Override
   public boolean hasNext() {
     return inner.hasTop();
   }
-  
+
   @Override
   public Entry<Key,Value> next() {
     try {
@@ -49,7 +49,7 @@ public class IteratorAdapter implements Iterator<Entry<Key,Value>> {
       throw new NoSuchElementException();
     }
   }
-  
+
   @Override
   public void remove() {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
index c55c378..f171889 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockAccumulo.java
@@ -110,7 +110,7 @@ public class MockAccumulo {
 
   public void createTable(String username, String tableName, TimeType timeType, Map<String,String> properties) {
     String namespace = Tables.qualify(tableName).getFirst();
-    HashMap<String, String> props = new HashMap<>(properties);
+    HashMap<String,String> props = new HashMap<>(properties);
 
     if (!namespaceExists(namespace)) {
       return;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
index 6f321ff..bb9f2c8 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchDeleter.java
@@ -35,14 +35,14 @@ import org.apache.accumulo.core.security.ColumnVisibility;
  * <li>There is no waiting for memory to fill before flushing</li>
  * <li>Only one thread is used for writing</li>
  * </ol>
- * 
+ *
  * Otherwise, it behaves as expected.
  */
 public class MockBatchDeleter extends MockBatchScanner implements BatchDeleter {
-  
+
   private final MockAccumulo acc;
   private final String tableName;
-  
+
   /**
    * Create a {@link BatchDeleter} for the specified instance on the specified table where the writer uses the specified {@link Authorizations}.
    */
@@ -51,10 +51,10 @@ public class MockBatchDeleter extends MockBatchScanner implements BatchDeleter {
     this.acc = acc;
     this.tableName = tableName;
   }
-  
+
   @Override
   public void delete() throws MutationsRejectedException, TableNotFoundException {
-    
+
     BatchWriter writer = new MockBatchWriter(acc, tableName);
     try {
       Iterator<Entry<Key,Value>> iter = super.iterator();
@@ -69,5 +69,5 @@ public class MockBatchDeleter extends MockBatchScanner implements BatchDeleter {
       writer.close();
     }
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
index 4512006..4034271 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchScanner.java
@@ -33,22 +33,22 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.commons.collections.iterators.IteratorChain;
 
 public class MockBatchScanner extends MockScannerBase implements BatchScanner {
-  
+
   List<Range> ranges = null;
-  
+
   public MockBatchScanner(MockTable mockTable, Authorizations authorizations) {
     super(mockTable, authorizations);
   }
-  
+
   @Override
   public void setRanges(Collection<Range> ranges) {
     if (ranges == null || ranges.size() == 0) {
       throw new IllegalArgumentException("ranges must be non null and contain at least 1 range");
     }
-    
+
     this.ranges = new ArrayList<Range>(ranges);
   }
-  
+
   @SuppressWarnings("unchecked")
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
@@ -69,7 +69,7 @@ public class MockBatchScanner extends MockScannerBase implements BatchScanner {
     }
     return chain;
   }
-  
+
   @Override
   public void close() {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
index f2c5c85..163587f 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockBatchWriter.java
@@ -23,21 +23,21 @@ import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.data.Mutation;
 
 public class MockBatchWriter implements BatchWriter {
-  
+
   final String tablename;
   final MockAccumulo acu;
-  
+
   MockBatchWriter(MockAccumulo acu, String tablename) {
     this.acu = acu;
     this.tablename = tablename;
   }
-  
+
   @Override
   public void addMutation(Mutation m) throws MutationsRejectedException {
     checkArgument(m != null, "m is null");
     acu.addMutation(tablename, m);
   }
-  
+
   @Override
   public void addMutations(Iterable<Mutation> iterable) throws MutationsRejectedException {
     checkArgument(iterable != null, "iterable is null");
@@ -45,11 +45,11 @@ public class MockBatchWriter implements BatchWriter {
       acu.addMutation(tablename, m);
     }
   }
-  
+
   @Override
   public void flush() throws MutationsRejectedException {}
-  
+
   @Override
   public void close() throws MutationsRejectedException {}
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
index ce262a2..8c57c5e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConfiguration.java
@@ -24,15 +24,15 @@ import org.apache.accumulo.core.conf.Property;
 
 class MockConfiguration extends AccumuloConfiguration {
   Map<String,String> map;
-  
+
   MockConfiguration(Map<String,String> settings) {
     map = settings;
   }
-  
+
   public void put(String k, String v) {
     map.put(k, v);
   }
-  
+
   @Override
   public String get(Property property) {
     return map.get(property.getKey());

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
index 8613602..4d32093 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockConnector.java
@@ -31,9 +31,9 @@ import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.Scanner;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.admin.InstanceOperations;
+import org.apache.accumulo.core.client.admin.NamespaceOperations;
 import org.apache.accumulo.core.client.admin.ReplicationOperations;
 import org.apache.accumulo.core.client.admin.SecurityOperations;
-import org.apache.accumulo.core.client.admin.NamespaceOperations;
 import org.apache.accumulo.core.client.admin.TableOperations;
 import org.apache.accumulo.core.client.impl.thrift.SecurityErrorCode;
 import org.apache.accumulo.core.client.security.tokens.NullToken;
@@ -41,15 +41,15 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.Credentials;
 
 public class MockConnector extends Connector {
-  
+
   String username;
   private final MockAccumulo acu;
   private final Instance instance;
-  
+
   MockConnector(String username, MockInstance instance) throws AccumuloSecurityException {
     this(new Credentials(username, new NullToken()), new MockAccumulo(MockInstance.getDefaultFileSystem()), instance);
   }
-  
+
   MockConnector(Credentials credentials, MockAccumulo acu, MockInstance instance) throws AccumuloSecurityException {
     if (credentials.getToken().isDestroyed())
       throw new AccumuloSecurityException(credentials.getPrincipal(), SecurityErrorCode.TOKEN_EXPIRED);
@@ -57,14 +57,14 @@ public class MockConnector extends Connector {
     this.acu = acu;
     this.instance = instance;
   }
-  
+
   @Override
   public BatchScanner createBatchScanner(String tableName, Authorizations authorizations, int numQueryThreads) throws TableNotFoundException {
     if (acu.tables.get(tableName) == null)
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return acu.createBatchScanner(tableName, authorizations);
   }
-  
+
   @Deprecated
   @Override
   public BatchDeleter createBatchDeleter(String tableName, Authorizations authorizations, int numQueryThreads, long maxMemory, long maxLatency,
@@ -73,14 +73,14 @@ public class MockConnector extends Connector {
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return new MockBatchDeleter(acu, tableName, authorizations);
   }
-  
+
   @Override
   public BatchDeleter createBatchDeleter(String tableName, Authorizations authorizations, int numQueryThreads, BatchWriterConfig config)
       throws TableNotFoundException {
     return createBatchDeleter(tableName, authorizations, numQueryThreads, config.getMaxMemory(), config.getMaxLatency(TimeUnit.MILLISECONDS),
         config.getMaxWriteThreads());
   }
-  
+
   @Deprecated
   @Override
   public BatchWriter createBatchWriter(String tableName, long maxMemory, long maxLatency, int maxWriteThreads) throws TableNotFoundException {
@@ -88,23 +88,23 @@ public class MockConnector extends Connector {
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return new MockBatchWriter(acu, tableName);
   }
-  
+
   @Override
   public BatchWriter createBatchWriter(String tableName, BatchWriterConfig config) throws TableNotFoundException {
     return createBatchWriter(tableName, config.getMaxMemory(), config.getMaxLatency(TimeUnit.MILLISECONDS), config.getMaxWriteThreads());
   }
-  
+
   @Deprecated
   @Override
   public MultiTableBatchWriter createMultiTableBatchWriter(long maxMemory, long maxLatency, int maxWriteThreads) {
     return new MockMultiTableBatchWriter(acu);
   }
-  
+
   @Override
   public MultiTableBatchWriter createMultiTableBatchWriter(BatchWriterConfig config) {
     return createMultiTableBatchWriter(config.getMaxMemory(), config.getMaxLatency(TimeUnit.MILLISECONDS), config.getMaxWriteThreads());
   }
-  
+
   @Override
   public Scanner createScanner(String tableName, Authorizations authorizations) throws TableNotFoundException {
     MockTable table = acu.tables.get(tableName);
@@ -112,27 +112,27 @@ public class MockConnector extends Connector {
       throw new TableNotFoundException(tableName, tableName, "no such table");
     return new MockScanner(table, authorizations);
   }
-  
+
   @Override
   public Instance getInstance() {
     return instance;
   }
-  
+
   @Override
   public String whoami() {
     return username;
   }
-  
+
   @Override
   public TableOperations tableOperations() {
     return new MockTableOperations(acu, username);
   }
-  
+
   @Override
   public SecurityOperations securityOperations() {
     return new MockSecurityOperations(acu);
   }
-  
+
   @Override
   public InstanceOperations instanceOperations() {
     return new MockInstanceOperations(acu);
@@ -142,7 +142,7 @@ public class MockConnector extends Connector {
   public NamespaceOperations namespaceOperations() {
     return new MockNamespaceOperations(acu, username);
   }
-  
+
   @Override
   public ConditionalWriter createConditionalWriter(String tableName, ConditionalWriterConfig config) throws TableNotFoundException {
     // TODO add implementation
@@ -154,5 +154,5 @@ public class MockConnector extends Connector {
     // TODO add implementation
     throw new UnsupportedOperationException();
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
index 9b07d49..67435d2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstance.java
@@ -42,11 +42,11 @@ import org.apache.hadoop.io.Text;
  * Mock Accumulo provides an in memory implementation of the Accumulo client API. It is possible that the behavior of this implementation may differ subtly from
  * the behavior of Accumulo. This could result in unit tests that pass on Mock Accumulo and fail on Accumulo or visa-versa. Documenting the differences would be
  * difficult and is not done.
- * 
+ *
  * <p>
  * An alternative to Mock Accumulo called MiniAccumuloCluster was introduced in Accumulo 1.5. MiniAccumuloCluster spins up actual Accumulo server processes, can
  * be used for unit testing, and its behavior should match Accumulo. The drawback of MiniAccumuloCluster is that it starts more slowly than Mock Accumulo.
- * 
+ *
  */
 
 public class MockInstance implements Instance {

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
index 87359bc..48122b7 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockInstanceOperations.java
@@ -71,7 +71,7 @@ class MockInstanceOperations implements InstanceOperations {
     try {
       AccumuloVFSClassLoader.loadClass(className, Class.forName(asTypeName));
     } catch (ClassNotFoundException e) {
-      log.warn("Could not find class named '"+className+"' in testClassLoad.", e);
+      log.warn("Could not find class named '" + className + "' in testClassLoad.", e);
       return false;
     }
     return true;
@@ -88,6 +88,5 @@ class MockInstanceOperations implements InstanceOperations {
   }
 
   @Override
-  public void waitForBalance() throws AccumuloException {
-  }
+  public void waitForBalance() throws AccumuloException {}
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
index b4a7068..9cc3dfb 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockMultiTableBatchWriter.java
@@ -29,12 +29,12 @@ import org.apache.accumulo.core.client.TableNotFoundException;
 public class MockMultiTableBatchWriter implements MultiTableBatchWriter {
   MockAccumulo acu = null;
   Map<String,MockBatchWriter> bws = null;
-  
+
   public MockMultiTableBatchWriter(MockAccumulo acu) {
     this.acu = acu;
     bws = new HashMap<String,MockBatchWriter>();
   }
-  
+
   @Override
   public BatchWriter getBatchWriter(String table) throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     if (!bws.containsKey(table)) {
@@ -42,13 +42,13 @@ public class MockMultiTableBatchWriter implements MultiTableBatchWriter {
     }
     return bws.get(table);
   }
-  
+
   @Override
   public void flush() throws MutationsRejectedException {}
-  
+
   @Override
   public void close() throws MutationsRejectedException {}
-  
+
   @Override
   public boolean isClosed() {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
index 7e7eecb..ac581ab 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockNamespaceOperations.java
@@ -36,7 +36,7 @@ import org.apache.log4j.Logger;
 class MockNamespaceOperations extends NamespaceOperationsHelper {
 
   private static final Logger log = Logger.getLogger(MockNamespaceOperations.class);
-  
+
   final private MockAccumulo acu;
   final private String username;
 
@@ -125,7 +125,7 @@ class MockNamespaceOperations extends NamespaceOperationsHelper {
     try {
       AccumuloVFSClassLoader.loadClass(className, Class.forName(asTypeName));
     } catch (ClassNotFoundException e) {
-      log.warn("Could not load class '"+className+"' with type name '"+asTypeName+"' in testClassLoad()", e);
+      log.warn("Could not load class '" + className + "' with type name '" + asTypeName + "' in testClassLoad()", e);
       return false;
     }
     return true;

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
index e7c0ee0..a9b6fd5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScanner.java
@@ -31,14 +31,14 @@ import org.apache.accumulo.core.iterators.SortedMapIterator;
 import org.apache.accumulo.core.security.Authorizations;
 
 public class MockScanner extends MockScannerBase implements Scanner {
-  
+
   int batchSize = 0;
   Range range = new Range();
-  
+
   MockScanner(MockTable table, Authorizations auths) {
     super(table, auths);
   }
-  
+
   @Deprecated
   @Override
   public void setTimeOut(int timeOut) {
@@ -47,7 +47,7 @@ public class MockScanner extends MockScannerBase implements Scanner {
     else
       setTimeout(timeOut, TimeUnit.SECONDS);
   }
-  
+
   @Deprecated
   @Override
   public int getTimeOut() {
@@ -56,47 +56,47 @@ public class MockScanner extends MockScannerBase implements Scanner {
       return Integer.MAX_VALUE;
     return (int) timeout;
   }
-  
+
   @Override
   public void setRange(Range range) {
     this.range = range;
   }
-  
+
   @Override
   public Range getRange() {
     return this.range;
   }
-  
+
   @Override
   public void setBatchSize(int size) {
     this.batchSize = size;
   }
-  
+
   @Override
   public int getBatchSize() {
     return this.batchSize;
   }
-  
+
   @Override
   public void enableIsolation() {}
-  
+
   @Override
   public void disableIsolation() {}
-  
+
   static class RangeFilter extends Filter {
     Range range;
-    
+
     RangeFilter(SortedKeyValueIterator<Key,Value> i, Range range) {
       setSource(i);
       this.range = range;
     }
-    
+
     @Override
     public boolean accept(Key k, Value v) {
       return range.contains(k);
     }
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     SortedKeyValueIterator<Key,Value> i = new SortedMapIterator(table.table);
@@ -107,7 +107,7 @@ public class MockScanner extends MockScannerBase implements Scanner {
     } catch (IOException e) {
       throw new RuntimeException(e);
     }
-    
+
   }
 
   @Override
@@ -117,7 +117,7 @@ public class MockScanner extends MockScannerBase implements Scanner {
 
   @Override
   public void setReadaheadThreshold(long batches) {
-    
+
   }
-  
+
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
index 72cb863..d88c30a 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockScannerBase.java
@@ -44,15 +44,15 @@ import org.apache.accumulo.core.security.Authorizations;
 import org.apache.commons.lang.NotImplementedException;
 
 public class MockScannerBase extends ScannerOptions implements ScannerBase {
-  
+
   protected final MockTable table;
   protected final Authorizations auths;
-  
+
   MockScannerBase(MockTable mockTable, Authorizations authorizations) {
     this.table = mockTable;
     this.auths = authorizations;
   }
-  
+
   static HashSet<ByteSequence> createColumnBSS(Collection<Column> columns) {
     HashSet<ByteSequence> columnSet = new HashSet<ByteSequence>();
     for (Column c : columns) {
@@ -60,35 +60,35 @@ public class MockScannerBase extends ScannerOptions implements ScannerBase {
     }
     return columnSet;
   }
-  
+
   static class MockIteratorEnvironment implements IteratorEnvironment {
     @Override
     public SortedKeyValueIterator<Key,Value> reserveMapFileReader(String mapFileName) throws IOException {
       throw new NotImplementedException();
     }
-    
+
     @Override
     public AccumuloConfiguration getConfig() {
       return AccumuloConfiguration.getDefaultConfiguration();
     }
-    
+
     @Override
     public IteratorScope getIteratorScope() {
       return IteratorScope.scan;
     }
-    
+
     @Override
     public boolean isFullMajorCompaction() {
       return false;
     }
-    
+
     private ArrayList<SortedKeyValueIterator<Key,Value>> topLevelIterators = new ArrayList<SortedKeyValueIterator<Key,Value>>();
-    
+
     @Override
     public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
       topLevelIterators.add(iter);
     }
-    
+
     SortedKeyValueIterator<Key,Value> getTopLevelIterator(SortedKeyValueIterator<Key,Value> iter) {
       if (topLevelIterators.isEmpty())
         return iter;
@@ -97,7 +97,7 @@ public class MockScannerBase extends ScannerOptions implements ScannerBase {
       return new MultiIterator(allIters, false);
     }
   }
-  
+
   public SortedKeyValueIterator<Key,Value> createFilter(SortedKeyValueIterator<Key,Value> inner) throws IOException {
     byte[] defaultLabels = {};
     inner = new ColumnFamilySkippingIterator(new DeletingIterator(inner, false));
@@ -109,7 +109,7 @@ public class MockScannerBase extends ScannerOptions implements ScannerBase {
         serverSideIteratorList, serverSideIteratorOptions, iterEnv, false));
     return result;
   }
-  
+
   @Override
   public Iterator<Entry<Key,Value>> iterator() {
     throw new UnsupportedOperationException();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
index ee9244b..2244d20 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockTable.java
@@ -41,30 +41,30 @@ import org.apache.accumulo.core.security.TablePermission;
 import org.apache.hadoop.io.Text;
 
 public class MockTable {
-  
+
   static class MockMemKey extends Key {
     private int count;
-    
+
     MockMemKey(Key key, int count) {
       super(key);
       this.count = count;
     }
-    
+
     @Override
     public int hashCode() {
       return super.hashCode() + count;
     }
-    
+
     @Override
     public boolean equals(Object other) {
       return (other instanceof MockMemKey) && super.equals((MockMemKey) other) && count == ((MockMemKey) other).count;
     }
-    
+
     @Override
     public String toString() {
       return super.toString() + " count=" + count;
     }
-    
+
     @Override
     public int compareTo(Key o) {
       int compare = super.compareTo(o);
@@ -82,7 +82,7 @@ public class MockTable {
       return 0;
     }
   };
-  
+
   final SortedMap<Key,Value> table = new ConcurrentSkipListMap<Key,Value>();
   int mutationCount = 0;
   final Map<String,String> settings;
@@ -93,7 +93,7 @@ public class MockTable {
   private MockNamespace namespace;
   private String namespaceName;
   private String tableId;
-  
+
   MockTable(boolean limitVersion, TimeType timeType, String tableId) {
     this.timeType = timeType;
     this.tableId = tableId;
@@ -160,27 +160,27 @@ public class MockTable {
           key.setTimestamp(mutationCount);
         else
           key.setTimestamp(now);
-      
+
       table.put(new MockMemKey(key, mutationCount), new Value(u.getValue()));
     }
   }
-  
+
   public void addSplits(SortedSet<Text> partitionKeys) {
     splits.addAll(partitionKeys);
   }
-  
+
   public Collection<Text> getSplits() {
     return splits;
   }
-  
+
   public void setLocalityGroups(Map<String,Set<Text>> groups) {
     localityGroups = groups;
   }
-  
+
   public Map<String,Set<Text>> getLocalityGroups() {
     return localityGroups;
   }
-  
+
   public void merge(Text start, Text end) {
     boolean reAdd = false;
     if (splits.contains(start))
@@ -189,19 +189,19 @@ public class MockTable {
     if (reAdd)
       splits.add(start);
   }
-  
+
   public void setNamespaceName(String n) {
     this.namespaceName = n;
   }
-  
+
   public void setNamespace(MockNamespace n) {
     this.namespace = n;
   }
-  
+
   public String getNamespaceName() {
     return this.namespaceName;
   }
-  
+
   public MockNamespace getNamespace() {
     return this.namespace;
   }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java b/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
index b39791d..efc896e 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/mock/MockUser.java
@@ -27,7 +27,7 @@ public class MockUser {
   final String name;
   AuthenticationToken token;
   Authorizations authorizations;
-  
+
   MockUser(String principal, AuthenticationToken token, Authorizations auths) {
     this.name = principal;
     this.token = token.clone();

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
index 900ae5a..bdcc652 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystem.java
@@ -28,10 +28,15 @@ public interface ReplicaSystem {
 
   /**
    * Replicate the given status to the target peer
-   * @param p Path to the resource we're reading from
-   * @param status Information to replicate
-   * @param target The peer
-   * @param helper Instance of ReplicaSystemHelper
+   *
+   * @param p
+   *          Path to the resource we're reading from
+   * @param status
+   *          Information to replicate
+   * @param target
+   *          The peer
+   * @param helper
+   *          Instance of ReplicaSystemHelper
    * @return A new Status for the progress that was made
    */
   public Status replicate(Path p, Status status, ReplicationTarget target, ReplicaSystemHelper helper);
@@ -39,8 +44,7 @@ public interface ReplicaSystem {
   /**
    * Configure the implementation with necessary information from the system configuration
    * <p>
-   * For example, we only need one implementation for Accumulo, but, for each peer,
-   * we have a ZK quorum and instance name
+   * For example, we only need one implementation for Accumulo, but, for each peer, we have a ZK quorum and instance name
    */
   public void configure(String configuration);
 }

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
index e721278..d76b3d8 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/replication/ReplicaSystemFactory.java
@@ -22,7 +22,7 @@ import org.slf4j.LoggerFactory;
 import com.google.common.base.Preconditions;
 
 /**
- * 
+ *
  */
 public class ReplicaSystemFactory {
   private static final Logger log = LoggerFactory.getLogger(ReplicaSystemFactory.class);
@@ -64,7 +64,7 @@ public class ReplicaSystemFactory {
 
   /**
    * Generate the configuration value for a {@link ReplicaSystem} in the instance properties
-   * 
+   *
    * @param system
    *          The desired ReplicaSystem to use
    * @param configuration

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java b/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
index 30c60d5..b4027b2 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/security/SecurityErrorCode.java
@@ -17,7 +17,7 @@
 package org.apache.accumulo.core.client.security;
 
 /**
- * 
+ *
  */
 public enum SecurityErrorCode {
   DEFAULT_SECURITY_ERROR,

http://git-wip-us.apache.org/repos/asf/accumulo/blob/6bc67602/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
----------------------------------------------------------------------
diff --git a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
index 99cc721..7836ea5 100644
--- a/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
+++ b/core/src/main/java/org/apache/accumulo/core/client/security/tokens/AuthenticationToken.java
@@ -33,15 +33,15 @@ import javax.security.auth.Destroyable;
 import org.apache.hadoop.io.Writable;
 
 /**
- * 
+ *
  * @since 1.5.0
  */
 public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
-  
+
   /**
    * A utility class to serialize/deserialize {@link AuthenticationToken} objects.<br/>
    * Unfortunately, these methods are provided in an inner-class, to avoid breaking the interface API.
-   * 
+   *
    * @since 1.6.0
    */
   public static final class AuthenticationTokenSerializer {
@@ -49,7 +49,7 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
      * A convenience method to create tokens from serialized bytes, created by {@link #serialize(AuthenticationToken)}
      * <p>
      * The specified tokenType will be instantiated, and used to deserialize the decoded bytes. The resulting object will then be returned to the caller.
-     * 
+     *
      * @param tokenType
      *          the token class to use to deserialize the bytes
      * @param tokenBytes
@@ -78,10 +78,10 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       }
       return type;
     }
-    
+
     /**
      * An alternate version of {@link #deserialize(Class, byte[])} that accepts a token class name rather than a token class.
-     * 
+     *
      * @param tokenClassName
      *          the fully-qualified class name to be returned
      * @see #serialize(AuthenticationToken)
@@ -97,12 +97,12 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       }
       return deserialize(tokenType, tokenBytes);
     }
-    
+
     /**
      * A convenience method to serialize tokens.
      * <p>
      * The provided {@link AuthenticationToken} will be serialized to bytes by its own implementation and returned to the caller.
-     * 
+     *
      * @param token
      *          the token to serialize
      * @return a serialized representation of the provided {@link AuthenticationToken}
@@ -125,17 +125,17 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       return bytes;
     }
   }
-  
+
   class Properties implements Destroyable, Map<String,char[]> {
-    
+
     private boolean destroyed = false;
     private HashMap<String,char[]> map = new HashMap<String,char[]>();
-    
+
     private void checkDestroyed() {
       if (destroyed)
         throw new IllegalStateException();
     }
-    
+
     public char[] put(String key, CharSequence value) {
       checkDestroyed();
       char[] toPut = new char[value.length()];
@@ -143,14 +143,14 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
         toPut[i] = value.charAt(i);
       return map.put(key, toPut);
     }
-    
+
     public void putAllStrings(Map<String,? extends CharSequence> map) {
       checkDestroyed();
       for (Map.Entry<String,? extends CharSequence> entry : map.entrySet()) {
         put(entry.getKey(), entry.getValue());
       }
     }
-    
+
     @Override
     public void destroy() throws DestroyFailedException {
       for (String key : this.keySet()) {
@@ -160,133 +160,133 @@ public interface AuthenticationToken extends Writable, Destroyable, Cloneable {
       this.clear();
       destroyed = true;
     }
-    
+
     @Override
     public boolean isDestroyed() {
       return destroyed;
     }
-    
+
     @Override
     public int size() {
       checkDestroyed();
       return map.size();
     }
-    
+
     @Override
     public boolean isEmpty() {
       checkDestroyed();
       return map.isEmpty();
     }
-    
+
     @Override
     public boolean containsKey(Object key) {
       checkDestroyed();
       return map.containsKey(key);
     }
-    
+
     @Override
     public boolean containsValue(Object value) {
       checkDestroyed();
       return map.containsValue(value);
     }
-    
+
     @Override
     public char[] get(Object key) {
       checkDestroyed();
       return map.get(key);
     }
-    
+
     @Override
     public char[] put(String key, char[] value) {
       checkDestroyed();
       return map.put(key, value);
     }
-    
+
     @Override
     public char[] remove(Object key) {
       checkDestroyed();
       return map.remove(key);
     }
-    
+
     @Override
     public void putAll(Map<? extends String,? extends char[]> m) {
       checkDestroyed();
       map.putAll(m);
     }
-    
+
     @Override
     public void clear() {
       checkDestroyed();
       map.clear();
     }
-    
+
     @Override
     public Set<String> keySet() {
       checkDestroyed();
       return map.keySet();
     }
-    
+
     @Override
     public Collection<char[]> values() {
       checkDestroyed();
       return map.values();
     }
-    
+
     @Override
     public Set<Map.Entry<String,char[]>> entrySet() {
       checkDestroyed();
       return map.entrySet();
     }
   }
-  
+
   static class TokenProperty implements Comparable<TokenProperty> {
     private String key, description;
     private boolean masked;
-    
+
     public TokenProperty(String name, String description, boolean mask) {
       this.key = name;
       this.description = description;
       this.masked = mask;
     }
-    
+
     @Override
     public String toString() {
       return this.key + " - " + description;
     }
-    
+
     public String getKey() {
       return this.key;
     }
-    
+
     public String getDescription() {
       return this.description;
     }
-    
+
     public boolean getMask() {
       return this.masked;
     }
-    
+
     @Override
     public int hashCode() {
       return key.hashCode();
     }
-    
+
     @Override
     public boolean equals(Object o) {
       if (o instanceof TokenProperty)
         return ((TokenProperty) o).key.equals(key);
       return false;
     }
-    
+
     @Override
     public int compareTo(TokenProperty o) {
       return key.compareTo(o.key);
     }
   }
-  
+
   public void init(Properties properties);
-  
+
   public Set<TokenProperty> getProperties();
-  
+
   public AuthenticationToken clone();
 }