You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by ct...@apache.org on 2019/03/07 23:04:01 UTC

[accumulo-testing] branch master updated (8f6c24c -> 48d255d)

This is an automated email from the ASF dual-hosted git repository.

ctubbsii pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/accumulo-testing.git.


    from 8f6c24c  Creates continuous BulkIngest (#63)
     new 31ee5bd  Update pom.xml
     new 48d255d  Apply formatter changes on build

The 2 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 pom.xml                                            | 314 +++++++++++++--------
 .../java/org/apache/accumulo/testing/TestEnv.java  |  14 +-
 .../apache/accumulo/testing/cli/ClientOpts.java    |   4 +-
 .../accumulo/testing/continuous/BulkIngest.java    |  20 +-
 .../testing/continuous/ContinuousIngest.java       |   4 +-
 .../testing/continuous/ContinuousOpts.java         |   4 +-
 .../testing/continuous/ContinuousScanner.java      |   2 +-
 .../testing/continuous/ContinuousVerify.java       |   8 +-
 .../accumulo/testing/continuous/CreateTable.java   |   8 +-
 .../accumulo/testing/continuous/TimeBinner.java    |   8 +-
 .../testing/continuous/UndefinedAnalyzer.java      |  23 +-
 .../testing/ingest/BulkImportDirectory.java        |  12 +-
 .../apache/accumulo/testing/ingest/TestIngest.java |  33 ++-
 .../accumulo/testing/ingest/VerifyIngest.java      |  31 +-
 .../apache/accumulo/testing/mapreduce/RowHash.java |  13 +-
 .../accumulo/testing/mapreduce/TeraSortIngest.java |  14 +-
 .../apache/accumulo/testing/merkle/MerkleTree.java |   5 +-
 .../accumulo/testing/merkle/MerkleTreeNode.java    |   8 +-
 .../accumulo/testing/merkle/cli/CompareTables.java |  16 +-
 .../testing/merkle/cli/ComputeRootHash.java        |   4 +-
 .../testing/merkle/cli/GenerateHashes.java         |  17 +-
 .../testing/merkle/cli/ManualComparison.java       |   2 +-
 .../accumulo/testing/merkle/package-info.java      |  33 ++-
 .../accumulo/testing/performance/Report.java       |   8 +-
 .../accumulo/testing/performance/impl/Csv.java     |  25 +-
 .../testing/performance/impl/MergeSiteConfig.java  |   9 +-
 .../testing/performance/impl/PerfTestRunner.java   |   4 +-
 .../performance/tests/ProbabilityFilter.java       |   4 +-
 .../performance/tests/RandomCachedLookupsPT.java   |  21 +-
 .../testing/performance/tests/ScanExecutorPT.java  |  12 +-
 .../performance/tests/ScanFewFamiliesPT.java       |  13 +-
 .../performance/tests/YieldingScanExecutorPT.java  |  17 +-
 .../testing/performance/util/TestData.java         |   4 +-
 .../testing/performance/util/TestExecutor.java     |  12 +-
 .../apache/accumulo/testing/randomwalk/Module.java |   8 +-
 .../accumulo/testing/randomwalk/RandWalkEnv.java   |   4 +-
 .../testing/randomwalk/bulk/BulkPlusOne.java       |   6 +-
 .../accumulo/testing/randomwalk/bulk/Compact.java  |   4 +-
 .../testing/randomwalk/bulk/ConsistencyCheck.java  |   4 +-
 .../accumulo/testing/randomwalk/bulk/Merge.java    |   4 +-
 .../accumulo/testing/randomwalk/bulk/Split.java    |   4 +-
 .../accumulo/testing/randomwalk/bulk/Verify.java   |   7 +-
 .../testing/randomwalk/concurrent/BulkImport.java  |   8 +-
 .../concurrent/ChangeAuthorizations.java           |   4 +-
 .../testing/randomwalk/concurrent/CloneTable.java  |   3 +-
 .../randomwalk/concurrent/ConcurrentFixture.java   |   2 +-
 .../testing/randomwalk/concurrent/Config.java      |  28 +-
 .../randomwalk/concurrent/IsolatedScan.java        |   4 +-
 .../testing/randomwalk/concurrent/Replication.java |   4 +-
 .../testing/randomwalk/conditional/Setup.java      |   4 +-
 .../testing/randomwalk/conditional/Transfer.java   |   4 +-
 .../testing/randomwalk/image/ImageFixture.java     |   2 +-
 .../testing/randomwalk/image/ScanMeta.java         |   4 +-
 .../accumulo/testing/randomwalk/image/Write.java   |   4 +-
 .../testing/randomwalk/multitable/CopyTool.java    |   4 +-
 .../randomwalk/multitable/MultiTableFixture.java   |   2 +-
 .../testing/randomwalk/security/AlterTable.java    |  13 +-
 .../randomwalk/security/AlterTablePerm.java        |   8 +-
 .../testing/randomwalk/security/Authenticate.java  |   7 +-
 .../testing/randomwalk/security/ChangePass.java    |   8 +-
 .../testing/randomwalk/security/CreateUser.java    |   8 +-
 .../testing/randomwalk/security/DropUser.java      |   4 +-
 .../randomwalk/security/SecurityFixture.java       |   2 +-
 .../testing/randomwalk/security/TableOp.java       |  32 +--
 .../testing/randomwalk/security/Validate.java      |   4 +-
 .../randomwalk/security/WalkingSecurity.java       |   4 +-
 .../testing/randomwalk/sequential/BatchVerify.java |  11 +-
 .../randomwalk/sequential/MapRedVerifyTool.java    |   8 +-
 .../randomwalk/sequential/SequentialFixture.java   |   2 +-
 .../testing/randomwalk/shard/BulkInsert.java       |  13 +-
 .../testing/randomwalk/shard/CloneIndex.java       |   6 +-
 .../testing/randomwalk/shard/CompactFilter.java    |  12 +-
 .../testing/randomwalk/shard/ExportIndex.java      |  24 +-
 .../accumulo/testing/randomwalk/shard/Grep.java    |  10 +-
 .../accumulo/testing/randomwalk/shard/Merge.java   |   4 +-
 .../testing/randomwalk/shard/ShardFixture.java     |  10 +-
 .../testing/randomwalk/shard/VerifyIndex.java      |   4 +-
 .../apache/accumulo/testing/scalability/Run.java   |   5 +-
 .../accumulo/testing/scalability/ScaleTest.java    |   4 +-
 .../apache/accumulo/testing/stress/ScanOpts.java   |   3 +-
 .../org/apache/accumulo/testing/stress/Write.java  |   6 +-
 .../accumulo/testing/stress/WriteOptions.java      |  15 +-
 .../accumulo/testing/stress/package-info.java      |  28 +-
 .../accumulo/testing/randomwalk/FrameworkTest.java |   8 +-
 84 files changed, 615 insertions(+), 492 deletions(-)


[accumulo-testing] 02/02: Apply formatter changes on build

Posted by ct...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

ctubbsii pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/accumulo-testing.git

commit 48d255d7be9c8ba4f10a212751128954f77f7b9b
Author: Christopher Tubbs <ct...@apache.org>
AuthorDate: Thu Mar 7 18:03:40 2019 -0500

    Apply formatter changes on build
---
 .../java/org/apache/accumulo/testing/TestEnv.java  | 14 ++++-----
 .../apache/accumulo/testing/cli/ClientOpts.java    |  4 +--
 .../accumulo/testing/continuous/BulkIngest.java    | 20 ++++++-------
 .../testing/continuous/ContinuousIngest.java       |  4 +--
 .../testing/continuous/ContinuousOpts.java         |  4 +--
 .../testing/continuous/ContinuousScanner.java      |  2 +-
 .../testing/continuous/ContinuousVerify.java       |  8 +++---
 .../accumulo/testing/continuous/CreateTable.java   |  8 +++---
 .../accumulo/testing/continuous/TimeBinner.java    |  8 +++++-
 .../testing/continuous/UndefinedAnalyzer.java      | 23 +++++++++------
 .../testing/ingest/BulkImportDirectory.java        | 12 ++++----
 .../apache/accumulo/testing/ingest/TestIngest.java | 33 +++++++++++-----------
 .../accumulo/testing/ingest/VerifyIngest.java      | 31 ++++++++++----------
 .../apache/accumulo/testing/mapreduce/RowHash.java | 13 ++++-----
 .../accumulo/testing/mapreduce/TeraSortIngest.java | 14 ++++-----
 .../apache/accumulo/testing/merkle/MerkleTree.java |  5 ++--
 .../accumulo/testing/merkle/MerkleTreeNode.java    |  8 +++---
 .../accumulo/testing/merkle/cli/CompareTables.java | 16 +++++------
 .../testing/merkle/cli/ComputeRootHash.java        |  4 +--
 .../testing/merkle/cli/GenerateHashes.java         | 17 ++++++-----
 .../testing/merkle/cli/ManualComparison.java       |  2 +-
 .../accumulo/testing/merkle/package-info.java      | 33 ++++++++++++----------
 .../accumulo/testing/performance/Report.java       |  8 +++---
 .../accumulo/testing/performance/impl/Csv.java     | 25 +++++++++-------
 .../testing/performance/impl/MergeSiteConfig.java  |  9 +++---
 .../testing/performance/impl/PerfTestRunner.java   |  4 +--
 .../performance/tests/ProbabilityFilter.java       |  4 +--
 .../performance/tests/RandomCachedLookupsPT.java   | 21 ++++++++------
 .../testing/performance/tests/ScanExecutorPT.java  | 12 ++++----
 .../performance/tests/ScanFewFamiliesPT.java       | 13 +++++----
 .../performance/tests/YieldingScanExecutorPT.java  | 17 +++++------
 .../testing/performance/util/TestData.java         |  4 +--
 .../testing/performance/util/TestExecutor.java     | 12 ++++----
 .../apache/accumulo/testing/randomwalk/Module.java |  8 +++---
 .../accumulo/testing/randomwalk/RandWalkEnv.java   |  4 +--
 .../testing/randomwalk/bulk/BulkPlusOne.java       |  6 ++--
 .../accumulo/testing/randomwalk/bulk/Compact.java  |  4 +--
 .../testing/randomwalk/bulk/ConsistencyCheck.java  |  4 +--
 .../accumulo/testing/randomwalk/bulk/Merge.java    |  4 +--
 .../accumulo/testing/randomwalk/bulk/Split.java    |  4 +--
 .../accumulo/testing/randomwalk/bulk/Verify.java   |  7 +++--
 .../testing/randomwalk/concurrent/BulkImport.java  |  8 +++---
 .../concurrent/ChangeAuthorizations.java           |  4 +--
 .../testing/randomwalk/concurrent/CloneTable.java  |  3 +-
 .../randomwalk/concurrent/ConcurrentFixture.java   |  2 +-
 .../testing/randomwalk/concurrent/Config.java      | 28 +++++++++---------
 .../randomwalk/concurrent/IsolatedScan.java        |  4 +--
 .../testing/randomwalk/concurrent/Replication.java |  4 +--
 .../testing/randomwalk/conditional/Setup.java      |  4 +--
 .../testing/randomwalk/conditional/Transfer.java   |  4 +--
 .../testing/randomwalk/image/ImageFixture.java     |  2 +-
 .../testing/randomwalk/image/ScanMeta.java         |  4 +--
 .../accumulo/testing/randomwalk/image/Write.java   |  4 +--
 .../testing/randomwalk/multitable/CopyTool.java    |  4 +--
 .../randomwalk/multitable/MultiTableFixture.java   |  2 +-
 .../testing/randomwalk/security/AlterTable.java    | 13 +++++----
 .../randomwalk/security/AlterTablePerm.java        |  8 +++---
 .../testing/randomwalk/security/Authenticate.java  |  7 ++---
 .../testing/randomwalk/security/ChangePass.java    |  8 +++---
 .../testing/randomwalk/security/CreateUser.java    |  8 +++---
 .../testing/randomwalk/security/DropUser.java      |  4 +--
 .../randomwalk/security/SecurityFixture.java       |  2 +-
 .../testing/randomwalk/security/TableOp.java       | 32 ++++++++++-----------
 .../testing/randomwalk/security/Validate.java      |  4 +--
 .../randomwalk/security/WalkingSecurity.java       |  4 +--
 .../testing/randomwalk/sequential/BatchVerify.java | 11 ++++----
 .../randomwalk/sequential/MapRedVerifyTool.java    |  8 +++---
 .../randomwalk/sequential/SequentialFixture.java   |  2 +-
 .../testing/randomwalk/shard/BulkInsert.java       | 13 +++++----
 .../testing/randomwalk/shard/CloneIndex.java       |  6 ++--
 .../testing/randomwalk/shard/CompactFilter.java    | 12 ++++----
 .../testing/randomwalk/shard/ExportIndex.java      | 24 ++++++++--------
 .../accumulo/testing/randomwalk/shard/Grep.java    | 10 +++----
 .../accumulo/testing/randomwalk/shard/Merge.java   |  4 +--
 .../testing/randomwalk/shard/ShardFixture.java     | 10 +++----
 .../testing/randomwalk/shard/VerifyIndex.java      |  4 +--
 .../apache/accumulo/testing/scalability/Run.java   |  5 ++--
 .../accumulo/testing/scalability/ScaleTest.java    |  4 +--
 .../apache/accumulo/testing/stress/ScanOpts.java   |  3 +-
 .../org/apache/accumulo/testing/stress/Write.java  |  6 ++--
 .../accumulo/testing/stress/WriteOptions.java      | 15 +++++-----
 .../accumulo/testing/stress/package-info.java      | 28 ++++++++++--------
 .../accumulo/testing/randomwalk/FrameworkTest.java |  8 +++---
 83 files changed, 414 insertions(+), 379 deletions(-)

diff --git a/src/main/java/org/apache/accumulo/testing/TestEnv.java b/src/main/java/org/apache/accumulo/testing/TestEnv.java
index d36fa1a..74c0dd3 100644
--- a/src/main/java/org/apache/accumulo/testing/TestEnv.java
+++ b/src/main/java/org/apache/accumulo/testing/TestEnv.java
@@ -23,20 +23,20 @@ public class TestEnv implements AutoCloseable {
 
   public TestEnv(String[] args) {
 
-    Map<String, String> options = new HashMap<>();
+    Map<String,String> options = new HashMap<>();
     List<String> arguments = new ArrayList<>();
 
     for (int i = 0; i < args.length; i++) {
-      if(args[i].equals("-o")) {
+      if (args[i].equals("-o")) {
         i++;
-        String[] tokens = args[i].split("=",2);
+        String[] tokens = args[i].split("=", 2);
         options.put(tokens[0], tokens[1]);
       } else {
         arguments.add(args[i]);
       }
     }
 
-    if(arguments.size() != 2) {
+    if (arguments.size() != 2) {
       throw new IllegalArgumentException("Expected <testPropsPath> <clientPropsPath> arguments.");
     }
 
@@ -46,7 +46,7 @@ public class TestEnv implements AutoCloseable {
     this.testProps = TestProps.loadFromFile(testPropsPath);
     this.clientProps = Accumulo.newClientProperties().from(clientPropsPath).build();
 
-    options.forEach((k,v) -> testProps.setProperty(k, v));
+    options.forEach((k, v) -> testProps.setProperty(k, v));
   }
 
   public TestEnv(String testPropsPath, String clientPropsPath) {
@@ -118,8 +118,8 @@ public class TestEnv implements AutoCloseable {
       hadoopConfig.set("fs.defaultFS", getHdfsRoot());
       // Below is required due to bundled jar breaking default config.
       // See http://stackoverflow.com/questions/17265002/hadoop-no-filesystem-for-scheme-file
-      hadoopConfig
-          .set("fs.hdfs.impl", org.apache.hadoop.hdfs.DistributedFileSystem.class.getName());
+      hadoopConfig.set("fs.hdfs.impl",
+          org.apache.hadoop.hdfs.DistributedFileSystem.class.getName());
       hadoopConfig.set("fs.file.impl", org.apache.hadoop.fs.LocalFileSystem.class.getName());
       hadoopConfig.set("mapreduce.framework.name", "yarn");
       hadoopConfig.set("yarn.resourcemanager.hostname", getYarnResourceManager());
diff --git a/src/main/java/org/apache/accumulo/testing/cli/ClientOpts.java b/src/main/java/org/apache/accumulo/testing/cli/ClientOpts.java
index 6643d44..2689b97 100644
--- a/src/main/java/org/apache/accumulo/testing/cli/ClientOpts.java
+++ b/src/main/java/org/apache/accumulo/testing/cli/ClientOpts.java
@@ -160,8 +160,8 @@ public class ClientOpts extends Help {
 
   public String getClientConfigFile() {
     if (clientConfigFile == null) {
-      URL clientPropsUrl = ClientOpts.class.getClassLoader().getResource(
-          "accumulo-client.properties");
+      URL clientPropsUrl = ClientOpts.class.getClassLoader()
+          .getResource("accumulo-client.properties");
       if (clientPropsUrl != null) {
         clientConfigFile = clientPropsUrl.getFile();
       }
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/BulkIngest.java b/src/main/java/org/apache/accumulo/testing/continuous/BulkIngest.java
index 35ff688..4b2e20b 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/BulkIngest.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/BulkIngest.java
@@ -41,8 +41,8 @@ import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
 /**
- * Bulk import a million random key value pairs. Same format as ContinuousIngest and can be
- * verified by running ContinuousVerify.
+ * Bulk import a million random key value pairs. Same format as ContinuousIngest and can be verified
+ * by running ContinuousVerify.
  */
 public class BulkIngest extends Configured implements Tool {
   public static final int NUM_KEYS = 1_000_000;
@@ -88,8 +88,8 @@ public class BulkIngest extends Configured implements Tool {
       try (AccumuloClient client = env.getAccumuloClient()) {
 
         // make sure splits file is closed before continuing
-        try (PrintStream out = new PrintStream(new BufferedOutputStream(fs.create(new Path(
-            splitsFile))))) {
+        try (PrintStream out = new PrintStream(
+            new BufferedOutputStream(fs.create(new Path(splitsFile))))) {
           Collection<Text> splits = client.tableOperations().listSplits(tableName, 100);
           for (Text split : splits) {
             out.println(Base64.getEncoder().encodeToString(split.copyBytes()));
@@ -151,8 +151,8 @@ public class BulkIngest extends Configured implements Tool {
     }
 
     @Override
-    protected void map(LongWritable key, LongWritable value, Context context) throws IOException,
-        InterruptedException {
+    protected void map(LongWritable key, LongWritable value, Context context)
+        throws IOException, InterruptedException {
       currentRow.set(ContinuousIngest.genRow(key.get()));
 
       // hack since we can't pass null - don't set first val (prevRow), we want it to be null
@@ -162,16 +162,16 @@ public class BulkIngest extends Configured implements Tool {
       }
 
       Key outputKey = new Key(currentRow, emptyCfCq, emptyCfCq);
-      Value outputValue = ContinuousIngest.createValue(uuid.getBytes(), 0,
-          currentValue.copyBytes(), null);
+      Value outputValue = ContinuousIngest.createValue(uuid.getBytes(), 0, currentValue.copyBytes(),
+          null);
 
       context.write(outputKey, outputValue);
     }
   }
 
   /**
-   * Generates a million LongWritable keys.  The LongWritable value points to the previous key.
-   * The first key value pair has a value of 1L.  This is translated to null in RandomMapper
+   * Generates a million LongWritable keys. The LongWritable value points to the previous key. The
+   * first key value pair has a value of 1L. This is translated to null in RandomMapper
    */
   public static class RandomInputFormat extends InputFormat {
 
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousIngest.java b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousIngest.java
index 957ce3e..0499b09 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousIngest.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousIngest.java
@@ -136,8 +136,8 @@ public class ContinuousIngest {
 
       byte[] ingestInstanceId = UUID.randomUUID().toString().getBytes(UTF_8);
 
-      log.info(String.format("UUID %d %s", System.currentTimeMillis(), new String(ingestInstanceId,
-          UTF_8)));
+      log.info(String.format("UUID %d %s", System.currentTimeMillis(),
+          new String(ingestInstanceId, UTF_8)));
 
       long count = 0;
       final int flushInterval = getFlushEntries(env.getTestProperties());
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousOpts.java b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousOpts.java
index cda9f4d..847942b 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousOpts.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousOpts.java
@@ -38,8 +38,8 @@ public class ContinuousOpts {
       logger.setLevel(Level.TRACE);
       logger.setAdditivity(false);
       try {
-        logger.addAppender(new FileAppender(new PatternLayout(
-            "%d{dd HH:mm:ss,SSS} [%-8c{2}] %-5p: %m%n"), debugLog, true));
+        logger.addAppender(new FileAppender(
+            new PatternLayout("%d{dd HH:mm:ss,SSS} [%-8c{2}] %-5p: %m%n"), debugLog, true));
       } catch (IOException ex) {
         throw new RuntimeException(ex);
       }
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousScanner.java b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousScanner.java
index 9fd4a80..d453ec7 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousScanner.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousScanner.java
@@ -88,7 +88,7 @@ public class ContinuousScanner {
           }
 
           // System.out.println("P2 "+delta
-          // +" "+numToScan+" "+distance+"  "+((double)numToScan/count ));
+          // +" "+numToScan+" "+distance+" "+((double)numToScan/count ));
         }
 
         System.out.printf("SCN %d %s %d %d%n", t1, new String(scanStart, UTF_8), (t2 - t1), count);
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousVerify.java b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousVerify.java
index 7604dda..880643d 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/ContinuousVerify.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/ContinuousVerify.java
@@ -145,14 +145,14 @@ public class ContinuousVerify extends Configured implements Tool {
 
     String tableName = env.getAccumuloTableName();
 
-    Job job = Job.getInstance(getConf(), this.getClass().getSimpleName() + "_" + tableName + "_"
-        + System.currentTimeMillis());
+    Job job = Job.getInstance(getConf(),
+        this.getClass().getSimpleName() + "_" + tableName + "_" + System.currentTimeMillis());
     job.setJarByClass(this.getClass());
 
     job.setInputFormatClass(AccumuloInputFormat.class);
 
-    boolean scanOffline = Boolean.parseBoolean(env
-        .getTestProperty(TestProps.CI_VERIFY_SCAN_OFFLINE));
+    boolean scanOffline = Boolean
+        .parseBoolean(env.getTestProperty(TestProps.CI_VERIFY_SCAN_OFFLINE));
     int maxMaps = Integer.parseInt(env.getTestProperty(TestProps.CI_VERIFY_MAX_MAPS));
     int reducers = Integer.parseInt(env.getTestProperty(TestProps.CI_VERIFY_REDUCERS));
     String outputDir = env.getTestProperty(TestProps.CI_VERIFY_OUTPUT_DIR);
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/CreateTable.java b/src/main/java/org/apache/accumulo/testing/continuous/CreateTable.java
index 9373144..fc3440a 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/CreateTable.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/CreateTable.java
@@ -39,8 +39,8 @@ public class CreateTable {
         System.exit(-1);
       }
 
-      int numTablets = Integer.parseInt(env
-          .getTestProperty(TestProps.CI_COMMON_ACCUMULO_NUM_TABLETS));
+      int numTablets = Integer
+          .parseInt(env.getTestProperty(TestProps.CI_COMMON_ACCUMULO_NUM_TABLETS));
 
       if (numTablets < 1) {
         System.err.println("ERROR: numTablets < 1");
@@ -70,8 +70,8 @@ public class CreateTable {
 
       client.tableOperations().create(tableName, ntc);
 
-      System.out.println("Created Accumulo table '" + tableName + "' with " + numTablets
-          + " tablets");
+      System.out
+          .println("Created Accumulo table '" + tableName + "' with " + numTablets + " tablets");
     }
   }
 
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/TimeBinner.java b/src/main/java/org/apache/accumulo/testing/continuous/TimeBinner.java
index 48b009f..83e42aa 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/TimeBinner.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/TimeBinner.java
@@ -35,7 +35,13 @@ import com.beust.jcommander.Parameter;
 public class TimeBinner {
 
   enum Operation {
-    AVG, SUM, MIN, MAX, COUNT, CUMULATIVE, AMM, // avg,min,max
+    AVG,
+    SUM,
+    MIN,
+    MAX,
+    COUNT,
+    CUMULATIVE,
+    AMM, // avg,min,max
     AMM_HACK1 // special case
   }
 
diff --git a/src/main/java/org/apache/accumulo/testing/continuous/UndefinedAnalyzer.java b/src/main/java/org/apache/accumulo/testing/continuous/UndefinedAnalyzer.java
index 6920a7a..1c61bdb 100644
--- a/src/main/java/org/apache/accumulo/testing/continuous/UndefinedAnalyzer.java
+++ b/src/main/java/org/apache/accumulo/testing/continuous/UndefinedAnalyzer.java
@@ -16,6 +16,8 @@
  */
 package org.apache.accumulo.testing.continuous;
 
+import static java.nio.charset.StandardCharsets.UTF_8;
+
 import java.io.BufferedReader;
 import java.io.File;
 import java.io.InputStreamReader;
@@ -31,7 +33,6 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.TreeMap;
 
-import com.beust.jcommander.Parameter;
 import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.BatchScanner;
 import org.apache.accumulo.core.data.Key;
@@ -40,7 +41,7 @@ import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.testing.cli.ClientOpts;
 import org.apache.hadoop.io.Text;
 
-import static java.nio.charset.StandardCharsets.UTF_8;
+import com.beust.jcommander.Parameter;
 
 /**
  * BUGS This code does not handle the fact that these files could include log events from previous
@@ -197,14 +198,16 @@ public class UndefinedAnalyzer {
 
               if (pos1 > 0 && pos2 > 0 && pos3 == -1) {
                 String tid = tablet.substring(0, pos1);
-                String endRow = tablet.charAt(pos1) == '<' ? "8000000000000000" : tablet.substring(pos1 + 1, pos2);
+                String endRow = tablet.charAt(pos1) == '<' ? "8000000000000000"
+                    : tablet.substring(pos1 + 1, pos2);
                 String prevEndRow = tablet.charAt(pos2) == '<' ? "" : tablet.substring(pos2 + 1);
                 if (tid.equals(tableId)) {
                   // System.out.println(" "+server+" "+tid+" "+endRow+" "+prevEndRow);
                   Date date = sdf.parse(day + " " + time);
                   // System.out.println(" "+date);
 
-                  assignments.add(new TabletAssignment(tablet, endRow, prevEndRow, server, date.getTime()));
+                  assignments.add(
+                      new TabletAssignment(tablet, endRow, prevEndRow, server, date.getTime()));
 
                 }
               } else if (!tablet.startsWith("!0")) {
@@ -262,7 +265,7 @@ public class UndefinedAnalyzer {
     }
 
     try (AccumuloClient client = opts.createClient();
-         BatchScanner bscanner = client.createBatchScanner(opts.tableName, opts.auths)) {
+        BatchScanner bscanner = client.createBatchScanner(opts.tableName, opts.auths)) {
       List<Range> refs = new ArrayList<>();
 
       for (UndefinedNode undefinedNode : undefs)
@@ -270,9 +273,9 @@ public class UndefinedAnalyzer {
 
       bscanner.setRanges(refs);
 
-      HashMap<String, List<String>> refInfo = new HashMap<>();
+      HashMap<String,List<String>> refInfo = new HashMap<>();
 
-      for (Entry<Key, Value> entry : bscanner) {
+      for (Entry<Key,Value> entry : bscanner) {
         String ref = entry.getKey().getRow().toString();
         List<String> vals = refInfo.computeIfAbsent(ref, k -> new ArrayList<>());
         vals.add(entry.getValue().toString());
@@ -313,9 +316,11 @@ public class UndefinedAnalyzer {
             }
 
             if (ta == null)
-              System.out.println(undefinedNode.undef + " " + undefinedNode.ref + " " + uuid + " " + t1 + " " + t2);
+              System.out.println(
+                  undefinedNode.undef + " " + undefinedNode.ref + " " + uuid + " " + t1 + " " + t2);
             else
-              System.out.println(undefinedNode.undef + " " + undefinedNode.ref + " " + ta.tablet + " " + ta.server + " " + uuid + " " + t1 + " " + t2);
+              System.out.println(undefinedNode.undef + " " + undefinedNode.ref + " " + ta.tablet
+                  + " " + ta.server + " " + uuid + " " + t1 + " " + t2);
 
           }
         } else {
diff --git a/src/main/java/org/apache/accumulo/testing/ingest/BulkImportDirectory.java b/src/main/java/org/apache/accumulo/testing/ingest/BulkImportDirectory.java
index 1cd0a39..a75364c 100644
--- a/src/main/java/org/apache/accumulo/testing/ingest/BulkImportDirectory.java
+++ b/src/main/java/org/apache/accumulo/testing/ingest/BulkImportDirectory.java
@@ -39,16 +39,16 @@ public class BulkImportDirectory {
     String failures = null;
   }
 
-  public static void main(String[] args) throws IOException, AccumuloException,
-      AccumuloSecurityException, TableNotFoundException {
+  public static void main(String[] args)
+      throws IOException, AccumuloException, AccumuloSecurityException, TableNotFoundException {
     final FileSystem fs = FileSystem.get(new Configuration());
     Opts opts = new Opts();
-    System.err
-        .println("Deprecated syntax for BulkImportDirectory, please use the new style (see --help)");
+    System.err.println(
+        "Deprecated syntax for BulkImportDirectory, please use the new style (see --help)");
     opts.parseArgs(BulkImportDirectory.class.getName(), args);
     fs.delete(new Path(opts.failures), true);
     fs.mkdirs(new Path(opts.failures));
-    opts.createClient().tableOperations()
-        .importDirectory(opts.tableName, opts.source, opts.failures, false);
+    opts.createClient().tableOperations().importDirectory(opts.tableName, opts.source,
+        opts.failures, false);
   }
 }
diff --git a/src/main/java/org/apache/accumulo/testing/ingest/TestIngest.java b/src/main/java/org/apache/accumulo/testing/ingest/TestIngest.java
index a7d392c..a478659 100644
--- a/src/main/java/org/apache/accumulo/testing/ingest/TestIngest.java
+++ b/src/main/java/org/apache/accumulo/testing/ingest/TestIngest.java
@@ -31,9 +31,9 @@ import org.apache.accumulo.core.client.BatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
-import org.apache.accumulo.core.client.security.SecurityErrorCode;
 import org.apache.accumulo.core.client.rfile.RFile;
 import org.apache.accumulo.core.client.rfile.RFileWriter;
+import org.apache.accumulo.core.client.security.SecurityErrorCode;
 import org.apache.accumulo.core.clientImpl.TabletServerBatchWriter;
 import org.apache.accumulo.core.data.ConstraintViolationSummary;
 import org.apache.accumulo.core.data.Key;
@@ -76,8 +76,7 @@ public class TestIngest {
     @Parameter(names = "--cols", description = "the number of columns to ingest per row")
     public int cols = 1;
 
-    @Parameter(
-        names = "--random",
+    @Parameter(names = "--random",
         description = "insert random rows and use the given number to seed the psuedo-random number generator")
     public Integer random = null;
 
@@ -108,8 +107,8 @@ public class TestIngest {
     public FileSystem fs = null;
   }
 
-  public static void createTable(AccumuloClient client, Opts args) throws AccumuloException,
-      AccumuloSecurityException, TableExistsException {
+  public static void createTable(AccumuloClient client, Opts args)
+      throws AccumuloException, AccumuloSecurityException, TableExistsException {
     if (args.createTable) {
       TreeSet<Text> splits = getSplitPoints(args.startRow, args.startRow + args.rows,
           args.numsplits);
@@ -256,8 +255,8 @@ public class TestIngest {
           } else {
             byte value[];
             if (opts.random != null) {
-              value = genRandomValue(random, randomValue, opts.random.intValue(), rowid
-                  + opts.startRow, j);
+              value = genRandomValue(random, randomValue, opts.random.intValue(),
+                  rowid + opts.startRow, j);
             } else {
               value = bytevals[j % bytevals.length];
             }
@@ -279,8 +278,8 @@ public class TestIngest {
           } else {
             byte value[];
             if (opts.random != null) {
-              value = genRandomValue(random, randomValue, opts.random.intValue(), rowid
-                  + opts.startRow, j);
+              value = genRandomValue(random, randomValue, opts.random.intValue(),
+                  rowid + opts.startRow, j);
             } else {
               value = bytevals[j % bytevals.length];
             }
@@ -305,9 +304,10 @@ public class TestIngest {
         bw.close();
       } catch (MutationsRejectedException e) {
         if (e.getSecurityErrorCodes().size() > 0) {
-          for (Entry<TabletId,Set<SecurityErrorCode>> entry : e.getSecurityErrorCodes().entrySet()) {
-            System.err.println("ERROR : Not authorized to write to : " + entry.getKey()
-                + " due to " + entry.getValue());
+          for (Entry<TabletId,Set<SecurityErrorCode>> entry : e.getSecurityErrorCodes()
+              .entrySet()) {
+            System.err.println("ERROR : Not authorized to write to : " + entry.getKey() + " due to "
+                + entry.getValue());
           }
         }
 
@@ -325,11 +325,10 @@ public class TestIngest {
     int totalValues = opts.rows * opts.cols;
     double elapsed = (stopTime - startTime) / 1000.0;
 
-    System.out
-        .printf(
-            "%,12d records written | %,8d records/sec | %,12d bytes written | %,8d bytes/sec | %6.3f secs   %n",
-            totalValues, (int) (totalValues / elapsed), bytesWritten,
-            (int) (bytesWritten / elapsed), elapsed);
+    System.out.printf(
+        "%,12d records written | %,8d records/sec | %,12d bytes written | %,8d bytes/sec | %6.3f secs   %n",
+        totalValues, (int) (totalValues / elapsed), bytesWritten, (int) (bytesWritten / elapsed),
+        elapsed);
   }
 
   public static void ingest(AccumuloClient c, Opts opts, Configuration conf)
diff --git a/src/main/java/org/apache/accumulo/testing/ingest/VerifyIngest.java b/src/main/java/org/apache/accumulo/testing/ingest/VerifyIngest.java
index db7e828..9b01bc1 100644
--- a/src/main/java/org/apache/accumulo/testing/ingest/VerifyIngest.java
+++ b/src/main/java/org/apache/accumulo/testing/ingest/VerifyIngest.java
@@ -75,8 +75,8 @@ public class VerifyIngest {
     }
   }
 
-  private static void verifyIngest(AccumuloClient client, Opts opts) throws AccumuloException,
-      AccumuloSecurityException, TableNotFoundException {
+  private static void verifyIngest(AccumuloClient client, Opts opts)
+      throws AccumuloException, AccumuloSecurityException, TableNotFoundException {
     byte[][] bytevals = TestIngest.generateValues(opts.dataSize);
 
     Authorizations labelAuths = new Authorizations("L1", "L2", "G1", "GROUP2");
@@ -119,8 +119,8 @@ public class VerifyIngest {
 
         byte ev[];
         if (opts.random != null) {
-          ev = TestIngest
-              .genRandomValue(random, randomValue, opts.random, expectedRow, expectedCol);
+          ev = TestIngest.genRandomValue(random, randomValue, opts.random, expectedRow,
+              expectedCol);
         } else {
           ev = bytevals[expectedCol % bytevals.length];
         }
@@ -175,23 +175,23 @@ public class VerifyIngest {
           }
 
           if (colNum != expectedCol) {
-            log.error("colNum != expectedCol  " + colNum + " != " + expectedCol + "  rowNum : "
-                + rowNum);
+            log.error(
+                "colNum != expectedCol  " + colNum + " != " + expectedCol + "  rowNum : " + rowNum);
             errors++;
           }
 
           if (expectedRow >= (opts.rows + opts.startRow)) {
-            log.error("expectedRow (" + expectedRow
-                + ") >= (ingestArgs.rows + ingestArgs.startRow)  (" + (opts.rows + opts.startRow)
-                + "), get batch returned data passed end key");
+            log.error(
+                "expectedRow (" + expectedRow + ") >= (ingestArgs.rows + ingestArgs.startRow)  ("
+                    + (opts.rows + opts.startRow) + "), get batch returned data passed end key");
             errors++;
             break;
           }
 
           byte value[];
           if (opts.random != null) {
-            value = TestIngest
-                .genRandomValue(random, randomValue, opts.random, expectedRow, colNum);
+            value = TestIngest.genRandomValue(random, randomValue, opts.random, expectedRow,
+                colNum);
           } else {
             value = bytevals[colNum % bytevals.length];
           }
@@ -235,11 +235,10 @@ public class VerifyIngest {
       throw new AccumuloException("Did not read expected number of rows. Saw "
           + (expectedRow - opts.startRow) + " expected " + opts.rows);
     } else {
-      System.out
-          .printf(
-              "%,12d records read | %,8d records/sec | %,12d bytes read | %,8d bytes/sec | %6.3f secs   %n",
-              recsRead, (int) ((recsRead) / ((t2 - t1) / 1000.0)), bytesRead,
-              (int) (bytesRead / ((t2 - t1) / 1000.0)), (t2 - t1) / 1000.0);
+      System.out.printf(
+          "%,12d records read | %,8d records/sec | %,12d bytes read | %,8d bytes/sec | %6.3f secs   %n",
+          recsRead, (int) ((recsRead) / ((t2 - t1) / 1000.0)), bytesRead,
+          (int) (bytesRead / ((t2 - t1) / 1000.0)), (t2 - t1) / 1000.0);
     }
   }
 
diff --git a/src/main/java/org/apache/accumulo/testing/mapreduce/RowHash.java b/src/main/java/org/apache/accumulo/testing/mapreduce/RowHash.java
index 086025c..f3a7328 100644
--- a/src/main/java/org/apache/accumulo/testing/mapreduce/RowHash.java
+++ b/src/main/java/org/apache/accumulo/testing/mapreduce/RowHash.java
@@ -20,17 +20,14 @@ import java.io.IOException;
 import java.util.Base64;
 import java.util.Collection;
 import java.util.Collections;
-import java.util.List;
-import java.util.Set;
 
 import org.apache.accumulo.core.client.IteratorSetting;
-import org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat;
-import org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat;
-import org.apache.accumulo.hadoopImpl.mapreduce.lib.MapReduceClientOnRequiredTable;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
-import org.apache.accumulo.core.util.Pair;
+import org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat;
+import org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat;
+import org.apache.accumulo.hadoopImpl.mapreduce.lib.MapReduceClientOnRequiredTable;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.conf.Configured;
 import org.apache.hadoop.io.MD5Hash;
@@ -50,8 +47,8 @@ public class RowHash extends Configured implements Tool {
     @Override
     public void map(Key row, Value data, Context context) throws IOException, InterruptedException {
       Mutation m = new Mutation(row.getRow());
-      m.put(new Text("cf-HASHTYPE"), new Text("cq-MD5BASE64"), new Value(Base64.getEncoder()
-          .encode(MD5Hash.digest(data.toString()).getDigest())));
+      m.put(new Text("cf-HASHTYPE"), new Text("cq-MD5BASE64"),
+          new Value(Base64.getEncoder().encode(MD5Hash.digest(data.toString()).getDigest())));
       context.write(null, m);
       context.progress();
     }
diff --git a/src/main/java/org/apache/accumulo/testing/mapreduce/TeraSortIngest.java b/src/main/java/org/apache/accumulo/testing/mapreduce/TeraSortIngest.java
index b4ae57c..e2fa4f5 100644
--- a/src/main/java/org/apache/accumulo/testing/mapreduce/TeraSortIngest.java
+++ b/src/main/java/org/apache/accumulo/testing/mapreduce/TeraSortIngest.java
@@ -27,9 +27,9 @@ import java.util.Random;
 
 import org.apache.accumulo.core.client.BatchWriterConfig;
 import org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat;
-import org.apache.accumulo.hadoopImpl.mapreduce.lib.MapReduceClientOnRequiredTable;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
+import org.apache.accumulo.hadoopImpl.mapreduce.lib.MapReduceClientOnRequiredTable;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.conf.Configured;
 import org.apache.hadoop.io.LongWritable;
@@ -141,8 +141,8 @@ public class TeraSortIngest extends Configured implements Tool {
       }
 
       @Override
-      public void initialize(InputSplit split, TaskAttemptContext context) throws IOException,
-          InterruptedException {}
+      public void initialize(InputSplit split, TaskAttemptContext context)
+          throws IOException, InterruptedException {}
 
       @Override
       public boolean nextKeyValue() throws IOException, InterruptedException {
@@ -169,8 +169,8 @@ public class TeraSortIngest extends Configured implements Tool {
       long totalRows = job.getConfiguration().getLong(NUMROWS, 0);
       int numSplits = job.getConfiguration().getInt(NUMSPLITS, 1);
       long rowsPerSplit = totalRows / numSplits;
-      System.out.println("Generating " + totalRows + " using " + numSplits + " maps with step of "
-          + rowsPerSplit);
+      System.out.println(
+          "Generating " + totalRows + " using " + numSplits + " maps with step of " + rowsPerSplit);
       ArrayList<InputSplit> splits = new ArrayList<>(numSplits);
       long currentRow = 0;
       for (int split = 0; split < numSplits - 1; ++split) {
@@ -318,8 +318,8 @@ public class TeraSortIngest extends Configured implements Tool {
     }
 
     @Override
-    public void map(LongWritable row, NullWritable ignored, Context context) throws IOException,
-        InterruptedException {
+    public void map(LongWritable row, NullWritable ignored, Context context)
+        throws IOException, InterruptedException {
       context.setStatus("Entering");
       long rowId = row.get();
       if (rand == null) {
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/MerkleTree.java b/src/main/java/org/apache/accumulo/testing/merkle/MerkleTree.java
index bf384b8..cc7f810 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/MerkleTree.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/MerkleTree.java
@@ -46,8 +46,9 @@ public class MerkleTree {
       Pair<Integer,Integer> pairToJoin = findNextPair(buffer);
 
       // Make a parent node from them
-      MerkleTreeNode parent = new MerkleTreeNode(Arrays.asList(buffer.get(pairToJoin.getFirst()),
-          buffer.get(pairToJoin.getSecond())), digestAlgorithm);
+      MerkleTreeNode parent = new MerkleTreeNode(
+          Arrays.asList(buffer.get(pairToJoin.getFirst()), buffer.get(pairToJoin.getSecond())),
+          digestAlgorithm);
 
       // Insert it back into the "tree" at the position of the first child
       buffer.set(pairToJoin.getFirst(), parent);
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/MerkleTreeNode.java b/src/main/java/org/apache/accumulo/testing/merkle/MerkleTreeNode.java
index a83e13c..52b3684 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/MerkleTreeNode.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/MerkleTreeNode.java
@@ -71,12 +71,12 @@ public class MerkleTreeNode {
       if (null == childrenRange) {
         childrenRange = child.getRange();
       } else {
-        List<Range> overlappingRanges = Range.mergeOverlapping(Arrays.asList(childrenRange,
-            child.getRange()));
+        List<Range> overlappingRanges = Range
+            .mergeOverlapping(Arrays.asList(childrenRange, child.getRange()));
         if (1 != overlappingRanges.size()) {
           log.error("Tried to merge non-contiguous ranges: {} {}", childrenRange, child.getRange());
-          throw new IllegalArgumentException("Ranges must be contiguous: " + childrenRange + ", "
-              + child.getRange());
+          throw new IllegalArgumentException(
+              "Ranges must be contiguous: " + childrenRange + ", " + child.getRange());
         }
 
         childrenRange = overlappingRanges.get(0);
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/cli/CompareTables.java b/src/main/java/org/apache/accumulo/testing/merkle/cli/CompareTables.java
index 70c1469..3d1111e 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/cli/CompareTables.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/cli/CompareTables.java
@@ -24,9 +24,9 @@ import java.util.List;
 import java.util.Map;
 import java.util.Map.Entry;
 
+import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.AccumuloException;
 import org.apache.accumulo.core.client.AccumuloSecurityException;
-import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.TableExistsException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.data.Range;
@@ -95,9 +95,9 @@ public class CompareTables {
     this.opts = opts;
   }
 
-  private Map<String,String> computeAllHashes() throws AccumuloException,
-      AccumuloSecurityException, TableExistsException, NoSuchAlgorithmException,
-      TableNotFoundException, FileNotFoundException {
+  private Map<String,String> computeAllHashes()
+      throws AccumuloException, AccumuloSecurityException, TableExistsException,
+      NoSuchAlgorithmException, TableNotFoundException, FileNotFoundException {
     try (AccumuloClient client = opts.createClient()) {
       final Map<String,String> hashesByTable = new HashMap<>();
 
@@ -105,8 +105,8 @@ public class CompareTables {
         final String outputTableName = table + "_merkle";
 
         if (client.tableOperations().exists(outputTableName)) {
-          throw new IllegalArgumentException("Expected output table name to not yet exist: "
-              + outputTableName);
+          throw new IllegalArgumentException(
+              "Expected output table name to not yet exist: " + outputTableName);
         }
 
         client.tableOperations().create(outputTableName);
@@ -123,8 +123,8 @@ public class CompareTables {
         }
 
         ComputeRootHash computeRootHash = new ComputeRootHash();
-        String hash = Hex.encodeHexString(computeRootHash.getHash(client, outputTableName,
-            opts.getHashName()));
+        String hash = Hex
+            .encodeHexString(computeRootHash.getHash(client, outputTableName, opts.getHashName()));
 
         hashesByTable.put(table, hash);
       }
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/cli/ComputeRootHash.java b/src/main/java/org/apache/accumulo/testing/merkle/cli/ComputeRootHash.java
index c08bd3d..7d06a6e 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/cli/ComputeRootHash.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/cli/ComputeRootHash.java
@@ -51,8 +51,8 @@ public class ComputeRootHash {
     String hashName;
   }
 
-  private byte[] getHash(ComputeRootHashOpts opts) throws TableNotFoundException,
-      NoSuchAlgorithmException {
+  private byte[] getHash(ComputeRootHashOpts opts)
+      throws TableNotFoundException, NoSuchAlgorithmException {
     try (AccumuloClient client = opts.createClient()) {
       return getHash(client, opts.tableName, opts.hashName);
     }
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/cli/GenerateHashes.java b/src/main/java/org/apache/accumulo/testing/merkle/cli/GenerateHashes.java
index ae8931c..9b21037 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/cli/GenerateHashes.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/cli/GenerateHashes.java
@@ -135,8 +135,8 @@ public class GenerateHashes {
     }
   }
 
-  public void run(GenerateHashesOpts opts) throws TableNotFoundException,
-      AccumuloSecurityException, AccumuloException, NoSuchAlgorithmException, FileNotFoundException {
+  public void run(GenerateHashesOpts opts) throws TableNotFoundException, AccumuloSecurityException,
+      AccumuloException, NoSuchAlgorithmException, FileNotFoundException {
     try (AccumuloClient client = opts.createClient()) {
       Collection<Range> ranges = getRanges(client, opts.tableName, opts.getSplitsFile());
       run(client, opts.tableName, opts.getOutputTableName(), opts.getHashName(),
@@ -144,9 +144,10 @@ public class GenerateHashes {
     }
   }
 
-  public void run(final AccumuloClient client, final String inputTableName, final String outputTableName, final String digestName, int numThreads,
-      final boolean iteratorPushdown, final Collection<Range> ranges) throws TableNotFoundException, AccumuloException,
-      NoSuchAlgorithmException {
+  public void run(final AccumuloClient client, final String inputTableName,
+      final String outputTableName, final String digestName, int numThreads,
+      final boolean iteratorPushdown, final Collection<Range> ranges)
+      throws TableNotFoundException, AccumuloException, NoSuchAlgorithmException {
     if (!client.tableOperations().exists(outputTableName)) {
       throw new IllegalArgumentException(outputTableName + " does not exist, please create it");
     }
@@ -203,7 +204,8 @@ public class GenerateHashes {
           }
 
           // Log some progress
-          log.info("{} computed digest for {} of {}", Thread.currentThread().getName(), range, Hex.encodeHexString(v.get()));
+          log.info("{} computed digest for {} of {}", Thread.currentThread().getName(), range,
+              Hex.encodeHexString(v.get()));
 
           try {
             bw.addMutation(m);
@@ -221,7 +223,8 @@ public class GenerateHashes {
         try {
           Thread.sleep(1000);
         } catch (InterruptedException e) {
-          log.error("Interrupted while waiting for executor service to gracefully complete. Exiting now");
+          log.error(
+              "Interrupted while waiting for executor service to gracefully complete. Exiting now");
           svc.shutdownNow();
           return;
         }
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/cli/ManualComparison.java b/src/main/java/org/apache/accumulo/testing/merkle/cli/ManualComparison.java
index ef4c97e..49129fe 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/cli/ManualComparison.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/cli/ManualComparison.java
@@ -24,9 +24,9 @@ import org.apache.accumulo.core.client.Scanner;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.security.Authorizations;
+import org.apache.accumulo.testing.cli.ClientOpts;
 
 import com.beust.jcommander.Parameter;
-import org.apache.accumulo.testing.cli.ClientOpts;
 
 /**
  * Accepts two table names and enumerates all key-values pairs in both checking for correctness. All
diff --git a/src/main/java/org/apache/accumulo/testing/merkle/package-info.java b/src/main/java/org/apache/accumulo/testing/merkle/package-info.java
index 2469d6f..6b5c7c2 100644
--- a/src/main/java/org/apache/accumulo/testing/merkle/package-info.java
+++ b/src/main/java/org/apache/accumulo/testing/merkle/package-info.java
@@ -15,25 +15,28 @@
  * limitations under the License.
  */
 /**
- * A <a href="http://en.wikipedia.org/wiki/Merkle_tree">Merkle tree</a> is a hash tree and can be used to evaluate equality over large
- * files with the ability to ascertain what portions of the files differ. Each leaf of the Merkle tree is some hash of a
- * portion of the file, with each leaf corresponding to some "range" within the source file. As such, if all leaves are
- * considered as ranges of the source file, the "sum" of all leaves creates a contiguous range over the entire file.
+ * A <a href="http://en.wikipedia.org/wiki/Merkle_tree">Merkle tree</a> is a hash tree and can be
+ * used to evaluate equality over large files with the ability to ascertain what portions of the
+ * files differ. Each leaf of the Merkle tree is some hash of a portion of the file, with each leaf
+ * corresponding to some "range" within the source file. As such, if all leaves are considered as
+ * ranges of the source file, the "sum" of all leaves creates a contiguous range over the entire
+ * file.
  * <p>
- * The parent of any nodes (typically, a binary tree; however this is not required) is the concatenation of the hashes of
- * the children. We can construct a full tree by walking up the tree, creating parents from children, until we have a root
- * node. To check equality of two files that each have a merkle tree built, we can very easily compare the value of at the
- * root of the Merkle tree to know whether or not the files are the same.
+ * The parent of any nodes (typically, a binary tree; however this is not required) is the
+ * concatenation of the hashes of the children. We can construct a full tree by walking up the tree,
+ * creating parents from children, until we have a root node. To check equality of two files that
+ * each have a merkle tree built, we can very easily compare the value of at the root of the Merkle
+ * tree to know whether or not the files are the same.
  * <p>
- * Additionally, in the situation where we have two files with we expect to be the same but are not, we can walk back down
- * the tree, finding subtrees that are equal and subtrees that are not. Subtrees that are equal correspond to portions of
- * the files which are identical, where subtrees that are not equal correspond to discrepancies between the two files.
+ * Additionally, in the situation where we have two files with we expect to be the same but are not,
+ * we can walk back down the tree, finding subtrees that are equal and subtrees that are not.
+ * Subtrees that are equal correspond to portions of the files which are identical, where subtrees
+ * that are not equal correspond to discrepancies between the two files.
  * <p>
- * We can apply this concept to Accumulo, treating a table as a file, and ranges within a file as an Accumulo Range. We can
- * then compute the hashes over each of these Ranges and compute the entire Merkle tree to determine if two tables are
- * equivalent.
+ * We can apply this concept to Accumulo, treating a table as a file, and ranges within a file as an
+ * Accumulo Range. We can then compute the hashes over each of these Ranges and compute the entire
+ * Merkle tree to determine if two tables are equivalent.
  *
  * @since 1.7.0
  */
 package org.apache.accumulo.testing.merkle;
-
diff --git a/src/main/java/org/apache/accumulo/testing/performance/Report.java b/src/main/java/org/apache/accumulo/testing/performance/Report.java
index a8bb1c8..f80a80b 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/Report.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/Report.java
@@ -56,8 +56,8 @@ public class Report {
     }
 
     public Builder result(String id, LongSummaryStatistics stats, String description) {
-      results.add(new Result(id, new Stats(stats.getMin(), stats.getMax(), stats.getSum(), stats
-          .getAverage(), stats.getCount()), description, Purpose.COMPARISON));
+      results.add(new Result(id, new Stats(stats.getMin(), stats.getMax(), stats.getSum(),
+          stats.getAverage(), stats.getCount()), description, Purpose.COMPARISON));
       return this;
     }
 
@@ -72,8 +72,8 @@ public class Report {
     }
 
     public Builder info(String id, LongSummaryStatistics stats, String description) {
-      results.add(new Result(id, new Stats(stats.getMin(), stats.getMax(), stats.getSum(), stats
-          .getAverage(), stats.getCount()), description, Purpose.INFORMATIONAL));
+      results.add(new Result(id, new Stats(stats.getMin(), stats.getMax(), stats.getSum(),
+          stats.getAverage(), stats.getCount()), description, Purpose.INFORMATIONAL));
       return this;
     }
 
diff --git a/src/main/java/org/apache/accumulo/testing/performance/impl/Csv.java b/src/main/java/org/apache/accumulo/testing/performance/impl/Csv.java
index a371d11..2af9e1c 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/impl/Csv.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/impl/Csv.java
@@ -79,24 +79,27 @@ public class Csv {
   }
 
   public static void main(String[] args) throws Exception {
-    Map<RowId, Map<String, Double>> rows = new TreeMap<>();
+    Map<RowId,Map<String,Double>> rows = new TreeMap<>();
 
     for (String file : args) {
       Collection<ContextualReport> reports = Compare.readReports(file);
 
-      Instant minStart = reports.stream().map(cr -> cr.startTime).map(Instant::parse).min(Instant::compareTo).get();
+      Instant minStart = reports.stream().map(cr -> cr.startTime).map(Instant::parse)
+          .min(Instant::compareTo).get();
 
-      String version = Iterables.getOnlyElement(reports.stream().map(cr -> cr.accumuloVersion).collect(toSet()));
+      String version = Iterables
+          .getOnlyElement(reports.stream().map(cr -> cr.accumuloVersion).collect(toSet()));
 
-      Map<String, Double> row = new HashMap<>();
+      Map<String,Double> row = new HashMap<>();
 
       for (ContextualReport report : reports) {
 
-        String id = report.id != null ? report.id :  report.testClass.substring(report.testClass.lastIndexOf('.')+1);
+        String id = report.id != null ? report.id
+            : report.testClass.substring(report.testClass.lastIndexOf('.') + 1);
 
         for (Result result : report.results) {
-          if(result.purpose == Result.Purpose.COMPARISON) {
-            row.put(id+"."+result.id, result.data.doubleValue());
+          if (result.purpose == Result.Purpose.COMPARISON) {
+            row.put(id + "." + result.id, result.data.doubleValue());
           }
         }
       }
@@ -104,12 +107,14 @@ public class Csv {
       rows.put(new RowId(minStart, version), row);
     }
 
-    List<String> allCols = rows.values().stream().flatMap(row -> row.keySet().stream()).distinct().sorted().collect(toList());
+    List<String> allCols = rows.values().stream().flatMap(row -> row.keySet().stream()).distinct()
+        .sorted().collect(toList());
 
     // print header
-    print(Stream.concat(Stream.of("Start Time","Version"), allCols.stream()).collect(joining(",")));
+    print(
+        Stream.concat(Stream.of("Start Time", "Version"), allCols.stream()).collect(joining(",")));
 
-    rows.forEach((id, row)->{
+    rows.forEach((id, row) -> {
       StringJoiner joiner = new StringJoiner(",");
       joiner.add(id.startTime.toString());
       joiner.add(id.accumuloVersion);
diff --git a/src/main/java/org/apache/accumulo/testing/performance/impl/MergeSiteConfig.java b/src/main/java/org/apache/accumulo/testing/performance/impl/MergeSiteConfig.java
index 0e98f85..f01afdd 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/impl/MergeSiteConfig.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/impl/MergeSiteConfig.java
@@ -31,17 +31,18 @@ public class MergeSiteConfig {
     String className = args[0];
     Path confFile = Paths.get(args[1], "accumulo.properties");
 
-    PerformanceTest perfTest = Class.forName(className).asSubclass(PerformanceTest.class).newInstance();
+    PerformanceTest perfTest = Class.forName(className).asSubclass(PerformanceTest.class)
+        .newInstance();
 
     Properties props = new Properties();
 
-    try(Reader in = Files.newBufferedReader(confFile)){
+    try (Reader in = Files.newBufferedReader(confFile)) {
       props.load(in);
     }
 
-    perfTest.getSystemConfig().getAccumuloConfig().forEach((k,v) -> props.setProperty(k, v));
+    perfTest.getSystemConfig().getAccumuloConfig().forEach((k, v) -> props.setProperty(k, v));
 
-    try(Writer out = Files.newBufferedWriter(confFile)){
+    try (Writer out = Files.newBufferedWriter(confFile)) {
       props.store(out, "Modified by performance test");
     }
   }
diff --git a/src/main/java/org/apache/accumulo/testing/performance/impl/PerfTestRunner.java b/src/main/java/org/apache/accumulo/testing/performance/impl/PerfTestRunner.java
index 2f230cd..c85b267 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/impl/PerfTestRunner.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/impl/PerfTestRunner.java
@@ -63,8 +63,8 @@ public class PerfTestRunner {
 
     DateTimeFormatter formatter = DateTimeFormatter.ofPattern("yyyyMMddHHmmss");
     String time = Instant.now().atZone(ZoneId.systemDefault()).format(formatter);
-    Path outputFile = Paths.get(outputDir, perfTest.getClass().getSimpleName() + "_" + time
-        + ".json");
+    Path outputFile = Paths.get(outputDir,
+        perfTest.getClass().getSimpleName() + "_" + time + ".json");
 
     try (Writer writer = Files.newBufferedWriter(outputFile)) {
       gson.toJson(report, writer);
diff --git a/src/main/java/org/apache/accumulo/testing/performance/tests/ProbabilityFilter.java b/src/main/java/org/apache/accumulo/testing/performance/tests/ProbabilityFilter.java
index 5c37201..6924f5b 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/tests/ProbabilityFilter.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/tests/ProbabilityFilter.java
@@ -9,9 +9,9 @@ import org.apache.accumulo.core.data.Value;
 
 public class ProbabilityFilter extends YieldingFilter {
   @Override
-  protected BiPredicate<Key, Value> createPredicate(Map<String,String> options) {
+  protected BiPredicate<Key,Value> createPredicate(Map<String,String> options) {
     double matchProbability = Double.parseDouble(options.get("probability"));
     Random rand = new Random();
-    return (k,v) -> rand.nextDouble() < matchProbability;
+    return (k, v) -> rand.nextDouble() < matchProbability;
   }
 }
diff --git a/src/main/java/org/apache/accumulo/testing/performance/tests/RandomCachedLookupsPT.java b/src/main/java/org/apache/accumulo/testing/performance/tests/RandomCachedLookupsPT.java
index 8fcd200..c09fe2e 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/tests/RandomCachedLookupsPT.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/tests/RandomCachedLookupsPT.java
@@ -90,8 +90,8 @@ public class RandomCachedLookupsPT implements PerformanceTest {
     long d128 = doLookups(env.getClient(), 128, NUM_LOOKUPS_PER_THREAD);
 
     reportBuilder.id("smalls");
-    reportBuilder
-        .description("Runs multiple threads each doing lots of small random scans.  For this test data and index cache are enabled.");
+    reportBuilder.description(
+        "Runs multiple threads each doing lots of small random scans.  For this test data and index cache are enabled.");
     reportBuilder.info("warmup", 32 * NUM_LOOKUPS_PER_THREAD, warmup,
         "Random lookup per sec for 32 threads");
     reportBuilder.info("lookups_1", NUM_LOOKUPS_PER_THREAD, d1,
@@ -127,9 +127,11 @@ public class RandomCachedLookupsPT implements PerformanceTest {
     return reportBuilder.build();
   }
 
-  public static void writeData(Report.Builder reportBuilder, AccumuloClient client, int numRows) throws Exception {
+  public static void writeData(Report.Builder reportBuilder, AccumuloClient client, int numRows)
+      throws Exception {
 
-    reportBuilder.parameter("rows", numRows, "Number of random rows written.  Each row has 4 columns.");
+    reportBuilder.parameter("rows", numRows,
+        "Number of random rows written.  Each row has 4 columns.");
 
     NewTableConfiguration ntc = new NewTableConfiguration();
     Map<String,String> props = new HashMap<>();
@@ -150,7 +152,8 @@ public class RandomCachedLookupsPT implements PerformanceTest {
     long t2 = System.currentTimeMillis();
 
     SortedSet<Text> partitionKeys = new TreeSet<>(
-        Stream.of("1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f").map(Text::new).collect(toList()));
+        Stream.of("1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f")
+            .map(Text::new).collect(toList()));
     client.tableOperations().addSplits("scanpt", partitionKeys);
 
     long t3 = System.currentTimeMillis();
@@ -197,10 +200,12 @@ public class RandomCachedLookupsPT implements PerformanceTest {
     reportBuilder.info("split", t3 - t2, "Time to split table in ms");
     reportBuilder.info("write", 4 * numRows, t4 - t3, "Rate to write data in entries/sec");
     reportBuilder.info("compact", 4 * numRows, t5 - t4, "Rate to compact table in entries/sec");
-    reportBuilder.info("fullScan", 4 * numRows, t6 - t5, "Rate to do full table scan in entries/sec");
+    reportBuilder.info("fullScan", 4 * numRows, t6 - t5,
+        "Rate to do full table scan in entries/sec");
   }
 
-  private static long doLookups(AccumuloClient client, int numThreads, int numScansPerThread) throws Exception {
+  private static long doLookups(AccumuloClient client, int numThreads, int numScansPerThread)
+      throws Exception {
 
     ExecutorService es = Executors.newFixedThreadPool(numThreads);
 
@@ -220,7 +225,7 @@ public class RandomCachedLookupsPT implements PerformanceTest {
 
     es.shutdown();
 
-    return t2 -t1;
+    return t2 - t1;
   }
 
   private static void doLookups(AccumuloClient client, int numScans) {
diff --git a/src/main/java/org/apache/accumulo/testing/performance/tests/ScanExecutorPT.java b/src/main/java/org/apache/accumulo/testing/performance/tests/ScanExecutorPT.java
index 0844683..20911cc 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/tests/ScanExecutorPT.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/tests/ScanExecutorPT.java
@@ -72,12 +72,12 @@ public class ScanExecutorPT implements PerformanceTest {
     siteCfg.put(Property.TSERV_MINTHREADS.getKey(), "200");
     siteCfg.put(Property.TSERV_SCAN_EXECUTORS_PREFIX.getKey() + "se1.threads",
         SCAN_EXECUTOR_THREADS);
-    siteCfg
-        .put(Property.TSERV_SCAN_EXECUTORS_PREFIX.getKey() + "se1.prioritizer", SCAN_PRIORITIZER);
+    siteCfg.put(Property.TSERV_SCAN_EXECUTORS_PREFIX.getKey() + "se1.prioritizer",
+        SCAN_PRIORITIZER);
     siteCfg.put(Property.TSERV_SCAN_EXECUTORS_PREFIX.getKey() + "se2.threads",
         SCAN_EXECUTOR_THREADS);
-    siteCfg
-        .put(Property.TSERV_SCAN_EXECUTORS_PREFIX.getKey() + "se2.prioritizer", SCAN_PRIORITIZER);
+    siteCfg.put(Property.TSERV_SCAN_EXECUTORS_PREFIX.getKey() + "se2.prioritizer",
+        SCAN_PRIORITIZER);
 
     return new SystemConfiguration().setAccumuloConfig(siteCfg);
   }
@@ -155,8 +155,8 @@ public class ScanExecutorPT implements PerformanceTest {
     return System.currentTimeMillis() - t1;
   }
 
-  private long scan(String tableName, AccumuloClient c, AtomicBoolean stop, Map<String,String> hints)
-      throws TableNotFoundException {
+  private long scan(String tableName, AccumuloClient c, AtomicBoolean stop,
+      Map<String,String> hints) throws TableNotFoundException {
     long count = 0;
     while (!stop.get()) {
       try (Scanner scanner = c.createScanner(tableName, Authorizations.EMPTY)) {
diff --git a/src/main/java/org/apache/accumulo/testing/performance/tests/ScanFewFamiliesPT.java b/src/main/java/org/apache/accumulo/testing/performance/tests/ScanFewFamiliesPT.java
index d2bddf4..4278b2c 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/tests/ScanFewFamiliesPT.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/tests/ScanFewFamiliesPT.java
@@ -69,10 +69,10 @@ public class ScanFewFamiliesPT implements PerformanceTest {
     for (int numFams : new int[] {1, 2, 4, 8, 16}) {
       LongSummaryStatistics stats = runScans(env, tableName, numFams);
       String fams = Strings.padStart(numFams + "", 2, '0');
-      builder.info("f" + fams + "_stats", stats, "Times in ms to fetch " + numFams
-          + " families from all rows");
-      builder.result("f" + fams, stats.getAverage(), "Average time in ms to fetch " + numFams
-          + " families from all rows");
+      builder.info("f" + fams + "_stats", stats,
+          "Times in ms to fetch " + numFams + " families from all rows");
+      builder.result("f" + fams, stats.getAverage(),
+          "Average time in ms to fetch " + numFams + " families from all rows");
     }
 
     builder.id("sfewfam");
@@ -96,10 +96,11 @@ public class ScanFewFamiliesPT implements PerformanceTest {
     return stats;
   }
 
-  private static long scan(String tableName, AccumuloClient c, Random rand, int numFamilies) throws TableNotFoundException {
+  private static long scan(String tableName, AccumuloClient c, Random rand, int numFamilies)
+      throws TableNotFoundException {
 
     Set<Text> families = new HashSet<>(numFamilies);
-    while(families.size() < numFamilies) {
+    while (families.size() < numFamilies) {
       families.add(new Text(TestData.fam(rand.nextInt(NUM_FAMS))));
     }
 
diff --git a/src/main/java/org/apache/accumulo/testing/performance/tests/YieldingScanExecutorPT.java b/src/main/java/org/apache/accumulo/testing/performance/tests/YieldingScanExecutorPT.java
index 5d50007..59571fb 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/tests/YieldingScanExecutorPT.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/tests/YieldingScanExecutorPT.java
@@ -135,18 +135,19 @@ public class YieldingScanExecutorPT implements PerformanceTest {
     builder.parameter("short_threads", NUM_SHORT_SCANS_THREADS, "Threads used to run short scans.");
     builder.parameter("long_threads", NUM_LONG_SCANS,
         "Threads running long fileter scans.  Each thread repeatedly scans entire table for "
-        + "duration of test randomly returning a few of the keys.");
+            + "duration of test randomly returning a few of the keys.");
     builder.parameter("rows", NUM_ROWS, "Rows in test table");
     builder.parameter("familes", NUM_FAMS, "Families per row in test table");
     builder.parameter("qualifiers", NUM_QUALS, "Qualifiers per family in test table");
     builder.parameter("server_scan_threads", SCAN_EXECUTOR_THREADS,
         "Server side scan handler threads that each executor has.  There are 2 executors.");
 
-    builder.parameter("filter_probabilities", FILTER_PROBABILITIES, "The chances that one of the long "
-        + "filter scans will return any key it sees. The probabilites are cycled through when "
-        + "starting long scans.");
-    builder.parameter("filter_yield_time", FILTER_YIELD_TIME, "The time in ms after which one of "
-        + "the long filter scans will yield.");
+    builder.parameter("filter_probabilities", FILTER_PROBABILITIES,
+        "The chances that one of the long "
+            + "filter scans will return any key it sees. The probabilites are cycled through when "
+            + "starting long scans.");
+    builder.parameter("filter_yield_time", FILTER_YIELD_TIME,
+        "The time in ms after which one of " + "the long filter scans will yield.");
     builder.parameter("quick_scan_time", QUICK_SCAN_TIME, "The threshold time in ms for deciding "
         + "what is a quick vs long scan.  Times less than this are sent to one executor and longer "
         + "times are sent to another.");
@@ -168,8 +169,8 @@ public class YieldingScanExecutorPT implements PerformanceTest {
     return System.currentTimeMillis() - t1;
   }
 
-  private long scan(String tableName, AccumuloClient c, AtomicBoolean stop, String filterProbability)
-      throws TableNotFoundException {
+  private long scan(String tableName, AccumuloClient c, AtomicBoolean stop,
+      String filterProbability) throws TableNotFoundException {
     long count = 0;
     while (!stop.get()) {
       try (Scanner scanner = c.createScanner(tableName, Authorizations.EMPTY)) {
diff --git a/src/main/java/org/apache/accumulo/testing/performance/util/TestData.java b/src/main/java/org/apache/accumulo/testing/performance/util/TestData.java
index 2351084..66a7f7e 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/util/TestData.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/util/TestData.java
@@ -42,8 +42,8 @@ public class TestData {
     return FastFormat.toZeroPaddedString(v, 9, 16, EMPTY);
   }
 
-  public static void generate(AccumuloClient client, String tableName, int rows, int fams, int quals)
-      throws Exception {
+  public static void generate(AccumuloClient client, String tableName, int rows, int fams,
+      int quals) throws Exception {
     try (BatchWriter writer = client.createBatchWriter(tableName)) {
       int v = 0;
       for (int r = 0; r < rows; r++) {
diff --git a/src/main/java/org/apache/accumulo/testing/performance/util/TestExecutor.java b/src/main/java/org/apache/accumulo/testing/performance/util/TestExecutor.java
index b6c4609..42b95d0 100644
--- a/src/main/java/org/apache/accumulo/testing/performance/util/TestExecutor.java
+++ b/src/main/java/org/apache/accumulo/testing/performance/util/TestExecutor.java
@@ -45,11 +45,13 @@ public class TestExecutor<T> implements Iterable<T>, AutoCloseable {
   }
 
   public Stream<T> stream() {
-    return futures.stream().map(f -> {try {
-      return f.get();
-    } catch (InterruptedException | ExecutionException e) {
-      throw new RuntimeException(e);
-    }});
+    return futures.stream().map(f -> {
+      try {
+        return f.get();
+      } catch (InterruptedException | ExecutionException e) {
+        throw new RuntimeException(e);
+      }
+    });
   }
 
   @Override
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/Module.java b/src/main/java/org/apache/accumulo/testing/randomwalk/Module.java
index e15b9f5..c832bac 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/Module.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/Module.java
@@ -272,8 +272,8 @@ public class Module extends Node {
         numHops++;
 
         if (!adjMap.containsKey(curNodeId) && !curNodeId.startsWith("alias.")) {
-          throw new Exception("Reached node(" + curNodeId + ") without outgoing edges in module("
-              + this + ")");
+          throw new Exception(
+              "Reached node(" + curNodeId + ") without outgoing edges in module(" + this + ")");
         }
         AdjList adj = adjMap.get(curNodeId);
         String nextNodeId = adj.randomNeighbor();
@@ -519,8 +519,8 @@ public class Module extends Node {
 
     // set the schema
     SchemaFactory sf = SchemaFactory.newInstance(XMLConstants.W3C_XML_SCHEMA_NS_URI);
-    Schema moduleSchema = sf.newSchema(this.getClass().getClassLoader()
-        .getResource("randomwalk/module.xsd"));
+    Schema moduleSchema = sf
+        .newSchema(this.getClass().getClassLoader().getResource("randomwalk/module.xsd"));
     dbf.setSchema(moduleSchema);
 
     // parse the document
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/RandWalkEnv.java b/src/main/java/org/apache/accumulo/testing/randomwalk/RandWalkEnv.java
index 216b995..e3ed805 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/RandWalkEnv.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/RandWalkEnv.java
@@ -48,8 +48,8 @@ public class RandWalkEnv extends TestEnv {
    * @throws NumberFormatException
    *           if any configuration property cannot be parsed
    */
-  public MultiTableBatchWriter getMultiTableBatchWriter() throws AccumuloException,
-      AccumuloSecurityException {
+  public MultiTableBatchWriter getMultiTableBatchWriter()
+      throws AccumuloException, AccumuloSecurityException {
     if (mtbw == null) {
       mtbw = getAccumuloClient().createMultiTableBatchWriter();
     }
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/BulkPlusOne.java b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/BulkPlusOne.java
index 0ab337e..64db3c6 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/BulkPlusOne.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/BulkPlusOne.java
@@ -24,10 +24,10 @@ import java.util.UUID;
 import java.util.concurrent.atomic.AtomicLong;
 
 import org.apache.accumulo.core.client.IteratorSetting.Column;
+import org.apache.accumulo.core.client.rfile.RFile;
 import org.apache.accumulo.core.client.rfile.RFileWriter;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Value;
-import org.apache.accumulo.core.client.rfile.RFile;
 import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 import org.apache.hadoop.fs.FileStatus;
@@ -96,8 +96,8 @@ public class BulkPlusOne extends BulkImportTest {
       }
       writer.close();
     }
-    env.getAccumuloClient().tableOperations()
-        .importDirectory(Setup.getTableName(), dir.toString(), fail.toString(), true);
+    env.getAccumuloClient().tableOperations().importDirectory(Setup.getTableName(), dir.toString(),
+        fail.toString(), true);
     fs.delete(dir, true);
     FileStatus[] failures = fs.listStatus(fail);
     if (failures != null && failures.length > 0) {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Compact.java b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Compact.java
index 329a397..4db7c1a 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Compact.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Compact.java
@@ -27,8 +27,8 @@ public class Compact extends SelectiveBulkTest {
     final Text[] points = Merge.getRandomTabletRange(state);
     final String rangeString = Merge.rangeToString(points);
     log.info("Compacting " + rangeString);
-    env.getAccumuloClient().tableOperations()
-        .compact(Setup.getTableName(), points[0], points[1], false, true);
+    env.getAccumuloClient().tableOperations().compact(Setup.getTableName(), points[0], points[1],
+        false, true);
     log.info("Compaction " + rangeString + " finished");
   }
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/ConsistencyCheck.java b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/ConsistencyCheck.java
index 119a3fe..f011e0a 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/ConsistencyCheck.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/ConsistencyCheck.java
@@ -38,8 +38,8 @@ public class ConsistencyCheck extends SelectiveBulkTest {
     log.info("Checking " + row);
     String user = env.getAccumuloClient().whoami();
     Authorizations auths = env.getAccumuloClient().securityOperations().getUserAuthorizations(user);
-    try (Scanner scanner = new IsolatedScanner(env.getAccumuloClient().createScanner(
-        Setup.getTableName(), auths))) {
+    try (Scanner scanner = new IsolatedScanner(
+        env.getAccumuloClient().createScanner(Setup.getTableName(), auths))) {
       scanner.setRange(new Range(row));
       scanner.fetchColumnFamily(BulkPlusOne.CHECK_COLUMN_FAMILY);
       Value v = null;
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Merge.java b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Merge.java
index c9835d6..cc898c3 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Merge.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Merge.java
@@ -39,8 +39,8 @@ public class Merge extends SelectiveBulkTest {
   }
 
   public static Text getRandomRow(Random rand) {
-    return new Text(String.format(BulkPlusOne.FMT, (rand.nextLong() & 0x7fffffffffffffffl)
-        % BulkPlusOne.LOTS));
+    return new Text(
+        String.format(BulkPlusOne.FMT, (rand.nextLong() & 0x7fffffffffffffffl) % BulkPlusOne.LOTS));
   }
 
   public static Text[] getRandomTabletRange(State state) {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Split.java b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Split.java
index 1aac330..2588cb1 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Split.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Split.java
@@ -32,8 +32,8 @@ public class Split extends SelectiveBulkTest {
     Random rand = (Random) state.get("rand");
     int count = rand.nextInt(20);
     for (int i = 0; i < count; i++)
-      splits.add(new Text(String.format(BulkPlusOne.FMT, (rand.nextLong() & 0x7fffffffffffffffl)
-          % BulkPlusOne.LOTS)));
+      splits.add(new Text(String.format(BulkPlusOne.FMT,
+          (rand.nextLong() & 0x7fffffffffffffffl) % BulkPlusOne.LOTS)));
     log.info("splitting " + splits);
     env.getAccumuloClient().tableOperations().addSplits(Setup.getTableName(), splits);
     log.info("split for " + splits + " finished");
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Verify.java b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Verify.java
index 8c7ecba..c5e5b1a 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Verify.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/bulk/Verify.java
@@ -25,7 +25,6 @@ import java.util.Properties;
 import java.util.concurrent.ThreadPoolExecutor;
 import java.util.concurrent.TimeUnit;
 
-import com.beust.jcommander.Parameter;
 import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.RowIterator;
 import org.apache.accumulo.core.client.Scanner;
@@ -38,6 +37,8 @@ import org.apache.accumulo.testing.randomwalk.State;
 import org.apache.accumulo.testing.randomwalk.Test;
 import org.apache.hadoop.io.Text;
 
+import com.beust.jcommander.Parameter;
+
 public class Verify extends Test {
 
   private static byte[] zero = new byte[] {'0'};
@@ -88,8 +89,8 @@ public class Verify extends Test {
         long curr = Long.parseLong(entry.getKey().getColumnQualifier().toString());
 
         if (curr - 1 != prev)
-          throw new Exception("Bad marker count " + entry.getKey() + " " + entry.getValue() + " "
-              + prev);
+          throw new Exception(
+              "Bad marker count " + entry.getKey() + " " + entry.getValue() + " " + prev);
 
         if (!entry.getValue().toString().equals("1"))
           throw new Exception("Bad marker value " + entry.getKey() + " " + entry.getValue());
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/BulkImport.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/BulkImport.java
index b618ca2..df45f87 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/BulkImport.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/BulkImport.java
@@ -29,12 +29,12 @@ import org.apache.accumulo.core.client.BatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableNotFoundException;
 import org.apache.accumulo.core.client.TableOfflineException;
+import org.apache.accumulo.core.client.rfile.RFile;
+import org.apache.accumulo.core.client.rfile.RFileWriter;
 import org.apache.accumulo.core.data.ColumnUpdate;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
-import org.apache.accumulo.core.client.rfile.RFile;
-import org.apache.accumulo.core.client.rfile.RFileWriter;
 import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 import org.apache.accumulo.testing.randomwalk.Test;
@@ -109,8 +109,8 @@ public class BulkImport extends Test {
     fs.mkdirs(new Path(bulkDir + "_f"));
 
     try {
-      BatchWriter bw = new RFileBatchWriter(env.getHadoopConfiguration(), fs, bulkDir
-          + "/file01.rf");
+      BatchWriter bw = new RFileBatchWriter(env.getHadoopConfiguration(), fs,
+          bulkDir + "/file01.rf");
       try {
         TreeSet<Long> rows = new TreeSet<>();
         int numRows = rand.nextInt(100000);
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ChangeAuthorizations.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ChangeAuthorizations.java
index 0fbf24f..5a0a9fc 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ChangeAuthorizations.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ChangeAuthorizations.java
@@ -43,8 +43,8 @@ public class ChangeAuthorizations extends Test {
 
     String userName = userNames.get(rand.nextInt(userNames.size()));
     try {
-      List<byte[]> auths = new ArrayList<>(client.securityOperations()
-          .getUserAuthorizations(userName).getAuthorizations());
+      List<byte[]> auths = new ArrayList<>(
+          client.securityOperations().getUserAuthorizations(userName).getAuthorizations());
 
       if (rand.nextBoolean()) {
         String authorization = String.format("a%d", rand.nextInt(5000));
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/CloneTable.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/CloneTable.java
index 811ef7b..f97ba01 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/CloneTable.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/CloneTable.java
@@ -59,7 +59,8 @@ public class CloneTable extends Test {
     } catch (AccumuloException e) {
       Throwable cause = e.getCause();
       if (cause != null && cause instanceof NamespaceNotFoundException)
-        log.debug("Clone: " + srcTableName + " to " + newTableName + " failed, namespace not found");
+        log.debug(
+            "Clone: " + srcTableName + " to " + newTableName + " failed, namespace not found");
       else
         throw e;
     }
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ConcurrentFixture.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ConcurrentFixture.java
index edaf632..71595b2 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ConcurrentFixture.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/ConcurrentFixture.java
@@ -20,8 +20,8 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Random;
 
-import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.Fixture;
+import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 import org.apache.hadoop.io.Text;
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Config.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Config.java
index 48d8eea..7c35899 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Config.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Config.java
@@ -115,8 +115,8 @@ public class Config extends Test {
       int choice = Integer.parseInt(lastSetting.toString());
       Property property = settings[choice].property;
       log.debug("Setting " + property.getKey() + " back to " + property.getDefaultValue());
-      env.getAccumuloClient().instanceOperations()
-          .setProperty(property.getKey(), property.getDefaultValue());
+      env.getAccumuloClient().instanceOperations().setProperty(property.getKey(),
+          property.getDefaultValue());
     }
     lastSetting = state.getOkIfAbsent(LAST_TABLE_SETTING);
     if (lastSetting != null) {
@@ -128,8 +128,8 @@ public class Config extends Test {
         log.debug("Setting " + property.getKey() + " on " + table + " back to "
             + property.getDefaultValue());
         try {
-          env.getAccumuloClient().tableOperations()
-              .setProperty(table, property.getKey(), property.getDefaultValue());
+          env.getAccumuloClient().tableOperations().setProperty(table, property.getKey(),
+              property.getDefaultValue());
         } catch (AccumuloException ex) {
           if (ex.getCause() instanceof ThriftTableOperationException) {
             ThriftTableOperationException ttoe = (ThriftTableOperationException) ex.getCause();
@@ -150,8 +150,8 @@ public class Config extends Test {
         log.debug("Setting " + property.getKey() + " on " + namespace + " back to "
             + property.getDefaultValue());
         try {
-          env.getAccumuloClient().namespaceOperations()
-              .setProperty(namespace, property.getKey(), property.getDefaultValue());
+          env.getAccumuloClient().namespaceOperations().setProperty(namespace, property.getKey(),
+              property.getDefaultValue());
         } catch (AccumuloException ex) {
           if (ex.getCause() instanceof ThriftTableOperationException) {
             ThriftTableOperationException ttoe = (ThriftTableOperationException) ex.getCause();
@@ -194,8 +194,8 @@ public class Config extends Test {
     state.set(LAST_TABLE_SETTING, table + "," + choice);
     log.debug("Setting " + setting.property.getKey() + " on table " + table + " to " + newValue);
     try {
-      env.getAccumuloClient().tableOperations()
-          .setProperty(table, setting.property.getKey(), "" + newValue);
+      env.getAccumuloClient().tableOperations().setProperty(table, setting.property.getKey(),
+          "" + newValue);
     } catch (AccumuloException ex) {
       if (ex.getCause() instanceof ThriftTableOperationException) {
         ThriftTableOperationException ttoe = (ThriftTableOperationException) ex.getCause();
@@ -222,11 +222,11 @@ public class Config extends Test {
     // generate a random value
     long newValue = random.nextLong(setting.min, setting.max);
     state.set(LAST_NAMESPACE_SETTING, namespace + "," + choice);
-    log.debug("Setting " + setting.property.getKey() + " on namespace " + namespace + " to "
-        + newValue);
+    log.debug(
+        "Setting " + setting.property.getKey() + " on namespace " + namespace + " to " + newValue);
     try {
-      env.getAccumuloClient().namespaceOperations()
-          .setProperty(namespace, setting.property.getKey(), "" + newValue);
+      env.getAccumuloClient().namespaceOperations().setProperty(namespace,
+          setting.property.getKey(), "" + newValue);
     } catch (AccumuloException ex) {
       if (ex.getCause() instanceof ThriftTableOperationException) {
         ThriftTableOperationException ttoe = (ThriftTableOperationException) ex.getCause();
@@ -246,8 +246,8 @@ public class Config extends Test {
     long newValue = random.nextLong(setting.min, setting.max);
     state.set(LAST_SETTING, "" + choice);
     log.debug("Setting " + setting.property.getKey() + " to " + newValue);
-    env.getAccumuloClient().instanceOperations()
-        .setProperty(setting.property.getKey(), "" + newValue);
+    env.getAccumuloClient().instanceOperations().setProperty(setting.property.getKey(),
+        "" + newValue);
   }
 
 }
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/IsolatedScan.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/IsolatedScan.java
index 14fb2ee..43112ce 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/IsolatedScan.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/IsolatedScan.java
@@ -51,8 +51,8 @@ public class IsolatedScan extends Test {
     String tableName = tableNames.get(rand.nextInt(tableNames.size()));
 
     try {
-      RowIterator iter = new RowIterator(new IsolatedScanner(client.createScanner(tableName,
-          Authorizations.EMPTY)));
+      RowIterator iter = new RowIterator(
+          new IsolatedScanner(client.createScanner(tableName, Authorizations.EMPTY)));
 
       while (iter.hasNext()) {
         PeekingIterator<Entry<Key,Value>> row = Iterators.peekingIterator(iter.next());
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Replication.java b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Replication.java
index d77675d..cb60e11 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Replication.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/concurrent/Replication.java
@@ -178,8 +178,8 @@ public class Replication extends Test {
   // junit isn't a dependency
   private void assertEquals(int expected, int actual) {
     if (expected != actual)
-      throw new RuntimeException(String.format("%d fails to match expected value %d", actual,
-          expected));
+      throw new RuntimeException(
+          String.format("%d fails to match expected value %d", actual, expected));
   }
 
   // junit isn't a dependency
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Setup.java b/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Setup.java
index 631d31f..b4fbb99 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Setup.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Setup.java
@@ -49,8 +49,8 @@ public class Setup extends Test {
       env.getAccumuloClient().tableOperations().create(tableName);
       log.debug("created table " + tableName);
       boolean blockCache = rand.nextBoolean();
-      env.getAccumuloClient().tableOperations()
-          .setProperty(tableName, Property.TABLE_BLOCKCACHE_ENABLED.getKey(), blockCache + "");
+      env.getAccumuloClient().tableOperations().setProperty(tableName,
+          Property.TABLE_BLOCKCACHE_ENABLED.getKey(), blockCache + "");
       log.debug("set " + Property.TABLE_BLOCKCACHE_ENABLED.getKey() + " " + blockCache);
     } catch (TableExistsException tee) {}
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Transfer.java b/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Transfer.java
index d16d283..a8fa018 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Transfer.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/conditional/Transfer.java
@@ -111,8 +111,8 @@ public class Transfer extends Test {
 
       int amt = rand.nextInt(50);
 
-      log.debug("transfer req " + bank + " " + amt + " " + acct1 + " " + a1 + " " + acct2 + " "
-          + a2);
+      log.debug(
+          "transfer req " + bank + " " + amt + " " + acct1 + " " + a1 + " " + acct2 + " " + a2);
 
       if (a1.bal >= amt) {
         ConditionalMutation cm = new ConditionalMutation(bank,
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/image/ImageFixture.java b/src/main/java/org/apache/accumulo/testing/randomwalk/image/ImageFixture.java
index 6e6a9b7..02c7add 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/image/ImageFixture.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/image/ImageFixture.java
@@ -29,8 +29,8 @@ import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableExistsException;
-import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.Fixture;
+import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 import org.apache.hadoop.io.Text;
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/image/ScanMeta.java b/src/main/java/org/apache/accumulo/testing/randomwalk/image/ScanMeta.java
index 1355381..f9a47f3 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/image/ScanMeta.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/image/ScanMeta.java
@@ -102,8 +102,8 @@ public class ScanMeta extends Test {
     if (!hashes.equals(hashes2)) {
       log.error("uuids from doc table : " + hashes.values());
       log.error("uuids from index     : " + hashes2.values());
-      throw new Exception("Mismatch between document table and index " + indexTableName + " "
-          + imageTableName);
+      throw new Exception(
+          "Mismatch between document table and index " + indexTableName + " " + imageTableName);
     }
 
     indexScanner.close();
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/image/Write.java b/src/main/java/org/apache/accumulo/testing/randomwalk/image/Write.java
index ae3be22..4cf1390 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/image/Write.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/image/Write.java
@@ -78,8 +78,8 @@ public class Write extends Test {
     state.set("totalWrites", totalWrites);
 
     // set count
-    m.put(META_COLUMN_FAMILY, COUNT_COLUMN_QUALIFIER, new Value(String.format("%d", totalWrites)
-        .getBytes(UTF_8)));
+    m.put(META_COLUMN_FAMILY, COUNT_COLUMN_QUALIFIER,
+        new Value(String.format("%d", totalWrites).getBytes(UTF_8)));
 
     // add mutation
     imagesBW.addMutation(m);
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/CopyTool.java b/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/CopyTool.java
index 2ac69fb..54866f2 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/CopyTool.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/CopyTool.java
@@ -20,12 +20,12 @@ import java.io.IOException;
 import java.util.Properties;
 
 import org.apache.accumulo.core.client.Accumulo;
-import org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat;
-import org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
 import org.apache.accumulo.core.security.Authorizations;
+import org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat;
+import org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat;
 import org.apache.hadoop.conf.Configured;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.mapreduce.Job;
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/MultiTableFixture.java b/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/MultiTableFixture.java
index 59ee619..1c65118 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/MultiTableFixture.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/multitable/MultiTableFixture.java
@@ -24,8 +24,8 @@ import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableNotFoundException;
-import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.Fixture;
+import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 
 public class MultiTableFixture extends Fixture {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTable.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTable.java
index d01a6fc..c24b961 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTable.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTable.java
@@ -36,8 +36,8 @@ public class AlterTable extends Test {
   @Override
   public void visit(State state, RandWalkEnv env, Properties props) throws Exception {
     String systemUser = WalkingSecurity.get(state, env).getSysUserName();
-    try (AccumuloClient client = env.createClient(systemUser, WalkingSecurity.get(state, env)
-        .getSysToken())) {
+    try (AccumuloClient client = env.createClient(systemUser,
+        WalkingSecurity.get(state, env).getSysToken())) {
 
       String tableName = WalkingSecurity.get(state, env).getTableName();
 
@@ -59,8 +59,9 @@ public class AlterTable extends Test {
           throw new AccumuloException("Got unexpected ae error code", ae);
         }
       }
-      String newTableName = String.format("security_%s_%s_%d", InetAddress.getLocalHost()
-          .getHostName().replaceAll("[-.]", "_"), env.getPid(), System.currentTimeMillis());
+      String newTableName = String.format("security_%s_%s_%d",
+          InetAddress.getLocalHost().getHostName().replaceAll("[-.]", "_"), env.getPid(),
+          System.currentTimeMillis());
 
       renameTable(client, state, env, tableName, newTableName, hasPermission, exists);
     }
@@ -74,8 +75,8 @@ public class AlterTable extends Test {
     } catch (AccumuloSecurityException ae) {
       if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
         if (hasPermission)
-          throw new AccumuloException(
-              "Got a security exception when I should have had permission.", ae);
+          throw new AccumuloException("Got a security exception when I should have had permission.",
+              ae);
         else
           return;
       } else if (ae.getSecurityErrorCode().equals(SecurityErrorCode.BAD_CREDENTIALS)) {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTablePerm.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTablePerm.java
index e70d486..27b73a6 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTablePerm.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/AlterTablePerm.java
@@ -62,8 +62,8 @@ public class AlterTablePerm extends Test {
     } else
       tabPerm = TablePermission.valueOf(perm);
     String tableName = WalkingSecurity.get(state, env).getTableName();
-    boolean hasPerm = WalkingSecurity.get(state, env)
-        .hasTablePermission(target, tableName, tabPerm);
+    boolean hasPerm = WalkingSecurity.get(state, env).hasTablePermission(target, tableName,
+        tabPerm);
     boolean canGive;
     String sourceUser;
     AuthenticationToken sourceToken;
@@ -195,8 +195,8 @@ public class AlterTablePerm extends Test {
       if (!tableExists)
         throw new AccumuloException("Table shouldn't have existed, but apparently does");
       if (!canGive)
-        throw new AccumuloException(client.whoami()
-            + " shouldn't have been able to grant privilege");
+        throw new AccumuloException(
+            client.whoami() + " shouldn't have been able to grant privilege");
     }
   }
 }
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/Authenticate.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/Authenticate.java
index 89f3290..db9c076 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/Authenticate.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/Authenticate.java
@@ -33,8 +33,8 @@ public class Authenticate extends Test {
 
   @Override
   public void visit(State state, RandWalkEnv env, Properties props) throws Exception {
-    authenticate(WalkingSecurity.get(state, env).getSysUserName(), WalkingSecurity.get(state, env)
-        .getSysToken(), state, env, props);
+    authenticate(WalkingSecurity.get(state, env).getSysUserName(),
+        WalkingSecurity.get(state, env).getSysToken(), state, env, props);
   }
 
   public static void authenticate(String principal, AuthenticationToken token, State state,
@@ -56,8 +56,7 @@ public class Authenticate extends Test {
       byte[] password = Arrays.copyOf(WalkingSecurity.get(state, env).getUserPassword(target),
           WalkingSecurity.get(state, env).getUserPassword(target).length);
       boolean hasPermission = client.securityOperations().hasSystemPermission(principal,
-          SystemPermission.SYSTEM)
-          || principal.equals(target);
+          SystemPermission.SYSTEM) || principal.equals(target);
 
       if (!success)
         for (int i = 0; i < password.length; i++)
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/ChangePass.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/ChangePass.java
index b75f58a..b2d0221 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/ChangePass.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/ChangePass.java
@@ -57,8 +57,7 @@ public class ChangePass extends Test {
       targetExists = WalkingSecurity.get(state, env).userExists(target);
 
       hasPerm = client.securityOperations().hasSystemPermission(principal,
-          SystemPermission.ALTER_USER)
-          || principal.equals(target);
+          SystemPermission.ALTER_USER) || principal.equals(target);
 
       Random r = new Random();
 
@@ -73,8 +72,9 @@ public class ChangePass extends Test {
         switch (ae.getSecurityErrorCode()) {
           case PERMISSION_DENIED:
             if (hasPerm)
-              throw new AccumuloException("Change failed when it should have succeeded to change "
-                  + target + "'s password", ae);
+              throw new AccumuloException(
+                  "Change failed when it should have succeeded to change " + target + "'s password",
+                  ae);
             return;
           case USER_DOESNT_EXIST:
             if (targetExists)
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/CreateUser.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/CreateUser.java
index b6b1ca2..a4b242b 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/CreateUser.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/CreateUser.java
@@ -32,8 +32,8 @@ public class CreateUser extends Test {
   @Override
   public void visit(State state, RandWalkEnv env, Properties props) throws Exception {
     String sysPrincipal = WalkingSecurity.get(state, env).getSysUserName();
-    try (AccumuloClient client = env.createClient(sysPrincipal, WalkingSecurity.get(state, env)
-        .getSysToken())) {
+    try (AccumuloClient client = env.createClient(sysPrincipal,
+        WalkingSecurity.get(state, env).getSysToken())) {
 
       String tableUserName = WalkingSecurity.get(state, env).getTabUserName();
 
@@ -52,8 +52,8 @@ public class CreateUser extends Test {
             else {
               // create user anyway for sake of state
               if (!exists) {
-                env.getAccumuloClient().securityOperations()
-                    .createLocalUser(tableUserName, tabUserPass);
+                env.getAccumuloClient().securityOperations().createLocalUser(tableUserName,
+                    tabUserPass);
                 WalkingSecurity.get(state, env).createUser(tableUserName, tabUserPass);
                 Thread.sleep(1000);
               }
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/DropUser.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/DropUser.java
index 9718dde..d5e96b0 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/DropUser.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/DropUser.java
@@ -31,8 +31,8 @@ public class DropUser extends Test {
   @Override
   public void visit(State state, RandWalkEnv env, Properties props) throws Exception {
     String sysPrincipal = WalkingSecurity.get(state, env).getSysUserName();
-    try (AccumuloClient client = env.createClient(sysPrincipal, WalkingSecurity.get(state, env)
-        .getSysToken())) {
+    try (AccumuloClient client = env.createClient(sysPrincipal,
+        WalkingSecurity.get(state, env).getSysToken())) {
 
       String tableUserName = WalkingSecurity.get(state, env).getTabUserName();
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/SecurityFixture.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/SecurityFixture.java
index 9d34282..ecbd807 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/SecurityFixture.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/SecurityFixture.java
@@ -25,8 +25,8 @@ import org.apache.accumulo.core.conf.ClientProperty;
 import org.apache.accumulo.core.security.Authorizations;
 import org.apache.accumulo.core.security.SystemPermission;
 import org.apache.accumulo.core.security.TablePermission;
-import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.Fixture;
+import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 
 public class SecurityFixture extends Fixture {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/TableOp.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/TableOp.java
index 8e671c4..7e217c3 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/TableOp.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/TableOp.java
@@ -57,8 +57,8 @@ public class TableOp extends Test {
   @Override
   public void visit(State state, RandWalkEnv env, Properties props) throws Exception {
     String tablePrincipal = WalkingSecurity.get(state, env).getTabUserName();
-    try (AccumuloClient client = env.createClient(tablePrincipal, WalkingSecurity.get(state, env)
-        .getTabToken())) {
+    try (AccumuloClient client = env.createClient(tablePrincipal,
+        WalkingSecurity.get(state, env).getTabToken())) {
       TableOperations tableOps = client.tableOperations();
       SecurityOperations secOps = client.securityOperations();
 
@@ -88,8 +88,8 @@ public class TableOp extends Test {
           Authorizations auths = secOps.getUserAuthorizations(tablePrincipal);
           boolean ambiguousZone = WalkingSecurity.get(state, env).inAmbiguousZone(client.whoami(),
               tp);
-          boolean ambiguousAuths = WalkingSecurity.get(state, env).ambiguousAuthorizations(
-              client.whoami());
+          boolean ambiguousAuths = WalkingSecurity.get(state, env)
+              .ambiguousAuthorizations(client.whoami());
 
           Scanner scan = null;
           try {
@@ -101,8 +101,8 @@ public class TableOp extends Test {
               Key k = entry.getKey();
               seen++;
               if (!auths.contains(k.getColumnVisibilityData()) && !ambiguousAuths)
-                throw new AccumuloException("Got data I should not be capable of seeing: " + k
-                    + " table " + tableName);
+                throw new AccumuloException(
+                    "Got data I should not be capable of seeing: " + k + " table " + tableName);
             }
             if (!canRead && !ambiguousZone)
               throw new AccumuloException(
@@ -117,8 +117,8 @@ public class TableOp extends Test {
               throw new AccumuloException("Got mismatched amounts of data");
           } catch (TableNotFoundException tnfe) {
             if (tableExists)
-              throw new AccumuloException(
-                  "Accumulo and test suite out of sync: table " + tableName, tnfe);
+              throw new AccumuloException("Accumulo and test suite out of sync: table " + tableName,
+                  tnfe);
             return;
           } catch (AccumuloSecurityException ae) {
             if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
@@ -137,8 +137,8 @@ public class TableOp extends Test {
             throw new AccumuloException("Unexpected exception!", ae);
           } catch (RuntimeException re) {
             if (re.getCause() instanceof AccumuloSecurityException
-                && ((AccumuloSecurityException) re.getCause()).getSecurityErrorCode().equals(
-                    SecurityErrorCode.PERMISSION_DENIED)) {
+                && ((AccumuloSecurityException) re.getCause()).getSecurityErrorCode()
+                    .equals(SecurityErrorCode.PERMISSION_DENIED)) {
               if (canRead && !ambiguousZone)
                 throw new AccumuloException(
                     "Table read permission out of sync with Accumulo: table " + tableName,
@@ -147,8 +147,8 @@ public class TableOp extends Test {
                 return;
             }
             if (re.getCause() instanceof AccumuloSecurityException
-                && ((AccumuloSecurityException) re.getCause()).getSecurityErrorCode().equals(
-                    SecurityErrorCode.BAD_AUTHORIZATIONS)) {
+                && ((AccumuloSecurityException) re.getCause()).getSecurityErrorCode()
+                    .equals(SecurityErrorCode.BAD_AUTHORIZATIONS)) {
               if (ambiguousAuths)
                 return;
               else
@@ -248,8 +248,8 @@ public class TableOp extends Test {
           } catch (AccumuloSecurityException ae) {
             if (ae.getSecurityErrorCode().equals(SecurityErrorCode.PERMISSION_DENIED)) {
               if (secOps.hasTablePermission(tablePrincipal, tableName, TablePermission.BULK_IMPORT))
-                throw new AccumuloException("Bulk Import failed when it should have worked: "
-                    + tableName);
+                throw new AccumuloException(
+                    "Bulk Import failed when it should have worked: " + tableName);
               return;
             } else if (ae.getSecurityErrorCode().equals(SecurityErrorCode.BAD_CREDENTIALS)) {
               if (WalkingSecurity.get(state, env).userPassTransient(client.whoami()))
@@ -263,8 +263,8 @@ public class TableOp extends Test {
           fs.delete(fail, true);
 
           if (!secOps.hasTablePermission(tablePrincipal, tableName, TablePermission.BULK_IMPORT))
-            throw new AccumuloException("Bulk Import succeeded when it should have failed: " + dir
-                + " table " + tableName);
+            throw new AccumuloException(
+                "Bulk Import succeeded when it should have failed: " + dir + " table " + tableName);
           break;
         case ALTER_TABLE:
           boolean tablePerm;
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/Validate.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/Validate.java
index 466c163..2874f3a 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/Validate.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/Validate.java
@@ -45,8 +45,8 @@ public class Validate extends Test {
     if (tableExists != cloudTableExists)
       throw new AccumuloException("Table existance out of sync");
 
-    boolean tableUserExists = WalkingSecurity.get(state, env).userExists(
-        WalkingSecurity.get(state, env).getTabUserName());
+    boolean tableUserExists = WalkingSecurity.get(state, env)
+        .userExists(WalkingSecurity.get(state, env).getTabUserName());
     boolean cloudTableUserExists = client.securityOperations().listLocalUsers()
         .contains(WalkingSecurity.get(state, env).getTabUserName());
     if (tableUserExists != cloudTableUserExists)
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/security/WalkingSecurity.java b/src/main/java/org/apache/accumulo/testing/randomwalk/security/WalkingSecurity.java
index 695fdd1..9d81b00 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/security/WalkingSecurity.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/security/WalkingSecurity.java
@@ -162,8 +162,8 @@ public class WalkingSecurity {
     setTabPerm(state, user, permission, table, false);
   }
 
-  public void cleanTablePermissions(String table) throws AccumuloSecurityException,
-      TableNotFoundException {
+  public void cleanTablePermissions(String table)
+      throws AccumuloSecurityException, TableNotFoundException {
     for (String user : new String[] {getSysUserName(), getTabUserName()}) {
       for (TablePermission tp : TablePermission.values()) {
         revokeTablePermission(user, table, tp);
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/BatchVerify.java b/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/BatchVerify.java
index 3788a4f..badef95 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/BatchVerify.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/BatchVerify.java
@@ -64,8 +64,8 @@ public class BatchVerify extends Test {
           rangeEnd = numWrites - 1;
         }
         count += rangeEnd - rangeStart + 1;
-        ranges.add(new Range(new Text(String.format("%010d", rangeStart)), new Text(String.format(
-            "%010d", rangeEnd))));
+        ranges.add(new Range(new Text(String.format("%010d", rangeStart)),
+            new Text(String.format("%010d", rangeEnd))));
       }
 
       ranges = Range.mergeOverlapping(ranges);
@@ -76,7 +76,8 @@ public class BatchVerify extends Test {
       if (count == 0 || ranges.size() == 0)
         return;
 
-      log.debug(String.format("scanning %d rows in the following %d ranges:", count, ranges.size()));
+      log.debug(
+          String.format("scanning %d rows in the following %d ranges:", count, ranges.size()));
       for (Range r : ranges) {
         log.debug(r.toString());
       }
@@ -97,8 +98,8 @@ public class BatchVerify extends Test {
       boolean done = false;
       for (Range r : ranges) {
         int start = Integer.parseInt(r.getStartKey().getRow().toString());
-        int end = Integer.parseInt(String.copyValueOf(r.getEndKey().getRow().toString()
-            .toCharArray(), 0, 10));
+        int end = Integer
+            .parseInt(String.copyValueOf(r.getEndKey().getRow().toString().toCharArray(), 0, 10));
         for (int i = start; i <= end; i++) {
 
           if (done) {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/MapRedVerifyTool.java b/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/MapRedVerifyTool.java
index 465e63f..f1e560c 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/MapRedVerifyTool.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/MapRedVerifyTool.java
@@ -21,11 +21,11 @@ import java.util.Iterator;
 import java.util.Properties;
 
 import org.apache.accumulo.core.client.Accumulo;
-import org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat;
-import org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat;
 import org.apache.accumulo.core.data.Key;
 import org.apache.accumulo.core.data.Mutation;
 import org.apache.accumulo.core.data.Value;
+import org.apache.accumulo.hadoop.mapreduce.AccumuloInputFormat;
+import org.apache.accumulo.hadoop.mapreduce.AccumuloOutputFormat;
 import org.apache.hadoop.conf.Configured;
 import org.apache.hadoop.io.IntWritable;
 import org.apache.hadoop.io.NullWritable;
@@ -71,8 +71,8 @@ public class MapRedVerifyTool extends Configured implements Tool {
       writeMutation(output, start, index);
     }
 
-    private void writeMutation(Context output, int start, int end) throws IOException,
-        InterruptedException {
+    private void writeMutation(Context output, int start, int end)
+        throws IOException, InterruptedException {
       Mutation m = new Mutation(new Text(String.format("%010d", start)));
       m.put(new Text(String.format("%010d", end)), new Text(""), new Value(new byte[0]));
       output.write(null, m);
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/SequentialFixture.java b/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/SequentialFixture.java
index f949cce..937458a 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/SequentialFixture.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/sequential/SequentialFixture.java
@@ -22,8 +22,8 @@ import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.client.TableExistsException;
-import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.Fixture;
+import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 
 public class SequentialFixture extends Fixture {
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/BulkInsert.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/BulkInsert.java
index ee59a4c..cbc6589 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/BulkInsert.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/BulkInsert.java
@@ -131,10 +131,10 @@ public class BulkInsert extends Test {
     dataWriter.close();
     indexWriter.close();
 
-    sort(state, env, fs, dataTableName, rootDir + "/data.seq", rootDir + "/data_bulk", rootDir
-        + "/data_work", maxSplits);
-    sort(state, env, fs, indexTableName, rootDir + "/index.seq", rootDir + "/index_bulk", rootDir
-        + "/index_work", maxSplits);
+    sort(state, env, fs, dataTableName, rootDir + "/data.seq", rootDir + "/data_bulk",
+        rootDir + "/data_work", maxSplits);
+    sort(state, env, fs, indexTableName, rootDir + "/index.seq", rootDir + "/index_bulk",
+        rootDir + "/index_work", maxSplits);
 
     bulkImport(fs, state, env, dataTableName, rootDir, "data");
     bulkImport(fs, state, env, indexTableName, rootDir, "index");
@@ -171,8 +171,9 @@ public class BulkInsert extends Test {
   private void sort(State state, RandWalkEnv env, FileSystem fs, String tableName, String seqFile,
       String outputDir, String workDir, int maxSplits) throws Exception {
 
-    PrintStream out = new PrintStream(new BufferedOutputStream(fs.create(new Path(workDir
-        + "/splits.txt"))), false, UTF_8.name());
+    PrintStream out = new PrintStream(
+        new BufferedOutputStream(fs.create(new Path(workDir + "/splits.txt"))), false,
+        UTF_8.name());
 
     AccumuloClient client = env.getAccumuloClient();
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CloneIndex.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CloneIndex.java
index 4d232b3..334beae 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CloneIndex.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CloneIndex.java
@@ -35,10 +35,8 @@ public class CloneIndex extends Test {
     long t1 = System.currentTimeMillis();
     env.getAccumuloClient().tableOperations().flush(indexTableName, null, null, true);
     long t2 = System.currentTimeMillis();
-    env.getAccumuloClient()
-        .tableOperations()
-        .clone(indexTableName, tmpIndexTableName, false, new HashMap<String,String>(),
-            new HashSet<String>());
+    env.getAccumuloClient().tableOperations().clone(indexTableName, tmpIndexTableName, false,
+        new HashMap<String,String>(), new HashSet<String>());
     long t3 = System.currentTimeMillis();
 
     log.debug("Cloned " + tmpIndexTableName + " from " + indexTableName + " flush: " + (t2 - t1)
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CompactFilter.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CompactFilter.java
index 1406acd..3c09149 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CompactFilter.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/CompactFilter.java
@@ -57,8 +57,8 @@ public class CompactFilter extends Test {
     documentFilters.add(is);
 
     long t1 = System.currentTimeMillis();
-    env.getAccumuloClient().tableOperations()
-        .compact(docTableName, null, null, documentFilters, true, true);
+    env.getAccumuloClient().tableOperations().compact(docTableName, null, null, documentFilters,
+        true, true);
     long t2 = System.currentTimeMillis();
     long t3 = t2 - t1;
 
@@ -70,12 +70,12 @@ public class CompactFilter extends Test {
     indexFilters.add(is);
 
     t1 = System.currentTimeMillis();
-    env.getAccumuloClient().tableOperations()
-        .compact(indexTableName, null, null, indexFilters, true, true);
+    env.getAccumuloClient().tableOperations().compact(indexTableName, null, null, indexFilters,
+        true, true);
     t2 = System.currentTimeMillis();
 
-    log.debug("Filtered documents using compaction iterators " + regex + " " + (t3) + " "
-        + (t2 - t1));
+    log.debug(
+        "Filtered documents using compaction iterators " + regex + " " + (t3) + " " + (t2 - t1));
 
     BatchScanner bscanner = env.getAccumuloClient().createBatchScanner(docTableName,
         new Authorizations(), 10);
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ExportIndex.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ExportIndex.java
index f55b363..7c81e3c 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ExportIndex.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ExportIndex.java
@@ -55,8 +55,8 @@ public class ExportIndex extends Test {
 
     // disable spits, so that splits can be compared later w/o worrying one
     // table splitting and the other not
-    env.getAccumuloClient().tableOperations()
-        .setProperty(indexTableName, Property.TABLE_SPLIT_THRESHOLD.getKey(), "20G");
+    env.getAccumuloClient().tableOperations().setProperty(indexTableName,
+        Property.TABLE_SPLIT_THRESHOLD.getKey(), "20G");
 
     long t1 = System.currentTimeMillis();
 
@@ -70,8 +70,8 @@ public class ExportIndex extends Test {
     long t3 = System.currentTimeMillis();
 
     // copy files
-    BufferedReader reader = new BufferedReader(new InputStreamReader(fs.open(new Path(exportDir,
-        "distcp.txt")), UTF_8));
+    BufferedReader reader = new BufferedReader(
+        new InputStreamReader(fs.open(new Path(exportDir, "distcp.txt")), UTF_8));
     String file = null;
     while ((file = reader.readLine()) != null) {
       Path src = new Path(file);
@@ -91,10 +91,10 @@ public class ExportIndex extends Test {
     fs.delete(new Path(exportDir), true);
     fs.delete(new Path(copyDir), true);
 
-    HashSet<Text> splits1 = new HashSet<>(env.getAccumuloClient().tableOperations()
-        .listSplits(indexTableName));
-    HashSet<Text> splits2 = new HashSet<>(env.getAccumuloClient().tableOperations()
-        .listSplits(tmpIndexTableName));
+    HashSet<Text> splits1 = new HashSet<>(
+        env.getAccumuloClient().tableOperations().listSplits(indexTableName));
+    HashSet<Text> splits2 = new HashSet<>(
+        env.getAccumuloClient().tableOperations().listSplits(tmpIndexTableName));
 
     if (!splits1.equals(splits2))
       throw new Exception("Splits not equals " + indexTableName + " " + tmpIndexTableName);
@@ -113,10 +113,10 @@ public class ExportIndex extends Test {
       throw new Exception("Props not equals " + indexTableName + " " + tmpIndexTableName);
 
     // unset the split threshold
-    env.getAccumuloClient().tableOperations()
-        .removeProperty(indexTableName, Property.TABLE_SPLIT_THRESHOLD.getKey());
-    env.getAccumuloClient().tableOperations()
-        .removeProperty(tmpIndexTableName, Property.TABLE_SPLIT_THRESHOLD.getKey());
+    env.getAccumuloClient().tableOperations().removeProperty(indexTableName,
+        Property.TABLE_SPLIT_THRESHOLD.getKey());
+    env.getAccumuloClient().tableOperations().removeProperty(tmpIndexTableName,
+        Property.TABLE_SPLIT_THRESHOLD.getKey());
 
     log.debug("Imported " + tmpIndexTableName + " from " + indexTableName + " flush: " + (t2 - t1)
         + "ms export: " + (t3 - t2) + "ms copy:" + (t4 - t3) + "ms import:" + (t5 - t4) + "ms");
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Grep.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Grep.java
index eb1dc08..3f289d0 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Grep.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Grep.java
@@ -72,8 +72,8 @@ public class Grep extends Test {
 
     for (int i = 0; i < words.length; i++) {
       IteratorSetting more = new IteratorSetting(20 + i, "ii" + i, RegExFilter.class);
-      RegExFilter
-          .setRegexs(more, null, null, null, "(^|(.*\\s))" + words[i] + "($|(\\s.*))", false);
+      RegExFilter.setRegexs(more, null, null, null, "(^|(.*\\s))" + words[i] + "($|(\\s.*))",
+          false);
       bs.addScanIterator(more);
     }
 
@@ -88,9 +88,9 @@ public class Grep extends Test {
     bs.close();
 
     if (!documentsFoundInIndex.equals(documentsFoundByGrep)) {
-      throw new Exception("Set of documents found not equal for words "
-          + Arrays.asList(words).toString() + " " + documentsFoundInIndex + " "
-          + documentsFoundByGrep);
+      throw new Exception(
+          "Set of documents found not equal for words " + Arrays.asList(words).toString() + " "
+              + documentsFoundInIndex + " " + documentsFoundByGrep);
     }
 
     log.debug("Grep and index agree " + Arrays.asList(words).toString() + " "
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Merge.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Merge.java
index eaa9253..0420eb0 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Merge.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/Merge.java
@@ -45,8 +45,8 @@ public class Merge extends Test {
       // throw an excpetion so that test will die an no further changes to
       // table will occur...
       // this way table is left as is for debugging.
-      throw new Exception("There are more tablets after a merge: " + splits.size() + " was "
-          + splitSet.size());
+      throw new Exception(
+          "There are more tablets after a merge: " + splits.size() + " was " + splitSet.size());
     }
   }
 
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ShardFixture.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ShardFixture.java
index e451ef1..cdcd0e6 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ShardFixture.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/ShardFixture.java
@@ -25,8 +25,8 @@ import org.apache.accumulo.core.client.AccumuloClient;
 import org.apache.accumulo.core.client.MultiTableBatchWriter;
 import org.apache.accumulo.core.client.MutationsRejectedException;
 import org.apache.accumulo.core.conf.Property;
-import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.Fixture;
+import org.apache.accumulo.testing.randomwalk.RandWalkEnv;
 import org.apache.accumulo.testing.randomwalk.State;
 import org.apache.hadoop.io.Text;
 import org.slf4j.Logger;
@@ -66,10 +66,10 @@ public class ShardFixture extends Fixture {
     log.info("Added " + splits.size() + " splits to " + name);
 
     if (enableCache) {
-      client.tableOperations()
-          .setProperty(name, Property.TABLE_INDEXCACHE_ENABLED.getKey(), "true");
-      client.tableOperations()
-          .setProperty(name, Property.TABLE_BLOCKCACHE_ENABLED.getKey(), "true");
+      client.tableOperations().setProperty(name, Property.TABLE_INDEXCACHE_ENABLED.getKey(),
+          "true");
+      client.tableOperations().setProperty(name, Property.TABLE_BLOCKCACHE_ENABLED.getKey(),
+          "true");
 
       log.info("Enabled caching for table " + name);
     }
diff --git a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/VerifyIndex.java b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/VerifyIndex.java
index 9b2741b..554d541 100644
--- a/src/main/java/org/apache/accumulo/testing/randomwalk/shard/VerifyIndex.java
+++ b/src/main/java/org/apache/accumulo/testing/randomwalk/shard/VerifyIndex.java
@@ -63,8 +63,8 @@ public class VerifyIndex extends Test {
     }
 
     if (iter.hasNext())
-      throw new Exception("index rebuild mismatch " + iter.next().getKey() + " "
-          + tmpIndexTableName);
+      throw new Exception(
+          "index rebuild mismatch " + iter.next().getKey() + " " + tmpIndexTableName);
 
     log.debug("Verified " + count + " index entries ");
 
diff --git a/src/main/java/org/apache/accumulo/testing/scalability/Run.java b/src/main/java/org/apache/accumulo/testing/scalability/Run.java
index 1c93390..f8d8479 100644
--- a/src/main/java/org/apache/accumulo/testing/scalability/Run.java
+++ b/src/main/java/org/apache/accumulo/testing/scalability/Run.java
@@ -78,8 +78,9 @@ public class Run {
       log.error("Error loading config file.", e);
     }
 
-    ScaleTest test = (ScaleTest) Class.forName(
-        String.format("org.apache.accumulo.test.scalability.%s", opts.testId)).newInstance();
+    ScaleTest test = (ScaleTest) Class
+        .forName(String.format("org.apache.accumulo.test.scalability.%s", opts.testId))
+        .newInstance();
 
     test.init(scaleProps, testProps, opts.numTabletServers);
 
diff --git a/src/main/java/org/apache/accumulo/testing/scalability/ScaleTest.java b/src/main/java/org/apache/accumulo/testing/scalability/ScaleTest.java
index f36ba40..fe57264 100644
--- a/src/main/java/org/apache/accumulo/testing/scalability/ScaleTest.java
+++ b/src/main/java/org/apache/accumulo/testing/scalability/ScaleTest.java
@@ -48,8 +48,8 @@ public abstract class ScaleTest {
     String password = this.scaleProps.getProperty("PASSWORD");
     System.out.println(password);
 
-    client = Accumulo.newClient().to(instanceName, zookeepers)
-        .as(user, new PasswordToken(password)).build();
+    client = Accumulo.newClient().to(instanceName, zookeepers).as(user, new PasswordToken(password))
+        .build();
   }
 
   protected void startTimer() {
diff --git a/src/main/java/org/apache/accumulo/testing/stress/ScanOpts.java b/src/main/java/org/apache/accumulo/testing/stress/ScanOpts.java
index b403e98..05e832d 100644
--- a/src/main/java/org/apache/accumulo/testing/stress/ScanOpts.java
+++ b/src/main/java/org/apache/accumulo/testing/stress/ScanOpts.java
@@ -16,9 +16,10 @@
  */
 package org.apache.accumulo.testing.stress;
 
-import com.beust.jcommander.Parameter;
 import org.apache.accumulo.testing.cli.ClientOpts;
 
+import com.beust.jcommander.Parameter;
+
 class ScanOpts extends ClientOpts {
   @Parameter(names = {"-t", "--table"}, description = "table to use")
   String tableName = WriteOptions.DEFAULT_TABLE;
diff --git a/src/main/java/org/apache/accumulo/testing/stress/Write.java b/src/main/java/org/apache/accumulo/testing/stress/Write.java
index ec07156..2b92b59 100644
--- a/src/main/java/org/apache/accumulo/testing/stress/Write.java
+++ b/src/main/java/org/apache/accumulo/testing/stress/Write.java
@@ -52,15 +52,15 @@ public class Write {
       }
 
       DataWriter dw = new DataWriter(c.createBatchWriter(opts.tableName), new RandomMutations(
-      // rows
+          // rows
           new RandomByteArrays(new RandomWithinRange(opts.row_seed, opts.rowMin(), opts.rowMax())),
           // cfs
           new RandomByteArrays(new RandomWithinRange(opts.cf_seed, opts.cfMin(), opts.cfMax())),
           // cqs
           new RandomByteArrays(new RandomWithinRange(opts.cq_seed, opts.cqMin(), opts.cqMax())),
           // vals
-          new RandomByteArrays(new RandomWithinRange(opts.value_seed, opts.valueMin(),
-              opts.valueMax())),
+          new RandomByteArrays(
+              new RandomWithinRange(opts.value_seed, opts.valueMin(), opts.valueMax())),
           // number of cells per row
           new RandomWithinRange(opts.row_width_seed, opts.rowWidthMin(), opts.rowWidthMax()),
           // max cells per mutation
diff --git a/src/main/java/org/apache/accumulo/testing/stress/WriteOptions.java b/src/main/java/org/apache/accumulo/testing/stress/WriteOptions.java
index 33ec845..a8f8321 100644
--- a/src/main/java/org/apache/accumulo/testing/stress/WriteOptions.java
+++ b/src/main/java/org/apache/accumulo/testing/stress/WriteOptions.java
@@ -16,9 +16,10 @@
  */
 package org.apache.accumulo.testing.stress;
 
-import com.beust.jcommander.Parameter;
 import org.apache.accumulo.testing.cli.ClientOpts;
 
+import com.beust.jcommander.Parameter;
+
 class WriteOptions extends ClientOpts {
   static final String DEFAULT_TABLE = "stress_test";
   static final int DEFAULT_MIN = 1, DEFAULT_MAX = 128, DEFAULT_SPREAD = DEFAULT_MAX - DEFAULT_MIN;
@@ -123,17 +124,15 @@ class WriteOptions extends ClientOpts {
 
     if (min_ref == null && max_ref != null) {
       // we don't support just specifying a max yet
-      throw new IllegalArgumentException(
-          String
-              .format(
-                  "[%s] Maximum value supplied, but no minimum. Must supply a minimum with a maximum value.",
-                  label));
+      throw new IllegalArgumentException(String.format(
+          "[%s] Maximum value supplied, but no minimum. Must supply a minimum with a maximum value.",
+          label));
     } else if (min_ref != null && max_ref != null) {
       // if a user supplied lower and upper bounds, we need to verify
       // that min <= max
       if (min_ref.compareTo(max_ref) > 0) {
-        throw new IllegalArgumentException(String.format(
-            "[%s] Min value (%d) is greater than max value (%d)", label, min_ref, max_ref));
+        throw new IllegalArgumentException(String
+            .format("[%s] Min value (%d) is greater than max value (%d)", label, min_ref, max_ref));
       }
     }
   }
diff --git a/src/main/java/org/apache/accumulo/testing/stress/package-info.java b/src/main/java/org/apache/accumulo/testing/stress/package-info.java
index 658938a..765a7fc 100644
--- a/src/main/java/org/apache/accumulo/testing/stress/package-info.java
+++ b/src/main/java/org/apache/accumulo/testing/stress/package-info.java
@@ -15,22 +15,26 @@
  * limitations under the License.
  */
 /**
- * This package contains utility classes designed to test Accumulo when large cells are being written. This is an attempt to observe the behavior Accumulo
- * displays when compacting and reading these cells.
+ * This package contains utility classes designed to test Accumulo when large cells are being
+ * written. This is an attempt to observe the behavior Accumulo displays when compacting and reading
+ * these cells.
  *
- * There are two components to this package: {@link org.apache.accumulo.testing.stress.Write} and {@link org.apache.accumulo.testing.stress.Scan}.
+ * There are two components to this package: {@link org.apache.accumulo.testing.stress.Write} and
+ * {@link org.apache.accumulo.testing.stress.Scan}.
  *
- * The {@link org.apache.accumulo.testing.stress.Write} provides facilities for writing random sized cells. Users can configure minimum and maximum
- * sized portions of a cell. The portions users can configure are the row, column family, column qualifier and value. Note that the sizes are uniformly
- * distributed between the minimum and maximum values. See {@link org.apache.accumulo.testing.stress.WriteOptions} for available options and default sizing
+ * The {@link org.apache.accumulo.testing.stress.Write} provides facilities for writing random sized
+ * cells. Users can configure minimum and maximum sized portions of a cell. The portions users can
+ * configure are the row, column family, column qualifier and value. Note that the sizes are
+ * uniformly distributed between the minimum and maximum values. See
+ * {@link org.apache.accumulo.testing.stress.WriteOptions} for available options and default sizing
  * information.
  *
- * The Scan provides users with the ability to query tables generated by the Write. It will pick a tablet at random and scan the entire range. The
- * amount of times this process is done is user configurable. By default, it happens 1,024 times. Users can also specify whether or not the scan should be
- * isolated or not.
+ * The Scan provides users with the ability to query tables generated by the Write. It will pick a
+ * tablet at random and scan the entire range. The amount of times this process is done is user
+ * configurable. By default, it happens 1,024 times. Users can also specify whether or not the scan
+ * should be isolated or not.
  *
- * There is no shared state intended by either of these services. This allows multiple clients to be run in parallel, either on the same host or distributed
- * across hosts.
+ * There is no shared state intended by either of these services. This allows multiple clients to be
+ * run in parallel, either on the same host or distributed across hosts.
  */
 package org.apache.accumulo.testing.stress;
-
diff --git a/src/test/java/org/apache/accumulo/testing/randomwalk/FrameworkTest.java b/src/test/java/org/apache/accumulo/testing/randomwalk/FrameworkTest.java
index de78794..d56eed5 100644
--- a/src/test/java/org/apache/accumulo/testing/randomwalk/FrameworkTest.java
+++ b/src/test/java/org/apache/accumulo/testing/randomwalk/FrameworkTest.java
@@ -39,8 +39,8 @@ public class FrameworkTest {
   // Need to use fully qualified name here because of conflict with
   // org.apache.accumulo.testing.randomwalk.Test
   @org.junit.Test
-  public void testXML() throws SAXException, URISyntaxException, ParserConfigurationException,
-      IOException {
+  public void testXML()
+      throws SAXException, URISyntaxException, ParserConfigurationException, IOException {
     SchemaFactory sf = SchemaFactory.newInstance(XMLConstants.W3C_XML_SCHEMA_NS_URI);
     Schema moduleSchema = sf.newSchema(getFile("/randomwalk/module.xsd"));
 
@@ -50,8 +50,8 @@ public class FrameworkTest {
     DocumentBuilder docbuilder = dbf.newDocumentBuilder();
     Document document = docbuilder.parse(getFile("/randomwalk/modules/unit/Basic.xml"));
 
-    assertNotEquals("Parsing randomwalk xml should result in nodes.", 0, document.getChildNodes()
-        .getLength());
+    assertNotEquals("Parsing randomwalk xml should result in nodes.", 0,
+        document.getChildNodes().getLength());
   }
 
   private File getFile(String resource) throws URISyntaxException {


[accumulo-testing] 01/02: Update pom.xml

Posted by ct...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

ctubbsii pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/accumulo-testing.git

commit 31ee5bd47226b5bd31437ecfe63a99f98738a22e
Author: Christopher Tubbs <ct...@apache.org>
AuthorDate: Thu Mar 7 17:59:14 2019 -0500

    Update pom.xml
    
    Update apache parent pom and update formatting plugins
---
 pom.xml | 314 +++++++++++++++++++++++++++++++++++++++++-----------------------
 1 file changed, 201 insertions(+), 113 deletions(-)

diff --git a/pom.xml b/pom.xml
index b1a606c..8ec588a 100644
--- a/pom.xml
+++ b/pom.xml
@@ -17,31 +17,26 @@
 -->
 <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
   <modelVersion>4.0.0</modelVersion>
-
   <parent>
     <groupId>org.apache</groupId>
     <artifactId>apache</artifactId>
-    <version>18</version>
+    <version>21</version>
   </parent>
-
   <groupId>org.apache.accumulo</groupId>
   <artifactId>accumulo-testing</artifactId>
   <version>2.0.0-SNAPSHOT</version>
-  <packaging>jar</packaging>
-
   <name>Apache Accumulo Testing</name>
   <description>Testing tools for Apache Accumulo</description>
-
   <properties>
     <accumulo.version>2.0.0-alpha-2</accumulo.version>
-    <hadoop.version>3.0.3</hadoop.version>
-    <zookeeper.version>3.4.9</zookeeper.version>
-    <slf4j.version>1.7.21</slf4j.version>
+    <eclipseFormatterStyle>${project.basedir}/contrib/Eclipse-Accumulo-Codestyle.xml</eclipseFormatterStyle>
+    <hadoop.version>3.1.1</hadoop.version>
     <maven.compiler.source>1.8</maven.compiler.source>
     <maven.compiler.target>1.8</maven.compiler.target>
-    <formatter.config>${project.basedir}/contrib/Eclipse-Accumulo-Codestyle.xml</formatter.config>
+    <slf4j.version>1.7.25</slf4j.version>
+    <spotbugs.version>3.1.7</spotbugs.version>
+    <zookeeper.version>3.4.13</zookeeper.version>
   </properties>
-
   <dependencyManagement>
     <dependencies>
       <dependency>
@@ -51,7 +46,82 @@
       </dependency>
     </dependencies>
   </dependencyManagement>
-
+  <dependencies>
+    <dependency>
+      <groupId>com.github.spotbugs</groupId>
+      <artifactId>spotbugs-annotations</artifactId>
+      <version>${spotbugs.version}</version>
+      <optional>true</optional>
+    </dependency>
+    <dependency>
+      <groupId>com.google.guava</groupId>
+      <artifactId>guava</artifactId>
+      <version>26.0-jre</version>
+    </dependency>
+    <dependency>
+      <groupId>commons-codec</groupId>
+      <artifactId>commons-codec</artifactId>
+      <version>1.11</version>
+    </dependency>
+    <dependency>
+      <groupId>commons-configuration</groupId>
+      <artifactId>commons-configuration</artifactId>
+      <version>1.10</version>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.accumulo</groupId>
+      <artifactId>accumulo-core</artifactId>
+      <version>${accumulo.version}</version>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.accumulo</groupId>
+      <artifactId>accumulo-hadoop-mapreduce</artifactId>
+      <version>${accumulo.version}</version>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.commons</groupId>
+      <artifactId>commons-math3</artifactId>
+      <version>3.6.1</version>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.hadoop</groupId>
+      <artifactId>hadoop-client-api</artifactId>
+      <version>${hadoop.version}</version>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.hadoop</groupId>
+      <artifactId>hadoop-client-runtime</artifactId>
+      <version>${hadoop.version}</version>
+    </dependency>
+    <dependency>
+      <groupId>org.slf4j</groupId>
+      <artifactId>slf4j-api</artifactId>
+      <version>${slf4j.version}</version>
+    </dependency>
+    <dependency>
+      <groupId>org.slf4j</groupId>
+      <artifactId>slf4j-log4j12</artifactId>
+      <version>${slf4j.version}</version>
+    </dependency>
+    <dependency>
+      <groupId>junit</groupId>
+      <artifactId>junit</artifactId>
+      <version>4.12</version>
+      <scope>test</scope>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.accumulo</groupId>
+      <artifactId>accumulo-minicluster</artifactId>
+      <version>${accumulo.version}</version>
+      <scope>test</scope>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.accumulo</groupId>
+      <artifactId>accumulo-test</artifactId>
+      <version>${accumulo.version}</version>
+      <scope>test</scope>
+    </dependency>
+  </dependencies>
   <build>
     <pluginManagement>
       <plugins>
@@ -61,17 +131,73 @@
           <artifactId>maven-bundle-plugin</artifactId>
           <version>3.0.1</version>
         </plugin>
+        <plugin>
+          <groupId>com.github.ekryd.sortpom</groupId>
+          <artifactId>sortpom-maven-plugin</artifactId>
+          <version>2.8.0</version>
+          <configuration>
+            <predefinedSortOrder>recommended_2008_06</predefinedSortOrder>
+            <createBackupFile>false</createBackupFile>
+            <lineSeparator>\n</lineSeparator>
+            <expandEmptyElements>false</expandEmptyElements>
+            <nrOfIndentSpace>2</nrOfIndentSpace>
+            <sortDependencies>scope,groupId,artifactId</sortDependencies>
+            <sortProperties>true</sortProperties>
+            <verifyFail>Stop</verifyFail>
+          </configuration>
+        </plugin>
+        <plugin>
+          <groupId>net.revelc.code.formatter</groupId>
+          <artifactId>formatter-maven-plugin</artifactId>
+          <version>2.8.1</version>
+          <configuration>
+            <configFile>${eclipseFormatterStyle}</configFile>
+            <compilerCompliance>${maven.compiler.source}</compilerCompliance>
+            <compilerSource>${maven.compiler.source}</compilerSource>
+            <compilerTargetPlatform>${maven.compiler.target}</compilerTargetPlatform>
+            <excludes>
+              <exclude>**/thrift/*.java</exclude>
+              <exclude>**/proto/*.java</exclude>
+            </excludes>
+            <lineEnding>LF</lineEnding>
+            <overrideConfigCompilerVersion>true</overrideConfigCompilerVersion>
+            <skipJsFormatting>true</skipJsFormatting>
+            <skipHtmlFormatting>true</skipHtmlFormatting>
+            <skipXmlFormatting>true</skipXmlFormatting>
+            <skipJsonFormatting>true</skipJsonFormatting>
+            <skipCssFormatting>true</skipCssFormatting>
+          </configuration>
+        </plugin>
+        <plugin>
+          <groupId>net.revelc.code</groupId>
+          <artifactId>impsort-maven-plugin</artifactId>
+          <version>1.2.0</version>
+          <configuration>
+            <removeUnused>true</removeUnused>
+            <groups>java.,javax.,org.,com.</groups>
+            <excludes>
+              <exclude>**/thrift/*.java</exclude>
+            </excludes>
+          </configuration>
+        </plugin>
       </plugins>
     </pluginManagement>
     <plugins>
       <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
         <artifactId>maven-compiler-plugin</artifactId>
-        <version>3.1</version>
         <configuration>
-          <source>${maven.compiler.source}</source>
-          <target>${maven.compiler.target}</target>
           <optimize>true</optimize>
-          <encoding>UTF-8</encoding>
+          <showDeprecation>true</showDeprecation>
+          <showWarnings>true</showWarnings>
+          <compilerArgs>
+            <arg>-Xlint:all</arg>
+            <arg>-Xlint:-processing</arg>
+            <!-- suppress try to ignore unused variable in try-with-resources -->
+            <arg>-Xlint:-try</arg>
+            <arg>-Xmaxwarns</arg>
+            <arg>5</arg>
+          </compilerArgs>
         </configuration>
       </plugin>
       <plugin>
@@ -103,21 +229,6 @@
         </configuration>
       </plugin>
       <plugin>
-        <groupId>net.revelc.code</groupId>
-        <artifactId>formatter-maven-plugin</artifactId>
-        <version>0.5.2</version>
-        <configuration>
-          <configFile>${formatter.config}</configFile>
-        </configuration>
-        <executions>
-          <execution>
-            <goals>
-              <goal>format</goal>
-            </goals>
-          </execution>
-        </executions>
-      </plugin>
-      <plugin>
         <!-- This was added to ensure project only uses public API -->
         <groupId>org.apache.maven.plugins</groupId>
         <artifactId>maven-checkstyle-plugin</artifactId>
@@ -125,14 +236,14 @@
         <configuration>
           <configLocation>contrib/checkstyle.xml</configLocation>
         </configuration>
-	<dependencies>
-            <dependency>
-              <groupId>com.puppycrawl.tools</groupId>
-              <artifactId>checkstyle</artifactId>
-              <version>8.17</version>
-            </dependency>
-	</dependencies>
-	<executions>
+        <dependencies>
+          <dependency>
+            <groupId>com.puppycrawl.tools</groupId>
+            <artifactId>checkstyle</artifactId>
+            <version>8.17</version>
+          </dependency>
+        </dependencies>
+        <executions>
           <execution>
             <id>check-style</id>
             <goals>
@@ -143,80 +254,58 @@
       </plugin>
     </plugins>
   </build>
-
-  <dependencies>
-    <dependency>
-      <groupId>com.google.guava</groupId>
-      <artifactId>guava</artifactId>
-      <version>26.0-jre</version>
-    </dependency>
-    <dependency>
-      <groupId>commons-configuration</groupId>
-      <artifactId>commons-configuration</artifactId>
-      <version>1.10</version>
-    </dependency>
-    <dependency>
-      <groupId>commons-codec</groupId>
-      <artifactId>commons-codec</artifactId>
-      <version>1.11</version>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.commons</groupId>
-      <artifactId>commons-math3</artifactId>
-      <version>3.6.1</version>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-core</artifactId>
-      <version>${accumulo.version}</version>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-hadoop-mapreduce</artifactId>
-      <version>${accumulo.version}</version>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-minicluster</artifactId>
-      <version>${accumulo.version}</version>
-      <scope>test</scope>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.accumulo</groupId>
-      <artifactId>accumulo-test</artifactId>
-      <version>${accumulo.version}</version>
-      <scope>test</scope>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.hadoop</groupId>
-      <artifactId>hadoop-client-api</artifactId>
-      <version>${hadoop.version}</version>
-    </dependency>
-    <dependency>
-      <groupId>org.apache.hadoop</groupId>
-      <artifactId>hadoop-client-runtime</artifactId>
-      <version>${hadoop.version}</version>
-    </dependency>
-    <dependency>
-      <groupId>org.slf4j</groupId>
-      <artifactId>slf4j-api</artifactId>
-      <version>${slf4j.version}</version>
-    </dependency>
-    <dependency>
-      <groupId>org.slf4j</groupId>
-      <artifactId>slf4j-log4j12</artifactId>
-      <version>${slf4j.version}</version>
-    </dependency>
-    <dependency>
-      <groupId>junit</groupId>
-      <artifactId>junit</artifactId>
-      <version>4.12</version>
-      <scope>test</scope>
-    </dependency>
-  </dependencies>
-
   <profiles>
     <profile>
+      <!-- on by default, but disable with '-P !autoformat' or '-DskipFormat' -->
+      <id>autoformat</id>
+      <activation>
+        <property>
+          <name>!skipFormat</name>
+        </property>
+      </activation>
+      <build>
+        <plugins>
+          <plugin>
+            <groupId>com.github.ekryd.sortpom</groupId>
+            <artifactId>sortpom-maven-plugin</artifactId>
+            <executions>
+              <execution>
+                <id>sort-pom</id>
+                <goals>
+                  <goal>sort</goal>
+                </goals>
+                <phase>process-sources</phase>
+              </execution>
+            </executions>
+          </plugin>
+          <plugin>
+            <groupId>net.revelc.code.formatter</groupId>
+            <artifactId>formatter-maven-plugin</artifactId>
+            <executions>
+              <execution>
+                <id>format-java-source</id>
+                <goals>
+                  <goal>format</goal>
+                </goals>
+              </execution>
+            </executions>
+          </plugin>
+          <plugin>
+            <groupId>net.revelc.code</groupId>
+            <artifactId>impsort-maven-plugin</artifactId>
+            <executions>
+              <execution>
+                <id>sort-imports</id>
+                <goals>
+                  <goal>sort</goal>
+                </goals>
+              </execution>
+            </executions>
+          </plugin>
+        </plugins>
+      </build>
+    </profile>
+    <profile>
       <id>create-shade-jar</id>
       <build>
         <plugins>
@@ -252,7 +341,7 @@
                   </filters>
                   <transformers>
                     <!-- Hadoop uses service loader to find filesystem impls, without this may not find them -->
-                    <transformer implementation="org.apache.maven.plugins.shade.resource.ServicesResourceTransformer"/>
+                    <transformer implementation="org.apache.maven.plugins.shade.resource.ServicesResourceTransformer" />
                   </transformers>
                 </configuration>
               </execution>
@@ -262,5 +351,4 @@
       </build>
     </profile>
   </profiles>
-
 </project>