You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by mb...@apache.org on 2015/09/23 19:00:57 UTC

[2/2] hbase git commit: HBASE-14464 Removed unused fs code

HBASE-14464 Removed unused fs code


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/0fa48ae5
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/0fa48ae5
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/0fa48ae5

Branch: refs/heads/branch-1
Commit: 0fa48ae56a97d414d732dc81c6b8c9899a58c74e
Parents: 8ed3729
Author: Matteo Bertozzi <ma...@cloudera.com>
Authored: Wed Sep 23 09:40:09 2015 -0700
Committer: Matteo Bertozzi <ma...@cloudera.com>
Committed: Wed Sep 23 09:56:24 2015 -0700

----------------------------------------------------------------------
 .../hadoop/hbase/master/MasterFileSystem.java   |  98 ---------------
 .../master/snapshot/SnapshotLogCleaner.java     | 103 ----------------
 .../hbase/master/snapshot/SnapshotManager.java  |   3 +-
 .../hadoop/hbase/snapshot/ExportSnapshot.java   |  16 +--
 .../hbase/snapshot/RestoreSnapshotHelper.java   |   2 +-
 .../hadoop/hbase/snapshot/SnapshotInfo.java     |  19 ---
 .../hadoop/hbase/snapshot/SnapshotManifest.java |   2 +-
 .../hbase/snapshot/SnapshotManifestV1.java      |   4 +-
 .../hbase/snapshot/SnapshotManifestV2.java      |   4 +-
 .../hbase/snapshot/SnapshotReferenceUtil.java   |  53 +-------
 .../org/apache/hadoop/hbase/util/FSUtils.java   |  38 ------
 .../org/apache/hadoop/hbase/util/FSVisitor.java | 121 -------------------
 .../org/apache/hadoop/hbase/TestNamespace.java  |  28 -----
 .../hbase/coprocessor/TestMasterObserver.java   |   5 -
 .../master/snapshot/TestSnapshotFileCache.java  |  40 ------
 .../master/snapshot/TestSnapshotLogCleaner.java |  85 -------------
 .../master/snapshot/TestSnapshotManager.java    |   2 -
 .../hbase/snapshot/TestExportSnapshot.java      |   7 --
 .../apache/hadoop/hbase/util/TestFSVisitor.java | 102 ----------------
 19 files changed, 12 insertions(+), 720 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
index e648c2e..02d18d6 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
@@ -42,7 +42,6 @@ import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.HRegionInfo;
 import org.apache.hadoop.hbase.HTableDescriptor;
-import org.apache.hadoop.hbase.InvalidFamilyOperationException;
 import org.apache.hadoop.hbase.RemoteExceptionHandler;
 import org.apache.hadoop.hbase.Server;
 import org.apache.hadoop.hbase.ServerName;
@@ -560,42 +559,6 @@ public class MasterFileSystem {
     }
   }
 
-  public void deleteRegion(HRegionInfo region) throws IOException {
-    HFileArchiver.archiveRegion(conf, fs, region);
-  }
-
-  public void deleteTable(TableName tableName) throws IOException {
-    fs.delete(FSUtils.getTableDir(rootdir, tableName), true);
-  }
-
-  /**
-   * Move the specified table to the hbase temp directory
-   * @param tableName Table name to move
-   * @return The temp location of the table moved
-   * @throws IOException in case of file-system failure
-   */
-  public Path moveTableToTemp(TableName tableName) throws IOException {
-    Path srcPath = FSUtils.getTableDir(rootdir, tableName);
-    Path tempPath = FSUtils.getTableDir(this.tempdir, tableName);
-
-    // Ensure temp exists
-    if (!fs.exists(tempPath.getParent()) && !fs.mkdirs(tempPath.getParent())) {
-      throw new IOException("HBase temp directory '" + tempPath.getParent() + "' creation failure.");
-    }
-
-    if (!fs.rename(srcPath, tempPath)) {
-      throw new IOException("Unable to move '" + srcPath + "' to temp '" + tempPath + "'");
-    }
-
-    return tempPath;
-  }
-
-  public void updateRegionInfo(HRegionInfo region) {
-    // TODO implement this.  i think this is currently broken in trunk i don't
-    //      see this getting updated.
-    //      @see HRegion.checkRegioninfoOnFilesystem()
-  }
-
   public void deleteFamilyFromFS(HRegionInfo region, byte[] familyName)
       throws IOException {
     // archive family store files
@@ -622,67 +585,6 @@ public class MasterFileSystem {
   }
 
   /**
-   * Delete column of a table
-   * @param tableName
-   * @param familyName
-   * @return Modified HTableDescriptor with requested column deleted.
-   * @throws IOException
-   */
-  public HTableDescriptor deleteColumn(TableName tableName, byte[] familyName)
-      throws IOException {
-    LOG.info("DeleteColumn. Table = " + tableName
-        + " family = " + Bytes.toString(familyName));
-    HTableDescriptor htd = this.services.getTableDescriptors().get(tableName);
-    htd.removeFamily(familyName);
-    this.services.getTableDescriptors().add(htd);
-    return htd;
-  }
-
-  /**
-   * Modify Column of a table
-   * @param tableName
-   * @param hcd HColumnDesciptor
-   * @return Modified HTableDescriptor with the column modified.
-   * @throws IOException
-   */
-  public HTableDescriptor modifyColumn(TableName tableName, HColumnDescriptor hcd)
-      throws IOException {
-    LOG.info("AddModifyColumn. Table = " + tableName
-        + " HCD = " + hcd.toString());
-
-    HTableDescriptor htd = this.services.getTableDescriptors().get(tableName);
-    byte [] familyName = hcd.getName();
-    if(!htd.hasFamily(familyName)) {
-      throw new InvalidFamilyOperationException("Family '" +
-        Bytes.toString(familyName) + "' doesn't exists so cannot be modified");
-    }
-    htd.modifyFamily(hcd);
-    this.services.getTableDescriptors().add(htd);
-    return htd;
-  }
-
-  /**
-   * Add column to a table
-   * @param tableName
-   * @param hcd
-   * @return Modified HTableDescriptor with new column added.
-   * @throws IOException
-   */
-  public HTableDescriptor addColumn(TableName tableName, HColumnDescriptor hcd)
-      throws IOException {
-    LOG.info("AddColumn. Table = " + tableName + " HCD = " +
-      hcd.toString());
-    HTableDescriptor htd = this.services.getTableDescriptors().get(tableName);
-    if (htd == null) {
-      throw new InvalidFamilyOperationException("Family '" +
-        hcd.getNameAsString() + "' cannot be modified as HTD is null");
-    }
-    htd.addFamily(hcd);
-    this.services.getTableDescriptors().add(htd);
-    return htd;
-  }
-
-  /**
    * The function is used in SSH to set recovery mode based on configuration after all outstanding
    * log split tasks drained.
    * @throws IOException

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotLogCleaner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotLogCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotLogCleaner.java
deleted file mode 100644
index 28aa800..0000000
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotLogCleaner.java
+++ /dev/null
@@ -1,103 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.master.snapshot;
-
-import java.io.IOException;
-import java.util.Collection;
-import java.util.Collections;
-
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.classification.InterfaceStability;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileStatus;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.HBaseInterfaceAudience;
-import org.apache.hadoop.hbase.master.cleaner.BaseLogCleanerDelegate;
-import org.apache.hadoop.hbase.snapshot.SnapshotReferenceUtil;
-import org.apache.hadoop.hbase.util.FSUtils;
-
-/**
- * Implementation of a log cleaner that checks if a log is still used by
- * snapshots of HBase tables.
- */
-@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CONFIG)
-@InterfaceStability.Evolving
-public class SnapshotLogCleaner extends BaseLogCleanerDelegate {
-  private static final Log LOG = LogFactory.getLog(SnapshotLogCleaner.class);
-
-  /**
-   * Conf key for the frequency to attempt to refresh the cache of hfiles currently used in
-   * snapshots (ms)
-   */
-  static final String WAL_CACHE_REFRESH_PERIOD_CONF_KEY =
-      "hbase.master.hlogcleaner.plugins.snapshot.period";
-
-  /** Refresh cache, by default, every 5 minutes */
-  private static final long DEFAULT_WAL_CACHE_REFRESH_PERIOD = 300000;
-
-  private SnapshotFileCache cache;
-
-  @Override
-  public synchronized Iterable<FileStatus> getDeletableFiles(Iterable<FileStatus> files) {
-    try {
-      return cache.getUnreferencedFiles(files);
-    } catch (IOException e) {
-      LOG.error("Exception while checking if files were valid, keeping them just in case.", e);
-      return Collections.emptyList();
-    }
-  }
-
-  /**
-   * This method should only be called <b>once</b>, as it starts a thread to keep the cache
-   * up-to-date.
-   * <p>
-   * {@inheritDoc}
-   */
-  @Override
-  public void setConf(Configuration conf) {
-    super.setConf(conf);
-    try {
-      long cacheRefreshPeriod = conf.getLong(
-        WAL_CACHE_REFRESH_PERIOD_CONF_KEY, DEFAULT_WAL_CACHE_REFRESH_PERIOD);
-      final FileSystem fs = FSUtils.getCurrentFileSystem(conf);
-      Path rootDir = FSUtils.getRootDir(conf);
-      cache = new SnapshotFileCache(fs, rootDir, cacheRefreshPeriod, cacheRefreshPeriod,
-          "snapshot-log-cleaner-cache-refresher", new SnapshotFileCache.SnapshotFileInspector() {
-            public Collection<String> filesUnderSnapshot(final Path snapshotDir)
-                throws IOException {
-              return SnapshotReferenceUtil.getWALNames(fs, snapshotDir);
-            }
-          });
-    } catch (IOException e) {
-      LOG.error("Failed to create snapshot log cleaner", e);
-    }
-  }
-
-  @Override
-  public void stop(String why) {
-    this.cache.stop(why);
-  }
-
-  @Override
-  public boolean isStopped() {
-    return this.cache.isStopped();
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
index 853e3cd..2c04209 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotManager.java
@@ -1028,7 +1028,6 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable
       // Inject snapshot cleaners, if snapshot.enable is true
       hfileCleaners.add(SnapshotHFileCleaner.class.getName());
       hfileCleaners.add(HFileLinkCleaner.class.getName());
-      logCleaners.add(SnapshotLogCleaner.class.getName());
 
       // Set cleaners conf
       conf.setStrings(HFileCleaner.MASTER_HFILE_CLEANER_PLUGINS,
@@ -1037,7 +1036,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable
         logCleaners.toArray(new String[logCleaners.size()]));
     } else {
       // Verify if cleaners are present
-      snapshotEnabled = logCleaners.contains(SnapshotLogCleaner.class.getName()) &&
+      snapshotEnabled =
         hfileCleaners.contains(SnapshotHFileCleaner.class.getName()) &&
         hfileCleaners.contains(HFileLinkCleaner.class.getName());
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
index 9d0f760..c5f97c3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java
@@ -209,8 +209,7 @@ public class ExportSnapshot extends Configured implements Tool {
               new Path(region, new Path(family, hfile)));
           break;
         case WAL:
-          Path oldLogsDir = new Path(outputRoot, HConstants.HREGION_OLDLOGDIR_NAME);
-          path = new Path(oldLogsDir, inputInfo.getWalName());
+          LOG.warn("snapshot does not keeps WALs: " + inputInfo);
           break;
         default:
           throw new IOException("Invalid File Type: " + inputInfo.getType().toString());
@@ -550,19 +549,6 @@ public class ExportSnapshot extends Configured implements Tool {
             files.add(new Pair<SnapshotFileInfo, Long>(fileInfo, size));
           }
         }
-
-        @Override
-        public void logFile (final String server, final String logfile)
-            throws IOException {
-          SnapshotFileInfo fileInfo = SnapshotFileInfo.newBuilder()
-            .setType(SnapshotFileInfo.Type.WAL)
-            .setWalServer(server)
-            .setWalName(logfile)
-            .build();
-
-          long size = new WALLink(conf, server, logfile).getFileStatus(fs).getLen();
-          files.add(new Pair<SnapshotFileInfo, Long>(fileInfo, size));
-        }
     });
 
     return files;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
index bac3347..f927709 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.java
@@ -39,8 +39,8 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HRegionInfo;
+import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HTableDescriptor;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.backup.HFileArchiver;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
index 606b9c9..36909d5 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotInfo.java
@@ -413,20 +413,6 @@ public final class SnapshotInfo extends Configured implements Tool {
               state == null ? "" : "(" + state + ")");
           }
         }
-
-        @Override
-        public void logFile (final String server, final String logfile)
-            throws IOException {
-          SnapshotStats.FileInfo info = stats.addLogFile(server, logfile);
-
-          if (showFiles) {
-            String state = info.getStateToString();
-            System.out.printf("%8s log %s on server %s (%s)%n",
-              (info.isMissing() ? "-" : fileSizeToString(info.getSize())),
-              logfile, server,
-              state == null ? "" : "(" + state + ")");
-          }
-        }
     });
 
     // Dump the stats
@@ -497,11 +483,6 @@ public final class SnapshotInfo extends Configured implements Tool {
             stats.addStoreFile(regionInfo, family, storeFile);
           }
         }
-
-        @Override
-        public void logFile (final String server, final String logfile) throws IOException {
-          stats.addLogFile(server, logfile);
-        }
     });
     return stats;
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
index 007dbfa..0204cd8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifest.java
@@ -61,7 +61,7 @@ import org.apache.hadoop.hbase.util.Threads;
  * On open() the snapshot will be loaded to the current in-memory format.
  */
 @InterfaceAudience.Private
-public class SnapshotManifest {
+public final class SnapshotManifest {
   private static final Log LOG = LogFactory.getLog(SnapshotManifest.class);
 
   private static final String DATA_MANIFEST_NAME = "data.manifest";

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
index 137acf3..328c998 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV1.java
@@ -54,7 +54,7 @@ import org.apache.hadoop.hbase.util.FSUtils;
  *      /snapshotName/regionName/familyName/fileName
  */
 @InterfaceAudience.Private
-public class SnapshotManifestV1 {
+public final class SnapshotManifestV1 {
   private static final Log LOG = LogFactory.getLog(SnapshotManifestV1.class);
 
   public static final int DESCRIPTOR_VERSION = 0;
@@ -154,7 +154,7 @@ public class SnapshotManifestV1 {
     fs.delete(new Path(snapshotDir, regionName), true);
   }
 
-  static SnapshotRegionManifest buildManifestFromDisk (final Configuration conf,
+  static SnapshotRegionManifest buildManifestFromDisk(final Configuration conf,
       final FileSystem fs, final Path tableDir, final HRegionInfo regionInfo) throws IOException {
     HRegionFileSystem regionFs = HRegionFileSystem.openRegionFromFileSystem(conf, fs,
           tableDir, regionInfo, true);

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
index dccbeb5..2df9bef 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotManifestV2.java
@@ -53,13 +53,15 @@ import org.apache.hadoop.hbase.util.FSUtils;
  *      /snapshotName/manifest.regionName
  */
 @InterfaceAudience.Private
-public class SnapshotManifestV2 {
+public final class SnapshotManifestV2 {
   private static final Log LOG = LogFactory.getLog(SnapshotManifestV2.class);
 
   public static final int DESCRIPTOR_VERSION = 2;
 
   private static final String SNAPSHOT_MANIFEST_PREFIX = "region-manifest.";
 
+  private SnapshotManifestV2() {}
+
   static class ManifestBuilder implements SnapshotManifest.RegionVisitor<
                     SnapshotRegionManifest.Builder, SnapshotRegionManifest.FamilyFiles.Builder> {
     private final Configuration conf;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
index d1e0236..c62d53d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/SnapshotReferenceUtil.java
@@ -41,9 +41,8 @@ import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.io.HFileLink;
 import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
 import org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotRegionManifest;
-import org.apache.hadoop.hbase.wal.DefaultWALProvider;
 import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
-import org.apache.hadoop.hbase.util.FSVisitor;
+import org.apache.hadoop.hbase.util.HFileArchiveUtil;
 
 /**
  * Utility methods for interacting with the snapshot referenced files.
@@ -57,8 +56,7 @@ public final class SnapshotReferenceUtil {
        final SnapshotRegionManifest.StoreFile storeFile) throws IOException;
   }
 
-  public interface SnapshotVisitor extends StoreFileVisitor,
-    FSVisitor.LogFileVisitor {
+  public interface SnapshotVisitor extends StoreFileVisitor {
   }
 
   private SnapshotReferenceUtil() {
@@ -66,18 +64,7 @@ public final class SnapshotReferenceUtil {
   }
 
   /**
-   * Get log directory for a server in a snapshot.
-   *
-   * @param snapshotDir directory where the specific snapshot is stored
-   * @param serverName name of the parent regionserver for the log files
-   * @return path to the log home directory for the archive files.
-   */
-  public static Path getLogsDir(Path snapshotDir, String serverName) {
-    return new Path(snapshotDir, DefaultWALProvider.getWALDirectoryName(serverName));
-  }
-
-  /**
-   * Iterate over the snapshot store files, restored.edits and logs
+   * Iterate over the snapshot store files
    *
    * @param conf The current {@link Configuration} instance.
    * @param fs {@link FileSystem}
@@ -106,7 +93,6 @@ public final class SnapshotReferenceUtil {
       final Path snapshotDir, final SnapshotDescription desc, final SnapshotVisitor visitor)
       throws IOException {
     visitTableStoreFiles(conf, fs, snapshotDir, desc, visitor);
-    visitLogFiles(fs, snapshotDir, visitor);
   }
 
   /**©
@@ -153,19 +139,6 @@ public final class SnapshotReferenceUtil {
   }
 
   /**
-   * Iterate over the snapshot log files
-   *
-   * @param fs {@link FileSystem}
-   * @param snapshotDir {@link Path} to the Snapshot directory
-   * @param visitor callback object to get the log files
-   * @throws IOException if an error occurred while scanning the directory
-   */
-  public static void visitLogFiles(final FileSystem fs, final Path snapshotDir,
-      final FSVisitor.LogFileVisitor visitor) throws IOException {
-    FSVisitor.visitLogFiles(fs, snapshotDir, visitor);
-  }
-
-  /**
    * Verify the validity of the snapshot
    *
    * @param conf The current {@link Configuration} instance.
@@ -357,24 +330,4 @@ public final class SnapshotReferenceUtil {
     });
     return names;
   }
-
-  /**
-   * Returns the log file names available in the snapshot.
-   *
-   * @param fs {@link FileSystem}
-   * @param snapshotDir {@link Path} to the Snapshot directory
-   * @throws IOException if an error occurred while scanning the directory
-   * @return the names of wals in the specified snaphot
-   */
-  public static Set<String> getWALNames(final FileSystem fs, final Path snapshotDir)
-      throws IOException {
-    final Set<String> names = new HashSet<String>();
-    visitLogFiles(fs, snapshotDir, new FSVisitor.LogFileVisitor() {
-      @Override
-      public void logFile (final String server, final String logfile) throws IOException {
-        names.add(logfile);
-      }
-    });
-    return names;
-  }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
index 8717948..87c50d9 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
@@ -1051,44 +1051,6 @@ public abstract class FSUtils {
     return blocksDistribution;
   }
 
-
-
-  /**
-   * Runs through the hbase rootdir and checks all stores have only
-   * one file in them -- that is, they've been major compacted.  Looks
-   * at root and meta tables too.
-   * @param fs filesystem
-   * @param hbaseRootDir hbase root directory
-   * @return True if this hbase install is major compacted.
-   * @throws IOException e
-   */
-  public static boolean isMajorCompacted(final FileSystem fs,
-      final Path hbaseRootDir)
-  throws IOException {
-    List<Path> tableDirs = getTableDirs(fs, hbaseRootDir);
-    PathFilter regionFilter = new RegionDirFilter(fs);
-    PathFilter familyFilter = new FamilyDirFilter(fs);
-    for (Path d : tableDirs) {
-      FileStatus[] regionDirs = fs.listStatus(d, regionFilter);
-      for (FileStatus regionDir : regionDirs) {
-        Path dd = regionDir.getPath();
-        // Else its a region name.  Now look in region for families.
-        FileStatus[] familyDirs = fs.listStatus(dd, familyFilter);
-        for (FileStatus familyDir : familyDirs) {
-          Path family = familyDir.getPath();
-          // Now in family make sure only one file.
-          FileStatus[] familyStatus = fs.listStatus(family);
-          if (familyStatus.length > 1) {
-            LOG.debug(family.toString() + " has " + familyStatus.length +
-                " files.");
-            return false;
-          }
-        }
-      }
-    }
-    return true;
-  }
-
   // TODO move this method OUT of FSUtils. No dependencies to HMaster
   /**
    * Returns the total overall fragmentation percentage. Includes hbase:meta and

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSVisitor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSVisitor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSVisitor.java
index f0cc0c1..6b81bfa 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSVisitor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSVisitor.java
@@ -19,7 +19,6 @@
 package org.apache.hadoop.hbase.util;
 
 import java.io.IOException;
-import java.util.NavigableSet;
 
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
@@ -28,8 +27,6 @@ import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.fs.PathFilter;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.wal.WALSplitter;
 
 /**
  * Utility methods for interacting with the hbase.root file system.
@@ -38,25 +35,11 @@ import org.apache.hadoop.hbase.wal.WALSplitter;
 public final class FSVisitor {
   private static final Log LOG = LogFactory.getLog(FSVisitor.class);
 
-  public interface RegionVisitor {
-    void region(final String region) throws IOException;
-  }
-
   public interface StoreFileVisitor {
     void storeFile(final String region, final String family, final String hfileName)
        throws IOException;
   }
 
-  public interface RecoveredEditsVisitor {
-    void recoveredEdits (final String region, final String logfile)
-      throws IOException;
-  }
-
-  public interface LogFileVisitor {
-    void logFile (final String server, final String logfile)
-      throws IOException;
-  }
-
   private FSVisitor() {
     // private constructor for utility class
   }
@@ -69,29 +52,6 @@ public final class FSVisitor {
    * @param visitor callback object to get the store files
    * @throws IOException if an error occurred while scanning the directory
    */
-  public static void visitRegions(final FileSystem fs, final Path tableDir,
-      final RegionVisitor visitor) throws IOException {
-    FileStatus[] regions = FSUtils.listStatus(fs, tableDir, new FSUtils.RegionDirFilter(fs));
-    if (regions == null) {
-      if (LOG.isTraceEnabled()) {
-        LOG.trace("No regions under directory:" + tableDir);
-      }
-      return;
-    }
-
-    for (FileStatus region: regions) {
-      visitor.region(region.getPath().getName());
-    }
-  }
-
-  /**
-   * Iterate over the table store files
-   *
-   * @param fs {@link FileSystem}
-   * @param tableDir {@link Path} to the table directory
-   * @param visitor callback object to get the store files
-   * @throws IOException if an error occurred while scanning the directory
-   */
   public static void visitTableStoreFiles(final FileSystem fs, final Path tableDir,
       final StoreFileVisitor visitor) throws IOException {
     FileStatus[] regions = FSUtils.listStatus(fs, tableDir, new FSUtils.RegionDirFilter(fs));
@@ -145,85 +105,4 @@ public final class FSVisitor {
       }
     }
   }
-
-  /**
-   * Iterate over each region in the table and inform about recovered.edits
-   *
-   * @param fs {@link FileSystem}
-   * @param tableDir {@link Path} to the table directory
-   * @param visitor callback object to get the recovered.edits files
-   * @throws IOException if an error occurred while scanning the directory
-   */
-  public static void visitTableRecoveredEdits(final FileSystem fs, final Path tableDir,
-      final FSVisitor.RecoveredEditsVisitor visitor) throws IOException {
-    FileStatus[] regions = FSUtils.listStatus(fs, tableDir, new FSUtils.RegionDirFilter(fs));
-    if (regions == null) {
-      if (LOG.isTraceEnabled()) {
-        LOG.trace("No recoveredEdits regions under directory:" + tableDir);
-      }
-      return;
-    }
-
-    for (FileStatus region: regions) {
-      visitRegionRecoveredEdits(fs, region.getPath(), visitor);
-    }
-  }
-
-  /**
-   * Iterate over recovered.edits of the specified region
-   *
-   * @param fs {@link FileSystem}
-   * @param regionDir {@link Path} to the Region directory
-   * @param visitor callback object to get the recovered.edits files
-   * @throws IOException if an error occurred while scanning the directory
-   */
-  public static void visitRegionRecoveredEdits(final FileSystem fs, final Path regionDir,
-      final FSVisitor.RecoveredEditsVisitor visitor) throws IOException {
-    NavigableSet<Path> files = WALSplitter.getSplitEditFilesSorted(fs, regionDir);
-    if (files == null || files.size() == 0) return;
-
-    for (Path source: files) {
-      // check to see if the file is zero length, in which case we can skip it
-      FileStatus stat = fs.getFileStatus(source);
-      if (stat.getLen() <= 0) continue;
-
-      visitor.recoveredEdits(regionDir.getName(), source.getName());
-    }
-  }
-
-  /**
-   * Iterate over hbase log files
-   *
-   * @param fs {@link FileSystem}
-   * @param rootDir {@link Path} to the HBase root folder
-   * @param visitor callback object to get the log files
-   * @throws IOException if an error occurred while scanning the directory
-   */
-  public static void visitLogFiles(final FileSystem fs, final Path rootDir,
-      final LogFileVisitor visitor) throws IOException {
-    Path logsDir = new Path(rootDir, HConstants.HREGION_LOGDIR_NAME);
-    FileStatus[] logServerDirs = FSUtils.listStatus(fs, logsDir);
-    if (logServerDirs == null) {
-      if (LOG.isTraceEnabled()) {
-        LOG.trace("No logs under directory:" + logsDir);
-      }
-      return;
-    }
-
-    for (FileStatus serverLogs: logServerDirs) {
-      String serverName = serverLogs.getPath().getName();
-
-      FileStatus[] wals = FSUtils.listStatus(fs, serverLogs.getPath());
-      if (wals == null) {
-        if (LOG.isTraceEnabled()) {
-          LOG.trace("No wals found for server: " + serverName + ", skipping.");
-        }
-        continue;
-      }
-
-      for (FileStatus walRef: wals) {
-        visitor.logFile(serverName, walRef.getPath().getName());
-      }
-    }
-  }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java
index d72b176..f0dc132 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java
@@ -69,7 +69,6 @@ public class TestNamespace {
   @BeforeClass
   public static void setUp() throws Exception {
     TEST_UTIL = new HBaseTestingUtility();
-    TEST_UTIL.getConfiguration().setInt("hbase.namespacejanitor.interval", 5000);
     TEST_UTIL.startMiniCluster(NUM_SLAVES_BASE);
     admin = TEST_UTIL.getHBaseAdmin();
     cluster = TEST_UTIL.getHBaseCluster();
@@ -289,33 +288,6 @@ public class TestNamespace {
     admin.deleteTable(desc.getTableName());
   }
 
-  @Ignore @Test
-  public void testNamespaceJanitor() throws Exception {
-    FileSystem fs = TEST_UTIL.getTestFileSystem();
-
-    int fsCount = fs.listStatus(new Path(FSUtils.getRootDir(TEST_UTIL.getConfiguration()),
-        HConstants.BASE_NAMESPACE_DIR)).length;
-    Path fakeNSPath =
-        FSUtils.getNamespaceDir(FSUtils.getRootDir(TEST_UTIL.getConfiguration()), "foo");
-    assertTrue(fs.mkdirs(fakeNSPath));
-
-    String fakeZnode = ZKUtil.joinZNode(ZooKeeperWatcher.namespaceZNode, "foo");
-    int zkCount = ZKUtil.listChildrenNoWatch(TEST_UTIL.getZooKeeperWatcher(),
-        ZooKeeperWatcher.namespaceZNode).size();
-    ZKUtil.createWithParents(TEST_UTIL.getZooKeeperWatcher(), fakeZnode);
-    Thread.sleep(10000);
-
-    //verify namespace count is the same and orphan is removed
-    assertFalse(fs.exists(fakeNSPath));
-    assertEquals(fsCount, fs.listStatus(new Path(FSUtils.getRootDir(TEST_UTIL.getConfiguration()),
-            HConstants.BASE_NAMESPACE_DIR)).length);
-
-    assertEquals(-1, ZKUtil.checkExists(TEST_UTIL.getZooKeeperWatcher(), fakeZnode));
-    assertEquals(zkCount,
-        ZKUtil.listChildrenNoWatch(TEST_UTIL.getZooKeeperWatcher(),
-            ZooKeeperWatcher.namespaceZNode).size());
-  }
-
   @Test(timeout = 60000)
   public void testNamespaceOperations() throws IOException {
     admin.createNamespace(NamespaceDescriptor.create(prefix + "ns1").build());

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java
index 46f9976..be61f5e 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestMasterObserver.java
@@ -1177,11 +1177,6 @@ public class TestMasterObserver {
     Configuration conf = UTIL.getConfiguration();
     conf.set(CoprocessorHost.MASTER_COPROCESSOR_CONF_KEY,
         CPMasterObserver.class.getName());
-    conf.set("hbase.master.hfilecleaner.plugins",
-      "org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner," +
-      "org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner");
-    conf.set("hbase.master.logcleaner.plugins",
-      "org.apache.hadoop.hbase.master.snapshot.SnapshotLogCleaner");
     // We need more than one data server on this test
     UTIL.startMiniCluster(2);
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotFileCache.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotFileCache.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotFileCache.java
index f1732bb..4db44aa 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotFileCache.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotFileCache.java
@@ -95,41 +95,6 @@ public class TestSnapshotFileCache {
   }
 
   @Test
-  public void testJustFindLogsDirectory() throws Exception {
-    // don't refresh the cache unless we tell it to
-    long period = Long.MAX_VALUE;
-    Path snapshotDir = SnapshotDescriptionUtils.getSnapshotsDir(rootDir);
-    SnapshotFileCache cache = new SnapshotFileCache(fs, rootDir, period, 10000000,
-        "test-snapshot-file-cache-refresh", new SnapshotFileCache.SnapshotFileInspector() {
-            public Collection<String> filesUnderSnapshot(final Path snapshotDir)
-                throws IOException {
-              return SnapshotReferenceUtil.getWALNames(fs, snapshotDir);
-            }
-        });
-
-    // create a file in a 'completed' snapshot
-    SnapshotDescription desc = SnapshotDescription.newBuilder().setName("snapshot").build();
-    Path snapshot = SnapshotDescriptionUtils.getCompletedSnapshotDir(desc, rootDir);
-    SnapshotDescriptionUtils.writeSnapshotInfo(desc, snapshot, fs);
-    Path file1 = new Path(new Path(new Path(snapshot, "7e91021"), "fam"), "file1");
-    fs.createNewFile(file1);
-
-    // and another file in the logs directory
-    Path logs = SnapshotReferenceUtil.getLogsDir(snapshot, "server");
-    Path log = new Path(logs, "me.hbase.com%2C58939%2C1350424310315.1350424315552");
-    fs.createNewFile(log);
-
-    FSUtils.logFileSystemState(fs, rootDir, LOG);
-
-    // then make sure the cache only finds the log files
-    Iterable<FileStatus> notSnapshot = getNonSnapshotFiles(cache, file1);
-    assertFalse("Cache found '" + file1 + "', but it shouldn't have.",
-        Iterables.contains(notSnapshot, file1.getName()));
-    notSnapshot = getNonSnapshotFiles(cache, log);
-    assertTrue("Cache didn't find:" + log, !Iterables.contains(notSnapshot, log));
-  }
-
-  @Test
   public void testReloadModifiedDirectory() throws IOException {
     // don't refresh the cache unless we tell it to
     long period = Long.MAX_VALUE;
@@ -214,10 +179,6 @@ public class TestSnapshotFileCache {
     SnapshotReferenceUtil
         .visitReferencedFiles(UTIL.getConfiguration(), fs, builder.getSnapshotsDir(),
             new SnapshotReferenceUtil.SnapshotVisitor() {
-              @Override public void logFile(String server, String logfile) throws IOException {
-                // do nothing.
-              }
-
               @Override public void storeFile(HRegionInfo regionInfo, String familyName,
                   SnapshotProtos.SnapshotRegionManifest.StoreFile storeFile) throws IOException {
                 FileStatus status = mockStoreFile(storeFile.getName());
@@ -238,7 +199,6 @@ public class TestSnapshotFileCache {
   class SnapshotFiles implements SnapshotFileCache.SnapshotFileInspector {
     public Collection<String> filesUnderSnapshot(final Path snapshotDir) throws IOException {
       Collection<String> files =  new HashSet<String>();
-      files.addAll(SnapshotReferenceUtil.getWALNames(fs, snapshotDir));
       files.addAll(SnapshotReferenceUtil.getHFileNames(UTIL.getConfiguration(), fs, snapshotDir));
       return files;
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotLogCleaner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotLogCleaner.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotLogCleaner.java
deleted file mode 100644
index 734ce8c..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotLogCleaner.java
+++ /dev/null
@@ -1,85 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hbase.master.snapshot;
-
-import static org.junit.Assert.assertFalse;
-
-import java.io.IOException;
-
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.hbase.HBaseTestingUtility;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.testclassification.SmallTests;
-import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.FSUtils;
-import org.junit.AfterClass;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-
-/**
- * Test that the snapshot log cleaner finds logs referenced in a snapshot
- */
-@Category(SmallTests.class)
-public class TestSnapshotLogCleaner {
-
-  private final static HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
-
-  @AfterClass
-  public static void cleanup() throws IOException {
-    Configuration conf = TEST_UTIL.getConfiguration();
-    Path rootDir = FSUtils.getRootDir(conf);
-    FileSystem fs = FileSystem.get(conf);
-    // cleanup
-    fs.delete(rootDir, true);
-  }
-
-  @Test
-  public void testFindsSnapshotFilesWhenCleaning() throws IOException {
-    Configuration conf = TEST_UTIL.getConfiguration();
-    FSUtils.setRootDir(conf, TEST_UTIL.getDataTestDir());
-    Path rootDir = FSUtils.getRootDir(conf);
-    FileSystem fs = FileSystem.get(conf);
-    SnapshotLogCleaner cleaner = new SnapshotLogCleaner();
-    cleaner.setConf(conf);
-
-    // write an hfile to the snapshot directory
-    String snapshotName = "snapshot";
-    byte[] snapshot = Bytes.toBytes(snapshotName);
-    Path snapshotDir = SnapshotDescriptionUtils.getCompletedSnapshotDir(snapshotName, rootDir);
-    Path snapshotLogDir = new Path(snapshotDir, HConstants.HREGION_LOGDIR_NAME);
-    String timestamp = "1339643343027";
-    String hostFromMaster = "localhost%2C59648%2C1339643336601";
-
-    Path hostSnapshotLogDir = new Path(snapshotLogDir, hostFromMaster);
-    String snapshotlogfile = hostFromMaster + "." + timestamp + ".hbase";
-
-    // add the reference to log in the snapshot
-    fs.create(new Path(hostSnapshotLogDir, snapshotlogfile));
-
-    // now check to see if that log file would get deleted.
-    Path oldlogDir = new Path(rootDir, HConstants.HREGION_OLDLOGDIR_NAME);
-    Path logFile = new Path(oldlogDir, snapshotlogfile);
-    fs.create(logFile);
-
-    // make sure that the file isn't deletable
-    assertFalse(cleaner.isFileDeletable(fs.getFileStatus(logFile)));
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotManager.java
index ffb0bb2..8066ec3 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotManager.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/snapshot/TestSnapshotManager.java
@@ -120,7 +120,6 @@ public class TestSnapshotManager {
     conf = new Configuration();
     conf.setStrings(HFileCleaner.MASTER_HFILE_CLEANER_PLUGINS,
       SnapshotHFileCleaner.class.getName(), HFileLinkCleaner.class.getName());
-    conf.set(HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS, SnapshotLogCleaner.class.getName());
     conf.setBoolean(SnapshotManager.HBASE_SNAPSHOT_ENABLED, false);
     manager = getNewManager(conf);
     assertFalse("Snapshot should be disabled", isSnapshotSupported(manager));
@@ -129,7 +128,6 @@ public class TestSnapshotManager {
     conf = new Configuration();
     conf.setStrings(HFileCleaner.MASTER_HFILE_CLEANER_PLUGINS,
       SnapshotHFileCleaner.class.getName(), HFileLinkCleaner.class.getName());
-    conf.set(HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS, SnapshotLogCleaner.class.getName());
     manager = getNewManager(conf);
     assertTrue("Snapshot should be enabled, because cleaners are present",
       isSnapshotSupported(manager));

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java
index 7e629cd..6a8a5a4 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.java
@@ -380,13 +380,6 @@ public class TestExportSnapshot {
           }
         }
 
-        @Override
-        public void logFile (final String server, final String logfile)
-            throws IOException {
-          snapshotFiles.add(logfile);
-          verifyNonEmptyFile(new Path(exportedSnapshot, new Path(server, logfile)));
-        }
-
         private void verifyNonEmptyFile(final Path path) throws IOException {
           assertTrue(path + " should exists", fs.exists(path));
           assertTrue(path + " should not be empty", fs.getFileStatus(path).getLen() > 0);

http://git-wip-us.apache.org/repos/asf/hbase/blob/0fa48ae5/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
index 3029601..2710f98 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestFSVisitor.java
@@ -54,31 +54,21 @@ public class TestFSVisitor {
 
   private Set<String> tableFamilies;
   private Set<String> tableRegions;
-  private Set<String> recoveredEdits;
   private Set<String> tableHFiles;
-  private Set<String> regionServers;
-  private Set<String> serverLogs;
 
   private FileSystem fs;
   private Path tableDir;
-  private Path logsDir;
   private Path rootDir;
 
   @Before
   public void setUp() throws Exception {
     fs = FileSystem.get(TEST_UTIL.getConfiguration());
     rootDir = TEST_UTIL.getDataTestDir("hbase");
-    logsDir = new Path(rootDir, HConstants.HREGION_LOGDIR_NAME);
 
     tableFamilies = new HashSet<String>();
     tableRegions = new HashSet<String>();
-    recoveredEdits = new HashSet<String>();
     tableHFiles = new HashSet<String>();
-    regionServers = new HashSet<String>();
-    serverLogs = new HashSet<String>();
     tableDir = createTableFiles(rootDir, TABLE_NAME, tableRegions, tableFamilies, tableHFiles);
-    createRecoverEdits(tableDir, tableRegions, recoveredEdits);
-    createLogs(logsDir, regionServers, serverLogs);
     FSUtils.logFileSystemState(fs, rootDir, LOG);
   }
 
@@ -105,36 +95,6 @@ public class TestFSVisitor {
     assertEquals(tableHFiles, hfiles);
   }
 
-  @Test
-  public void testVisitRecoveredEdits() throws IOException {
-    final Set<String> regions = new HashSet<String>();
-    final Set<String> edits = new HashSet<String>();
-    FSVisitor.visitTableRecoveredEdits(fs, tableDir, new FSVisitor.RecoveredEditsVisitor() {
-      public void recoveredEdits (final String region, final String logfile)
-          throws IOException {
-        regions.add(region);
-        edits.add(logfile);
-      }
-    });
-    assertEquals(tableRegions, regions);
-    assertEquals(recoveredEdits, edits);
-  }
-
-  @Test
-  public void testVisitLogFiles() throws IOException {
-    final Set<String> servers = new HashSet<String>();
-    final Set<String> logs = new HashSet<String>();
-    FSVisitor.visitLogFiles(fs, rootDir, new FSVisitor.LogFileVisitor() {
-      public void logFile (final String server, final String logfile) throws IOException {
-        servers.add(server);
-        logs.add(logfile);
-      }
-    });
-    assertEquals(regionServers, servers);
-    assertEquals(serverLogs, logs);
-  }
-
-
   /*
    * |-testtb/
    * |----f1d3ff8443297732862df21dc4e57262/
@@ -166,66 +126,4 @@ public class TestFSVisitor {
     }
     return tableDir;
   }
-
-  /*
-   * |-testtb/
-   * |----f1d3ff8443297732862df21dc4e57262/
-   * |-------recovered.edits/
-   * |----------0000001351969633479
-   * |----------0000001351969633481
-   */
-  private void createRecoverEdits(final Path tableDir, final Set<String> tableRegions,
-      final Set<String> recoverEdits) throws IOException {
-    for (String region: tableRegions) {
-      Path regionEditsDir = WALSplitter.getRegionDirRecoveredEditsDir(new Path(tableDir, region));
-      long seqId = System.currentTimeMillis();
-      for (int i = 0; i < 3; ++i) {
-        String editName = String.format("%019d", seqId + i);
-        recoverEdits.add(editName);
-        FSDataOutputStream stream = fs.create(new Path(regionEditsDir, editName));
-        stream.write(Bytes.toBytes("test"));
-        stream.close();
-      }
-    }
-  }
-
-  /*
-   * Old style
-   * |-.logs/
-   * |----server5,5,1351969633508/
-   * |-------server5,5,1351969633508.0
-   * |----server6,6,1351969633512/
-   * |-------server6,6,1351969633512.0
-   * |-------server6,6,1351969633512.3
-   * New style
-   * |-.logs/
-   * |----server3,5,1351969633508/
-   * |-------server3,5,1351969633508.default.0
-   * |----server4,6,1351969633512/
-   * |-------server4,6,1351969633512.default.0
-   * |-------server4,6,1351969633512.some_provider.3
-   */
-  private void createLogs(final Path logDir, final Set<String> servers,
-      final Set<String> logs) throws IOException {
-    for (int s = 0; s < 7; ++s) {
-      String server = String.format("server%d,%d,%d", s, s, System.currentTimeMillis());
-      servers.add(server);
-      Path serverLogDir = new Path(logDir, server);
-      if (s % 2 == 0) {
-        if (s % 3 == 0) {
-          server += ".default";
-        } else {
-          server += "." + s;
-        }
-      }
-      fs.mkdirs(serverLogDir);
-      for (int i = 0; i < 5; ++i) {
-        String logfile = server + '.' + i;
-        logs.add(logfile);
-        FSDataOutputStream stream = fs.create(new Path(serverLogDir, logfile));
-        stream.write(Bytes.toBytes("test"));
-        stream.close();
-      }
-    }
-  }
 }