You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by ar...@apache.org on 2017/08/25 19:03:59 UTC
[1/3] hadoop git commit: HDFS-7764. DirectoryScanner shouldn't abort
the scan if one directory had an error (Rakesh R via cmccabe)
Repository: hadoop
Updated Branches:
refs/heads/branch-2.8.2 310fa63cc -> 2a25ddc0b
HDFS-7764. DirectoryScanner shouldn't abort the scan if one directory had an error (Rakesh R via cmccabe)
(cherry picked from commit f67149ab08bb49381def6c535ab4c4610e0a4221)
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/86db22f8
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/86db22f8
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/86db22f8
Branch: refs/heads/branch-2.8.2
Commit: 86db22f8905921678fb14a7f2bc6335467afb074
Parents: 310fa63
Author: Colin Patrick Mccabe <cm...@cloudera.com>
Authored: Thu Jan 28 19:54:50 2016 -0800
Committer: Arpit Agarwal <ar...@apache.org>
Committed: Fri Aug 25 12:03:37 2017 -0700
----------------------------------------------------------------------
.../hdfs/server/datanode/DirectoryScanner.java | 72 +++++++++++++-------
.../server/datanode/TestDirectoryScanner.java | 50 ++++++++++++++
2 files changed, 97 insertions(+), 25 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/86db22f8/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
index 490321e..64fa94e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
@@ -19,6 +19,7 @@ package org.apache.hadoop.hdfs.server.datanode;
import com.google.common.annotations.VisibleForTesting;
import java.io.File;
+import java.io.FilenameFilter;
import java.io.IOException;
import java.util.Arrays;
import java.util.Collections;
@@ -45,12 +46,13 @@ import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileUtil;
-import org.apache.hadoop.util.AutoCloseableLock;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
+import org.apache.hadoop.io.IOUtils;
+import org.apache.hadoop.util.AutoCloseableLock;
import org.apache.hadoop.util.Daemon;
import org.apache.hadoop.util.StopWatch;
import org.apache.hadoop.util.Time;
@@ -731,18 +733,20 @@ public class DirectoryScanner implements Runnable {
for (Entry<Integer, Future<ScanInfoPerBlockPool>> report :
compilersInProgress.entrySet()) {
+ Integer index = report.getKey();
try {
- dirReports[report.getKey()] = report.getValue().get();
+ dirReports[index] = report.getValue().get();
// If our compiler threads were interrupted, give up on this run
- if (dirReports[report.getKey()] == null) {
+ if (dirReports[index] == null) {
dirReports = null;
break;
}
} catch (Exception ex) {
- LOG.error("Error compiling report", ex);
- // Propagate ex to DataBlockScanner to deal with
- throw new RuntimeException(ex);
+ FsVolumeSpi fsVolumeSpi = volumes.get(index);
+ LOG.error("Error compiling report for the volume, StorageId: "
+ + fsVolumeSpi.getStorageID(), ex);
+ // Continue scanning the other volumes
}
}
} catch (IOException e) {
@@ -751,7 +755,9 @@ public class DirectoryScanner implements Runnable {
if (dirReports != null) {
// Compile consolidated report for all the volumes
for (ScanInfoPerBlockPool report : dirReports) {
- list.addAll(report);
+ if(report != null){
+ list.addAll(report);
+ }
}
}
return list.toSortedArrays();
@@ -841,12 +847,11 @@ public class DirectoryScanner implements Runnable {
File bpFinalizedDir, File dir, LinkedList<ScanInfo> report)
throws InterruptedException {
- File[] files;
-
throttle();
+ List <String> fileNames;
try {
- files = FileUtil.listFiles(dir);
+ fileNames = IOUtils.listDirectory(dir, BlockDirFilter.INSTANCE);
} catch (IOException ioe) {
LOG.warn("Exception occured while compiling report: ", ioe);
// Initiate a check on disk failure.
@@ -854,44 +859,50 @@ public class DirectoryScanner implements Runnable {
// Ignore this directory and proceed.
return report;
}
- Arrays.sort(files);
+ Collections.sort(fileNames);
+
/*
* Assumption: In the sorted list of files block file appears immediately
* before block metadata file. This is true for the current naming
* convention for block file blk_<blockid> and meta file
* blk_<blockid>_<genstamp>.meta
*/
- for (int i = 0; i < files.length; i++) {
+ for (int i = 0; i < fileNames.size(); i++) {
// Make sure this thread can make a timely exit. With a low throttle
// rate, completing a run can take a looooong time.
if (Thread.interrupted()) {
throw new InterruptedException();
}
- if (files[i].isDirectory()) {
- compileReport(vol, bpFinalizedDir, files[i], report);
+ File file = new File(dir, fileNames.get(i));
+ if (file.isDirectory()) {
+ compileReport(vol, bpFinalizedDir, file, report);
continue;
}
- if (!Block.isBlockFilename(files[i])) {
- if (isBlockMetaFile(Block.BLOCK_FILE_PREFIX, files[i].getName())) {
- long blockId = Block.getBlockId(files[i].getName());
- verifyFileLocation(files[i].getParentFile(), bpFinalizedDir,
+ if (!Block.isBlockFilename(file)) {
+ if (isBlockMetaFile(Block.BLOCK_FILE_PREFIX, file.getName())) {
+ long blockId = Block.getBlockId(file.getName());
+ verifyFileLocation(file.getParentFile(), bpFinalizedDir,
blockId);
- report.add(new ScanInfo(blockId, null, files[i], vol));
+ report.add(new ScanInfo(blockId, null, file, vol));
}
continue;
}
- File blockFile = files[i];
- long blockId = Block.filename2id(blockFile.getName());
+ File blockFile = file;
+ long blockId = Block.filename2id(file.getName());
File metaFile = null;
// Skip all the files that start with block name until
// getting to the metafile for the block
- while (i + 1 < files.length && files[i + 1].isFile()
- && files[i + 1].getName().startsWith(blockFile.getName())) {
+ while (i + 1 < fileNames.size()) {
+ File blkMetaFile = new File(dir, fileNames.get(i + 1));
+ if (!(blkMetaFile.isFile()
+ && blkMetaFile.getName().startsWith(blockFile.getName()))) {
+ break;
+ }
i++;
- if (isBlockMetaFile(blockFile.getName(), files[i].getName())) {
- metaFile = files[i];
+ if (isBlockMetaFile(blockFile.getName(), blkMetaFile.getName())) {
+ metaFile = blkMetaFile;
break;
}
}
@@ -952,4 +963,15 @@ public class DirectoryScanner implements Runnable {
perfTimer.reset().start();
}
}
+
+ private enum BlockDirFilter implements FilenameFilter {
+ INSTANCE;
+
+ @Override
+ public boolean accept(File dir, String name) {
+ return name.startsWith(DataStorage.BLOCK_SUBDIR_PREFIX)
+ || name.startsWith(DataStorage.STORAGE_DIR_FINALIZED)
+ || name.startsWith(Block.BLOCK_FILE_PREFIX);
+ }
+ }
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/86db22f8/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
index 01cc192..35bd7e8 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
@@ -30,6 +30,8 @@ import java.io.FileOutputStream;
import java.io.IOException;
import java.nio.channels.ClosedChannelException;
import java.nio.channels.FileChannel;
+import java.util.ArrayList;
+import java.util.Iterator;
import java.util.LinkedList;
import java.util.List;
import java.util.Random;
@@ -54,6 +56,7 @@ import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
+import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi.FsVolumeReferences;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeReference;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetTestUtil;
@@ -64,6 +67,7 @@ import org.apache.hadoop.util.AutoCloseableLock;
import org.apache.hadoop.util.Time;
import org.junit.Before;
import org.junit.Test;
+import org.mockito.Mockito;
/**
* Tests {@link DirectoryScanner} handling of differences
@@ -941,4 +945,50 @@ public class TestDirectoryScanner {
new File(TEST_VOLUME.getFinalizedDir(BPID_2).getAbsolutePath(),
"blk_567__1004.meta"));
}
+
+ /**
+ * Test the behavior of exception handling during directory scan operation.
+ * Directory scanner shouldn't abort the scan on every directory just because
+ * one had an error.
+ */
+ @Test(timeout = 60000)
+ public void testExceptionHandlingWhileDirectoryScan() throws Exception {
+ cluster = new MiniDFSCluster.Builder(CONF).build();
+ try {
+ cluster.waitActive();
+ bpid = cluster.getNamesystem().getBlockPoolId();
+ fds = DataNodeTestUtils.getFSDataset(cluster.getDataNodes().get(0));
+ client = cluster.getFileSystem().getClient();
+ CONF.setInt(DFSConfigKeys.DFS_DATANODE_DIRECTORYSCAN_THREADS_KEY, 1);
+ DataNode dataNode = cluster.getDataNodes().get(0);
+
+ // Add files with 2 blocks
+ createFile(GenericTestUtils.getMethodName(), BLOCK_LENGTH * 2, false);
+
+ // Inject error on #getFinalizedDir() so that ReportCompiler#call() will
+ // hit exception while preparing the block info report list.
+ List<FsVolumeSpi> volumes = new ArrayList<>();
+ Iterator<FsVolumeSpi> iterator = fds.getFsVolumeReferences().iterator();
+ while (iterator.hasNext()) {
+ FsVolumeSpi volume = iterator.next();
+ FsVolumeSpi spy = Mockito.spy(volume);
+ Mockito.doThrow(new IOException("Error while getFinalizedDir"))
+ .when(spy).getFinalizedDir(volume.getBlockPoolList()[0]);
+ volumes.add(spy);
+ }
+ FsVolumeReferences volReferences = new FsVolumeReferences(volumes);
+ FsDatasetSpi<? extends FsVolumeSpi> spyFds = Mockito.spy(fds);
+ Mockito.doReturn(volReferences).when(spyFds).getFsVolumeReferences();
+
+ scanner = new DirectoryScanner(dataNode, spyFds, CONF);
+ scanner.setRetainDiffs(true);
+ scanner.reconcile();
+ } finally {
+ if (scanner != null) {
+ scanner.shutdown();
+ scanner = null;
+ }
+ cluster.shutdown();
+ }
+ }
}
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org
[2/3] hadoop git commit: HDFS-12319. DirectoryScanner will throw
IllegalStateException when Multiple BP's are present. Contributed by Brahma
Reddy Battula.
Posted by ar...@apache.org.
HDFS-12319. DirectoryScanner will throw IllegalStateException when Multiple BP's are present. Contributed by Brahma Reddy Battula.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/4b85d38f
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/4b85d38f
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/4b85d38f
Branch: refs/heads/branch-2.8.2
Commit: 4b85d38f6052ed59b848f9a8f5caa5be2a1595f8
Parents: 86db22f
Author: Arpit Agarwal <ar...@apache.org>
Authored: Fri Aug 25 10:41:52 2017 -0700
Committer: Arpit Agarwal <ar...@apache.org>
Committed: Fri Aug 25 12:03:39 2017 -0700
----------------------------------------------------------------------
.../hdfs/server/datanode/DirectoryScanner.java | 5 +-
.../server/datanode/TestDirectoryScanner.java | 66 ++++++++++++++++++--
2 files changed, 62 insertions(+), 9 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/4b85d38f/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
index 64fa94e..f96386a 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java
@@ -815,13 +815,12 @@ public class DirectoryScanner implements Runnable {
public ScanInfoPerBlockPool call() throws IOException {
String[] bpList = volume.getBlockPoolList();
ScanInfoPerBlockPool result = new ScanInfoPerBlockPool(bpList.length);
+ perfTimer.start();
+ throttleTimer.start();
for (String bpid : bpList) {
LinkedList<ScanInfo> report = new LinkedList<>();
File bpFinalizedDir = volume.getFinalizedDir(bpid);
- perfTimer.start();
- throttleTimer.start();
-
try {
result.put(bpid,
compileReport(volume, bpFinalizedDir, bpFinalizedDir, report));
http://git-wip-us.apache.org/repos/asf/hadoop/blob/4b85d38f/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
index 35bd7e8..c7fa7f0 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
@@ -47,16 +47,19 @@ import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.StorageType;
-import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.DFSClient;
+import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.DFSTestUtil;
import org.apache.hadoop.hdfs.HdfsConfiguration;
import org.apache.hadoop.hdfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSNNTopology;
import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
-import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
+import org.apache.hadoop.hdfs.server.datanode.checker.VolumeCheckResult;
+import org.apache.hadoop.hdfs.server.datanode.fsdataset.DataNodeVolumeMetrics;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi.FsVolumeReferences;
+import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeReference;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetTestUtil;
@@ -307,15 +310,22 @@ public class TestDirectoryScanner {
missingMemoryBlocks, mismatchBlocks, 0);
}
- private void scan(long totalBlocks, int diffsize, long missingMetaFile, long missingBlockFile,
- long missingMemoryBlocks, long mismatchBlocks, long duplicateBlocks) throws IOException {
+ private void scan(long totalBlocks, int diffsize, long missingMetaFile,
+ long missingBlockFile, long missingMemoryBlocks, long mismatchBlocks,
+ long duplicateBlocks) throws IOException {
scanner.reconcile();
-
+ verifyStats(totalBlocks, diffsize, missingMetaFile, missingBlockFile,
+ missingMemoryBlocks, mismatchBlocks, duplicateBlocks);
+ }
+
+ private void verifyStats(long totalBlocks, int diffsize, long missingMetaFile,
+ long missingBlockFile, long missingMemoryBlocks, long mismatchBlocks,
+ long duplicateBlocks) {
assertTrue(scanner.diffs.containsKey(bpid));
LinkedList<DirectoryScanner.ScanInfo> diff = scanner.diffs.get(bpid);
assertTrue(scanner.stats.containsKey(bpid));
DirectoryScanner.Stats stats = scanner.stats.get(bpid);
-
+
assertEquals(diffsize, diff.size());
assertEquals(totalBlocks, stats.totalBlocks);
assertEquals(missingMetaFile, stats.missingMetaFile);
@@ -991,4 +1001,48 @@ public class TestDirectoryScanner {
cluster.shutdown();
}
}
+
+ @Test
+ public void testDirectoryScannerInFederatedCluster() throws Exception {
+ //Create Federated cluster with two nameservices and one DN
+ try (MiniDFSCluster cluster = new MiniDFSCluster.Builder(CONF)
+ .nnTopology(MiniDFSNNTopology.simpleHAFederatedTopology(2))
+ .numDataNodes(1).build()) {
+ cluster.waitActive();
+ cluster.transitionToActive(1);
+ cluster.transitionToActive(3);
+ DataNode dataNode = cluster.getDataNodes().get(0);
+ fds = DataNodeTestUtils.getFSDataset(cluster.getDataNodes().get(0));
+ //Create one block in first nameservice
+ FileSystem fs = cluster.getFileSystem(1);
+ int bp1Files = 1;
+ writeFile(fs, bp1Files);
+ //Create two blocks in second nameservice
+ FileSystem fs2 = cluster.getFileSystem(3);
+ int bp2Files = 2;
+ writeFile(fs2, bp2Files);
+ //Call the Directory scanner
+ scanner = new DirectoryScanner(dataNode, fds, CONF);
+ scanner.setRetainDiffs(true);
+ scanner.reconcile();
+ //Check blocks in corresponding BP
+ bpid = cluster.getNamesystem(1).getBlockPoolId();
+ verifyStats(bp1Files, 0, 0, 0, 0, 0, 0);
+ bpid = cluster.getNamesystem(3).getBlockPoolId();
+ verifyStats(bp2Files, 0, 0, 0, 0, 0, 0);
+ } finally {
+ if (scanner != null) {
+ scanner.shutdown();
+ scanner = null;
+ }
+ }
+ }
+
+ private void writeFile(FileSystem fs, int numFiles) throws IOException {
+ final String fileName = "/" + GenericTestUtils.getMethodName();
+ final Path filePath = new Path(fileName);
+ for (int i = 0; i < numFiles; i++) {
+ DFSTestUtil.createFile(fs, filePath, 1, (short) 1, 0);
+ }
+ }
}
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org
[3/3] hadoop git commit: HDFS-12319. Addendum commit for missed
changes.
Posted by ar...@apache.org.
HDFS-12319. Addendum commit for missed changes.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/2a25ddc0
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/2a25ddc0
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/2a25ddc0
Branch: refs/heads/branch-2.8.2
Commit: 2a25ddc0b108ae64484dfedffcccd828a07bab5d
Parents: 4b85d38
Author: Arpit Agarwal <ar...@apache.org>
Authored: Fri Aug 25 12:02:59 2017 -0700
Committer: Arpit Agarwal <ar...@apache.org>
Committed: Fri Aug 25 12:03:42 2017 -0700
----------------------------------------------------------------------
.../apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java | 2 --
1 file changed, 2 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/2a25ddc0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
index c7fa7f0..e1b4082 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDirectoryScanner.java
@@ -56,8 +56,6 @@ import org.apache.hadoop.hdfs.MiniDFSNNTopology;
import org.apache.hadoop.hdfs.protocol.Block;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
-import org.apache.hadoop.hdfs.server.datanode.checker.VolumeCheckResult;
-import org.apache.hadoop.hdfs.server.datanode.fsdataset.DataNodeVolumeMetrics;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi.FsVolumeReferences;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeReference;
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org