You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by ta...@apache.org on 2021/04/06 05:19:23 UTC
[hadoop] branch branch-3.3 updated: HADOOP-17622. Avoid usage of
deprecated IOUtils#cleanup API. (#2862)
This is an automated email from the ASF dual-hosted git repository.
tasanuma pushed a commit to branch branch-3.3
in repository https://gitbox.apache.org/repos/asf/hadoop.git
The following commit(s) were added to refs/heads/branch-3.3 by this push:
new 8b4b3d6 HADOOP-17622. Avoid usage of deprecated IOUtils#cleanup API. (#2862)
8b4b3d6 is described below
commit 8b4b3d6fe62d567975dfad2211134e9982c4d4b5
Author: Viraj Jasani <vj...@apache.org>
AuthorDate: Tue Apr 6 10:09:10 2021 +0530
HADOOP-17622. Avoid usage of deprecated IOUtils#cleanup API. (#2862)
Signed-off-by: Takanobu Asanuma <ta...@apache.org>
(cherry picked from commit 3f2682b92b540be3ce15642ab8be463df87a4e4e)
---
.../java/org/apache/hadoop/fs/TestLocalFileSystem.java | 2 +-
.../java/org/apache/hadoop/metrics2/sink/TestFileSink.java | 2 +-
.../java/org/apache/hadoop/net/unix/TestDomainSocket.java | 2 +-
.../apache/hadoop/hdfs/server/datanode/BPServiceActor.java | 6 +++---
.../apache/hadoop/hdfs/server/datanode/BlockReceiver.java | 2 +-
.../apache/hadoop/hdfs/server/datanode/BlockScanner.java | 2 +-
.../apache/hadoop/hdfs/server/datanode/DataXceiver.java | 4 ++--
.../apache/hadoop/hdfs/server/datanode/VolumeScanner.java | 8 ++++----
.../hdfs/server/datanode/checker/DatasetVolumeChecker.java | 6 +++---
.../server/datanode/erasurecode/StripedBlockReader.java | 2 +-
.../server/datanode/fsdataset/ReplicaInputStreams.java | 4 ++--
.../datanode/fsdataset/impl/FsDatasetAsyncDiskService.java | 2 +-
.../hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java | 14 +++++++-------
.../hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java | 2 +-
.../hdfs/server/datanode/fsdataset/impl/FsVolumeList.java | 2 +-
.../hadoop/hdfs/server/namenode/EditsDoubleBuffer.java | 2 +-
.../main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java | 4 ++--
.../hdfs/tools/offlineImageViewer/PBImageTextWriter.java | 10 +++++-----
.../org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java | 2 +-
.../src/test/java/org/apache/hadoop/fs/TestUnbuffer.java | 4 ++--
.../org/apache/hadoop/fs/permission/TestStickyBit.java | 6 +++---
.../org/apache/hadoop/hdfs/TestDataTransferKeepalive.java | 2 +-
.../src/test/java/org/apache/hadoop/hdfs/TestHFlush.java | 2 +-
.../src/test/java/org/apache/hadoop/hdfs/TestPread.java | 4 ++--
.../java/org/apache/hadoop/hdfs/TestRollingUpgrade.java | 2 +-
.../hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java | 2 +-
.../hadoop/hdfs/TestTrashWithSecureEncryptionZones.java | 2 +-
.../apache/hadoop/hdfs/TestWriteConfigurationToDFS.java | 2 +-
.../protocol/datatransfer/sasl/TestSaslDataTransfer.java | 2 +-
.../java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java | 2 +-
.../apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java | 4 ++--
.../hadoop/hdfs/server/datanode/TestBlockReplacement.java | 2 +-
.../hadoop/hdfs/server/datanode/TestCachingStrategy.java | 2 +-
.../apache/hadoop/hdfs/server/mover/TestStorageMover.java | 2 +-
.../apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java | 3 ++-
.../hadoop/hdfs/server/namenode/FSXAttrBaseTest.java | 2 +-
.../hadoop/hdfs/server/namenode/TestAclConfigFlag.java | 2 +-
.../hadoop/hdfs/server/namenode/TestAddStripedBlocks.java | 6 +++---
.../namenode/TestCommitBlockWithInvalidGenStamp.java | 2 +-
.../apache/hadoop/hdfs/server/namenode/TestEditLog.java | 6 ++++--
.../hdfs/server/namenode/TestEditLogFileOutputStream.java | 2 +-
.../org/apache/hadoop/hdfs/server/namenode/TestFsck.java | 2 +-
.../apache/hadoop/hdfs/server/namenode/TestMetaSave.java | 4 ++--
.../hdfs/server/namenode/TestQuotaWithStripedBlocks.java | 2 +-
.../hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java | 2 +-
.../hdfs/server/namenode/ha/TestRetryCacheWithHA.java | 2 +-
.../hdfs/server/namenode/snapshot/TestAclWithSnapshot.java | 2 +-
.../server/namenode/snapshot/TestXAttrWithSnapshot.java | 2 +-
.../hadoop/hdfs/util/TestAtomicFileOutputStream.java | 2 +-
.../java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java | 12 ++++++------
.../test/java/org/apache/hadoop/tools/TestDistCpSync.java | 2 +-
.../org/apache/hadoop/tools/TestDistCpSyncReverseBase.java | 2 +-
.../java/org/apache/hadoop/tools/TestDistCpWithAcls.java | 2 +-
.../org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java | 2 +-
.../java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java | 2 +-
.../org/apache/hadoop/tools/TestGlobbedCopyListing.java | 4 ++--
.../org/apache/hadoop/tools/mapred/TestCopyMapper.java | 2 +-
.../hadoop/tools/mapred/TestUniformSizeInputFormat.java | 2 +-
.../hadoop/tools/mapred/lib/TestDynamicInputFormat.java | 2 +-
.../apache/hadoop/mapred/gridmix/ReadRecordFactory.java | 2 +-
.../org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java | 2 +-
.../org/apache/hadoop/mapred/gridmix/SerialJobFactory.java | 2 +-
.../org/apache/hadoop/mapred/gridmix/StressJobFactory.java | 2 +-
.../main/java/org/apache/hadoop/tools/rumen/Folder.java | 2 +-
64 files changed, 102 insertions(+), 99 deletions(-)
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java
index 1384bb6..e7b72a9 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestLocalFileSystem.java
@@ -312,7 +312,7 @@ public class TestLocalFileSystem {
.new LocalFSFileInputStream(path), 1024);
assertNotNull(bis.getFileDescriptor());
} finally {
- IOUtils.cleanup(null, bis);
+ IOUtils.cleanupWithLogger(null, bis);
}
}
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java
index b20653e..6788940 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/sink/TestFileSink.java
@@ -115,7 +115,7 @@ public class TestFileSink {
IOUtils.copyBytes(is, baos, 1024, true);
outFileContent = new String(baos.toByteArray(), "UTF-8");
} finally {
- IOUtils.cleanup(null, baos, is);
+ IOUtils.cleanupWithLogger(null, baos, is);
}
// Check the out file content. Should be something like the following:
diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java
index 466c83e..61cbd85 100644
--- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java
+++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/net/unix/TestDomainSocket.java
@@ -759,6 +759,6 @@ public class TestDomainSocket {
readerThread.join();
Assert.assertFalse(failed.get());
Assert.assertEquals(3, bytesRead.get());
- IOUtils.cleanup(null, socks);
+ IOUtils.cleanupWithLogger(null, socks);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java
index 0cc9fac..3a5094d 100755
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BPServiceActor.java
@@ -617,8 +617,8 @@ class BPServiceActor implements Runnable {
private synchronized void cleanUp() {
shouldServiceRun = false;
- IOUtils.cleanup(null, bpNamenode);
- IOUtils.cleanup(null, lifelineSender);
+ IOUtils.cleanupWithLogger(null, bpNamenode);
+ IOUtils.cleanupWithLogger(null, lifelineSender);
bpos.shutdownActor(this);
}
@@ -989,7 +989,7 @@ class BPServiceActor implements Runnable {
} catch (InterruptedException e) {
Thread.currentThread().interrupt();
}
- IOUtils.cleanup(null, lifelineNamenode);
+ IOUtils.cleanupWithLogger(null, lifelineNamenode);
}
@Override
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
index cc9551d..e140394 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
@@ -369,7 +369,7 @@ class BlockReceiver implements Closeable {
streams.close();
}
if (replicaHandler != null) {
- IOUtils.cleanup(null, replicaHandler);
+ IOUtils.cleanupWithLogger(null, replicaHandler);
replicaHandler = null;
}
if (measuredFlushTime) {
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java
index 485cf00..6dcfad4 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockScanner.java
@@ -252,7 +252,7 @@ public class BlockScanner {
if (!success) {
// If we didn't create a new VolumeScanner object, we don't
// need this reference to the volume.
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java
index 7df5caf..c79a076 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java
@@ -431,7 +431,7 @@ class DataXceiver extends Receiver implements Runnable {
blk.getBlockId(), dnR.getDatanodeUuid(), success));
}
if (fis != null) {
- IOUtils.cleanup(null, fis);
+ IOUtils.cleanupWithLogger(null, fis);
}
}
}
@@ -554,7 +554,7 @@ class DataXceiver extends Receiver implements Runnable {
LOG.warn("Failed to shut down socket in error handler", e);
}
}
- IOUtils.cleanup(null, shmInfo);
+ IOUtils.cleanupWithLogger(null, shmInfo);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java
index 6bc25eb..0367b4a 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/VolumeScanner.java
@@ -451,7 +451,7 @@ public class VolumeScanner extends Thread {
} catch (IOException e) {
resultHandler.handle(block, e);
} finally {
- IOUtils.cleanup(null, blockSender);
+ IOUtils.cleanupWithLogger(null, blockSender);
}
metrics.incrBlockVerificationFailures();
return -1;
@@ -674,13 +674,13 @@ public class VolumeScanner extends Thread {
// Save the current position of all block iterators and close them.
for (BlockIterator iter : blockIters) {
saveBlockIterator(iter);
- IOUtils.cleanup(null, iter);
+ IOUtils.cleanupWithLogger(null, iter);
}
} finally {
VolumeScannerCBInjector.get().terminationCallBack(this);
// When the VolumeScanner exits, release the reference we were holding
// on the volume. This will allow the volume to be removed later.
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
}
}
@@ -767,7 +767,7 @@ public class VolumeScanner extends Thread {
if (iter.getBlockPoolId().equals(bpid)) {
LOG.trace("{}: disabling scanning on block pool {}", this, bpid);
i.remove();
- IOUtils.cleanup(null, iter);
+ IOUtils.cleanupWithLogger(null, iter);
if (curBlockIter == iter) {
curBlockIter = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
index d077d21..997a6d9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/checker/DatasetVolumeChecker.java
@@ -234,7 +234,7 @@ public class DatasetVolumeChecker {
}
}), MoreExecutors.directExecutor());
} else {
- IOUtils.cleanup(null, reference);
+ IOUtils.cleanupWithLogger(null, reference);
if (numVolumes.decrementAndGet() == 0) {
latch.countDown();
}
@@ -311,7 +311,7 @@ public class DatasetVolumeChecker {
);
return true;
} else {
- IOUtils.cleanup(null, volumeReference);
+ IOUtils.cleanupWithLogger(null, volumeReference);
}
return false;
}
@@ -404,7 +404,7 @@ public class DatasetVolumeChecker {
}
private void cleanup() {
- IOUtils.cleanup(null, reference);
+ IOUtils.cleanupWithLogger(null, reference);
invokeCallback();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java
index b1ad03f..54302e3 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/erasurecode/StripedBlockReader.java
@@ -158,7 +158,7 @@ class StripedBlockReader {
return peer;
} finally {
if (!success) {
- IOUtils.cleanup(null, peer);
+ IOUtils.cleanupWithLogger(null, peer);
IOUtils.closeSocket(sock);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
index f40315a..f8bd8c0 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/ReplicaInputStreams.java
@@ -130,7 +130,7 @@ public class ReplicaInputStreams implements Closeable {
dataInFd = null;
}
if (volumeRef != null) {
- IOUtils.cleanup(null, volumeRef);
+ IOUtils.cleanupWithLogger(null, volumeRef);
volumeRef = null;
}
// throw IOException if there is any
@@ -146,7 +146,7 @@ public class ReplicaInputStreams implements Closeable {
dataInFd = null;
IOUtils.closeStream(checksumIn);
checksumIn = null;
- IOUtils.cleanup(null, volumeRef);
+ IOUtils.cleanupWithLogger(null, volumeRef);
volumeRef = null;
}
}
\ No newline at end of file
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
index 81213a0..2a89a80d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetAsyncDiskService.java
@@ -335,7 +335,7 @@ class FsDatasetAsyncDiskService {
+ block.getLocalBlock() + " URI " + replicaToDelete.getBlockURI());
}
updateDeletedBlockId(block);
- IOUtils.cleanup(null, volumeRef);
+ IOUtils.cleanupWithLogger(null, volumeRef);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
index 4526eca..e3f922c 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java
@@ -905,11 +905,11 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
return new ReplicaInputStreams(
blockInStream, metaInStream, ref, datanode.getFileIoProvider());
} catch (IOException e) {
- IOUtils.cleanup(null, blockInStream);
+ IOUtils.cleanupWithLogger(null, blockInStream);
throw e;
}
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
}
@@ -1252,7 +1252,7 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
replica = append(b.getBlockPoolId(), replicaInfo, newGS,
b.getNumBytes());
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
return new ReplicaHandler(replica, ref);
@@ -1384,7 +1384,7 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
replica = (ReplicaInPipeline) replicaInfo;
}
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
return new ReplicaHandler(replica, ref);
@@ -1478,7 +1478,7 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
+ " for block " + b.getBlockId());
}
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
@@ -1570,7 +1570,7 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
// bump the replica's generation stamp to newGS
rbw.getReplicaInfo().bumpReplicaGS(newGS);
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
return new ReplicaHandler(rbw, ref);
@@ -1716,7 +1716,7 @@ class FsDatasetImpl implements FsDatasetSpi<FsVolumeImpl> {
try {
newReplicaInfo = v.createTemporary(b);
} catch (IOException e) {
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
throw e;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java
index b5d2b95e..b804de4 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java
@@ -117,7 +117,7 @@ public class FsDatasetUtil {
}
return raf.getFD();
} catch(IOException ioe) {
- IOUtils.cleanup(null, raf);
+ IOUtils.cleanupWithLogger(null, raf);
throw ioe;
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java
index 049654b..2f4f7cf 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsVolumeList.java
@@ -296,7 +296,7 @@ class FsVolumeList {
} else {
// If the volume is not put into a volume scanner, it does not need to
// hold the reference.
- IOUtils.cleanup(null, ref);
+ IOUtils.cleanupWithLogger(null, ref);
}
// If the volume is used to replace a failed volume, it needs to reset the
// volume failure info for this volume.
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java
index affba02..ccc233e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java
@@ -76,7 +76,7 @@ public class EditsDoubleBuffer {
+ " bytes still to be flushed and cannot be closed.");
}
- IOUtils.cleanup(null, bufCurrent, bufReady);
+ IOUtils.cleanupWithLogger(null, bufCurrent, bufReady);
bufCurrent = bufReady = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java
index 1784ea2..f5967e1 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/DebugAdmin.java
@@ -202,7 +202,7 @@ public class DebugAdmin extends Configured implements Tool {
blockFile);
return 0;
} finally {
- IOUtils.cleanup(null, metaStream, dataStream, checksumStream);
+ IOUtils.cleanupWithLogger(null, metaStream, dataStream, checksumStream);
}
}
}
@@ -287,7 +287,7 @@ public class DebugAdmin extends Configured implements Tool {
+ " saved metadata to meta file " + outFile);
return 0;
} finally {
- IOUtils.cleanup(null, metaOut);
+ IOUtils.cleanupWithLogger(null, metaOut);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java
index cd4047d..ccab7b0 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/PBImageTextWriter.java
@@ -319,10 +319,10 @@ abstract class PBImageTextWriter implements Closeable {
@Override
public void close() throws IOException {
if (batch != null) {
- IOUtils.cleanup(null, batch);
+ IOUtils.cleanupWithLogger(null, batch);
batch = null;
}
- IOUtils.cleanup(null, db);
+ IOUtils.cleanupWithLogger(null, db);
db = null;
}
@@ -388,13 +388,13 @@ abstract class PBImageTextWriter implements Closeable {
dirMap = new LevelDBStore(new File(dbDir, "dirMap"));
} catch (IOException e) {
LOG.error("Failed to open LevelDBs", e);
- IOUtils.cleanup(null, this);
+ IOUtils.cleanupWithLogger(null, this);
}
}
@Override
public void close() throws IOException {
- IOUtils.cleanup(null, dirChildMap, dirMap);
+ IOUtils.cleanupWithLogger(null, dirChildMap, dirMap);
dirChildMap = null;
dirMap = null;
}
@@ -515,7 +515,7 @@ abstract class PBImageTextWriter implements Closeable {
@Override
public void close() throws IOException {
out.flush();
- IOUtils.cleanup(null, metadataMap);
+ IOUtils.cleanupWithLogger(null, metadataMap);
}
void append(StringBuffer buffer, int field) {
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java
index f4c264c..99b1ddb 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestEnhancedByteBufferAccess.java
@@ -841,7 +841,7 @@ public class TestEnhancedByteBufferAccess {
if (buf2 != null) {
fsIn2.releaseBuffer(buf2);
}
- IOUtils.cleanup(null, fsIn, fsIn2);
+ IOUtils.cleanupWithLogger(null, fsIn, fsIn2);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java
index ef4c04d..feb77f8 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestUnbuffer.java
@@ -87,7 +87,7 @@ public class TestUnbuffer {
Assert.assertEquals(b, b2);
} finally {
if (stream != null) {
- IOUtils.cleanup(null, stream);
+ IOUtils.cleanupWithLogger(null, stream);
}
if (cluster != null) {
cluster.shutdown();
@@ -122,7 +122,7 @@ public class TestUnbuffer {
}
} finally {
for (FSDataInputStream stream : streams) {
- IOUtils.cleanup(null, stream);
+ IOUtils.cleanupWithLogger(null, stream);
}
if (cluster != null) {
cluster.shutdown();
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java
index a6409fd..96d15e5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/permission/TestStickyBit.java
@@ -96,7 +96,7 @@ public class TestStickyBit {
@AfterClass
public static void shutdown() throws Exception {
- IOUtils.cleanup(null, hdfs, hdfsAsUser1, hdfsAsUser2);
+ IOUtils.cleanupWithLogger(null, hdfs, hdfsAsUser1, hdfsAsUser2);
if (cluster != null) {
cluster.shutdown();
}
@@ -121,7 +121,7 @@ public class TestStickyBit {
h.close();
h = null;
} finally {
- IOUtils.cleanup(null, h);
+ IOUtils.cleanupWithLogger(null, h);
}
}
@@ -500,7 +500,7 @@ public class TestStickyBit {
o.close();
o = null;
} finally {
- IOUtils.cleanup(null, o);
+ IOUtils.cleanupWithLogger(null, o);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java
index 0bf21ee..9881f92 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDataTransferKeepalive.java
@@ -227,7 +227,7 @@ public class TestDataTransferKeepalive {
IOUtils.copyBytes(stm, new IOUtils.NullOutputStream(), 1024);
}
} finally {
- IOUtils.cleanup(null, stms);
+ IOUtils.cleanupWithLogger(null, stms);
}
assertEquals(5, peerCache.size());
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java
index ce49f69..dcc5bb6 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHFlush.java
@@ -177,7 +177,7 @@ public class TestHFlush {
blocks = fileSystem.dfs.getLocatedBlocks(path.toString(), 0);
assertEquals(3, blocks.getLocatedBlocks().size());
} finally {
- IOUtils.cleanup(null, stm, fileSystem);
+ IOUtils.cleanupWithLogger(null, stm, fileSystem);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java
index ea332c8..df6b7dc 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestPread.java
@@ -363,8 +363,8 @@ public class TestPread {
assertTrue(false);
} finally {
Mockito.reset(injector);
- IOUtils.cleanup(null, input);
- IOUtils.cleanup(null, output);
+ IOUtils.cleanupWithLogger(null, input);
+ IOUtils.cleanupWithLogger(null, output);
fileSys.close();
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java
index 46fab47..a716335 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestRollingUpgrade.java
@@ -741,7 +741,7 @@ public class TestRollingUpgrade {
// do checkpoint in SNN again
snn.doCheckpoint();
} finally {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (snn != null) {
snn.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java
index db97c02..fcf1333 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSecureEncryptionZoneWithKMS.java
@@ -273,7 +273,7 @@ public class TestSecureEncryptionZoneWithKMS {
@After
public void shutdown() throws IOException {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java
index cce1454..4728884 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestTrashWithSecureEncryptionZones.java
@@ -229,7 +229,7 @@ public class TestTrashWithSecureEncryptionZones {
@AfterClass
public static void destroy() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java
index 5503238..52abbe9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestWriteConfigurationToDFS.java
@@ -54,7 +54,7 @@ public class TestWriteConfigurationToDFS {
fs.close();
fs = null;
} finally {
- IOUtils.cleanup(null, os, fs);
+ IOUtils.cleanupWithLogger(null, os, fs);
cluster.shutdown();
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java
index d6612c1..3dd0b7e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/protocol/datatransfer/sasl/TestSaslDataTransfer.java
@@ -256,7 +256,7 @@ public class TestSaslDataTransfer extends SaslDataTransferTestCase {
} catch (SocketTimeoutException e) {
GenericTestUtils.assertExceptionContains("Read timed out", e);
} finally {
- IOUtils.cleanup(null, socket, serverSocket);
+ IOUtils.cleanupWithLogger(null, socket, serverSocket);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java
index 539f216..9168057 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/QJMTestUtil.java
@@ -173,7 +173,7 @@ public abstract class QJMTestUtil {
lastRecoveredTxn = elis.getLastTxId();
}
} finally {
- IOUtils.cleanup(null, streams.toArray(new Closeable[0]));
+ IOUtils.cleanupWithLogger(null, streams.toArray(new Closeable[0]));
}
return lastRecoveredTxn;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java
index 35bae49..eb64cf5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/TestSecureNNWithQJM.java
@@ -163,7 +163,7 @@ public class TestSecureNNWithQJM {
@After
public void shutdown() throws IOException {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
@@ -213,7 +213,7 @@ public class TestSecureNNWithQJM {
* @throws IOException if there is an I/O error
*/
private void restartNameNode() throws IOException {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
cluster.restartNameNode();
fs = cluster.getFileSystem();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java
index 67b41f8..3f99f1b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestBlockReplacement.java
@@ -479,7 +479,7 @@ public class TestBlockReplacement {
assertEquals("The block should be only on 1 datanode ", 1,
locatedBlocks1.get(0).getLocations().length);
} finally {
- IOUtils.cleanup(null, client);
+ IOUtils.cleanupWithLogger(null, client);
cluster.shutdown();
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java
index 459ef88..011df46 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestCachingStrategy.java
@@ -332,7 +332,7 @@ public class TestCachingStrategy {
stats = tracker.getStats(fadvisedFileName);
stats.assertNotDroppedInRange(0, TEST_PATH_LEN - WRITE_PACKET_SIZE);
} finally {
- IOUtils.cleanup(null, fis);
+ IOUtils.cleanupWithLogger(null, fis);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java
index fec04be..d95e76f 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/mover/TestStorageMover.java
@@ -230,7 +230,7 @@ public class TestStorageMover {
}
void shutdownCluster() throws Exception {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java
index 8a2c0e2..3a9ad25 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSAclBaseTest.java
@@ -108,7 +108,8 @@ public abstract class FSAclBaseTest {
@After
public void destroyFileSystems() {
- IOUtils.cleanup(null, fs, fsAsBruce, fsAsDiana, fsAsSupergroupMember);
+ IOUtils.cleanupWithLogger(null, fs, fsAsBruce, fsAsDiana,
+ fsAsSupergroupMember);
fs = fsAsBruce = fsAsDiana = fsAsSupergroupMember = fsAsBob = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java
index 019531c..44e86ae 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/FSXAttrBaseTest.java
@@ -127,7 +127,7 @@ public class FSXAttrBaseTest {
@After
public void destroyFileSystems() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
fs = null;
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java
index f45728f..8e01f73 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAclConfigFlag.java
@@ -55,7 +55,7 @@ public class TestAclConfigFlag {
@After
public void shutdown() throws Exception {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java
index d17a36f..ef91fc3 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestAddStripedBlocks.java
@@ -206,7 +206,7 @@ public class TestAddStripedBlocks {
assertTrue(blocks[0].isStriped());
checkStripedBlockUC((BlockInfoStriped) fileNode.getLastBlock(), false);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
@@ -271,7 +271,7 @@ public class TestAddStripedBlocks {
assertArrayEquals(indices, blockIndices);
assertArrayEquals(expectedDNs, datanodes);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
@@ -327,7 +327,7 @@ public class TestAddStripedBlocks {
assertTrue(storageIDs.contains(newstorage.getStorageID()));
}
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
// 3. restart the namenode. mimic the full block reports and check the
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java
index 77a3268..f7b1ea5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCommitBlockWithInvalidGenStamp.java
@@ -95,7 +95,7 @@ public class TestCommitBlockWithInvalidGenStamp {
dfs.getClient().getClientName(), previous, fileNode.getId());
Assert.assertTrue("should complete successfully", complete);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
index 0592b97..0fa8dfa 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
@@ -1548,7 +1548,8 @@ public class TestEditLog {
LOG.error("edit log failover didn't work", e);
fail("Edit log failover didn't work");
} finally {
- IOUtils.cleanup(null, streams.toArray(new EditLogInputStream[0]));
+ IOUtils.cleanupWithLogger(null,
+ streams.toArray(new EditLogInputStream[0]));
}
}
@@ -1598,7 +1599,8 @@ public class TestEditLog {
LOG.error("edit log failover didn't work", e);
fail("Edit log failover didn't work");
} finally {
- IOUtils.cleanup(null, streams.toArray(new EditLogInputStream[0]));
+ IOUtils.cleanupWithLogger(null,
+ streams.toArray(new EditLogInputStream[0]));
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java
index 67b48b1..9ea4548 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogFileOutputStream.java
@@ -154,7 +154,7 @@ public class TestEditLogFileOutputStream {
editLogStream.abort();
editLogStream.abort();
} finally {
- IOUtils.cleanup(null, editLogStream);
+ IOUtils.cleanupWithLogger(null, editLogStream);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java
index d89bb6b..f2630d2 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFsck.java
@@ -567,7 +567,7 @@ public class TestFsck {
}
}
} finally {
- IOUtils.cleanup(null, in);
+ IOUtils.cleanupWithLogger(null, in);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java
index 0c9b499..c88570b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestMetaSave.java
@@ -212,7 +212,7 @@ public class TestMetaSave {
line = rdr.readLine();
}
} finally {
- IOUtils.cleanup(null, rdr, isr, fis);
+ IOUtils.cleanupWithLogger(null, rdr, isr, fis);
}
}
@@ -271,7 +271,7 @@ public class TestMetaSave {
line = rdr.readLine();
}
} finally {
- IOUtils.cleanup(null, rdr, isr, fis);
+ IOUtils.cleanupWithLogger(null, rdr, isr, fis);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java
index 38b98a4..e89ed8d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestQuotaWithStripedBlocks.java
@@ -140,7 +140,7 @@ public class TestQuotaWithStripedBlocks {
Assert.assertEquals(cellSize * groupSize,
actualDiskUsed);
} finally {
- IOUtils.cleanup(null, out);
+ IOUtils.cleanupWithLogger(null, out);
}
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java
index 5064a03..c527837 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrConfigFlag.java
@@ -48,7 +48,7 @@ public class TestXAttrConfigFlag {
@After
public void shutdown() throws Exception {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
cluster = null;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java
index 9074247..e3e934b 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestRetryCacheWithHA.java
@@ -1315,7 +1315,7 @@ public class TestRetryCacheWithHA {
} catch (Exception e) {
LOG.info("Got Exception while calling " + op.name, e);
} finally {
- IOUtils.cleanup(null, op.client);
+ IOUtils.cleanupWithLogger(null, op.client);
}
}
}.start();
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java
index ea9c5e2..511607e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestAclWithSnapshot.java
@@ -86,7 +86,7 @@ public class TestAclWithSnapshot {
@AfterClass
public static void shutdown() throws Exception {
- IOUtils.cleanup(null, hdfs, fsAsBruce, fsAsDiana);
+ IOUtils.cleanupWithLogger(null, hdfs, fsAsBruce, fsAsDiana);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java
index 4b957bf..2c93e12 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/snapshot/TestXAttrWithSnapshot.java
@@ -79,7 +79,7 @@ public class TestXAttrWithSnapshot {
@AfterClass
public static void shutdown() throws Exception {
- IOUtils.cleanup(null, hdfs);
+ IOUtils.cleanupWithLogger(null, hdfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java
index bce92ef..144f990 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/util/TestAtomicFileOutputStream.java
@@ -142,7 +142,7 @@ public class TestAtomicFileOutputStream {
fos = null;
}
} finally {
- IOUtils.cleanup(null, fos);
+ IOUtils.cleanupWithLogger(null, fos);
FileUtil.setWritable(TEST_DIR, true);
}
}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java
index 2212939..1fc54d6 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHDFSForHA.java
@@ -95,7 +95,7 @@ public class TestWebHDFSForHA {
final Path dir2 = new Path("/test2");
Assert.assertTrue(fs.mkdirs(dir2));
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -130,7 +130,7 @@ public class TestWebHDFSForHA {
verify(fs).renewDelegationToken(token);
verify(fs).cancelDelegationToken(token);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -199,7 +199,7 @@ public class TestWebHDFSForHA {
Exception unwrapped = re.unwrapRemoteException(StandbyException.class);
Assert.assertTrue(unwrapped instanceof StandbyException);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -238,7 +238,7 @@ public class TestWebHDFSForHA {
IOUtils.readFully(in, buf, 0, buf.length);
Assert.assertArrayEquals(data, buf);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -264,7 +264,7 @@ public class TestWebHDFSForHA {
fs = (WebHdfsFileSystem)FileSystem.get(WEBHDFS_URI, conf);
Assert.assertEquals(2, fs.getResolvedNNAddr().length);
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
@@ -304,7 +304,7 @@ public class TestWebHDFSForHA {
} catch (IOException e) {
result = false;
} finally {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
}
synchronized (TestWebHDFSForHA.this) {
resultMap.put("mkdirs", result);
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java
index d6bbc25..c80a79b 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSync.java
@@ -83,7 +83,7 @@ public class TestDistCpSync {
@After
public void tearDown() throws Exception {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java
index cca1c53..50f5823 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpSyncReverseBase.java
@@ -153,7 +153,7 @@ public abstract class TestDistCpSyncReverseBase {
@After
public void tearDown() throws Exception {
- IOUtils.cleanup(null, dfs);
+ IOUtils.cleanupWithLogger(null, dfs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java
index 49613ba..38b7933 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithAcls.java
@@ -96,7 +96,7 @@ public class TestDistCpWithAcls {
@AfterClass
public static void shutdown() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java
index d0e3258..4637f6d 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithRawXAttrs.java
@@ -73,7 +73,7 @@ public class TestDistCpWithRawXAttrs {
@AfterClass
public static void shutdown() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java
index 77d428b..36c6e6a 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestDistCpWithXAttrs.java
@@ -114,7 +114,7 @@ public class TestDistCpWithXAttrs {
@AfterClass
public static void shutdown() {
- IOUtils.cleanup(null, fs);
+ IOUtils.cleanupWithLogger(null, fs);
if (cluster != null) {
cluster.shutdown();
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java
index 1c92a9c..389fe36 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/TestGlobbedCopyListing.java
@@ -72,7 +72,7 @@ public class TestGlobbedCopyListing {
recordInExpectedValues(path);
}
finally {
- IOUtils.cleanup(null, fileSystem);
+ IOUtils.cleanupWithLogger(null, fileSystem);
}
}
@@ -85,7 +85,7 @@ public class TestGlobbedCopyListing {
recordInExpectedValues(path);
}
finally {
- IOUtils.cleanup(null, fileSystem, outputStream);
+ IOUtils.cleanupWithLogger(null, fileSystem, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java
index bf31657..780d82d 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestCopyMapper.java
@@ -261,7 +261,7 @@ public class TestCopyMapper {
System.out.println(fileStatus.getReplication());
}
finally {
- IOUtils.cleanup(null, outputStream);
+ IOUtils.cleanupWithLogger(null, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java
index 5315137..622e391 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/TestUniformSizeInputFormat.java
@@ -91,7 +91,7 @@ public class TestUniformSizeInputFormat {
return size;
}
finally {
- IOUtils.cleanup(null, fileSystem, outputStream);
+ IOUtils.cleanupWithLogger(null, fileSystem, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java
index 61a1745..6a310ca 100644
--- a/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java
+++ b/hadoop-tools/hadoop-distcp/src/test/java/org/apache/hadoop/tools/mapred/lib/TestDynamicInputFormat.java
@@ -100,7 +100,7 @@ public class TestDynamicInputFormat {
new Path(path))[0].getPath().toString());
}
finally {
- IOUtils.cleanup(null, fileSystem, outputStream);
+ IOUtils.cleanupWithLogger(null, fileSystem, outputStream);
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java
index 2cb806e..f95c4b3 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReadRecordFactory.java
@@ -79,7 +79,7 @@ class ReadRecordFactory extends RecordFactory {
@Override
public void close() throws IOException {
- IOUtils.cleanup(null, src);
+ IOUtils.cleanupWithLogger(null, src);
factory.close();
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java
index bdbfc3b..fe3b5d3 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/ReplayJobFactory.java
@@ -112,7 +112,7 @@ import java.util.concurrent.TimeUnit;
} catch (InterruptedException e) {
// exit thread; ignore any jobs remaining in the trace
} finally {
- IOUtils.cleanup(null, jobProducer);
+ IOUtils.cleanupWithLogger(null, jobProducer);
}
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java
index e8c7d61..cb05ab6 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/SerialJobFactory.java
@@ -143,7 +143,7 @@ public class SerialJobFactory extends JobFactory<JobStats> {
} catch (InterruptedException e) {
return;
} finally {
- IOUtils.cleanup(null, jobProducer);
+ IOUtils.cleanupWithLogger(null, jobProducer);
}
}
diff --git a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java
index bd15c29..4e7fc9c 100644
--- a/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java
+++ b/hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/StressJobFactory.java
@@ -247,7 +247,7 @@ public class StressJobFactory extends JobFactory<Statistics.ClusterStats> {
LOG.error("[STRESS] Interrupted in the main block!", e);
return;
} finally {
- IOUtils.cleanup(null, jobProducer);
+ IOUtils.cleanupWithLogger(null, jobProducer);
}
}
}
diff --git a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java
index 424405a..683bcc8 100644
--- a/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java
+++ b/hadoop-tools/hadoop-rumen/src/main/java/org/apache/hadoop/tools/rumen/Folder.java
@@ -470,7 +470,7 @@ public class Folder extends Configured implements Tool {
next = heap.poll();
}
} finally {
- IOUtils.cleanup(null, reader);
+ IOUtils.cleanupWithLogger(null, reader);
if (outGen != null) {
outGen.close();
}
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org