You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by aa...@apache.org on 2017/05/29 08:31:17 UTC
hadoop git commit: HDFS-11832. Switch leftover logs to slf4j format
in BlockManager.java. Contributed by Hui Xu and Chen Liang.
Repository: hadoop
Updated Branches:
refs/heads/trunk 6c6a7a596 -> a7f085d6b
HDFS-11832. Switch leftover logs to slf4j format in BlockManager.java. Contributed by Hui Xu and Chen Liang.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/a7f085d6
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/a7f085d6
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/a7f085d6
Branch: refs/heads/trunk
Commit: a7f085d6bf499edf23e650a4f7211c53a442da0e
Parents: 6c6a7a5
Author: Akira Ajisaka <aa...@apache.org>
Authored: Mon May 29 17:30:23 2017 +0900
Committer: Akira Ajisaka <aa...@apache.org>
Committed: Mon May 29 17:30:23 2017 +0900
----------------------------------------------------------------------
.../server/blockmanagement/BlockManager.java | 153 ++++++++-----------
.../blockmanagement/InvalidateBlocks.java | 17 +--
2 files changed, 72 insertions(+), 98 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hadoop/blob/a7f085d6/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
index f0c12cd..42ee850 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
@@ -498,14 +498,13 @@ public class BlockManager implements BlockStatsMXBean {
bmSafeMode = new BlockManagerSafeMode(this, namesystem, haEnabled, conf);
- LOG.info("defaultReplication = " + defaultReplication);
- LOG.info("maxReplication = " + maxReplication);
- LOG.info("minReplication = " + minReplication);
- LOG.info("maxReplicationStreams = " + maxReplicationStreams);
- LOG.info("redundancyRecheckInterval = " + redundancyRecheckIntervalMs +
- "ms");
- LOG.info("encryptDataTransfer = " + encryptDataTransfer);
- LOG.info("maxNumBlocksToLog = " + maxNumBlocksToLog);
+ LOG.info("defaultReplication = {}", defaultReplication);
+ LOG.info("maxReplication = {}", maxReplication);
+ LOG.info("minReplication = {}", minReplication);
+ LOG.info("maxReplicationStreams = {}", maxReplicationStreams);
+ LOG.info("redundancyRecheckInterval = {}ms", redundancyRecheckIntervalMs);
+ LOG.info("encryptDataTransfer = {}", encryptDataTransfer);
+ LOG.info("maxNumBlocksToLog = {}", maxNumBlocksToLog);
}
private static BlockTokenSecretManager createBlockTokenSecretManager(
@@ -513,7 +512,8 @@ public class BlockManager implements BlockStatsMXBean {
final boolean isEnabled = conf.getBoolean(
DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY,
DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_DEFAULT);
- LOG.info(DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY + "=" + isEnabled);
+ LOG.info("{} = {}", DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY,
+ isEnabled);
if (!isEnabled) {
if (UserGroupInformation.isSecurityEnabled()) {
@@ -534,12 +534,10 @@ public class BlockManager implements BlockStatsMXBean {
DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_LIFETIME_DEFAULT);
final String encryptionAlgorithm = conf.get(
DFSConfigKeys.DFS_DATA_ENCRYPTION_ALGORITHM_KEY);
- LOG.info(DFSConfigKeys.DFS_BLOCK_ACCESS_KEY_UPDATE_INTERVAL_KEY
- + "=" + updateMin + " min(s), "
- + DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_LIFETIME_KEY
- + "=" + lifetimeMin + " min(s), "
- + DFSConfigKeys.DFS_DATA_ENCRYPTION_ALGORITHM_KEY
- + "=" + encryptionAlgorithm);
+ LOG.info("{}={} min(s), {}={} min(s), {}={}",
+ DFSConfigKeys.DFS_BLOCK_ACCESS_KEY_UPDATE_INTERVAL_KEY, updateMin,
+ DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_LIFETIME_KEY, lifetimeMin,
+ DFSConfigKeys.DFS_DATA_ENCRYPTION_ALGORITHM_KEY, encryptionAlgorithm);
String nsId = DFSUtil.getNamenodeNameServiceId(conf);
boolean isHaEnabled = HAUtil.isHAEnabled(conf, nsId);
@@ -692,8 +690,8 @@ public class BlockManager implements BlockStatsMXBean {
Collection<DatanodeDescriptor> corruptNodes =
corruptReplicas.getNodes(block);
if (corruptNodes == null) {
- LOG.warn(block.getBlockId() +
- " is corrupt but has no associated node.");
+ LOG.warn("{} is corrupt but has no associated node.",
+ block.getBlockId());
continue;
}
int numNodesToFind = corruptNodes.size();
@@ -1156,9 +1154,9 @@ public class BlockManager implements BlockStatsMXBean {
final int numCorruptNodes = numReplicas.corruptReplicas();
final int numCorruptReplicas = corruptReplicas.numCorruptReplicas(blk);
if (numCorruptNodes != numCorruptReplicas) {
- LOG.warn("Inconsistent number of corrupt replicas for "
- + blk + " blockMap has " + numCorruptNodes
- + " but corrupt replicas map has " + numCorruptReplicas);
+ LOG.warn("Inconsistent number of corrupt replicas for {}"
+ + " blockMap has {} but corrupt replicas map has {}",
+ blk, numCorruptNodes, numCorruptReplicas);
}
final int numNodes = blocksMap.numNodes(blk);
@@ -1232,7 +1230,7 @@ public class BlockManager implements BlockStatsMXBean {
Collections.<LocatedBlock> emptyList(), null, false, feInfo, ecPolicy);
} else {
if (LOG.isDebugEnabled()) {
- LOG.debug("blocks = " + java.util.Arrays.asList(blocks));
+ LOG.debug("blocks = {}", java.util.Arrays.asList(blocks));
}
final AccessMode mode = needBlockToken? BlockTokenIdentifier.AccessMode.READ: null;
final List<LocatedBlock> locatedblocks = createLocatedBlockList(
@@ -1829,8 +1827,7 @@ public class BlockManager implements BlockStatsMXBean {
numReplicas);
if(srcNodes == null || srcNodes.length == 0) {
// block can not be reconstructed from any node
- LOG.debug("Block " + block + " cannot be reconstructed " +
- "from any node");
+ LOG.debug("Block {} cannot be reconstructed from any node", block);
return null;
}
@@ -2272,11 +2269,9 @@ public class BlockManager implements BlockStatsMXBean {
removeBlock(b);
}
if (trackBlockCounts) {
- if (LOG.isDebugEnabled()) {
- LOG.debug("Adjusting safe-mode totals for deletion."
- + "decreasing safeBlocks by " + numRemovedSafe
- + ", totalBlocks by " + numRemovedComplete);
- }
+ LOG.debug("Adjusting safe-mode totals for deletion."
+ + "decreasing safeBlocks by {}, totalBlocks by {}",
+ numRemovedSafe, numRemovedComplete);
bmSafeMode.adjustBlockTotals(-numRemovedSafe, -numRemovedComplete);
}
}
@@ -2440,18 +2435,14 @@ public class BlockManager implements BlockStatsMXBean {
BlockInfo bi = getStoredBlock(b);
if (bi == null) {
- if (LOG.isDebugEnabled()) {
- LOG.debug("BLOCK* rescanPostponedMisreplicatedBlocks: " +
- "Postponed mis-replicated block " + b + " no longer found " +
- "in block map.");
- }
+ LOG.debug("BLOCK* rescanPostponedMisreplicatedBlocks: " +
+ "Postponed mis-replicated block {} no longer found " +
+ "in block map.", b);
continue;
}
MisReplicationResult res = processMisReplicatedBlock(bi);
- if (LOG.isDebugEnabled()) {
- LOG.debug("BLOCK* rescanPostponedMisreplicatedBlocks: " +
- "Re-scanned block " + b + ", result is " + res);
- }
+ LOG.debug("BLOCK* rescanPostponedMisreplicatedBlocks: " +
+ "Re-scanned block {}, result is {}", b, res);
if (res == MisReplicationResult.POSTPONE) {
rescannedMisreplicatedBlocks.add(b);
}
@@ -2461,10 +2452,9 @@ public class BlockManager implements BlockStatsMXBean {
rescannedMisreplicatedBlocks.clear();
long endSize = postponedMisreplicatedBlocks.size();
namesystem.writeUnlock();
- LOG.info("Rescan of postponedMisreplicatedBlocks completed in " +
- (Time.monotonicNow() - startTime) + " msecs. " +
- endSize + " blocks are left. " +
- (startSize - endSize) + " blocks were removed.");
+ LOG.info("Rescan of postponedMisreplicatedBlocks completed in {}" +
+ " msecs. {} blocks are left. {} blocks were removed.",
+ (Time.monotonicNow() - startTime), endSize, (startSize - endSize));
}
}
@@ -2603,9 +2593,9 @@ public class BlockManager implements BlockStatsMXBean {
ReplicaState reportedState = iblk.getState();
if (LOG.isDebugEnabled()) {
- LOG.debug("Initial report of block " + iblk.getBlockName()
- + " on " + storageInfo.getDatanodeDescriptor() + " size " +
- iblk.getNumBytes() + " replicaState = " + reportedState);
+ LOG.debug("Initial report of block {} on {} size {} replicaState = {}",
+ iblk.getBlockName(), storageInfo.getDatanodeDescriptor(),
+ iblk.getNumBytes(), reportedState);
}
if (shouldPostponeBlocksFromFuture && isGenStampInFuture(iblk)) {
queueReportedBlock(storageInfo, iblk, reportedState,
@@ -2684,11 +2674,8 @@ public class BlockManager implements BlockStatsMXBean {
ReplicaState reportedState = replica.getState();
- if (LOG.isDebugEnabled()) {
- LOG.debug("Reported block " + replica
- + " on " + dn + " size " + replica.getNumBytes()
- + " replicaState = " + reportedState);
- }
+ LOG.debug("Reported block {} on {} size {} replicaState = {}",
+ replica, dn, replica.getNumBytes(), reportedState);
if (shouldPostponeBlocksFromFuture
&& isGenStampInFuture(replica)) {
@@ -2754,9 +2741,7 @@ public class BlockManager implements BlockStatsMXBean {
BlockUCState ucState = storedBlock.getBlockUCState();
// Block is on the NN
- if (LOG.isDebugEnabled()) {
- LOG.debug("In memory blockUCState = " + ucState);
- }
+ LOG.debug("In memory blockUCState = {}", ucState);
// Ignore replicas already scheduled to be removed from the DN
if (invalidateBlocks.contains(dn, replica)) {
@@ -2798,13 +2783,10 @@ public class BlockManager implements BlockStatsMXBean {
private void queueReportedBlock(DatanodeStorageInfo storageInfo, Block block,
ReplicaState reportedState, String reason) {
assert shouldPostponeBlocksFromFuture;
-
- if (LOG.isDebugEnabled()) {
- LOG.debug("Queueing reported block " + block +
- " in state " + reportedState +
- " from datanode " + storageInfo.getDatanodeDescriptor() +
- " for later processing because " + reason + ".");
- }
+
+ LOG.debug("Queueing reported block {} in state {}" +
+ " from datanode {} for later processing because {}.",
+ block, reportedState, storageInfo.getDatanodeDescriptor(), reason);
pendingDNMessages.enqueueReportedBlock(storageInfo, block, reportedState);
}
@@ -2825,9 +2807,7 @@ public class BlockManager implements BlockStatsMXBean {
private void processQueuedMessages(Iterable<ReportedBlockInfo> rbis)
throws IOException {
for (ReportedBlockInfo rbi : rbis) {
- if (LOG.isDebugEnabled()) {
- LOG.debug("Processing previouly queued message " + rbi);
- }
+ LOG.debug("Processing previouly queued message {}", rbi);
if (rbi.getReportedState() == null) {
// This is a DELETE_BLOCK request
DatanodeStorageInfo storageInfo = rbi.getStorageInfo();
@@ -2852,8 +2832,8 @@ public class BlockManager implements BlockStatsMXBean {
"block postponement.";
int count = pendingDNMessages.count();
if (count > 0) {
- LOG.info("Processing " + count + " messages from DataNodes " +
- "that were previously queued during standby state");
+ LOG.info("Processing {} messages from DataNodes " +
+ "that were previously queued during standby state", count);
}
processQueuedMessages(pendingDNMessages.takeAll());
assert pendingDNMessages.count() == 0;
@@ -2935,9 +2915,8 @@ public class BlockManager implements BlockStatsMXBean {
// the block report got a little bit delayed after the pipeline
// closed. So, ignore this report, assuming we will get a
// FINALIZED replica later. See HDFS-2791
- LOG.info("Received an RBW replica for " + storedBlock +
- " on " + dn + ": ignoring it, since it is " +
- "complete with the same genstamp");
+ LOG.info("Received an RBW replica for {} on {}: ignoring it, since "
+ + "it is complete with the same genstamp", storedBlock, dn);
return null;
} else {
return new BlockToMarkCorrupt(new Block(reported), storedBlock,
@@ -2952,7 +2931,7 @@ public class BlockManager implements BlockStatsMXBean {
+ " for block: " + storedBlock +
" on " + dn + " size " + storedBlock.getNumBytes();
// log here at WARN level since this is really a broken HDFS invariant
- LOG.warn(msg);
+ LOG.warn("{}", msg);
return new BlockToMarkCorrupt(new Block(reported), storedBlock, msg,
Reason.INVALID_STATE);
}
@@ -3131,9 +3110,9 @@ public class BlockManager implements BlockStatsMXBean {
int corruptReplicasCount = corruptReplicas.numCorruptReplicas(storedBlock);
int numCorruptNodes = num.corruptReplicas();
if (numCorruptNodes != corruptReplicasCount) {
- LOG.warn("Inconsistent number of corrupt replicas for " +
- storedBlock + ". blockMap has " + numCorruptNodes +
- " but corrupt replicas map has " + corruptReplicasCount);
+ LOG.warn("Inconsistent number of corrupt replicas for {}" +
+ ". blockMap has {} but corrupt replicas map has {}",
+ storedBlock, numCorruptNodes, corruptReplicasCount);
}
if ((corruptReplicasCount > 0) && (numLiveReplicas >= fileRedundancy)) {
invalidateCorruptReplicas(storedBlock, reportedBlock, num);
@@ -3298,12 +3277,13 @@ public class BlockManager implements BlockStatsMXBean {
/ totalBlocks, 1.0);
if (!blocksItr.hasNext()) {
- LOG.info("Total number of blocks = " + blocksMap.size());
- LOG.info("Number of invalid blocks = " + nrInvalid);
- LOG.info("Number of under-replicated blocks = " + nrUnderReplicated);
- LOG.info("Number of over-replicated blocks = " + nrOverReplicated
- + ((nrPostponed > 0) ? (" (" + nrPostponed + " postponed)") : ""));
- LOG.info("Number of blocks being written = " + nrUnderConstruction);
+ LOG.info("Total number of blocks = {}", blocksMap.size());
+ LOG.info("Number of invalid blocks = {}", nrInvalid);
+ LOG.info("Number of under-replicated blocks = {}", nrUnderReplicated);
+ LOG.info("Number of over-replicated blocks = {}{}", nrOverReplicated,
+ ((nrPostponed > 0) ? (" (" + nrPostponed + " postponed)") : ""));
+ LOG.info("Number of blocks being written = {}",
+ nrUnderConstruction);
NameNode.stateChangeLog
.info("STATE* Replication Queue initialization "
+ "scan for invalid, over- and under-replicated blocks "
@@ -3730,11 +3710,8 @@ public class BlockManager implements BlockStatsMXBean {
final DatanodeDescriptor node = storageInfo.getDatanodeDescriptor();
- if(LOG.isDebugEnabled()) {
- LOG.debug("Reported block " + block
- + " on " + node + " size " + block.getNumBytes()
- + " replicaState = " + reportedState);
- }
+ LOG.debug("Reported block {} on {} size {} replicaState = {}",
+ block, node, block.getNumBytes(), reportedState);
if (shouldPostponeBlocksFromFuture &&
isGenStampInFuture(block)) {
@@ -3756,9 +3733,7 @@ public class BlockManager implements BlockStatsMXBean {
BlockUCState ucState = storedBlock.getBlockUCState();
// Block is on the NN
- if(LOG.isDebugEnabled()) {
- LOG.debug("In memory blockUCState = " + ucState);
- }
+ LOG.debug("In memory blockUCState = {}", ucState);
// Ignore replicas already scheduled to be removed from the DN
if(invalidateBlocks.contains(node, block)) {
@@ -4013,8 +3988,8 @@ public class BlockManager implements BlockStatsMXBean {
numExtraRedundancy++;
}
}
- LOG.info("Invalidated " + numExtraRedundancy
- + " extra redundancy blocks on " + srcNode + " after it is in service");
+ LOG.info("Invalidated {} extra redundancy blocks on {} after "
+ + "it is in service", numExtraRedundancy, srcNode);
}
/**
@@ -4175,8 +4150,8 @@ public class BlockManager implements BlockStatsMXBean {
try {
DatanodeDescriptor dnDescriptor = datanodeManager.getDatanode(dn);
if (dnDescriptor == null) {
- LOG.warn("DataNode " + dn + " cannot be found with UUID " +
- dn.getDatanodeUuid() + ", removing block invalidation work.");
+ LOG.warn("DataNode {} cannot be found with UUID {}" +
+ ", removing block invalidation work.", dn, dn.getDatanodeUuid());
invalidateBlocks.remove(dn);
return 0;
}
http://git-wip-us.apache.org/repos/asf/hadoop/blob/a7f085d6/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/InvalidateBlocks.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/InvalidateBlocks.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/InvalidateBlocks.java
index 694cc95..4e8bb58 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/InvalidateBlocks.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/InvalidateBlocks.java
@@ -70,13 +70,14 @@ class InvalidateBlocks {
}
private void printBlockDeletionTime(final Logger log) {
- log.info(DFSConfigKeys.DFS_NAMENODE_STARTUP_DELAY_BLOCK_DELETION_SEC_KEY
- + " is set to " + DFSUtil.durationToString(pendingPeriodInMs));
+ log.info("{} is set to {}",
+ DFSConfigKeys.DFS_NAMENODE_STARTUP_DELAY_BLOCK_DELETION_SEC_KEY,
+ DFSUtil.durationToString(pendingPeriodInMs));
SimpleDateFormat sdf = new SimpleDateFormat("yyyy MMM dd HH:mm:ss");
Calendar calendar = new GregorianCalendar();
calendar.add(Calendar.SECOND, (int) (this.pendingPeriodInMs / 1000));
- log.info("The block deletion will start around "
- + sdf.format(calendar.getTime()));
+ log.info("The block deletion will start around {}",
+ sdf.format(calendar.getTime()));
}
/** @return the number of blocks to be invalidated . */
@@ -173,11 +174,9 @@ class InvalidateBlocks {
synchronized List<Block> invalidateWork(final DatanodeDescriptor dn) {
final long delay = getInvalidationDelay();
if (delay > 0) {
- if (BlockManager.LOG.isDebugEnabled()) {
- BlockManager.LOG
- .debug("Block deletion is delayed during NameNode startup. "
- + "The deletion will start after " + delay + " ms.");
- }
+ BlockManager.LOG
+ .debug("Block deletion is delayed during NameNode startup. "
+ + "The deletion will start after {} ms.", delay);
return null;
}
final LightWeightHashSet<Block> set = node2blocks.get(dn);
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org