You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2011/10/29 01:20:14 UTC
svn commit: r1190669 - in /hbase/trunk: ./
src/main/java/org/apache/hadoop/hbase/io/hfile/
src/main/java/org/apache/hadoop/hbase/master/
src/main/java/org/apache/hadoop/hbase/regionserver/
src/main/java/org/apache/hadoop/hbase/regionserver/compactions/...
Author: stack
Date: Fri Oct 28 23:20:14 2011
New Revision: 1190669
URL: http://svn.apache.org/viewvc?rev=1190669&view=rev
Log:
HBASE-4694 Some cleanup of log messages in RS and M
Modified:
hbase/trunk/CHANGES.txt
hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java
hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java
Modified: hbase/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hbase/trunk/CHANGES.txt?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/CHANGES.txt (original)
+++ hbase/trunk/CHANGES.txt Fri Oct 28 23:20:14 2011
@@ -691,6 +691,7 @@ Release 0.92.0 - Unreleased
cluster instead of one per method (nkeywal)
HBASE-3929 Add option to HFile tool to produce basic stats (Matteo
Bertozzi and todd via todd)
+ HBASE-4694 Some cleanup of log messages in RS and M
TASKS
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java Fri Oct 28 23:20:14 2011
@@ -258,7 +258,6 @@ public class HFile {
public static final WriterFactory getWriterFactory(Configuration conf,
CacheConfig cacheConf) {
int version = getFormatVersion(conf);
- LOG.debug("Using HFile format version " + version);
switch (version) {
case 1:
return new HFileWriterV1.WriterFactoryV1(conf, cacheConf);
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java Fri Oct 28 23:20:14 2011
@@ -44,6 +44,7 @@ import org.apache.hadoop.hbase.util.Clas
import org.apache.hadoop.hbase.util.CompoundBloomFilterWriter;
import org.apache.hadoop.io.RawComparator;
import org.apache.hadoop.io.WritableUtils;
+import org.apache.hadoop.util.StringUtils;
/**
* Provides functionality to write ({@link BlockIndexWriter}) and read
@@ -750,12 +751,15 @@ public class HFileBlockIndex {
totalBlockUncompressedSize +=
blockWriter.getUncompressedSizeWithoutHeader();
- LOG.info("Wrote a " + numLevels + "-level index with root level at pos "
+ if (LOG.isDebugEnabled()) {
+ LOG.debug("Wrote a " + numLevels + "-level index with root level at pos "
+ out.getPos() + ", " + rootChunk.getNumEntries()
+ " root-level entries, " + totalNumEntries + " total entries, "
- + totalBlockOnDiskSize + " bytes total on-disk size, "
- + totalBlockUncompressedSize + " bytes total uncompressed size.");
-
+ + StringUtils.humanReadableInt(this.totalBlockOnDiskSize) +
+ " on-disk size, "
+ + StringUtils.humanReadableInt(totalBlockUncompressedSize) +
+ " total uncompressed size.");
+ }
return rootLevelIndexPos;
}
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java Fri Oct 28 23:20:14 2011
@@ -326,7 +326,7 @@ public class HFileReaderV2 extends Abstr
if (evictOnClose && cacheConf.isBlockCacheEnabled()) {
int numEvicted = cacheConf.getBlockCache().evictBlocksByPrefix(name
+ HFile.CACHE_KEY_SEPARATOR);
- LOG.debug("On close of file " + name + " evicted " + numEvicted
+ LOG.debug("On close, file=" + name + " evicted=" + numEvicted
+ " block(s)");
}
if (closeIStream && istream != null) {
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java Fri Oct 28 23:20:14 2011
@@ -178,7 +178,7 @@ public class HFileWriterV2 extends Abstr
// Meta data block index writer
metaBlockIndexWriter = new HFileBlockIndex.BlockIndexWriter();
- LOG.debug("HFileWriter initialized with " + cacheConf);
+ LOG.debug("Initialized with " + cacheConf);
}
/**
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java Fri Oct 28 23:20:14 2011
@@ -229,20 +229,6 @@ public class DefaultLoadBalancer impleme
NavigableMap<ServerAndLoad, List<HRegionInfo>> serversByLoad =
new TreeMap<ServerAndLoad, List<HRegionInfo>>();
int numRegions = 0;
- StringBuilder strBalanceParam = new StringBuilder("Server information: ");
- // Iterate so we can count regions as we build the map
- for (Map.Entry<ServerName, List<HRegionInfo>> server: clusterState.entrySet()) {
- List<HRegionInfo> regions = server.getValue();
- int sz = regions.size();
- if (sz == 0) emptyRegionServerPresent = true;
- numRegions += sz;
- serversByLoad.put(new ServerAndLoad(server.getKey(), sz), regions);
- strBalanceParam.append(server.getKey().getServerName()).append("=").
- append(server.getValue().size()).append(", ");
- }
- strBalanceParam.delete(strBalanceParam.length() - 2,
- strBalanceParam.length());
- LOG.debug(strBalanceParam.toString());
// Check if we even need to do any load balancing
float average = (float)numRegions / numServers; // for logging
@@ -262,13 +248,13 @@ public class DefaultLoadBalancer impleme
int min = numRegions / numServers;
int max = numRegions % numServers == 0 ? min : min + 1;
- // Using to check banance result.
- strBalanceParam.delete(0, strBalanceParam.length());
+ // Using to check balance result.
+ StringBuilder strBalanceParam = new StringBuilder();
strBalanceParam.append("Balance parameter: numRegions=").append(numRegions)
.append(", numServers=").append(numServers).append(", max=").append(max)
.append(", min=").append(min);
LOG.debug(strBalanceParam.toString());
-
+
// Balance the cluster
// TODO: Look at data block locality or a more complex load to do this
MinMaxPriorityQueue<RegionPlan> regionsToMove =
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java Fri Oct 28 23:20:14 2011
@@ -85,7 +85,7 @@ class LogRoller extends HasThread implem
LOG.debug("Hlog roll period " + this.rollperiod + "ms elapsed");
}
} else if (LOG.isDebugEnabled()) {
- LOG.debug("HLog roll manually triggered");
+ LOG.debug("HLog roll requested");
}
rollLock.lock(); // FindBugs UL_UNRELEASED_LOCK_EXCEPTION_PATH
try {
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java Fri Oct 28 23:20:14 2011
@@ -173,25 +173,27 @@ class MemStoreFlusher extends HasThread
HRegion regionToFlush;
if (bestFlushableRegion != null &&
- bestAnyRegion.memstoreSize.get() > 2 * bestFlushableRegion.memstoreSize.get()) {
+ bestAnyRegion.memstoreSize.get() > 2 * bestFlushableRegion.memstoreSize.get()) {
// Even if it's not supposed to be flushed, pick a region if it's more than twice
// as big as the best flushable one - otherwise when we're under pressure we make
// lots of little flushes and cause lots of compactions, etc, which just makes
// life worse!
- LOG.info("Under global heap pressure: " +
+ if (LOG.isDebugEnabled()) {
+ LOG.debug("Under global heap pressure: " +
"Region " + bestAnyRegion.getRegionNameAsString() + " has too many " +
"store files, but is " +
StringUtils.humanReadableInt(bestAnyRegion.memstoreSize.get()) +
" vs best flushable region's " +
StringUtils.humanReadableInt(bestFlushableRegion.memstoreSize.get()) +
". Choosing the bigger.");
- regionToFlush = bestAnyRegion;
+ }
+ regionToFlush = bestAnyRegion;
} else {
- if (bestFlushableRegion == null) {
- regionToFlush = bestAnyRegion;
- } else {
- regionToFlush = bestFlushableRegion;
- }
+ if (bestFlushableRegion == null) {
+ regionToFlush = bestAnyRegion;
+ } else {
+ regionToFlush = bestFlushableRegion;
+ }
}
Preconditions.checkState(regionToFlush.memstoreSize.get() > 0);
@@ -216,7 +218,8 @@ class MemStoreFlusher extends HasThread
fqe = flushQueue.poll(threadWakeFrequency, TimeUnit.MILLISECONDS);
if (fqe == null || fqe instanceof WakeupFlushThread) {
if (isAboveLowWaterMark()) {
- LOG.info("Flush thread woke up with memory above low water.");
+ LOG.debug("Flush thread woke up because memory above low water=" +
+ StringUtils.humanReadableInt(this.globalMemStoreLimitLowMark));
if (!flushOneForGlobalPressure()) {
// Wasn't able to flush any region, but we're above low water mark
// This is unlikely to happen, but might happen when closing the
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java Fri Oct 28 23:20:14 2011
@@ -390,7 +390,7 @@ public class Store implements HeapSize {
}
Path dstPath = StoreFile.getRandomFilename(fs, homedir);
- LOG.info("Renaming bulk load file " + srcPath + " to " + dstPath);
+ LOG.debug("Renaming bulk load file " + srcPath + " to " + dstPath);
StoreFile.rename(fs, srcPath, dstPath);
StoreFile sf = new StoreFile(fs, dstPath, this.conf, this.cacheConf,
@@ -573,7 +573,7 @@ public class Store implements HeapSize {
Path dstPath = new Path(homedir, fileName);
validateStoreFile(path);
String msg = "Renaming flushed file at " + path + " to " + dstPath;
- LOG.info(msg);
+ LOG.debug(msg);
status.setStatus("Flushing " + this + ": " + msg);
if (!fs.rename(path, dstPath)) {
LOG.warn("Unable to rename " + path + " to " + dstPath);
@@ -593,7 +593,7 @@ public class Store implements HeapSize {
// the flushing through the StoreFlusherImpl class
HRegion.incrNumericPersistentMetric("cf." + this.toString() + ".flushSize",
flushedSize.longValue());
- if(LOG.isInfoEnabled()) {
+ if (LOG.isInfoEnabled()) {
LOG.info("Added " + sf + ", entries=" + r.getEntries() +
", sequenceid=" + logCacheFlushId +
", filesize=" + StringUtils.humanReadableInt(r.length()));
@@ -720,7 +720,7 @@ public class Store implements HeapSize {
LOG.info("Starting compaction of " + filesToCompact.size() + " file(s) in "
+ this.storeNameStr + " of "
+ this.region.getRegionInfo().getRegionNameAsString()
- + " into " + region.getTmpDir() + ", seqid=" + maxId + ", totalSize="
+ + " into tmpdir=" + region.getTmpDir() + ", seqid=" + maxId + ", totalSize="
+ StringUtils.humanReadableInt(cr.getSize()));
StoreFile sf = null;
@@ -741,8 +741,9 @@ public class Store implements HeapSize {
LOG.info("Completed" + (cr.isMajor() ? " major " : " ") + "compaction of "
+ filesToCompact.size() + " file(s) in " + this.storeNameStr + " of "
+ this.region.getRegionInfo().getRegionNameAsString()
- + "; new storefile name=" + (sf == null ? "none" : sf.toString())
- + ", size=" + (sf == null ? "none" :
+ + " into " +
+ (sf == null ? "none" : sf.getPath().getName()) +
+ ", size=" + (sf == null ? "none" :
StringUtils.humanReadableInt(sf.getReader().length()))
+ "; total size for store is "
+ StringUtils.humanReadableInt(storeSize));
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java Fri Oct 28 23:20:14 2011
@@ -20,7 +20,6 @@
package org.apache.hadoop.hbase.regionserver.compactions;
import java.io.IOException;
-import java.util.Date;
import java.util.List;
import java.util.concurrent.RejectedExecutionHandler;
import java.util.concurrent.ThreadPoolExecutor;
@@ -199,7 +198,7 @@ public class CompactionRequest implement
server.checkFileSystem();
} finally {
s.finishRequest(this);
- LOG.debug("CompactSplitThread Status: " + server.compactSplitThread);
+ LOG.debug("CompactSplitThread status: " + server.compactSplitThread);
}
}
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java Fri Oct 28 23:20:14 2011
@@ -561,9 +561,6 @@ public class HLog implements Syncable {
}
this.filenum = System.currentTimeMillis();
Path newPath = computeFilename();
- if (LOG.isDebugEnabled()) {
- LOG.debug("Enabling new writer for "+FSUtils.getPath(newPath));
- }
// Tell our listeners that a new log is about to be created
if (!this.listeners.isEmpty()) {
@@ -597,7 +594,7 @@ public class HLog implements Syncable {
this.numEntries.get() +
", filesize=" +
this.fs.getFileStatus(oldFile).getLen() + ". ": "") +
- "New hlog " + FSUtils.getPath(newPath));
+ " for " + FSUtils.getPath(newPath));
this.numEntries.set(0);
}
// Can we delete any of the old log files?
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java Fri Oct 28 23:20:14 2011
@@ -93,8 +93,9 @@ public class SequenceFileLogWriter imple
this.writer_out = getSequenceFilePrivateFSDataOutputStreamAccessible();
this.syncFs = getSyncFs();
this.hflush = getHFlush();
- String msg =
- "syncFs=" + (this.syncFs != null) + ", hflush=" + (this.hflush != null);
+ String msg = "Path=" + path +
+ ", syncFs=" + (this.syncFs != null) +
+ ", hflush=" + (this.hflush != null);
if (this.syncFs != null || this.hflush != null) {
LOG.debug(msg);
} else {
Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java?rev=1190669&r1=1190668&r2=1190669&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java Fri Oct 28 23:20:14 2011
@@ -173,12 +173,12 @@ public final class BloomFilterFactory {
CacheConfig cacheConf, BloomType bloomType, int maxKeys,
HFile.Writer writer) {
if (!isGeneralBloomEnabled(conf)) {
- LOG.info("Bloom filters are disabled by configuration for "
+ LOG.debug("Bloom filters are disabled by configuration for "
+ writer.getPath()
+ (conf == null ? " (configuration is null)" : ""));
return null;
} else if (bloomType == BloomType.NONE) {
- LOG.info("Bloom filter is turned off for the column family");
+ LOG.debug("Bloom filter is turned off for the column family");
return null;
}