You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2011/10/25 08:57:12 UTC

svn commit: r1188539 [1/2] - in /hbase/branches/0.92: ./ src/main/java/org/apache/hadoop/hbase/ src/main/java/org/apache/hadoop/hbase/catalog/ src/main/java/org/apache/hadoop/hbase/client/ src/main/java/org/apache/hadoop/hbase/client/coprocessor/ src/m...

Author: stack
Date: Tue Oct 25 06:57:10 2011
New Revision: 1188539

URL: http://svn.apache.org/viewvc?rev=1188539&view=rev
Log:
HBASE-4670 Fix javadoc warnings

Modified:
    hbase/branches/0.92/CHANGES.txt
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterBase.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/Objects.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionServerTracker.java
    hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java

Modified: hbase/branches/0.92/CHANGES.txt
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/CHANGES.txt?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/CHANGES.txt (original)
+++ hbase/branches/0.92/CHANGES.txt Tue Oct 25 06:57:10 2011
@@ -369,6 +369,7 @@ Release 0.92.0 - Unreleased
                master (Eugene Koontz via apurtell)
    HBASE-3512  Shell support for listing currently loaded coprocessors (Eugene
                Koontz via apurtell)
+   HBASE-4670  Fix javadoc warnings
 
   TESTS
    HBASE-4492  TestRollingRestart fails intermittently

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java Tue Oct 25 06:57:10 2011
@@ -437,13 +437,13 @@ public final class HConstants {
   public static int DEFAULT_HBASE_META_SCANNER_CACHING = 100;
 
   /**
-   * Parameter name for unique identifier for this {@link Configuration}
-   * instance. If there are two or more {@link Configuration} instances that,
+   * Parameter name for unique identifier for this {@link org.apache.hadoop.conf.Configuration}
+   * instance. If there are two or more {@link org.apache.hadoop.conf.Configuration} instances that,
    * for all intents and purposes, are the same except for their instance ids,
-   * then they will not be able to share the same {@link Connection} instance.
+   * then they will not be able to share the same {@link org.apache.hadoop.hbase.client.HConnection} instance.
    * On the other hand, even if the instance ids are the same, it could result
-   * in non-shared {@link Connection} instances if some of the other connection
-   * parameters differ.
+   * in non-shared {@link org.apache.hadoop.hbase.client.HConnection}
+   * instances if some of the other connection parameters differ.
    */
   public static String HBASE_CLIENT_INSTANCE_ID = "hbase.client.instance.id";
 

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java Tue Oct 25 06:57:10 2011
@@ -90,7 +90,7 @@ public class HServerAddress implements W
    }
   
   /** @return Bind address -- the raw IP, the result of a call to
-   * {@link InetSocketAddress#getAddress()#getHostAddress()} --
+   * InetSocketAddress#getAddress()#getHostAddress() --
    * or null if cannot resolve */
   public String getBindAddress() {
     return getBindAddressInternal(address);

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java Tue Oct 25 06:57:10 2011
@@ -104,8 +104,6 @@ implements WritableComparable<HServerInf
 
   /**
    * @return ServerName and load concatenated.
-   * @see #getServerName()
-   * @see #getLoad()
    */
   @Override
   public synchronized String toString() {
@@ -150,4 +148,4 @@ implements WritableComparable<HServerInf
     if (this.startCode != o.getStartCode()) return (int)(this.startCode - o.getStartCode());
     return 0;
   }
-}
\ No newline at end of file
+}

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java Tue Oct 25 06:57:10 2011
@@ -126,8 +126,6 @@ public class HTableDescriptor implements
   /**
    * <em>INTERNAL</em> Used by HBase Shell interface to access this metadata 
    * attribute which denotes if the deferred log flush option is enabled
-   * 
-   * @see #getDeferredLogFlush()
    */
   public static final String DEFERRED_LOG_FLUSH = "DEFERRED_LOG_FLUSH";
   private static final ImmutableBytesWritable DEFERRED_LOG_FLUSH_KEY =
@@ -524,7 +522,7 @@ public class HTableDescriptor implements
    * before these deferred edits in memory are flushed onto the filesystem. 
    * </p>
    * 
-   * @param true if that deferred log flush is enabled on the table.
+   * @param isDeferredLogFlush
    */
   public void setDeferredLogFlush(final boolean isDeferredLogFlush) {
     setValue(DEFERRED_LOG_FLUSH_KEY, isDeferredLogFlush? TRUE: FALSE);
@@ -552,8 +550,8 @@ public class HTableDescriptor implements
   /**
    * This get the class associated with the region split policy which 
    * determines when a region split should occur.  The class used by
-   * default is {@link ConstantSizeRegionSplitPolicy} which split the
-   * region base on a constant {@link #getMaxFileSize()}
+   * default is {@link org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy}
+   * which split the region base on a constant {@link #getMaxFileSize()}
    * 
    * @return the class name of the region split policy for this table.
    * If this returns null, the default constant size based split policy
@@ -877,7 +875,7 @@ public class HTableDescriptor implements
   }
 
   /**
-   * Returns an unmodifiable collection of all the {@link HColumnDescriptors} 
+   * Returns an unmodifiable collection of all the {@link HColumnDescriptor} 
    * of all the column families of the table.
    *  
    * @return Immutable collection of {@link HColumnDescriptor} of all the

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java Tue Oct 25 06:57:10 2011
@@ -1749,7 +1749,6 @@ public class KeyValue implements Writabl
    * (the value part of the returned KV is always empty). Used in creating
    * "fake keys" for the multi-column Bloom filter optimization to skip the
    * row/column we already know is not in the file.
-   * @param kv the key-value pair to take row and column from
    * @return the last key on the row/column of the given key-value pair
    */
   public KeyValue createLastOnRowCol() {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java Tue Oct 25 06:57:10 2011
@@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.util.Byte
  * usually timestamp of server startup). The {@link #toString()} format of
  * ServerName is safe to use in the  filesystem and as znode name up in
  * ZooKeeper.  Its format is:
- * <code>&lt;hostname> '{@link #SERVERNAME_SEPARATOR"}' &lt;port> '{@ink #SERVERNAME_SEPARATOR"}' &lt;startcode></code>.
+ * <code>&lt;hostname> '{@link #SERVERNAME_SEPARATOR}' &lt;port> '{@link #SERVERNAME_SEPARATOR}' &lt;startcode></code>.
  * For example, if hostname is <code>example.org</code>, port is <code>1234</code>,
  * and the startcode for the regionserver is <code>1212121212</code>, then
  * the {@link #toString()} would be <code>example.org,1234,1212121212</code>.
@@ -220,7 +220,7 @@ public class ServerName implements Compa
 
   /**
    * @param left
-   * @param rigth
+   * @param right
    * @return True if <code>other</code> has same hostname and port.
    */
   public static boolean isSameHostnameAndPort(final ServerName left,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java Tue Oct 25 06:57:10 2011
@@ -49,8 +49,6 @@ public interface TableDescriptors {
   /**
    * Get Map of all HTableDescriptors. Populates the descriptor cache as a
    * side effect.
-   * @param fs
-   * @param rootdir
    * @return Map of all descriptors.
    * @throws IOException
    */
@@ -72,4 +70,4 @@ public interface TableDescriptors {
    */
   public HTableDescriptor remove(final String tablename)
   throws IOException;
-}
\ No newline at end of file
+}

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java Tue Oct 25 06:57:10 2011
@@ -145,7 +145,7 @@ public class CatalogTracker {
    * Does not timeout.
    * @param zk If zk is null, we'll create an instance (and shut it down
    * when {@link #stop()} is called) else we'll use what is passed.
-   * @param connection server connection
+   * @param conf
    * @param abortable If fatal exception we'll call abort on this.  May be null.
    * If it is we'll use the Connection associated with the passed
    * {@link Configuration} as our Abortable.

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java Tue Oct 25 06:57:10 2011
@@ -50,7 +50,7 @@ public class MetaMigrationRemovingHTD {
   /**
    * Update legacy META rows, removing HTD from HRI.
    * @param masterServices
-   * @return
+   * @return List of table descriptors.
    * @throws IOException
    */
   public static List<HTableDescriptor> updateMetaWithNewRegionInfo(
@@ -78,7 +78,7 @@ public class MetaMigrationRemovingHTD {
   /**
    * Update the ROOT with new HRI. (HRI with no HTD)
    * @param masterServices
-   * @return
+   * @return List of table descriptors
    * @throws IOException
    */
   public static List<HTableDescriptor> updateRootWithNewRegionInfo(
@@ -239,4 +239,4 @@ public class MetaMigrationRemovingHTD {
         "Master startup aborted.");
     }
   }
-}
\ No newline at end of file
+}

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java Tue Oct 25 06:57:10 2011
@@ -251,7 +251,7 @@ public class MetaReader {
    * @param metaServer connection to server hosting ROOT
    * @return location of META in ROOT where location, or null if not available
    * @throws IOException
-   * @deprecated Does not retry; use {@link #readRegionLocation(CatalogTracker, byte[])
+   * @deprecated Does not retry; use {@link #readRegionLocation(CatalogTracker, byte[])}
    */
   public static ServerName readMetaLocation(HRegionInterface metaServer)
   throws IOException {
@@ -589,7 +589,7 @@ public class MetaReader {
 
   /**
    * @param catalogTracker
-   * @param hsi Server specification
+   * @param serverName
    * @return List of user regions installed on this server (does not include
    * catalog regions).
    * @throws IOException
@@ -742,4 +742,4 @@ public class MetaReader {
       this.results.add(r);
     }
   }
-}
\ No newline at end of file
+}

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java Tue Oct 25 06:57:10 2011
@@ -21,7 +21,6 @@ package org.apache.hadoop.hbase.client;
 
 import java.io.Closeable;
 import java.io.IOException;
-import java.net.InetSocketAddress;
 import java.util.List;
 import java.util.Map;
 import java.util.concurrent.ExecutorService;
@@ -192,7 +191,7 @@ public interface HConnection extends Abo
    * @param regionServer - the server to connect to
    * @return proxy for HRegionServer
    * @throws IOException if a remote or network exception occurs
-   * @deprecated Use {@link #getHRegionConnection(InetSocketAddress)}
+   * @deprecated Use {@link #getHRegionConnection(String, int)}
    */
   public HRegionInterface getHRegionConnection(HServerAddress regionServer)
   throws IOException;
@@ -384,4 +383,4 @@ public interface HConnection extends Abo
   public HTableDescriptor[] getHTableDescriptors(List<String> tableNames)
   throws IOException;
 
-}
\ No newline at end of file
+}

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java Tue Oct 25 06:57:10 2011
@@ -208,10 +208,7 @@ public class HConnectionManager {
    * This will then close connection to
    * the zookeeper ensemble and let go of all resources.
    *
-   * @param conf
-   *          configuration whose identity is used to find {@link HConnection}
-   *          instance.
-   *          .
+   * @param connection
    */
   public static void deleteStaleConnection(HConnection connection) {
     deleteConnection(connection, true, true);
@@ -302,7 +299,7 @@ public class HConnectionManager {
    * of a {@link HConnection} instance based on the given {@link Configuration}.
    *
    * <p>
-   * If you find yourself wanting to use a {@link Connection} for a relatively
+   * If you find yourself wanting to use a {@link HConnection} for a relatively
    * short duration of time, and do not want to deal with the hassle of creating
    * and cleaning up that resource, then you should consider using this
    * convenience class.

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java Tue Oct 25 06:57:10 2011
@@ -279,7 +279,7 @@ public interface HTableInterface {
   /**
    * Atomically increments a column value.
    * <p>
-   * Equivalent to {@code {@link #incrementColumnValue(byte[], byte[], byte[],
+   * Equivalent to {@link #incrementColumnValue(byte[], byte[], byte[],
    * long, boolean) incrementColumnValue}(row, family, qualifier, amount,
    * <b>true</b>)}
    * @param row The row that contains the cell to increment.
@@ -377,7 +377,7 @@ public interface HTableInterface {
    *
    * @param protocol The class or interface defining the remote protocol
    * @param row The row key used to identify the remote region location
-   * @return
+   * @return A CoprocessorProtocol instance
    */
   <T extends CoprocessorProtocol> T coprocessorProxy(Class<T> protocol, byte[] row);
 
@@ -402,7 +402,7 @@ public interface HTableInterface {
    * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Call#call(Object)}
    * method
    * @return a <code>Map</code> of region names to
-   * {@link Batch.Call#call(Object)} return values
+   * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Call#call(Object)} return values
    */
   <T extends CoprocessorProtocol, R> Map<byte[],R> coprocessorExec(
       Class<T> protocol, byte[] startKey, byte[] endKey, Batch.Call<T,R> callable)
@@ -421,7 +421,7 @@ public interface HTableInterface {
    *
    * <p>
    * For each result, the given
-   * {@link Batch.Callback#update(byte[], byte[], Object)}
+   * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Callback#update(byte[], byte[], Object)}
    * method will be called.
    *</p>
    *
@@ -432,7 +432,7 @@ public interface HTableInterface {
    * @param callable wraps the CoprocessorProtocol implementation method calls
    * made per-region
    * @param callback an instance upon which
-   * {@link Batch.Callback#update(byte[], byte[], Object)} with the
+   * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Callback#update(byte[], byte[], Object)} with the
    * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Call#call(Object)}
    * return value for each region
    * @param <T> CoprocessorProtocol subclass for the remote invocation

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java Tue Oct 25 06:57:10 2011
@@ -106,8 +106,6 @@ public class HTablePool implements Close
    *          configuration
    * @param maxSize
    *          maximum number of references to keep for each table
-   * @param tableFactory
-   *          table factory
    * @param poolType
    *          pool type which is one of {@link PoolType#Reusable} or
    *          {@link PoolType#ThreadLocal}

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java Tue Oct 25 06:57:10 2011
@@ -131,7 +131,6 @@ public abstract class Mutation extends O
 
   /**
    * Method for setting the put's familyMap
-   * @return familyMap
    */
   public void setFamilyMap(Map<byte [], List<KeyValue>> map) {
     this.familyMap = map;

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java Tue Oct 25 06:57:10 2011
@@ -76,11 +76,6 @@ extends RetriesExhaustedException {
     return actions.get(i);
   }
 
-  /**
-   * @param i
-   * @return
-   * @deprecated
-   */
   public HServerAddress getAddress(int i) {
     return new HServerAddress(Addressing.createInetSocketAddressFromHostAndPortStr(getHostnamePort(i)));
   }

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java Tue Oct 25 06:57:10 2011
@@ -170,7 +170,7 @@ public class AggregationClient {
    * @param tableName
    * @param ci
    * @param scan
-   * @return
+   * @return <R, S>
    * @throws Throwable
    */
   public <R, S> long rowCount(final byte[] tableName,
@@ -282,7 +282,7 @@ public class AggregationClient {
    * @param tableName
    * @param ci
    * @param scan
-   * @return
+   * @return <R, S>
    * @throws Throwable
    */
   public <R, S> double avg(final byte[] tableName,
@@ -348,7 +348,7 @@ public class AggregationClient {
    * @param tableName
    * @param ci
    * @param scan
-   * @return
+   * @return <R, S>
    * @throws Throwable
    */
   public <R, S> double std(final byte[] tableName, ColumnInterpreter<R, S> ci,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java Tue Oct 25 06:57:10 2011
@@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.util.Byte
  * and its promoted data type is also a Long value. For computing aggregation
  * function, this class is used to find the datatype of the cell value. Client
  * is supposed to instantiate it and passed along as a parameter. See
- * {@link TestAggregateProtocol} methods for its sample usage.
+ * TestAggregateProtocol methods for its sample usage.
  * Its methods handle null arguments gracefully. 
  */
 public class LongColumnInterpreter implements ColumnInterpreter<Long, Long> {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java Tue Oct 25 06:57:10 2011
@@ -80,11 +80,10 @@ public interface AggregateProtocol exten
   <T, S> S getSum(ColumnInterpreter<T, S> ci, Scan scan) throws IOException;
 
   /**
-   * Gives the row count for the given column family and column qualifier, in
-   * the given row range as defined in the Scan object.
    * @param ci
    * @param scan
-   * @return
+   * @return Row count for the given column family and column qualifier, in
+   * the given row range as defined in the Scan object.
    * @throws IOException
    */
   <T, S> long getRowNum(ColumnInterpreter<T, S> ci, Scan scan)
@@ -104,7 +103,7 @@ public interface AggregateProtocol exten
    * type.
    * @param ci
    * @param scan
-   * @return
+   * @return Average
    * @throws IOException
    */
   <T, S> Pair<S, Long> getAvg(ColumnInterpreter<T, S> ci, Scan scan)
@@ -120,7 +119,7 @@ public interface AggregateProtocol exten
    * deviation is square root of variance.
    * @param ci
    * @param scan
-   * @return
+   * @return STD
    * @throws IOException
    */
   <T, S> Pair<List<S>, Long> getStd(ColumnInterpreter<T, S> ci, Scan scan)

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java Tue Oct 25 06:57:10 2011
@@ -42,14 +42,15 @@ import org.apache.hadoop.io.Writable;
  * <S>. There is a conversion method
  * {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and
  * returns a <S> type.
- * @param <T, S>: T - cell value data type, S - promoted data type
+ * @param <T> Cell value data type
+ * @param <S> Promoted data type
  */
 public interface ColumnInterpreter<T, S> extends Writable {
 
   /**
    * @param colFamily
    * @param colQualifier
-   * @param value
+   * @param kv
    * @return value of type T
    * @throws IOException
    */
@@ -57,44 +58,39 @@ public interface ColumnInterpreter<T, S>
       throws IOException;
 
   /**
-   * returns sum or non null value among (if either of them is null); otherwise
-   * returns a null.
    * @param l1
    * @param l2
-   * @return
+   * @return sum or non null value among (if either of them is null); otherwise
+   * returns a null.
    */
   public S add(S l1, S l2);
 
   /**
    * returns the maximum value for this type T
-   * @return
+   * @return max
    */
 
   T getMaxValue();
 
-  /**
-   * @return
-   */
-
   T getMinValue();
 
   /**
    * @param o1
    * @param o2
-   * @return
+   * @return multiplication
    */
   S multiply(S o1, S o2);
 
   /**
    * @param o
-   * @return
+   * @return increment
    */
   S increment(S o);
 
   /**
    * provides casting opportunity between the data types.
    * @param o
-   * @return
+   * @return cast
    */
   S castToReturnType(T o);
 
@@ -109,10 +105,10 @@ public interface ColumnInterpreter<T, S>
 
   /**
    * used for computing average of <S> data values. Not providing the divide
-   * method that takes two <S> values as it si not needed as of now.
+   * method that takes two <S> values as it is not needed as of now.
    * @param o
    * @param l
-   * @return
+   * @return Average
    */
   double divideForAvg(S o, Long l);
 }
\ No newline at end of file

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java Tue Oct 25 06:57:10 2011
@@ -220,7 +220,7 @@ public interface MasterObserver extends 
   /**
    * Called prior to unassigning a given region.
    * @param ctx the environment to interact with the framework and master
-   * @param regionName the name of the region
+   * @param regionInfo
    * @param force whether to force unassignment or not
    */
   void preUnassign(final ObserverContext<MasterCoprocessorEnvironment> ctx,
@@ -229,7 +229,7 @@ public interface MasterObserver extends 
   /**
    * Called after the region unassignment has been requested.
    * @param ctx the environment to interact with the framework and master
-   * @param regionName the name of the region
+   * @param regionInfo
    * @param force whether to force unassignment or not
    */
   void postUnassign(final ObserverContext<MasterCoprocessorEnvironment> ctx,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java Tue Oct 25 06:57:10 2011
@@ -154,7 +154,7 @@ import org.apache.hadoop.hbase.client.Ge
 // and intercept preXXX() method to check user privilege for the given table
 // and column family.
 public class AccessControlCoprocessor extends BaseRegionObserverCoprocessor {
-  @Override
+  // @Override
   public Get preGet(CoprocessorEnvironment e, Get get)
       throws CoprocessorException {
 
@@ -208,9 +208,9 @@ extends CoprocessorProtocol {
 // Aggregation implementation at a region.
 public static class ColumnAggregationEndpoint extends BaseEndpointCoprocessor
 implements ColumnAggregationProtocol {
-  @Override
-   // Scan the region by the given family and qualifier. Return the aggregation
-   // result.
+  // @Override
+  // Scan the region by the given family and qualifier. Return the aggregation
+  // result.
   public int sum(byte[] family, byte[] qualifier)
   throws IOException {
     // aggregate at each region

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java Tue Oct 25 06:57:10 2011
@@ -108,7 +108,7 @@ public class RegionTransitionData implem
    *
    * @param eventType type of event
    * @param regionName name of region as per <code>HRegionInfo#getRegionName()</code>
-   * @param origin Originating {@link ServerName}
+   * @param serverName Originating {@link ServerName}
    * @param payload Payload examples include the daughters involved in a
    * {@link EventType#RS_ZK_REGION_SPLIT}. Can be null
    */

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java Tue Oct 25 06:57:10 2011
@@ -47,7 +47,7 @@ public class BitComparator extends Writa
   /**
    * Constructor
    * @param value value
-   * @param BitwiseOp bitOperator - the operator to use on the bit comparison
+   * @param bitOperator operator to use on the bit comparison
    */
   public BitComparator(byte[] value, BitwiseOp bitOperator) {
     super(value);

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java Tue Oct 25 06:57:10 2011
@@ -19,26 +19,19 @@
  */
 package org.apache.hadoop.hbase.filter;
 
-import java.io.IOException;
+import java.lang.reflect.InvocationTargetException;
+import java.lang.reflect.Method;
 import java.nio.ByteBuffer;
 import java.nio.charset.CharacterCodingException;
-import java.util.TreeSet;
 import java.util.ArrayList;
-import java.util.Stack;
+import java.util.EmptyStackException;
 import java.util.HashMap;
 import java.util.Set;
+import java.util.Stack;
 
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.Writables;
 import org.apache.hadoop.hbase.KeyValue;
-import org.apache.hadoop.hbase.filter.ParseConstants;
-
-import org.apache.hadoop.hbase.filter.FilterList;
 import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-import java.lang.ArrayIndexOutOfBoundsException;
-import java.lang.ClassCastException;
-import java.lang.reflect.*;
-import java.util.EmptyStackException;
+import org.apache.hadoop.hbase.util.Bytes;
 
 /**
  * This class allows a user to specify a filter via a string
@@ -289,7 +282,7 @@ public class ParseFilter {
 /**
  * Returns the arguments of the filter from the filter string
  * <p>
- * @param filter_string filter string given by the user
+ * @param filterStringAsByteArray filter string given by the user
  * @return an ArrayList containing the arguments of the filter in the filter string
  */
   public static ArrayList<byte []> getFilterArguments (byte [] filterStringAsByteArray) {
@@ -361,7 +354,6 @@ public class ParseFilter {
  * @param operatorStack the stack containing the operators and parenthesis
  * @param filterStack the stack containing the filters
  * @param operator the operator found while parsing the filterString
- * @return returns the filterStack after evaluating the stack
  */
   public void reduce(Stack<ByteBuffer> operatorStack,
                      Stack<Filter> filterStack,
@@ -646,7 +638,7 @@ public class ParseFilter {
  * byte array representing abc
  * <p>
  * @param quotedByteArray the quoted byte array
- * @return
+ * @return Unquoted byte array
  */
   public static byte [] removeQuotesFromByteArray (byte [] quotedByteArray) {
     if (quotedByteArray == null ||
@@ -665,7 +657,7 @@ public class ParseFilter {
  * Converts an int expressed in a byte array to an actual int
  * <p>
  * This doesn't use Bytes.toInt because that assumes
- * that there will be {@link #SIZEOF_INT} bytes available.
+ * that there will be {@link Bytes#SIZEOF_INT} bytes available.
  * <p>
  * @param numberAsByteArray the int value expressed as a byte array
  * @return the int value
@@ -688,7 +680,7 @@ public class ParseFilter {
  * Converts a long expressed in a byte array to an actual long
  * <p>
  * This doesn't use Bytes.toLong because that assumes
- * that there will be {@link #SIZEOF_LONG} bytes available.
+ * that there will be {@link Bytes#SIZEOF_INT} bytes available.
  * <p>
  * @param numberAsByteArray the long value expressed as a byte array
  * @return the long value

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java Tue Oct 25 06:57:10 2011
@@ -20,11 +20,11 @@
 
 package org.apache.hadoop.hbase.filter;
 
-import org.apache.hadoop.hbase.KeyValue;
-
-import java.util.List;
 import java.util.ArrayList;
 
+import org.apache.hadoop.hbase.KeyValue;
+import org.apache.hadoop.hbase.client.Scan;
+
 /**
  * This filter is used to filter based on the key. It takes an operator
  * (equal, greater, not equal, etc) and a byte [] comparator for the row,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java Tue Oct 25 06:57:10 2011
@@ -55,7 +55,7 @@ public class HalfStoreFileReader extends
   /**
    * @param fs
    * @param p
-   * @param c
+   * @param cacheConf
    * @param r
    * @throws IOException
    */

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java Tue Oct 25 06:57:10 2011
@@ -198,7 +198,7 @@ public abstract class AbstractHFileReade
   }
 
   /**
-   * TODO left from {@HFile} version 1: move this to StoreFile after Ryan's
+   * TODO left from {@link HFile} version 1: move this to StoreFile after Ryan's
    * patch goes in to eliminate {@link KeyValue} here.
    *
    * @return the first row key, or null if the file is empty.
@@ -212,7 +212,7 @@ public abstract class AbstractHFileReade
   }
 
   /**
-   * TODO left from {@HFile} version 1: move this to StoreFile after
+   * TODO left from {@link HFile} version 1: move this to StoreFile after
    * Ryan's patch goes in to eliminate {@link KeyValue} here.
    *
    * @return the last row key, or null if the file is empty.

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java Tue Oct 25 06:57:10 2011
@@ -74,7 +74,7 @@ public interface BlockCache {
 
   /**
    * Get the statistics for this block cache.
-   * @return
+   * @return Stats
    */
   public CacheStats getStats();
 

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java Tue Oct 25 06:57:10 2011
@@ -235,18 +235,14 @@ public class HFile {
   }
 
   /**
-   * Returns the factory to be used to create {@link HFile} writers. Should
-   * always be {@link HFileWriterV2#WRITER_FACTORY_V2} in production, but
-   * can also be {@link HFileWriterV1#WRITER_FACTORY_V1} in testing.
+   * Returns the factory to be used to create {@link HFile} writers.
    */
   public static final WriterFactory getWriterFactory(Configuration conf) {
     return HFile.getWriterFactory(conf, new CacheConfig(conf));
   }
 
   /**
-   * Returns the factory to be used to create {@link HFile} writers. Should
-   * always be {@link HFileWriterV2#WRITER_FACTORY_V2} in production, but
-   * can also be {@link HFileWriterV1#WRITER_FACTORY_V1} in testing.
+   * Returns the factory to be used to create {@link HFile} writers
    */
   public static final WriterFactory getWriterFactory(Configuration conf,
       CacheConfig cacheConf) {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java Tue Oct 25 06:57:10 2011
@@ -485,10 +485,10 @@ public class HFileBlock implements Cache
    * <ul>
    * <li>Construct an {@link HFileBlock.Writer}, providing a compression
    * algorithm
-   * <li>Call {@link Writer#startWriting(BlockType)} and get a data stream to
+   * <li>Call {@link Writer#startWriting(BlockType, boolean)} and get a data stream to
    * write to
    * <li>Write your data into the stream
-   * <li>Call {@link Writer#writeHeaderAndData()} as many times as you need to
+   * <li>Call {@link Writer#writeHeaderAndData(FSDataOutputStream)} as many times as you need to
    * store the serialized block into an external stream, or call
    * {@link Writer#getHeaderAndData()} to get it as a byte array.
    * <li>Repeat to write more blocks
@@ -576,8 +576,6 @@ public class HFileBlock implements Cache
     private long prevOffset;
 
     /**
-     * @param blockType
-     *          block type to create
      * @param compressionAlgorithm
      *          compression algorithm to use
      */
@@ -707,7 +705,7 @@ public class HFileBlock implements Cache
     }
 
     /**
-     * Similar to {@link #writeHeaderAndData(DataOutputStream)}, but records
+     * Similar to {@link #writeHeaderAndData(FSDataOutputStream)}, but records
      * the offset of this block so that it can be referenced in the next block
      * of the same type.
      *
@@ -854,7 +852,7 @@ public class HFileBlock implements Cache
     }
 
     /**
-     * Similar to {@link #getUncompressedDataWithHeader()} but returns a byte
+     * Similar to {@link #getUncompressedBufferWithHeader()} but returns a byte
      * buffer.
      *
      * @return uncompressed block for caching on write in the form of a buffer
@@ -1074,20 +1072,15 @@ public class HFileBlock implements Cache
     /**
      * Decompresses data from the given stream using the configured compression
      * algorithm.
-     *
-     * @param boundedStream
+     * @param dest
+     * @param destOffset
+     * @param bufferedBoundedStream
      *          a stream to read compressed data from, bounded to the exact
      *          amount of compressed data
      * @param compressedSize
      *          compressed data size, header not included
      * @param uncompressedSize
      *          uncompressed data size, header not included
-     * @param header
-     *          the header to include before the decompressed data, or null.
-     *          Only the first {@link HFileBlock#HEADER_SIZE} bytes of the
-     *          buffer are included.
-     * @return the byte buffer containing the given header (optionally) and the
-     *         decompressed data
      * @throws IOException
      */
     protected void decompress(byte[] dest, int destOffset,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java Tue Oct 25 06:57:10 2011
@@ -52,7 +52,7 @@ import org.apache.hadoop.io.WritableUtil
  * Examples of how to use the block index writer can be found in
  * {@link CompoundBloomFilterWriter} and {@link HFileWriterV2}. Examples of how
  * to use the reader can be found in {@link HFileReaderV2} and
- * {@link TestHFileBlockIndex}.
+ * TestHFileBlockIndex.
  */
 public class HFileBlockIndex {
 
@@ -62,7 +62,7 @@ public class HFileBlockIndex {
 
   /**
    * The maximum size guideline for index blocks (both leaf, intermediate, and
-   * root). If not specified, {@link #DEFAULT_MAX_CHUNK_SIZE} is used.
+   * root). If not specified, <code>DEFAULT_MAX_CHUNK_SIZE</code> is used.
    */
   public static final String MAX_CHUNK_SIZE_KEY = "hfile.index.block.max.size";
 
@@ -913,8 +913,6 @@ public class HFileBlockIndex {
      * blocks, so the non-root index format is used.
      *
      * @param out
-     * @param position The beginning offset of the inline block in the file not
-     *          include the header.
      */
     @Override
     public void writeInlineBlock(DataOutput out) throws IOException {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java Tue Oct 25 06:57:10 2011
@@ -55,7 +55,6 @@ public class HFileReaderV1 extends Abstr
    * stream.
    * @param size Length of the stream.
    * @param cacheConf cache references and configuration
-   * @throws IOException
    */
   public HFileReaderV1(Path path, FixedFileTrailer trailer,
       final FSDataInputStream fsdis, final long size,
@@ -80,7 +79,7 @@ public class HFileReaderV1 extends Abstr
    * Read in the index and file info.
    *
    * @return A map of fileinfo data.
-   * @see {@link Writer#appendFileInfo(byte[], byte[])}.
+   * @see Writer#appendFileInfo(byte[], byte[])
    * @throws IOException
    */
   @Override

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java Tue Oct 25 06:57:10 2011
@@ -207,7 +207,8 @@ public class HFileReaderV2 extends Abstr
    * Read in a file block.
    *
    * @param dataBlockOffset offset to read.
-   * @param onDiskSize size of the block
+   * @param onDiskBlockSize size of the block
+   * @param cacheBlock
    * @param pread Use positional read instead of seek+read (positional is better
    *          doing random reads whereas seek+read is better scanning).
    * @param isCompaction is this block being read as part of a compaction

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java Tue Oct 25 06:57:10 2011
@@ -56,7 +56,6 @@ public interface InlineBlockWriter {
    * @param offset the offset of the block in the stream
    * @param onDiskSize the on-disk size of the block
    * @param uncompressedSize the uncompressed size of the block
-   * @param rawSize
    */
   void blockWritten(long offset, int onDiskSize, int uncompressedSize);
 
@@ -69,5 +68,4 @@ public interface InlineBlockWriter {
    * @return true if inline blocks produced by this writer should be cached
    */
   boolean cacheOnWrite();
-
-}
+}
\ No newline at end of file

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java Tue Oct 25 06:57:10 2011
@@ -103,7 +103,7 @@ public class SlabCache implements SlabIt
    * The second list is blocksize of the slabs in bytes. (E.g. the slab holds
    * blocks of this size).
    *
-   * @param Configuration file.
+   * @param conf Configuration file.
    */
   public void addSlabByConf(Configuration conf) {
     // Proportions we allocate to each slab of the total size.
@@ -229,7 +229,8 @@ public class SlabCache implements SlabIt
   /**
    * Get the buffer of the block with the specified name.
    *
-   * @param blockName block name
+   * @param key
+   * @param caching
    * @return buffer of specified block name, or null if not in cache
    */
   public Cacheable getBlock(String key, boolean caching) {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java Tue Oct 25 06:57:10 2011
@@ -48,7 +48,7 @@ public interface Delayable {
   /**
    * Signal that the  RPC server is now allowed to send the response.
    * @param result The value to return to the caller.  If the corresponding
-   * {@link #delayResponse(boolean)} specified that the return value should
+   * delay response specified that the return value should
    * not be delayed, this parameter must be null.
    * @throws IOException
    */

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java Tue Oct 25 06:57:10 2011
@@ -134,9 +134,9 @@ public abstract class HBaseServer implem
   }
 
   /** Returns the server instance called under or null.  May be called under
-   * {@link #call(Class, Writable, long)} implementations, and under {@link Writable}
-   * methods of paramters and return values.  Permits applications to access
-   * the server context.
+   * {@link #call(Class, Writable, long, MonitoredRPCHandler)} implementations,
+   * and under {@link Writable} methods of paramters and return values.
+   * Permits applications to access the server context.
    * @return HBaseServer
    */
   public static RpcServer get() {
@@ -860,7 +860,6 @@ public abstract class HBaseServer implem
     // Processes one response. Returns true if there are no more pending
     // data for this channel.
     //
-    @SuppressWarnings({"ConstantConditions"})
     private boolean processResponse(final LinkedList<Call> responseQueue,
                                     boolean inHandler) throws IOException {
       boolean error = true;
@@ -1280,14 +1279,15 @@ public abstract class HBaseServer implem
 
   }
 
+
+  private Function<Writable,Integer> qosFunction = null;
+
   /**
    * Gets the QOS level for this call.  If it is higher than the highPriorityLevel and there
    * are priorityHandlers available it will be processed in it's own thread set.
    *
-   * @param param
-   * @return priority, higher is better
+   * @param newFunc
    */
-  private Function<Writable,Integer> qosFunction = null;
   @Override
   public void setQosFunction(Function<Writable, Integer> newFunc) {
     qosFunction = newFunc;

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java Tue Oct 25 06:57:10 2011
@@ -159,7 +159,7 @@ public class InputSampler<K,V> extends C
    * native Hadoop ones (We'll throw a ClassNotFoundException if end up in
    * here when we should be using native hadoop TotalOrderPartitioner).
    * @param job
-   * @return
+   * @return Context
    * @throws IOException 
    */
   public static TaskAttemptContext getTaskAttemptContext(final Job job)

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java Tue Oct 25 06:57:10 2011
@@ -276,7 +276,7 @@ implements HMasterInterface, HMasterRegi
    * Main processing loop for the HMaster.
    * <ol>
    * <li>Block until becoming active master
-   * <li>Finish initialization via {@link #finishInitialization()}
+   * <li>Finish initialization via finishInitialization(MonitoredTask)
    * <li>Enter loop until we are stopped
    * <li>Stop services and perform cleanup once stopped
    * </ol>
@@ -856,7 +856,7 @@ implements HMasterInterface, HMasterRegi
   }
 
   /**
-   * Switch for the background {@link CatalogJanitor} thread.
+   * Switch for the background CatalogJanitor thread.
    * Used for testing.  The thread will continue to run.  It will just be a noop
    * if disabled.
    * @param b If false, the catalog janitor won't do anything.

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java Tue Oct 25 06:57:10 2011
@@ -61,7 +61,7 @@ public interface LoadBalancer extends Co
   /**
    * Perform the major balance operation
    * @param clusterState
-   * @return
+   * @return List of plans
    */
   public List<RegionPlan> balanceCluster(Map<ServerName, List<HRegionInfo>> clusterState);
 
@@ -69,7 +69,7 @@ public interface LoadBalancer extends Co
    * Perform a Round Robin assignment of regions.
    * @param regions
    * @param servers
-   * @return
+   * @return Map of servername to regioninfos
    */
   public Map<ServerName, List<HRegionInfo>> roundRobinAssignment(List<HRegionInfo> regions, List<ServerName> servers);
 
@@ -77,7 +77,7 @@ public interface LoadBalancer extends Co
    * Assign regions to the previously hosting region server
    * @param regions
    * @param servers
-   * @return
+   * @return List of plans
    */
   public Map<ServerName, List<HRegionInfo>> retainAssignment(Map<HRegionInfo, ServerName> regions, List<ServerName> servers);
 
@@ -85,14 +85,14 @@ public interface LoadBalancer extends Co
    * Sync assign a region
    * @param regions
    * @param servers
-   * @return
+    * @return Map regioninfos to servernames
    */
   public Map<HRegionInfo, ServerName> immediateAssignment(List<HRegionInfo> regions, List<ServerName> servers);
 
   /**
    * Get a random region server from the list
    * @param servers
-   * @return
+   * @return Servername
    */
   public ServerName randomAssignment(List<ServerName> servers);
 }

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java Tue Oct 25 06:57:10 2011
@@ -31,7 +31,7 @@ public class LoadBalancerFactory {
   /**
    * Create a loadblanacer from the given conf.
    * @param conf
-   * @return
+   * @return A {@link LoadBalancer}
    */
   public static LoadBalancer getLoadBalancer(Configuration conf) {
 

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java Tue Oct 25 06:57:10 2011
@@ -164,15 +164,13 @@ public class MasterFileSystem {
 
   /**
    * @return HBase root dir.
-   * @throws IOException
    */
   public Path getRootDir() {
     return this.rootdir;
   }
 
   /**
-   * Returns the unique identifier generated for this cluster
-   * @return
+   * @return The unique identifier generated for this cluster
    */
   public String getClusterId() {
     return clusterId;

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java Tue Oct 25 06:57:10 2011
@@ -243,7 +243,7 @@ public class ServerManager {
   }
 
   /**
-   * @param serverName
+   * @param address
    * @return HServerLoad if serverName is known else null
    * @deprecated Use {@link #getLoad(HServerAddress)}
    */

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java Tue Oct 25 06:57:10 2011
@@ -68,8 +68,8 @@ import static org.apache.hadoop.hbase.ma
  * znode is created under /hbase/splitlog. SplitLogWorkers race to grab a task.
  *
  * SplitLogManager monitors the task znodes that it creates using the
- * {@link #timeoutMonitor} thread. If a task's progress is slow then
- * {@link #resubmit(String, boolean)} will take away the task from the owner
+ * timeoutMonitor thread. If a task's progress is slow then
+ * resubmit(String, boolean) will take away the task from the owner
  * {@link SplitLogWorker} and the task will be
  * upforgrabs again. When the task is done then the task's znode is deleted by
  * SplitLogManager.
@@ -122,8 +122,6 @@ public class SplitLogManager extends Zoo
    * @param conf
    * @param stopper
    * @param serverName
-   * @param services
-   * @param service
    */
   public SplitLogManager(ZooKeeperWatcher zkw, final Configuration conf,
       Stoppable stopper, String serverName) {
@@ -205,7 +203,7 @@ public class SplitLogManager extends Zoo
    * @throws IOException
    *             if there was an error while splitting any log file
    * @return cumulative size of the logfiles split
-   * @throws KeeperException 
+   * @throws IOException 
    */
   public long splitLogDistributed(final Path logDir) throws IOException {
     List<Path> logDirs = new ArrayList<Path>();
@@ -218,8 +216,7 @@ public class SplitLogManager extends Zoo
    * available worker region server. This method must only be called after the
    * region servers have been brought online.
    *
-   * @param logDir
-   *          the log directory encoded with a region server name
+   * @param logDirs
    * @throws IOException
    *          if there was an error while splitting any log file
    * @return cumulative size of the logfiles split

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java Tue Oct 25 06:57:10 2011
@@ -56,7 +56,7 @@ public interface MonitoredTask extends C
   /**
    * Public exposure of Object.clone() in order to allow clients to easily 
    * capture current state.
-   * @returns a copy of the object whose references will not change
+   * @return a copy of the object whose references will not change
    */
   public abstract MonitoredTask clone();
 

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java Tue Oct 25 06:57:10 2011
@@ -48,8 +48,9 @@ public abstract class ThreadMonitoring {
   /**
    * Print all of the thread's information and stack traces.
    * 
-   * @param stream the stream to
-   * 
+   * @param sb
+   * @param info
+   * @param indent
    */
   public static void appendThreadInfo(StringBuilder sb,
                                       ThreadInfo info,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java Tue Oct 25 06:57:10 2011
@@ -295,7 +295,7 @@ public class HRegion implements HeapSize
   /**
    * HRegion constructor.  his constructor should only be used for testing and
    * extensions.  Instances of HRegion should be instantiated with the
-   * {@link HRegion#newHRegion(Path, HLog, FileSystem, Configuration, org.apache.hadoop.hbase.HRegionInfo, FlushRequester)} method.
+   * {@link HRegion#newHRegion(Path, HLog, FileSystem, Configuration, HRegionInfo, HTableDescriptor, RegionServerServices)} method.
    *
    *
    * @param tableDir qualified path of directory where region should be located,
@@ -313,7 +313,7 @@ public class HRegion implements HeapSize
    * is new), then read them from the supplied path.
    * @param rsServices reference to {@link RegionServerServices} or null
    *
-   * @see HRegion#newHRegion(Path, HLog, FileSystem, Configuration, org.apache.hadoop.hbase.HRegionInfo, FlushRequester)
+   * @see HRegion#newHRegion(Path, HLog, FileSystem, Configuration, HRegionInfo, HTableDescriptor, RegionServerServices)
    */
   public HRegion(Path tableDir, HLog log, FileSystem fs, Configuration conf,
       HRegionInfo regionInfo, final HTableDescriptor htd,
@@ -1110,7 +1110,7 @@ public class HRegion implements HeapSize
    * @param status 
    * @return true if the region needs compacting
    * @throws IOException
-   * @see #internalFlushcache()
+   * @see #internalFlushcache(MonitoredTask)
    */
   protected boolean internalFlushcache(
       final HLog wal, final long myseqid, MonitoredTask status)
@@ -2534,7 +2534,7 @@ public class HRegion implements HeapSize
   
   /**
    * Release the row lock!
-   * @param lockid  The lock ID to release.
+   * @param lockId  The lock ID to release.
    */
   public void releaseRowLock(final Integer lockId) {
     HashedBytes rowKey = lockIds.remove(lockId);
@@ -2944,13 +2944,14 @@ public class HRegion implements HeapSize
 
   /**
    * Open a Region.
-   * @param info Info for region to be opened.
+   * @param info Info for region to be opened
+   * @param htd
    * @param wal HLog for region to use. This method will call
    * HLog#setSequenceNumber(long) passing the result of the call to
    * HRegion#getMinSequenceId() to ensure the log id is properly kept
    * up.  HRegionStore does this every time it opens a new region.
    * @param conf
-   * @param flusher An interface we can request flushes against.
+   * @param rsServices An interface we can request flushes against.
    * @param reporter An interface we can report progress against.
    * @return new HRegion
    *

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java Tue Oct 25 06:57:10 2011
@@ -210,8 +210,8 @@ public class RegionCoprocessorHost
    * {@link org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost#postOpen()} are such hooks.
    *
    * See also {@link org.apache.hadoop.hbase.master.MasterCoprocessorHost#handleCoprocessorThrowable()}
-   * @param env: The coprocessor that threw the exception.
-   * @param e: The exception that was thrown.
+   * @param env The coprocessor that threw the exception.
+   * @param e The exception that was thrown.
    */
   private void handleCoprocessorThrowableNoRethrow(
       final CoprocessorEnvironment env, final Throwable e) {
@@ -566,7 +566,6 @@ public class RegionCoprocessorHost
   /**
    * @param get the Get request
    * @param results the result set
-   * @return the possibly transformed result set to use
    * @exception IOException Exception
    */
   public void postGet(final Get get, final List<KeyValue> results)

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java Tue Oct 25 06:57:10 2011
@@ -548,8 +548,7 @@ public class SplitLogWorker extends ZooK
    * acquired by a {@link SplitLogWorker}. Since there isn't a water-tight
    * guarantee that two workers will not be executing the same task therefore it
    * is better to have workers prepare the task and then have the
-   * {@link SplitLogManager} commit the work in
-   * {@link SplitLogManager.TaskFinisher}
+   * {@link SplitLogManager} commit the work in SplitLogManager.TaskFinisher
    */
   static public interface TaskExecutor {
     static public enum Status {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java Tue Oct 25 06:57:10 2011
@@ -61,7 +61,7 @@ import com.google.common.util.concurrent
 /**
  * Executes region split as a "transaction".  Call {@link #prepare()} to setup
  * the transaction, {@link #execute(Server, RegionServerServices)} to run the
- * transaction and {@link #rollback(OnlineRegions)} to cleanup if execute fails.
+ * transaction and {@link #rollback(Server, RegionServerServices)} to cleanup if execute fails.
  *
  * <p>Here is an example of how you would use this class:
  * <pre>
@@ -145,9 +145,6 @@ public class SplitTransaction {
 
   /**
    * Constructor
-   * @param services So we can online new regions.  If null, we'll skip onlining
-   * (Useful testing).
-   * @param c Configuration to use running split
    * @param r Region to split
    * @param splitrow Row to split around
    */
@@ -430,8 +427,7 @@ public class SplitTransaction {
    * @param services Used to online/offline regions.
    * @throws IOException If thrown, transaction failed. Call {@link #rollback(Server, RegionServerServices)}
    * @return Regions created
-   * @throws KeeperException
-   * @throws NodeExistsException 
+   * @throws IOException
    * @see #rollback(Server, RegionServerServices)
    */
   public PairOfSameType<HRegion> execute(final Server server,

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java Tue Oct 25 06:57:10 2011
@@ -67,7 +67,7 @@ import com.google.common.collect.Orderin
 /**
  * A Store data file.  Stores usually have one or more of these files.  They
  * are produced by flushing the memstore to disk.  To
- * create, call {@link #createWriter(FileSystem, Path, int, Configuration)}
+ * create, call {@link #createWriter(FileSystem, Path, int, Configuration, CacheConfig)}
  * and append data. Be sure to add any metadata before calling close on the
  * Writer (Use the appendMetadata convenience methods). On close, a StoreFile
  * is sitting in the Filesystem.  To refer to it, create a StoreFile instance
@@ -392,7 +392,7 @@ public class StoreFile {
    * helper function to compute HDFS blocks distribution of a given file.
    * For reference file, it is an estimate
    * @param fs  The FileSystem
-   * @param o  The path of the file
+   * @param p  The path of the file
    * @return HDFS blocks distribution
    */
   static public HDFSBlocksDistribution computeHDFSBlockDistribution(
@@ -518,7 +518,6 @@ public class StoreFile {
 
   /**
    * @return Current reader.  Must call createReader first else returns null.
-   * @throws IOException
    * @see #createReader()
    */
   public Reader getReader() {
@@ -526,7 +525,7 @@ public class StoreFile {
   }
 
   /**
-   * @param b 
+   * @param evictOnClose 
    * @throws IOException
    */
   public synchronized void closeReader(boolean evictOnClose)
@@ -819,7 +818,6 @@ public class StoreFile {
      * If the timeRangeTracker is not set,
      * update TimeRangeTracker to include the timestamp of this key
      * @param kv
-     * @throws IOException
      */
     public void includeInTimeRangeTracker(final KeyValue kv) {
       if (!isTimeRangeTrackerSet) {
@@ -1081,7 +1079,7 @@ public class StoreFile {
 
     /**
      * A method for checking Bloom filters. Called directly from
-     * {@link StoreFileScanner} in case of a multi-column query.
+     * StoreFileScanner in case of a multi-column query.
      *
      * @param row
      * @param rowOffset
@@ -1089,7 +1087,7 @@ public class StoreFile {
      * @param col
      * @param colOffset
      * @param colLen
-     * @return
+     * @return True if passes
      */
     public boolean passesBloomFilter(byte[] row, int rowOffset, int rowLen,
         byte[] col, int colOffset, int colLen) {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java Tue Oct 25 06:57:10 2011
@@ -1284,7 +1284,7 @@ public class HLog implements Syncable {
    *
    * In this method, by removing the entry in lastSeqWritten for the region
    * being flushed we ensure that the next edit inserted in this region will be
-   * correctly recorded in {@link #append(HRegionInfo, HLogKey, WALEdit)}. The
+   * correctly recorded in {@link #append(HRegionInfo, byte[], WALEdit, long, HTableDescriptor)} The
    * lsn of the earliest in-memory lsn - which is now in the memstore snapshot -
    * is saved temporarily in the lastSeqWritten map while the flush is active.
    *
@@ -1292,7 +1292,7 @@ public class HLog implements Syncable {
    *         {@link #completeCacheFlush(byte[], byte[], long, boolean)} (byte[],
    *         byte[], long)}
    * @see #completeCacheFlush(byte[], byte[], long, boolean)
-   * @see #abortCacheFlush()
+   * @see #abortCacheFlush(byte[])
    */
   public long startCacheFlush(final byte[] encodedRegionName) {
     this.cacheFlushLock.lock();
@@ -1594,29 +1594,6 @@ public class HLog implements Syncable {
     System.err.println("         For example: HLog --split hdfs://example.com:9000/hbase/.logs/DIR");
   }
 
-  private static void dump(final Configuration conf, final Path p)
-  throws IOException {
-    FileSystem fs = FileSystem.get(conf);
-    if (!fs.exists(p)) {
-      throw new FileNotFoundException(p.toString());
-    }
-    if (!fs.isFile(p)) {
-      throw new IOException(p + " is not a file");
-    }
-    Reader log = getReader(fs, p, conf);
-    try {
-      int count = 0;
-      HLog.Entry entry;
-      while ((entry = log.next()) != null) {
-        System.out.println("#" + count + ", pos=" + log.getPosition() + " " +
-          entry.toString());
-        count++;
-      }
-    } finally {
-      log.close();
-    }
-  }
-
   private static void split(final Configuration conf, final Path p)
   throws IOException {
     FileSystem fs = FileSystem.get(conf);

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java Tue Oct 25 06:57:10 2011
@@ -71,7 +71,7 @@ public class HLogKey implements Writable
    * @param tablename   - name of table
    * @param logSeqNum   - log sequence number
    * @param now Time at which this edit was written.
-   * @param UUID of the cluster (used in Replication)
+   * @param clusterId of the cluster (used in Replication)
    */
   public HLogKey(final byte [] encodedRegionName, final byte [] tablename,
       long logSeqNum, final long now, UUID clusterId) {

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java Tue Oct 25 06:57:10 2011
@@ -37,8 +37,7 @@ import java.nio.ByteBuffer;
  * When creating the filter, the sender can choose its desired point in a
  * trade-off between the false positive rate and the size.
  *
- * @see {@link BloomFilterWriter} for the ability to add elements to a Bloom
- *      filter
+ * @see BloomFilterWriter for the ability to add elements to a Bloom filter
  */
 public interface BloomFilter extends BloomFilterBase {
 
@@ -59,5 +58,4 @@ public interface BloomFilter extends Blo
    *         and thus allows a null byte buffer to be passed to contains()
    */
   boolean supportsAutoLoading();
-
-}
+}
\ No newline at end of file

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java Tue Oct 25 06:57:10 2011
@@ -127,12 +127,11 @@ public final class BloomFilterFactory {
    * {@link org.apache.hadoop.hbase.regionserver.StoreFile} writing.
    *
    * @param conf
+   * @param cacheConf
    * @param bloomType
    * @param maxKeys an estimate of the number of keys we expect to insert.
    *        Irrelevant if compound Bloom filters are enabled.
    * @param writer the HFile writer
-   * @param comparator the comparator to use for compound Bloom filters. This
-   *        has no effect if creating single-chunk version 1 Bloom filters.
    * @return the new Bloom filter, or null in case Bloom filters are disabled
    *         or when failed to create one.
    */

Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java Tue Oct 25 06:57:10 2011
@@ -164,8 +164,7 @@ public class ByteBloomFilter implements 
    * @param bitSize
    * @param errorRate
    * @return maximum number of keys that can be inserted into the Bloom filter
-   * @see {@link #computeMaxKeys(long, double, int)} for a more precise
-   *      estimate
+   * @see #computeMaxKeys(long, double, int) for a more precise estimate
    */
   public static long idealMaxKeys(long bitSize, double errorRate) {
     // The reason we need to use floor here is that otherwise we might put
@@ -227,7 +226,7 @@ public class ByteBloomFilter implements 
    *
    * @param bitSize
    * @param foldFactor
-   * @return
+   * @return Foldable byte size
    */
   public static int computeFoldableByteSize(long bitSize, int foldFactor) {
     long byteSizeLong = (bitSize + 7) / 8;