You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2011/10/25 08:57:12 UTC
svn commit: r1188539 [1/2] - in /hbase/branches/0.92: ./
src/main/java/org/apache/hadoop/hbase/
src/main/java/org/apache/hadoop/hbase/catalog/
src/main/java/org/apache/hadoop/hbase/client/
src/main/java/org/apache/hadoop/hbase/client/coprocessor/ src/m...
Author: stack
Date: Tue Oct 25 06:57:10 2011
New Revision: 1188539
URL: http://svn.apache.org/viewvc?rev=1188539&view=rev
Log:
HBASE-4670 Fix javadoc warnings
Modified:
hbase/branches/0.92/CHANGES.txt
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/Bytes.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterBase.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/Objects.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionServerTracker.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java
Modified: hbase/branches/0.92/CHANGES.txt
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/CHANGES.txt?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/CHANGES.txt (original)
+++ hbase/branches/0.92/CHANGES.txt Tue Oct 25 06:57:10 2011
@@ -369,6 +369,7 @@ Release 0.92.0 - Unreleased
master (Eugene Koontz via apurtell)
HBASE-3512 Shell support for listing currently loaded coprocessors (Eugene
Koontz via apurtell)
+ HBASE-4670 Fix javadoc warnings
TESTS
HBASE-4492 TestRollingRestart fails intermittently
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HConstants.java Tue Oct 25 06:57:10 2011
@@ -437,13 +437,13 @@ public final class HConstants {
public static int DEFAULT_HBASE_META_SCANNER_CACHING = 100;
/**
- * Parameter name for unique identifier for this {@link Configuration}
- * instance. If there are two or more {@link Configuration} instances that,
+ * Parameter name for unique identifier for this {@link org.apache.hadoop.conf.Configuration}
+ * instance. If there are two or more {@link org.apache.hadoop.conf.Configuration} instances that,
* for all intents and purposes, are the same except for their instance ids,
- * then they will not be able to share the same {@link Connection} instance.
+ * then they will not be able to share the same {@link org.apache.hadoop.hbase.client.HConnection} instance.
* On the other hand, even if the instance ids are the same, it could result
- * in non-shared {@link Connection} instances if some of the other connection
- * parameters differ.
+ * in non-shared {@link org.apache.hadoop.hbase.client.HConnection}
+ * instances if some of the other connection parameters differ.
*/
public static String HBASE_CLIENT_INSTANCE_ID = "hbase.client.instance.id";
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerAddress.java Tue Oct 25 06:57:10 2011
@@ -90,7 +90,7 @@ public class HServerAddress implements W
}
/** @return Bind address -- the raw IP, the result of a call to
- * {@link InetSocketAddress#getAddress()#getHostAddress()} --
+ * InetSocketAddress#getAddress()#getHostAddress() --
* or null if cannot resolve */
public String getBindAddress() {
return getBindAddressInternal(address);
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HServerInfo.java Tue Oct 25 06:57:10 2011
@@ -104,8 +104,6 @@ implements WritableComparable<HServerInf
/**
* @return ServerName and load concatenated.
- * @see #getServerName()
- * @see #getLoad()
*/
@Override
public synchronized String toString() {
@@ -150,4 +148,4 @@ implements WritableComparable<HServerInf
if (this.startCode != o.getStartCode()) return (int)(this.startCode - o.getStartCode());
return 0;
}
-}
\ No newline at end of file
+}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java Tue Oct 25 06:57:10 2011
@@ -126,8 +126,6 @@ public class HTableDescriptor implements
/**
* <em>INTERNAL</em> Used by HBase Shell interface to access this metadata
* attribute which denotes if the deferred log flush option is enabled
- *
- * @see #getDeferredLogFlush()
*/
public static final String DEFERRED_LOG_FLUSH = "DEFERRED_LOG_FLUSH";
private static final ImmutableBytesWritable DEFERRED_LOG_FLUSH_KEY =
@@ -524,7 +522,7 @@ public class HTableDescriptor implements
* before these deferred edits in memory are flushed onto the filesystem.
* </p>
*
- * @param true if that deferred log flush is enabled on the table.
+ * @param isDeferredLogFlush
*/
public void setDeferredLogFlush(final boolean isDeferredLogFlush) {
setValue(DEFERRED_LOG_FLUSH_KEY, isDeferredLogFlush? TRUE: FALSE);
@@ -552,8 +550,8 @@ public class HTableDescriptor implements
/**
* This get the class associated with the region split policy which
* determines when a region split should occur. The class used by
- * default is {@link ConstantSizeRegionSplitPolicy} which split the
- * region base on a constant {@link #getMaxFileSize()}
+ * default is {@link org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy}
+ * which split the region base on a constant {@link #getMaxFileSize()}
*
* @return the class name of the region split policy for this table.
* If this returns null, the default constant size based split policy
@@ -877,7 +875,7 @@ public class HTableDescriptor implements
}
/**
- * Returns an unmodifiable collection of all the {@link HColumnDescriptors}
+ * Returns an unmodifiable collection of all the {@link HColumnDescriptor}
* of all the column families of the table.
*
* @return Immutable collection of {@link HColumnDescriptor} of all the
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java Tue Oct 25 06:57:10 2011
@@ -1749,7 +1749,6 @@ public class KeyValue implements Writabl
* (the value part of the returned KV is always empty). Used in creating
* "fake keys" for the multi-column Bloom filter optimization to skip the
* row/column we already know is not in the file.
- * @param kv the key-value pair to take row and column from
* @return the last key on the row/column of the given key-value pair
*/
public KeyValue createLastOnRowCol() {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ServerName.java Tue Oct 25 06:57:10 2011
@@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.util.Byte
* usually timestamp of server startup). The {@link #toString()} format of
* ServerName is safe to use in the filesystem and as znode name up in
* ZooKeeper. Its format is:
- * <code><hostname> '{@link #SERVERNAME_SEPARATOR"}' <port> '{@ink #SERVERNAME_SEPARATOR"}' <startcode></code>.
+ * <code><hostname> '{@link #SERVERNAME_SEPARATOR}' <port> '{@link #SERVERNAME_SEPARATOR}' <startcode></code>.
* For example, if hostname is <code>example.org</code>, port is <code>1234</code>,
* and the startcode for the regionserver is <code>1212121212</code>, then
* the {@link #toString()} would be <code>example.org,1234,1212121212</code>.
@@ -220,7 +220,7 @@ public class ServerName implements Compa
/**
* @param left
- * @param rigth
+ * @param right
* @return True if <code>other</code> has same hostname and port.
*/
public static boolean isSameHostnameAndPort(final ServerName left,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java Tue Oct 25 06:57:10 2011
@@ -49,8 +49,6 @@ public interface TableDescriptors {
/**
* Get Map of all HTableDescriptors. Populates the descriptor cache as a
* side effect.
- * @param fs
- * @param rootdir
* @return Map of all descriptors.
* @throws IOException
*/
@@ -72,4 +70,4 @@ public interface TableDescriptors {
*/
public HTableDescriptor remove(final String tablename)
throws IOException;
-}
\ No newline at end of file
+}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java Tue Oct 25 06:57:10 2011
@@ -145,7 +145,7 @@ public class CatalogTracker {
* Does not timeout.
* @param zk If zk is null, we'll create an instance (and shut it down
* when {@link #stop()} is called) else we'll use what is passed.
- * @param connection server connection
+ * @param conf
* @param abortable If fatal exception we'll call abort on this. May be null.
* If it is we'll use the Connection associated with the passed
* {@link Configuration} as our Abortable.
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java Tue Oct 25 06:57:10 2011
@@ -50,7 +50,7 @@ public class MetaMigrationRemovingHTD {
/**
* Update legacy META rows, removing HTD from HRI.
* @param masterServices
- * @return
+ * @return List of table descriptors.
* @throws IOException
*/
public static List<HTableDescriptor> updateMetaWithNewRegionInfo(
@@ -78,7 +78,7 @@ public class MetaMigrationRemovingHTD {
/**
* Update the ROOT with new HRI. (HRI with no HTD)
* @param masterServices
- * @return
+ * @return List of table descriptors
* @throws IOException
*/
public static List<HTableDescriptor> updateRootWithNewRegionInfo(
@@ -239,4 +239,4 @@ public class MetaMigrationRemovingHTD {
"Master startup aborted.");
}
}
-}
\ No newline at end of file
+}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java Tue Oct 25 06:57:10 2011
@@ -251,7 +251,7 @@ public class MetaReader {
* @param metaServer connection to server hosting ROOT
* @return location of META in ROOT where location, or null if not available
* @throws IOException
- * @deprecated Does not retry; use {@link #readRegionLocation(CatalogTracker, byte[])
+ * @deprecated Does not retry; use {@link #readRegionLocation(CatalogTracker, byte[])}
*/
public static ServerName readMetaLocation(HRegionInterface metaServer)
throws IOException {
@@ -589,7 +589,7 @@ public class MetaReader {
/**
* @param catalogTracker
- * @param hsi Server specification
+ * @param serverName
* @return List of user regions installed on this server (does not include
* catalog regions).
* @throws IOException
@@ -742,4 +742,4 @@ public class MetaReader {
this.results.add(r);
}
}
-}
\ No newline at end of file
+}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java Tue Oct 25 06:57:10 2011
@@ -21,7 +21,6 @@ package org.apache.hadoop.hbase.client;
import java.io.Closeable;
import java.io.IOException;
-import java.net.InetSocketAddress;
import java.util.List;
import java.util.Map;
import java.util.concurrent.ExecutorService;
@@ -192,7 +191,7 @@ public interface HConnection extends Abo
* @param regionServer - the server to connect to
* @return proxy for HRegionServer
* @throws IOException if a remote or network exception occurs
- * @deprecated Use {@link #getHRegionConnection(InetSocketAddress)}
+ * @deprecated Use {@link #getHRegionConnection(String, int)}
*/
public HRegionInterface getHRegionConnection(HServerAddress regionServer)
throws IOException;
@@ -384,4 +383,4 @@ public interface HConnection extends Abo
public HTableDescriptor[] getHTableDescriptors(List<String> tableNames)
throws IOException;
-}
\ No newline at end of file
+}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java Tue Oct 25 06:57:10 2011
@@ -208,10 +208,7 @@ public class HConnectionManager {
* This will then close connection to
* the zookeeper ensemble and let go of all resources.
*
- * @param conf
- * configuration whose identity is used to find {@link HConnection}
- * instance.
- * .
+ * @param connection
*/
public static void deleteStaleConnection(HConnection connection) {
deleteConnection(connection, true, true);
@@ -302,7 +299,7 @@ public class HConnectionManager {
* of a {@link HConnection} instance based on the given {@link Configuration}.
*
* <p>
- * If you find yourself wanting to use a {@link Connection} for a relatively
+ * If you find yourself wanting to use a {@link HConnection} for a relatively
* short duration of time, and do not want to deal with the hassle of creating
* and cleaning up that resource, then you should consider using this
* convenience class.
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java Tue Oct 25 06:57:10 2011
@@ -279,7 +279,7 @@ public interface HTableInterface {
/**
* Atomically increments a column value.
* <p>
- * Equivalent to {@code {@link #incrementColumnValue(byte[], byte[], byte[],
+ * Equivalent to {@link #incrementColumnValue(byte[], byte[], byte[],
* long, boolean) incrementColumnValue}(row, family, qualifier, amount,
* <b>true</b>)}
* @param row The row that contains the cell to increment.
@@ -377,7 +377,7 @@ public interface HTableInterface {
*
* @param protocol The class or interface defining the remote protocol
* @param row The row key used to identify the remote region location
- * @return
+ * @return A CoprocessorProtocol instance
*/
<T extends CoprocessorProtocol> T coprocessorProxy(Class<T> protocol, byte[] row);
@@ -402,7 +402,7 @@ public interface HTableInterface {
* {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Call#call(Object)}
* method
* @return a <code>Map</code> of region names to
- * {@link Batch.Call#call(Object)} return values
+ * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Call#call(Object)} return values
*/
<T extends CoprocessorProtocol, R> Map<byte[],R> coprocessorExec(
Class<T> protocol, byte[] startKey, byte[] endKey, Batch.Call<T,R> callable)
@@ -421,7 +421,7 @@ public interface HTableInterface {
*
* <p>
* For each result, the given
- * {@link Batch.Callback#update(byte[], byte[], Object)}
+ * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Callback#update(byte[], byte[], Object)}
* method will be called.
*</p>
*
@@ -432,7 +432,7 @@ public interface HTableInterface {
* @param callable wraps the CoprocessorProtocol implementation method calls
* made per-region
* @param callback an instance upon which
- * {@link Batch.Callback#update(byte[], byte[], Object)} with the
+ * {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Callback#update(byte[], byte[], Object)} with the
* {@link org.apache.hadoop.hbase.client.coprocessor.Batch.Call#call(Object)}
* return value for each region
* @param <T> CoprocessorProtocol subclass for the remote invocation
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java Tue Oct 25 06:57:10 2011
@@ -106,8 +106,6 @@ public class HTablePool implements Close
* configuration
* @param maxSize
* maximum number of references to keep for each table
- * @param tableFactory
- * table factory
* @param poolType
* pool type which is one of {@link PoolType#Reusable} or
* {@link PoolType#ThreadLocal}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Mutation.java Tue Oct 25 06:57:10 2011
@@ -131,7 +131,6 @@ public abstract class Mutation extends O
/**
* Method for setting the put's familyMap
- * @return familyMap
*/
public void setFamilyMap(Map<byte [], List<KeyValue>> map) {
this.familyMap = map;
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java Tue Oct 25 06:57:10 2011
@@ -76,11 +76,6 @@ extends RetriesExhaustedException {
return actions.get(i);
}
- /**
- * @param i
- * @return
- * @deprecated
- */
public HServerAddress getAddress(int i) {
return new HServerAddress(Addressing.createInetSocketAddressFromHostAndPortStr(getHostnamePort(i)));
}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java Tue Oct 25 06:57:10 2011
@@ -170,7 +170,7 @@ public class AggregationClient {
* @param tableName
* @param ci
* @param scan
- * @return
+ * @return <R, S>
* @throws Throwable
*/
public <R, S> long rowCount(final byte[] tableName,
@@ -282,7 +282,7 @@ public class AggregationClient {
* @param tableName
* @param ci
* @param scan
- * @return
+ * @return <R, S>
* @throws Throwable
*/
public <R, S> double avg(final byte[] tableName,
@@ -348,7 +348,7 @@ public class AggregationClient {
* @param tableName
* @param ci
* @param scan
- * @return
+ * @return <R, S>
* @throws Throwable
*/
public <R, S> double std(final byte[] tableName, ColumnInterpreter<R, S> ci,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java Tue Oct 25 06:57:10 2011
@@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.util.Byte
* and its promoted data type is also a Long value. For computing aggregation
* function, this class is used to find the datatype of the cell value. Client
* is supposed to instantiate it and passed along as a parameter. See
- * {@link TestAggregateProtocol} methods for its sample usage.
+ * TestAggregateProtocol methods for its sample usage.
* Its methods handle null arguments gracefully.
*/
public class LongColumnInterpreter implements ColumnInterpreter<Long, Long> {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java Tue Oct 25 06:57:10 2011
@@ -80,11 +80,10 @@ public interface AggregateProtocol exten
<T, S> S getSum(ColumnInterpreter<T, S> ci, Scan scan) throws IOException;
/**
- * Gives the row count for the given column family and column qualifier, in
- * the given row range as defined in the Scan object.
* @param ci
* @param scan
- * @return
+ * @return Row count for the given column family and column qualifier, in
+ * the given row range as defined in the Scan object.
* @throws IOException
*/
<T, S> long getRowNum(ColumnInterpreter<T, S> ci, Scan scan)
@@ -104,7 +103,7 @@ public interface AggregateProtocol exten
* type.
* @param ci
* @param scan
- * @return
+ * @return Average
* @throws IOException
*/
<T, S> Pair<S, Long> getAvg(ColumnInterpreter<T, S> ci, Scan scan)
@@ -120,7 +119,7 @@ public interface AggregateProtocol exten
* deviation is square root of variance.
* @param ci
* @param scan
- * @return
+ * @return STD
* @throws IOException
*/
<T, S> Pair<List<S>, Long> getStd(ColumnInterpreter<T, S> ci, Scan scan)
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java Tue Oct 25 06:57:10 2011
@@ -42,14 +42,15 @@ import org.apache.hadoop.io.Writable;
* <S>. There is a conversion method
* {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and
* returns a <S> type.
- * @param <T, S>: T - cell value data type, S - promoted data type
+ * @param <T> Cell value data type
+ * @param <S> Promoted data type
*/
public interface ColumnInterpreter<T, S> extends Writable {
/**
* @param colFamily
* @param colQualifier
- * @param value
+ * @param kv
* @return value of type T
* @throws IOException
*/
@@ -57,44 +58,39 @@ public interface ColumnInterpreter<T, S>
throws IOException;
/**
- * returns sum or non null value among (if either of them is null); otherwise
- * returns a null.
* @param l1
* @param l2
- * @return
+ * @return sum or non null value among (if either of them is null); otherwise
+ * returns a null.
*/
public S add(S l1, S l2);
/**
* returns the maximum value for this type T
- * @return
+ * @return max
*/
T getMaxValue();
- /**
- * @return
- */
-
T getMinValue();
/**
* @param o1
* @param o2
- * @return
+ * @return multiplication
*/
S multiply(S o1, S o2);
/**
* @param o
- * @return
+ * @return increment
*/
S increment(S o);
/**
* provides casting opportunity between the data types.
* @param o
- * @return
+ * @return cast
*/
S castToReturnType(T o);
@@ -109,10 +105,10 @@ public interface ColumnInterpreter<T, S>
/**
* used for computing average of <S> data values. Not providing the divide
- * method that takes two <S> values as it si not needed as of now.
+ * method that takes two <S> values as it is not needed as of now.
* @param o
* @param l
- * @return
+ * @return Average
*/
double divideForAvg(S o, Long l);
}
\ No newline at end of file
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java Tue Oct 25 06:57:10 2011
@@ -220,7 +220,7 @@ public interface MasterObserver extends
/**
* Called prior to unassigning a given region.
* @param ctx the environment to interact with the framework and master
- * @param regionName the name of the region
+ * @param regionInfo
* @param force whether to force unassignment or not
*/
void preUnassign(final ObserverContext<MasterCoprocessorEnvironment> ctx,
@@ -229,7 +229,7 @@ public interface MasterObserver extends
/**
* Called after the region unassignment has been requested.
* @param ctx the environment to interact with the framework and master
- * @param regionName the name of the region
+ * @param regionInfo
* @param force whether to force unassignment or not
*/
void postUnassign(final ObserverContext<MasterCoprocessorEnvironment> ctx,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/coprocessor/package-info.java Tue Oct 25 06:57:10 2011
@@ -154,7 +154,7 @@ import org.apache.hadoop.hbase.client.Ge
// and intercept preXXX() method to check user privilege for the given table
// and column family.
public class AccessControlCoprocessor extends BaseRegionObserverCoprocessor {
- @Override
+ // @Override
public Get preGet(CoprocessorEnvironment e, Get get)
throws CoprocessorException {
@@ -208,9 +208,9 @@ extends CoprocessorProtocol {
// Aggregation implementation at a region.
public static class ColumnAggregationEndpoint extends BaseEndpointCoprocessor
implements ColumnAggregationProtocol {
- @Override
- // Scan the region by the given family and qualifier. Return the aggregation
- // result.
+ // @Override
+ // Scan the region by the given family and qualifier. Return the aggregation
+ // result.
public int sum(byte[] family, byte[] qualifier)
throws IOException {
// aggregate at each region
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java Tue Oct 25 06:57:10 2011
@@ -108,7 +108,7 @@ public class RegionTransitionData implem
*
* @param eventType type of event
* @param regionName name of region as per <code>HRegionInfo#getRegionName()</code>
- * @param origin Originating {@link ServerName}
+ * @param serverName Originating {@link ServerName}
* @param payload Payload examples include the daughters involved in a
* {@link EventType#RS_ZK_REGION_SPLIT}. Can be null
*/
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java Tue Oct 25 06:57:10 2011
@@ -47,7 +47,7 @@ public class BitComparator extends Writa
/**
* Constructor
* @param value value
- * @param BitwiseOp bitOperator - the operator to use on the bit comparison
+ * @param bitOperator operator to use on the bit comparison
*/
public BitComparator(byte[] value, BitwiseOp bitOperator) {
super(value);
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java Tue Oct 25 06:57:10 2011
@@ -19,26 +19,19 @@
*/
package org.apache.hadoop.hbase.filter;
-import java.io.IOException;
+import java.lang.reflect.InvocationTargetException;
+import java.lang.reflect.Method;
import java.nio.ByteBuffer;
import java.nio.charset.CharacterCodingException;
-import java.util.TreeSet;
import java.util.ArrayList;
-import java.util.Stack;
+import java.util.EmptyStackException;
import java.util.HashMap;
import java.util.Set;
+import java.util.Stack;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.Writables;
import org.apache.hadoop.hbase.KeyValue;
-import org.apache.hadoop.hbase.filter.ParseConstants;
-
-import org.apache.hadoop.hbase.filter.FilterList;
import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-import java.lang.ArrayIndexOutOfBoundsException;
-import java.lang.ClassCastException;
-import java.lang.reflect.*;
-import java.util.EmptyStackException;
+import org.apache.hadoop.hbase.util.Bytes;
/**
* This class allows a user to specify a filter via a string
@@ -289,7 +282,7 @@ public class ParseFilter {
/**
* Returns the arguments of the filter from the filter string
* <p>
- * @param filter_string filter string given by the user
+ * @param filterStringAsByteArray filter string given by the user
* @return an ArrayList containing the arguments of the filter in the filter string
*/
public static ArrayList<byte []> getFilterArguments (byte [] filterStringAsByteArray) {
@@ -361,7 +354,6 @@ public class ParseFilter {
* @param operatorStack the stack containing the operators and parenthesis
* @param filterStack the stack containing the filters
* @param operator the operator found while parsing the filterString
- * @return returns the filterStack after evaluating the stack
*/
public void reduce(Stack<ByteBuffer> operatorStack,
Stack<Filter> filterStack,
@@ -646,7 +638,7 @@ public class ParseFilter {
* byte array representing abc
* <p>
* @param quotedByteArray the quoted byte array
- * @return
+ * @return Unquoted byte array
*/
public static byte [] removeQuotesFromByteArray (byte [] quotedByteArray) {
if (quotedByteArray == null ||
@@ -665,7 +657,7 @@ public class ParseFilter {
* Converts an int expressed in a byte array to an actual int
* <p>
* This doesn't use Bytes.toInt because that assumes
- * that there will be {@link #SIZEOF_INT} bytes available.
+ * that there will be {@link Bytes#SIZEOF_INT} bytes available.
* <p>
* @param numberAsByteArray the int value expressed as a byte array
* @return the int value
@@ -688,7 +680,7 @@ public class ParseFilter {
* Converts a long expressed in a byte array to an actual long
* <p>
* This doesn't use Bytes.toLong because that assumes
- * that there will be {@link #SIZEOF_LONG} bytes available.
+ * that there will be {@link Bytes#SIZEOF_INT} bytes available.
* <p>
* @param numberAsByteArray the long value expressed as a byte array
* @return the long value
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java Tue Oct 25 06:57:10 2011
@@ -20,11 +20,11 @@
package org.apache.hadoop.hbase.filter;
-import org.apache.hadoop.hbase.KeyValue;
-
-import java.util.List;
import java.util.ArrayList;
+import org.apache.hadoop.hbase.KeyValue;
+import org.apache.hadoop.hbase.client.Scan;
+
/**
* This filter is used to filter based on the key. It takes an operator
* (equal, greater, not equal, etc) and a byte [] comparator for the row,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java Tue Oct 25 06:57:10 2011
@@ -55,7 +55,7 @@ public class HalfStoreFileReader extends
/**
* @param fs
* @param p
- * @param c
+ * @param cacheConf
* @param r
* @throws IOException
*/
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java Tue Oct 25 06:57:10 2011
@@ -198,7 +198,7 @@ public abstract class AbstractHFileReade
}
/**
- * TODO left from {@HFile} version 1: move this to StoreFile after Ryan's
+ * TODO left from {@link HFile} version 1: move this to StoreFile after Ryan's
* patch goes in to eliminate {@link KeyValue} here.
*
* @return the first row key, or null if the file is empty.
@@ -212,7 +212,7 @@ public abstract class AbstractHFileReade
}
/**
- * TODO left from {@HFile} version 1: move this to StoreFile after
+ * TODO left from {@link HFile} version 1: move this to StoreFile after
* Ryan's patch goes in to eliminate {@link KeyValue} here.
*
* @return the last row key, or null if the file is empty.
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java Tue Oct 25 06:57:10 2011
@@ -74,7 +74,7 @@ public interface BlockCache {
/**
* Get the statistics for this block cache.
- * @return
+ * @return Stats
*/
public CacheStats getStats();
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java Tue Oct 25 06:57:10 2011
@@ -235,18 +235,14 @@ public class HFile {
}
/**
- * Returns the factory to be used to create {@link HFile} writers. Should
- * always be {@link HFileWriterV2#WRITER_FACTORY_V2} in production, but
- * can also be {@link HFileWriterV1#WRITER_FACTORY_V1} in testing.
+ * Returns the factory to be used to create {@link HFile} writers.
*/
public static final WriterFactory getWriterFactory(Configuration conf) {
return HFile.getWriterFactory(conf, new CacheConfig(conf));
}
/**
- * Returns the factory to be used to create {@link HFile} writers. Should
- * always be {@link HFileWriterV2#WRITER_FACTORY_V2} in production, but
- * can also be {@link HFileWriterV1#WRITER_FACTORY_V1} in testing.
+ * Returns the factory to be used to create {@link HFile} writers
*/
public static final WriterFactory getWriterFactory(Configuration conf,
CacheConfig cacheConf) {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java Tue Oct 25 06:57:10 2011
@@ -485,10 +485,10 @@ public class HFileBlock implements Cache
* <ul>
* <li>Construct an {@link HFileBlock.Writer}, providing a compression
* algorithm
- * <li>Call {@link Writer#startWriting(BlockType)} and get a data stream to
+ * <li>Call {@link Writer#startWriting(BlockType, boolean)} and get a data stream to
* write to
* <li>Write your data into the stream
- * <li>Call {@link Writer#writeHeaderAndData()} as many times as you need to
+ * <li>Call {@link Writer#writeHeaderAndData(FSDataOutputStream)} as many times as you need to
* store the serialized block into an external stream, or call
* {@link Writer#getHeaderAndData()} to get it as a byte array.
* <li>Repeat to write more blocks
@@ -576,8 +576,6 @@ public class HFileBlock implements Cache
private long prevOffset;
/**
- * @param blockType
- * block type to create
* @param compressionAlgorithm
* compression algorithm to use
*/
@@ -707,7 +705,7 @@ public class HFileBlock implements Cache
}
/**
- * Similar to {@link #writeHeaderAndData(DataOutputStream)}, but records
+ * Similar to {@link #writeHeaderAndData(FSDataOutputStream)}, but records
* the offset of this block so that it can be referenced in the next block
* of the same type.
*
@@ -854,7 +852,7 @@ public class HFileBlock implements Cache
}
/**
- * Similar to {@link #getUncompressedDataWithHeader()} but returns a byte
+ * Similar to {@link #getUncompressedBufferWithHeader()} but returns a byte
* buffer.
*
* @return uncompressed block for caching on write in the form of a buffer
@@ -1074,20 +1072,15 @@ public class HFileBlock implements Cache
/**
* Decompresses data from the given stream using the configured compression
* algorithm.
- *
- * @param boundedStream
+ * @param dest
+ * @param destOffset
+ * @param bufferedBoundedStream
* a stream to read compressed data from, bounded to the exact
* amount of compressed data
* @param compressedSize
* compressed data size, header not included
* @param uncompressedSize
* uncompressed data size, header not included
- * @param header
- * the header to include before the decompressed data, or null.
- * Only the first {@link HFileBlock#HEADER_SIZE} bytes of the
- * buffer are included.
- * @return the byte buffer containing the given header (optionally) and the
- * decompressed data
* @throws IOException
*/
protected void decompress(byte[] dest, int destOffset,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java Tue Oct 25 06:57:10 2011
@@ -52,7 +52,7 @@ import org.apache.hadoop.io.WritableUtil
* Examples of how to use the block index writer can be found in
* {@link CompoundBloomFilterWriter} and {@link HFileWriterV2}. Examples of how
* to use the reader can be found in {@link HFileReaderV2} and
- * {@link TestHFileBlockIndex}.
+ * TestHFileBlockIndex.
*/
public class HFileBlockIndex {
@@ -62,7 +62,7 @@ public class HFileBlockIndex {
/**
* The maximum size guideline for index blocks (both leaf, intermediate, and
- * root). If not specified, {@link #DEFAULT_MAX_CHUNK_SIZE} is used.
+ * root). If not specified, <code>DEFAULT_MAX_CHUNK_SIZE</code> is used.
*/
public static final String MAX_CHUNK_SIZE_KEY = "hfile.index.block.max.size";
@@ -913,8 +913,6 @@ public class HFileBlockIndex {
* blocks, so the non-root index format is used.
*
* @param out
- * @param position The beginning offset of the inline block in the file not
- * include the header.
*/
@Override
public void writeInlineBlock(DataOutput out) throws IOException {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java Tue Oct 25 06:57:10 2011
@@ -55,7 +55,6 @@ public class HFileReaderV1 extends Abstr
* stream.
* @param size Length of the stream.
* @param cacheConf cache references and configuration
- * @throws IOException
*/
public HFileReaderV1(Path path, FixedFileTrailer trailer,
final FSDataInputStream fsdis, final long size,
@@ -80,7 +79,7 @@ public class HFileReaderV1 extends Abstr
* Read in the index and file info.
*
* @return A map of fileinfo data.
- * @see {@link Writer#appendFileInfo(byte[], byte[])}.
+ * @see Writer#appendFileInfo(byte[], byte[])
* @throws IOException
*/
@Override
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java Tue Oct 25 06:57:10 2011
@@ -207,7 +207,8 @@ public class HFileReaderV2 extends Abstr
* Read in a file block.
*
* @param dataBlockOffset offset to read.
- * @param onDiskSize size of the block
+ * @param onDiskBlockSize size of the block
+ * @param cacheBlock
* @param pread Use positional read instead of seek+read (positional is better
* doing random reads whereas seek+read is better scanning).
* @param isCompaction is this block being read as part of a compaction
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java Tue Oct 25 06:57:10 2011
@@ -56,7 +56,6 @@ public interface InlineBlockWriter {
* @param offset the offset of the block in the stream
* @param onDiskSize the on-disk size of the block
* @param uncompressedSize the uncompressed size of the block
- * @param rawSize
*/
void blockWritten(long offset, int onDiskSize, int uncompressedSize);
@@ -69,5 +68,4 @@ public interface InlineBlockWriter {
* @return true if inline blocks produced by this writer should be cached
*/
boolean cacheOnWrite();
-
-}
+}
\ No newline at end of file
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java Tue Oct 25 06:57:10 2011
@@ -103,7 +103,7 @@ public class SlabCache implements SlabIt
* The second list is blocksize of the slabs in bytes. (E.g. the slab holds
* blocks of this size).
*
- * @param Configuration file.
+ * @param conf Configuration file.
*/
public void addSlabByConf(Configuration conf) {
// Proportions we allocate to each slab of the total size.
@@ -229,7 +229,8 @@ public class SlabCache implements SlabIt
/**
* Get the buffer of the block with the specified name.
*
- * @param blockName block name
+ * @param key
+ * @param caching
* @return buffer of specified block name, or null if not in cache
*/
public Cacheable getBlock(String key, boolean caching) {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java Tue Oct 25 06:57:10 2011
@@ -48,7 +48,7 @@ public interface Delayable {
/**
* Signal that the RPC server is now allowed to send the response.
* @param result The value to return to the caller. If the corresponding
- * {@link #delayResponse(boolean)} specified that the return value should
+ * delay response specified that the return value should
* not be delayed, this parameter must be null.
* @throws IOException
*/
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java Tue Oct 25 06:57:10 2011
@@ -134,9 +134,9 @@ public abstract class HBaseServer implem
}
/** Returns the server instance called under or null. May be called under
- * {@link #call(Class, Writable, long)} implementations, and under {@link Writable}
- * methods of paramters and return values. Permits applications to access
- * the server context.
+ * {@link #call(Class, Writable, long, MonitoredRPCHandler)} implementations,
+ * and under {@link Writable} methods of paramters and return values.
+ * Permits applications to access the server context.
* @return HBaseServer
*/
public static RpcServer get() {
@@ -860,7 +860,6 @@ public abstract class HBaseServer implem
// Processes one response. Returns true if there are no more pending
// data for this channel.
//
- @SuppressWarnings({"ConstantConditions"})
private boolean processResponse(final LinkedList<Call> responseQueue,
boolean inHandler) throws IOException {
boolean error = true;
@@ -1280,14 +1279,15 @@ public abstract class HBaseServer implem
}
+
+ private Function<Writable,Integer> qosFunction = null;
+
/**
* Gets the QOS level for this call. If it is higher than the highPriorityLevel and there
* are priorityHandlers available it will be processed in it's own thread set.
*
- * @param param
- * @return priority, higher is better
+ * @param newFunc
*/
- private Function<Writable,Integer> qosFunction = null;
@Override
public void setQosFunction(Function<Writable, Integer> newFunc) {
qosFunction = newFunc;
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/mapreduce/hadoopbackport/InputSampler.java Tue Oct 25 06:57:10 2011
@@ -159,7 +159,7 @@ public class InputSampler<K,V> extends C
* native Hadoop ones (We'll throw a ClassNotFoundException if end up in
* here when we should be using native hadoop TotalOrderPartitioner).
* @param job
- * @return
+ * @return Context
* @throws IOException
*/
public static TaskAttemptContext getTaskAttemptContext(final Job job)
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/HMaster.java Tue Oct 25 06:57:10 2011
@@ -276,7 +276,7 @@ implements HMasterInterface, HMasterRegi
* Main processing loop for the HMaster.
* <ol>
* <li>Block until becoming active master
- * <li>Finish initialization via {@link #finishInitialization()}
+ * <li>Finish initialization via finishInitialization(MonitoredTask)
* <li>Enter loop until we are stopped
* <li>Stop services and perform cleanup once stopped
* </ol>
@@ -856,7 +856,7 @@ implements HMasterInterface, HMasterRegi
}
/**
- * Switch for the background {@link CatalogJanitor} thread.
+ * Switch for the background CatalogJanitor thread.
* Used for testing. The thread will continue to run. It will just be a noop
* if disabled.
* @param b If false, the catalog janitor won't do anything.
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java Tue Oct 25 06:57:10 2011
@@ -61,7 +61,7 @@ public interface LoadBalancer extends Co
/**
* Perform the major balance operation
* @param clusterState
- * @return
+ * @return List of plans
*/
public List<RegionPlan> balanceCluster(Map<ServerName, List<HRegionInfo>> clusterState);
@@ -69,7 +69,7 @@ public interface LoadBalancer extends Co
* Perform a Round Robin assignment of regions.
* @param regions
* @param servers
- * @return
+ * @return Map of servername to regioninfos
*/
public Map<ServerName, List<HRegionInfo>> roundRobinAssignment(List<HRegionInfo> regions, List<ServerName> servers);
@@ -77,7 +77,7 @@ public interface LoadBalancer extends Co
* Assign regions to the previously hosting region server
* @param regions
* @param servers
- * @return
+ * @return List of plans
*/
public Map<ServerName, List<HRegionInfo>> retainAssignment(Map<HRegionInfo, ServerName> regions, List<ServerName> servers);
@@ -85,14 +85,14 @@ public interface LoadBalancer extends Co
* Sync assign a region
* @param regions
* @param servers
- * @return
+ * @return Map regioninfos to servernames
*/
public Map<HRegionInfo, ServerName> immediateAssignment(List<HRegionInfo> regions, List<ServerName> servers);
/**
* Get a random region server from the list
* @param servers
- * @return
+ * @return Servername
*/
public ServerName randomAssignment(List<ServerName> servers);
}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java Tue Oct 25 06:57:10 2011
@@ -31,7 +31,7 @@ public class LoadBalancerFactory {
/**
* Create a loadblanacer from the given conf.
* @param conf
- * @return
+ * @return A {@link LoadBalancer}
*/
public static LoadBalancer getLoadBalancer(Configuration conf) {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java Tue Oct 25 06:57:10 2011
@@ -164,15 +164,13 @@ public class MasterFileSystem {
/**
* @return HBase root dir.
- * @throws IOException
*/
public Path getRootDir() {
return this.rootdir;
}
/**
- * Returns the unique identifier generated for this cluster
- * @return
+ * @return The unique identifier generated for this cluster
*/
public String getClusterId() {
return clusterId;
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java Tue Oct 25 06:57:10 2011
@@ -243,7 +243,7 @@ public class ServerManager {
}
/**
- * @param serverName
+ * @param address
* @return HServerLoad if serverName is known else null
* @deprecated Use {@link #getLoad(HServerAddress)}
*/
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java Tue Oct 25 06:57:10 2011
@@ -68,8 +68,8 @@ import static org.apache.hadoop.hbase.ma
* znode is created under /hbase/splitlog. SplitLogWorkers race to grab a task.
*
* SplitLogManager monitors the task znodes that it creates using the
- * {@link #timeoutMonitor} thread. If a task's progress is slow then
- * {@link #resubmit(String, boolean)} will take away the task from the owner
+ * timeoutMonitor thread. If a task's progress is slow then
+ * resubmit(String, boolean) will take away the task from the owner
* {@link SplitLogWorker} and the task will be
* upforgrabs again. When the task is done then the task's znode is deleted by
* SplitLogManager.
@@ -122,8 +122,6 @@ public class SplitLogManager extends Zoo
* @param conf
* @param stopper
* @param serverName
- * @param services
- * @param service
*/
public SplitLogManager(ZooKeeperWatcher zkw, final Configuration conf,
Stoppable stopper, String serverName) {
@@ -205,7 +203,7 @@ public class SplitLogManager extends Zoo
* @throws IOException
* if there was an error while splitting any log file
* @return cumulative size of the logfiles split
- * @throws KeeperException
+ * @throws IOException
*/
public long splitLogDistributed(final Path logDir) throws IOException {
List<Path> logDirs = new ArrayList<Path>();
@@ -218,8 +216,7 @@ public class SplitLogManager extends Zoo
* available worker region server. This method must only be called after the
* region servers have been brought online.
*
- * @param logDir
- * the log directory encoded with a region server name
+ * @param logDirs
* @throws IOException
* if there was an error while splitting any log file
* @return cumulative size of the logfiles split
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java Tue Oct 25 06:57:10 2011
@@ -56,7 +56,7 @@ public interface MonitoredTask extends C
/**
* Public exposure of Object.clone() in order to allow clients to easily
* capture current state.
- * @returns a copy of the object whose references will not change
+ * @return a copy of the object whose references will not change
*/
public abstract MonitoredTask clone();
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java Tue Oct 25 06:57:10 2011
@@ -48,8 +48,9 @@ public abstract class ThreadMonitoring {
/**
* Print all of the thread's information and stack traces.
*
- * @param stream the stream to
- *
+ * @param sb
+ * @param info
+ * @param indent
*/
public static void appendThreadInfo(StringBuilder sb,
ThreadInfo info,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java Tue Oct 25 06:57:10 2011
@@ -295,7 +295,7 @@ public class HRegion implements HeapSize
/**
* HRegion constructor. his constructor should only be used for testing and
* extensions. Instances of HRegion should be instantiated with the
- * {@link HRegion#newHRegion(Path, HLog, FileSystem, Configuration, org.apache.hadoop.hbase.HRegionInfo, FlushRequester)} method.
+ * {@link HRegion#newHRegion(Path, HLog, FileSystem, Configuration, HRegionInfo, HTableDescriptor, RegionServerServices)} method.
*
*
* @param tableDir qualified path of directory where region should be located,
@@ -313,7 +313,7 @@ public class HRegion implements HeapSize
* is new), then read them from the supplied path.
* @param rsServices reference to {@link RegionServerServices} or null
*
- * @see HRegion#newHRegion(Path, HLog, FileSystem, Configuration, org.apache.hadoop.hbase.HRegionInfo, FlushRequester)
+ * @see HRegion#newHRegion(Path, HLog, FileSystem, Configuration, HRegionInfo, HTableDescriptor, RegionServerServices)
*/
public HRegion(Path tableDir, HLog log, FileSystem fs, Configuration conf,
HRegionInfo regionInfo, final HTableDescriptor htd,
@@ -1110,7 +1110,7 @@ public class HRegion implements HeapSize
* @param status
* @return true if the region needs compacting
* @throws IOException
- * @see #internalFlushcache()
+ * @see #internalFlushcache(MonitoredTask)
*/
protected boolean internalFlushcache(
final HLog wal, final long myseqid, MonitoredTask status)
@@ -2534,7 +2534,7 @@ public class HRegion implements HeapSize
/**
* Release the row lock!
- * @param lockid The lock ID to release.
+ * @param lockId The lock ID to release.
*/
public void releaseRowLock(final Integer lockId) {
HashedBytes rowKey = lockIds.remove(lockId);
@@ -2944,13 +2944,14 @@ public class HRegion implements HeapSize
/**
* Open a Region.
- * @param info Info for region to be opened.
+ * @param info Info for region to be opened
+ * @param htd
* @param wal HLog for region to use. This method will call
* HLog#setSequenceNumber(long) passing the result of the call to
* HRegion#getMinSequenceId() to ensure the log id is properly kept
* up. HRegionStore does this every time it opens a new region.
* @param conf
- * @param flusher An interface we can request flushes against.
+ * @param rsServices An interface we can request flushes against.
* @param reporter An interface we can report progress against.
* @return new HRegion
*
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java Tue Oct 25 06:57:10 2011
@@ -210,8 +210,8 @@ public class RegionCoprocessorHost
* {@link org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost#postOpen()} are such hooks.
*
* See also {@link org.apache.hadoop.hbase.master.MasterCoprocessorHost#handleCoprocessorThrowable()}
- * @param env: The coprocessor that threw the exception.
- * @param e: The exception that was thrown.
+ * @param env The coprocessor that threw the exception.
+ * @param e The exception that was thrown.
*/
private void handleCoprocessorThrowableNoRethrow(
final CoprocessorEnvironment env, final Throwable e) {
@@ -566,7 +566,6 @@ public class RegionCoprocessorHost
/**
* @param get the Get request
* @param results the result set
- * @return the possibly transformed result set to use
* @exception IOException Exception
*/
public void postGet(final Get get, final List<KeyValue> results)
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java Tue Oct 25 06:57:10 2011
@@ -548,8 +548,7 @@ public class SplitLogWorker extends ZooK
* acquired by a {@link SplitLogWorker}. Since there isn't a water-tight
* guarantee that two workers will not be executing the same task therefore it
* is better to have workers prepare the task and then have the
- * {@link SplitLogManager} commit the work in
- * {@link SplitLogManager.TaskFinisher}
+ * {@link SplitLogManager} commit the work in SplitLogManager.TaskFinisher
*/
static public interface TaskExecutor {
static public enum Status {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java Tue Oct 25 06:57:10 2011
@@ -61,7 +61,7 @@ import com.google.common.util.concurrent
/**
* Executes region split as a "transaction". Call {@link #prepare()} to setup
* the transaction, {@link #execute(Server, RegionServerServices)} to run the
- * transaction and {@link #rollback(OnlineRegions)} to cleanup if execute fails.
+ * transaction and {@link #rollback(Server, RegionServerServices)} to cleanup if execute fails.
*
* <p>Here is an example of how you would use this class:
* <pre>
@@ -145,9 +145,6 @@ public class SplitTransaction {
/**
* Constructor
- * @param services So we can online new regions. If null, we'll skip onlining
- * (Useful testing).
- * @param c Configuration to use running split
* @param r Region to split
* @param splitrow Row to split around
*/
@@ -430,8 +427,7 @@ public class SplitTransaction {
* @param services Used to online/offline regions.
* @throws IOException If thrown, transaction failed. Call {@link #rollback(Server, RegionServerServices)}
* @return Regions created
- * @throws KeeperException
- * @throws NodeExistsException
+ * @throws IOException
* @see #rollback(Server, RegionServerServices)
*/
public PairOfSameType<HRegion> execute(final Server server,
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java Tue Oct 25 06:57:10 2011
@@ -67,7 +67,7 @@ import com.google.common.collect.Orderin
/**
* A Store data file. Stores usually have one or more of these files. They
* are produced by flushing the memstore to disk. To
- * create, call {@link #createWriter(FileSystem, Path, int, Configuration)}
+ * create, call {@link #createWriter(FileSystem, Path, int, Configuration, CacheConfig)}
* and append data. Be sure to add any metadata before calling close on the
* Writer (Use the appendMetadata convenience methods). On close, a StoreFile
* is sitting in the Filesystem. To refer to it, create a StoreFile instance
@@ -392,7 +392,7 @@ public class StoreFile {
* helper function to compute HDFS blocks distribution of a given file.
* For reference file, it is an estimate
* @param fs The FileSystem
- * @param o The path of the file
+ * @param p The path of the file
* @return HDFS blocks distribution
*/
static public HDFSBlocksDistribution computeHDFSBlockDistribution(
@@ -518,7 +518,6 @@ public class StoreFile {
/**
* @return Current reader. Must call createReader first else returns null.
- * @throws IOException
* @see #createReader()
*/
public Reader getReader() {
@@ -526,7 +525,7 @@ public class StoreFile {
}
/**
- * @param b
+ * @param evictOnClose
* @throws IOException
*/
public synchronized void closeReader(boolean evictOnClose)
@@ -819,7 +818,6 @@ public class StoreFile {
* If the timeRangeTracker is not set,
* update TimeRangeTracker to include the timestamp of this key
* @param kv
- * @throws IOException
*/
public void includeInTimeRangeTracker(final KeyValue kv) {
if (!isTimeRangeTrackerSet) {
@@ -1081,7 +1079,7 @@ public class StoreFile {
/**
* A method for checking Bloom filters. Called directly from
- * {@link StoreFileScanner} in case of a multi-column query.
+ * StoreFileScanner in case of a multi-column query.
*
* @param row
* @param rowOffset
@@ -1089,7 +1087,7 @@ public class StoreFile {
* @param col
* @param colOffset
* @param colLen
- * @return
+ * @return True if passes
*/
public boolean passesBloomFilter(byte[] row, int rowOffset, int rowLen,
byte[] col, int colOffset, int colLen) {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java Tue Oct 25 06:57:10 2011
@@ -1284,7 +1284,7 @@ public class HLog implements Syncable {
*
* In this method, by removing the entry in lastSeqWritten for the region
* being flushed we ensure that the next edit inserted in this region will be
- * correctly recorded in {@link #append(HRegionInfo, HLogKey, WALEdit)}. The
+ * correctly recorded in {@link #append(HRegionInfo, byte[], WALEdit, long, HTableDescriptor)} The
* lsn of the earliest in-memory lsn - which is now in the memstore snapshot -
* is saved temporarily in the lastSeqWritten map while the flush is active.
*
@@ -1292,7 +1292,7 @@ public class HLog implements Syncable {
* {@link #completeCacheFlush(byte[], byte[], long, boolean)} (byte[],
* byte[], long)}
* @see #completeCacheFlush(byte[], byte[], long, boolean)
- * @see #abortCacheFlush()
+ * @see #abortCacheFlush(byte[])
*/
public long startCacheFlush(final byte[] encodedRegionName) {
this.cacheFlushLock.lock();
@@ -1594,29 +1594,6 @@ public class HLog implements Syncable {
System.err.println(" For example: HLog --split hdfs://example.com:9000/hbase/.logs/DIR");
}
- private static void dump(final Configuration conf, final Path p)
- throws IOException {
- FileSystem fs = FileSystem.get(conf);
- if (!fs.exists(p)) {
- throw new FileNotFoundException(p.toString());
- }
- if (!fs.isFile(p)) {
- throw new IOException(p + " is not a file");
- }
- Reader log = getReader(fs, p, conf);
- try {
- int count = 0;
- HLog.Entry entry;
- while ((entry = log.next()) != null) {
- System.out.println("#" + count + ", pos=" + log.getPosition() + " " +
- entry.toString());
- count++;
- }
- } finally {
- log.close();
- }
- }
-
private static void split(final Configuration conf, final Path p)
throws IOException {
FileSystem fs = FileSystem.get(conf);
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java Tue Oct 25 06:57:10 2011
@@ -71,7 +71,7 @@ public class HLogKey implements Writable
* @param tablename - name of table
* @param logSeqNum - log sequence number
* @param now Time at which this edit was written.
- * @param UUID of the cluster (used in Replication)
+ * @param clusterId of the cluster (used in Replication)
*/
public HLogKey(final byte [] encodedRegionName, final byte [] tablename,
long logSeqNum, final long now, UUID clusterId) {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java Tue Oct 25 06:57:10 2011
@@ -37,8 +37,7 @@ import java.nio.ByteBuffer;
* When creating the filter, the sender can choose its desired point in a
* trade-off between the false positive rate and the size.
*
- * @see {@link BloomFilterWriter} for the ability to add elements to a Bloom
- * filter
+ * @see BloomFilterWriter for the ability to add elements to a Bloom filter
*/
public interface BloomFilter extends BloomFilterBase {
@@ -59,5 +58,4 @@ public interface BloomFilter extends Blo
* and thus allows a null byte buffer to be passed to contains()
*/
boolean supportsAutoLoading();
-
-}
+}
\ No newline at end of file
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java Tue Oct 25 06:57:10 2011
@@ -127,12 +127,11 @@ public final class BloomFilterFactory {
* {@link org.apache.hadoop.hbase.regionserver.StoreFile} writing.
*
* @param conf
+ * @param cacheConf
* @param bloomType
* @param maxKeys an estimate of the number of keys we expect to insert.
* Irrelevant if compound Bloom filters are enabled.
* @param writer the HFile writer
- * @param comparator the comparator to use for compound Bloom filters. This
- * has no effect if creating single-chunk version 1 Bloom filters.
* @return the new Bloom filter, or null in case Bloom filters are disabled
* or when failed to create one.
*/
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java?rev=1188539&r1=1188538&r2=1188539&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java Tue Oct 25 06:57:10 2011
@@ -164,8 +164,7 @@ public class ByteBloomFilter implements
* @param bitSize
* @param errorRate
* @return maximum number of keys that can be inserted into the Bloom filter
- * @see {@link #computeMaxKeys(long, double, int)} for a more precise
- * estimate
+ * @see #computeMaxKeys(long, double, int) for a more precise estimate
*/
public static long idealMaxKeys(long bitSize, double errorRate) {
// The reason we need to use floor here is that otherwise we might put
@@ -227,7 +226,7 @@ public class ByteBloomFilter implements
*
* @param bitSize
* @param foldFactor
- * @return
+ * @return Foldable byte size
*/
public static int computeFoldableByteSize(long bitSize, int foldFactor) {
long byteSizeLong = (bitSize + 7) / 8;