You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2011/10/29 01:37:21 UTC
svn commit: r1190672 - in /hbase/branches/0.92: ./
src/main/java/org/apache/hadoop/hbase/
src/main/java/org/apache/hadoop/hbase/client/
src/main/java/org/apache/hadoop/hbase/io/
src/main/java/org/apache/hadoop/hbase/ipc/ src/main/java/org/apache/hadoop...
Author: stack
Date: Fri Oct 28 23:37:20 2011
New Revision: 1190672
URL: http://svn.apache.org/viewvc?rev=1190672&view=rev
Log:
HBASE-4436 Remove trivial 0.90 deprecated code from 0.92 and trunk.
Modified:
hbase/branches/0.92/CHANGES.txt
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Delete.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Get.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTable.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Put.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Result.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java
hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java
hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TestSerialization.java
hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java
hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java
Modified: hbase/branches/0.92/CHANGES.txt
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/CHANGES.txt?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/CHANGES.txt (original)
+++ hbase/branches/0.92/CHANGES.txt Fri Oct 28 23:37:20 2011
@@ -676,6 +676,8 @@ Release 0.92.0 - Unreleased
HBASE-4296 Deprecate HTable[Interface].getRowOrBefore(...) (Lars Hofhansl)
HBASE-2195 Support cyclic replication (Lars Hofhansl)
HBASE-2196 Support more than one slave cluster (Lars Hofhansl)
+ HBASE-4435 Remove trivial 0.90 deprecated code from 0.92 and trunk
+ (Jonathan Hsieh)
NEW FEATURES
HBASE-2001 Coprocessors: Colocate user code with regions (Mingjie Lai via
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java Fri Oct 28 23:37:20 2011
@@ -55,22 +55,6 @@ public class HColumnDescriptor implement
// Version 8 -- reintroduction of bloom filters, changed from boolean to enum
private static final byte COLUMN_DESCRIPTOR_VERSION = (byte)8;
- /**
- * The type of compression.
- * @see org.apache.hadoop.io.SequenceFile.Writer
- * @deprecated Compression now means which compression library
- * rather than 'what' to compress.
- */
- @Deprecated
- public static enum CompressionType {
- /** Do not compress records. */
- NONE,
- /** Compress values only, each separately. */
- RECORD,
- /** Compress sequences of records together in blocks. */
- BLOCK
- }
-
public static final String COMPRESSION = "COMPRESSION";
public static final String COMPRESSION_COMPACT = "COMPRESSION_COMPACT";
public static final String BLOCKCACHE = "BLOCKCACHE";
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/KeyValue.java Fri Oct 28 23:37:20 2011
@@ -1654,20 +1654,6 @@ public class KeyValue implements Writabl
}
/**
- * @param row - row key (arbitrary byte array)
- * @param c column - {@link #parseColumn(byte[])} is called to split
- * the column.
- * @param ts - timestamp
- * @return First possible key on passed <code>row</code>, column and timestamp
- * @deprecated
- */
- public static KeyValue createFirstOnRow(final byte [] row, final byte [] c,
- final long ts) {
- byte [][] split = parseColumn(c);
- return new KeyValue(row, split[0], split[1], ts, Type.Maximum);
- }
-
- /**
* Create a KeyValue for the specified row, family and qualifier that would be
* smaller than all other possible KeyValues that have the same row,family,qualifier.
* Used for seeking.
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Delete.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Delete.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Delete.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Delete.java Fri Oct 28 23:37:20 2011
@@ -280,32 +280,4 @@ public class Delete extends Mutation
}
writeAttributes(out);
}
-
- /**
- * Delete all versions of the specified column, given in
- * <code>family:qualifier</code> notation, and with a timestamp less than
- * or equal to the specified timestamp.
- * @param column colon-delimited family and qualifier
- * @param timestamp maximum version timestamp
- * @deprecated use {@link #deleteColumn(byte[], byte[], long)} instead
- * @return this for invocation chaining
- */
- public Delete deleteColumns(byte [] column, long timestamp) {
- byte [][] parts = KeyValue.parseColumn(column);
- this.deleteColumns(parts[0], parts[1], timestamp);
- return this;
- }
-
- /**
- * Delete the latest version of the specified column, given in
- * <code>family:qualifier</code> notation.
- * @param column colon-delimited family and qualifier
- * @deprecated use {@link #deleteColumn(byte[], byte[])} instead
- * @return this for invocation chaining
- */
- public Delete deleteColumn(byte [] column) {
- byte [][] parts = KeyValue.parseColumn(column);
- this.deleteColumn(parts[0], parts[1], HConstants.LATEST_TIMESTAMP);
- return this;
- }
}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Get.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Get.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Get.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Get.java Fri Oct 28 23:37:20 2011
@@ -457,41 +457,4 @@ public class Get extends OperationWithAt
throw new RuntimeException("Can't find class " + className);
}
}
-
- /**
- * Adds an array of columns specified the old format, family:qualifier.
- * <p>
- * Overrides previous calls to addFamily for any families in the input.
- * @param columns array of columns, formatted as <pre>family:qualifier</pre>
- * @deprecated issue multiple {@link #addColumn(byte[], byte[])} instead
- * @return this for invocation chaining
- */
- @SuppressWarnings({"deprecation"})
- public Get addColumns(byte [][] columns) {
- if (columns == null) return this;
- for (byte[] column : columns) {
- try {
- addColumn(column);
- } catch (Exception ignored) {
- }
- }
- return this;
- }
-
- /**
- *
- * @param column Old format column.
- * @return This.
- * @deprecated use {@link #addColumn(byte[], byte[])} instead
- */
- public Get addColumn(final byte [] column) {
- if (column == null) return this;
- byte [][] split = KeyValue.parseColumn(column);
- if (split.length > 1 && split[1] != null && split[1].length > 0) {
- addColumn(split[0], split[1]);
- } else {
- addFamily(split[0]);
- }
- return this;
- }
}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java Fri Oct 28 23:37:20 2011
@@ -40,7 +40,6 @@ import org.apache.hadoop.hbase.HColumnDe
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.HRegionLocation;
-import org.apache.hadoop.hbase.HServerAddress;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.MasterNotRunningException;
import org.apache.hadoop.hbase.NotServingRegionException;
@@ -956,22 +955,6 @@ public class HBaseAdmin implements Abort
* Asynchronous operation.
*
* @param tableName name of table
- * @param columnName name of column to be modified
- * @param descriptor new column descriptor to use
- * @throws IOException if a remote or network exception occurs
- * @deprecated The <code>columnName</code> is redundant. Use {@link #addColumn(String, HColumnDescriptor)}
- */
- public void modifyColumn(final String tableName, final String columnName,
- HColumnDescriptor descriptor)
- throws IOException {
- modifyColumn(tableName, descriptor);
- }
-
- /**
- * Modify an existing column family on a table.
- * Asynchronous operation.
- *
- * @param tableName name of table
* @param descriptor new column descriptor to use
* @throws IOException if a remote or network exception occurs
*/
@@ -985,22 +968,6 @@ public class HBaseAdmin implements Abort
* Asynchronous operation.
*
* @param tableName name of table
- * @param columnName name of column to be modified
- * @param descriptor new column descriptor to use
- * @throws IOException if a remote or network exception occurs
- * @deprecated The <code>columnName</code> is redundant. Use {@link #modifyColumn(byte[], HColumnDescriptor)}
- */
- public void modifyColumn(final byte [] tableName, final byte [] columnName,
- HColumnDescriptor descriptor)
- throws IOException {
- modifyColumn(tableName, descriptor);
- }
-
- /**
- * Modify an existing column family on a table.
- * Asynchronous operation.
- *
- * @param tableName name of table
* @param descriptor new column descriptor to use
* @throws IOException if a remote or network exception occurs
*/
@@ -1309,24 +1276,6 @@ public class HBaseAdmin implements Abort
}
/**
- * Tries to assign a region. Region could be reassigned to the same server.
- *
- * @param regionName
- * Region name to assign.
- * @param force
- * True to force assign.
- * @throws MasterNotRunningException
- * @throws ZooKeeperConnectionException
- * @throws IOException
- * @deprecated The <code>force</code> is unused.Use {@link #assign(byte[])}
- */
- public void assign(final byte[] regionName, final boolean force)
- throws MasterNotRunningException, ZooKeeperConnectionException,
- IOException {
- getMaster().assign(regionName, force);
- }
-
- /**
* @param regionName
* Region name to assign.
* @throws MasterNotRunningException
@@ -1525,18 +1474,6 @@ public class HBaseAdmin implements Abort
}
/**
- * Stop the designated regionserver.
- * @throws IOException if a remote or network exception occurs
- * @deprecated Use {@link #stopRegionServer(String)}
- */
- public synchronized void stopRegionServer(final HServerAddress hsa)
- throws IOException {
- HRegionInterface rs =
- this.connection.getHRegionConnection(hsa);
- rs.stop("Called by admin client " + this.connection.toString());
- }
-
- /**
* Stop the designated regionserver
* @param hostnamePort Hostname and port delimited by a <code>:</code> as in
* <code>example.org:1234</code>
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnection.java Fri Oct 28 23:37:20 2011
@@ -283,7 +283,7 @@ public interface HConnection extends Abo
* @throws IOException if there are problems talking to META. Per-item
* exceptions are stored in the results array.
*/
- public void processBatch(List<Row> actions, final byte[] tableName,
+ public void processBatch(List<? extends Row> actions, final byte[] tableName,
ExecutorService pool, Object[] results)
throws IOException, InterruptedException;
@@ -328,20 +328,6 @@ public interface HConnection extends Abo
final Batch.Callback<R> callback) throws IOException, Throwable;
/**
- * Process a batch of Puts.
- *
- * @param list The collection of actions. The list is mutated: all successful Puts
- * are removed from the list.
- * @param tableName Name of the hbase table
- * @param pool Thread pool for parallel execution
- * @throws IOException
- * @deprecated Use HConnectionManager::processBatch instead.
- */
- public void processBatchOfPuts(List<Put> list,
- final byte[] tableName, ExecutorService pool)
- throws IOException;
-
- /**
* Enable or disable region cache prefetch for the table. It will be
* applied for the given table's all HTable instances within this
* connection. By default, the cache prefetch is enabled.
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java Fri Oct 28 23:37:20 2011
@@ -1301,7 +1301,7 @@ public class HConnectionManager {
};
}
- public void processBatch(List<Row> list,
+ public void processBatch(List<? extends Row> list,
final byte[] tableName,
ExecutorService pool,
Object[] results) throws IOException, InterruptedException {
@@ -1549,31 +1549,6 @@ public class HConnectionManager {
}
}
- /**
- * @deprecated Use HConnectionManager::processBatch instead.
- */
- public void processBatchOfPuts(List<Put> list,
- final byte[] tableName,
- ExecutorService pool) throws IOException {
- Object[] results = new Object[list.size()];
- try {
- processBatch((List) list, tableName, pool, results);
- } catch (InterruptedException e) {
- throw new IOException(e);
- } finally {
-
- // mutate list so that it is empty for complete success, or contains only failed records
- // results are returned in the same order as the requests in list
- // walk the list backwards, so we can remove from list without impacting the indexes of earlier members
- for (int i = results.length - 1; i>=0; i--) {
- if (results[i] instanceof Result) {
- // successful Puts are removed from the list here.
- list.remove(i);
- }
- }
- }
- }
-
private Throwable translateException(Throwable t) throws IOException {
if (t instanceof UndeclaredThrowableException) {
t = t.getCause();
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTable.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTable.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTable.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/HTable.java Fri Oct 28 23:37:20 2011
@@ -852,7 +852,23 @@ public class HTable implements HTableInt
@Override
public void flushCommits() throws IOException {
try {
- this.connection.processBatchOfPuts(writeBuffer, tableName, pool);
+ Object[] results = new Object[writeBuffer.size()];
+ try {
+ this.connection.processBatch(writeBuffer, tableName, pool, results);
+ } catch (InterruptedException e) {
+ throw new IOException(e);
+ } finally {
+ // mutate list so that it is empty for complete success, or contains
+ // only failed records results are returned in the same order as the
+ // requests in list walk the list backwards, so we can remove from list
+ // without impacting the indexes of earlier members
+ for (int i = results.length - 1; i>=0; i--) {
+ if (results[i] instanceof Result) {
+ // successful Puts are removed from the list here.
+ writeBuffer.remove(i);
+ }
+ }
+ }
} finally {
if (clearBufferOnFail) {
writeBuffer.clear();
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Put.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Put.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Put.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Put.java Fri Oct 28 23:37:20 2011
@@ -424,19 +424,4 @@ public class Put extends Mutation
}
writeAttributes(out);
}
-
- /**
- * Add the specified column and value, with the specified timestamp as
- * its version to this Put operation.
- * @param column Old style column name with family and qualifier put together
- * with a colon.
- * @param ts version timestamp
- * @param value column value
- * @deprecated use {@link #add(byte[], byte[], long, byte[])} instead
- * @return true
- */
- public Put add(byte [] column, long ts, byte [] value) {
- byte [][] parts = KeyValue.parseColumn(column);
- return add(parts[0], parts[1], ts, value);
- }
}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Result.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Result.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Result.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/client/Result.java Fri Oct 28 23:37:20 2011
@@ -20,18 +20,6 @@
package org.apache.hadoop.hbase.client;
-import org.apache.hadoop.hbase.HConstants;
-import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.KeyValue;
-import org.apache.hadoop.hbase.KeyValue.SplitKeyValue;
-import org.apache.hadoop.hbase.ServerName;
-import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
-import org.apache.hadoop.hbase.io.WritableWithSize;
-import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.Pair;
-import org.apache.hadoop.hbase.util.Writables;
-import org.apache.hadoop.io.Writable;
-
import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
@@ -43,6 +31,13 @@ import java.util.Map;
import java.util.NavigableMap;
import java.util.TreeMap;
+import org.apache.hadoop.hbase.KeyValue;
+import org.apache.hadoop.hbase.KeyValue.SplitKeyValue;
+import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
+import org.apache.hadoop.hbase.io.WritableWithSize;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.io.Writable;
+
/**
* Single row result of a {@link Get} or {@link Scan} query.<p>
*
@@ -169,19 +164,6 @@ public class Result implements Writable,
}
/**
- * Returns a sorted array of KeyValues in this Result.
- * <p>
- * Since HBase 0.20.5 this is equivalent to {@link #raw}. Use
- * {@link #raw} instead.
- *
- * @return sorted array of KeyValues
- * @deprecated
- */
- public KeyValue[] sorted() {
- return raw(); // side effect of loading this.kvs
- }
-
- /**
* Return the KeyValues for the specific column. The KeyValues are sorted in
* the {@link KeyValue#COMPARATOR} order. That implies the first entry in
* the list is the most recent column. If the query (Scan or Get) only
@@ -399,32 +381,6 @@ public class Result implements Writable,
return returnMap;
}
- private Map.Entry<Long,byte[]> getKeyValue(byte[] family, byte[] qualifier) {
- if(this.familyMap == null) {
- getMap();
- }
- if(isEmpty()) {
- return null;
- }
- NavigableMap<byte [], NavigableMap<Long, byte[]>> qualifierMap =
- familyMap.get(family);
- if(qualifierMap == null) {
- return null;
- }
- NavigableMap<Long, byte[]> versionMap =
- getVersionMap(qualifierMap, qualifier);
- if(versionMap == null) {
- return null;
- }
- return versionMap.firstEntry();
- }
-
- private NavigableMap<Long, byte[]> getVersionMap(
- NavigableMap<byte [], NavigableMap<Long, byte[]>> qualifierMap, byte [] qualifier) {
- return qualifier != null?
- qualifierMap.get(qualifier): qualifierMap.get(new byte[0]);
- }
-
/**
* Returns the value of the first column in the Result.
* @return value of the first column
@@ -675,8 +631,8 @@ public class Result implements Writable,
throw new Exception("This row doesn't have the same number of KVs: "
+ res1.toString() + " compared to " + res2.toString());
}
- KeyValue[] ourKVs = res1.sorted();
- KeyValue[] replicatedKVs = res2.sorted();
+ KeyValue[] ourKVs = res1.raw();
+ KeyValue[] replicatedKVs = res2.raw();
for (int i = 0; i < res1.size(); i++) {
if (!ourKVs[i].equals(replicatedKVs[i]) &&
!Bytes.equals(ourKVs[i].getValue(), replicatedKVs[i].getValue())) {
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java Fri Oct 28 23:37:20 2011
@@ -50,8 +50,6 @@ import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Delete;
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.Increment;
-import org.apache.hadoop.hbase.client.MultiPut;
-import org.apache.hadoop.hbase.client.MultiPutResponse;
import org.apache.hadoop.hbase.client.MultiAction;
import org.apache.hadoop.hbase.client.Action;
import org.apache.hadoop.hbase.client.MultiResponse;
@@ -202,9 +200,6 @@ public class HbaseObjectWritable impleme
addToMap(Delete [].class, code++);
- addToMap(MultiPut.class, code++);
- addToMap(MultiPutResponse.class, code++);
-
addToMap(HLog.Entry.class, code++);
addToMap(HLog.Entry[].class, code++);
addToMap(HLogKey.class, code++);
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java Fri Oct 28 23:37:20 2011
@@ -32,8 +32,6 @@ import org.apache.hadoop.hbase.client.De
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.Increment;
import org.apache.hadoop.hbase.client.MultiAction;
-import org.apache.hadoop.hbase.client.MultiPut;
-import org.apache.hadoop.hbase.client.MultiPutResponse;
import org.apache.hadoop.hbase.client.MultiResponse;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Result;
@@ -311,15 +309,6 @@ public interface HRegionInterface extend
public <R> MultiResponse multi(MultiAction<R> multi) throws IOException;
/**
- * Multi put for putting multiple regions worth of puts at once.
- *
- * @param puts the request
- * @return the reply
- * @throws IOException e
- */
- public MultiPutResponse multiPut(MultiPut puts) throws IOException;
-
- /**
* Bulk load an HFile into an open region
*/
public void bulkLoadHFile(String hfilePath, byte[] regionName, byte[] familyName)
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java Fri Oct 28 23:37:20 2011
@@ -84,8 +84,6 @@ import org.apache.hadoop.hbase.client.Ge
import org.apache.hadoop.hbase.client.HConnectionManager;
import org.apache.hadoop.hbase.client.Increment;
import org.apache.hadoop.hbase.client.MultiAction;
-import org.apache.hadoop.hbase.client.MultiPut;
-import org.apache.hadoop.hbase.client.MultiPutResponse;
import org.apache.hadoop.hbase.client.MultiResponse;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Result;
@@ -2990,25 +2988,6 @@ public class HRegionServer implements HR
}
/**
- * @deprecated Use HRegionServer.multi( MultiAction action) instead
- */
- @Override
- public MultiPutResponse multiPut(MultiPut puts) throws IOException {
- checkOpen();
- MultiPutResponse resp = new MultiPutResponse();
-
- // do each region as it's own.
- for (Map.Entry<byte[], List<Put>> e : puts.puts.entrySet()) {
- int result = put(e.getKey(), e.getValue());
- resp.addResult(e.getKey(), result);
-
- e.getValue().clear(); // clear some RAM
- }
-
- return resp;
- }
-
- /**
* Executes a single {@link org.apache.hadoop.hbase.ipc.CoprocessorProtocol}
* method using the registered protocol handlers.
* {@link CoprocessorProtocol} implementations must be registered per-region
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java Fri Oct 28 23:37:20 2011
@@ -44,7 +44,6 @@ import org.apache.hadoop.hbase.HBaseConf
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.HRegionInfo;
-import org.apache.hadoop.hbase.HServerAddress;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Delete;
@@ -316,7 +315,7 @@ public class ThriftServer {
get.addColumn(family, qualifier);
}
Result result = table.get(get);
- return ThriftUtilities.cellFromHBase(result.sorted());
+ return ThriftUtilities.cellFromHBase(result.raw());
} catch (IOException e) {
throw new IOError(e.getMessage());
}
@@ -344,7 +343,7 @@ public class ThriftServer {
get.addColumn(family, qualifier);
get.setMaxVersions(numVersions);
Result result = table.get(get);
- return ThriftUtilities.cellFromHBase(result.sorted());
+ return ThriftUtilities.cellFromHBase(result.raw());
} catch (IOException e) {
throw new IOError(e.getMessage());
}
@@ -376,7 +375,7 @@ public class ThriftServer {
get.setTimeRange(Long.MIN_VALUE, timestamp);
get.setMaxVersions(numVersions);
Result result = table.get(get);
- return ThriftUtilities.cellFromHBase(result.sorted());
+ return ThriftUtilities.cellFromHBase(result.raw());
} catch (IOException e) {
throw new IOError(e.getMessage());
}
Modified: hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java (original)
+++ hbase/branches/0.92/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java Fri Oct 28 23:37:20 2011
@@ -135,7 +135,7 @@ public class ThriftUtilities {
TRowResult result = new TRowResult();
result.row = ByteBuffer.wrap(result_.getRow());
result.columns = new TreeMap<ByteBuffer, TCell>();
- for(KeyValue kv : result_.sorted()) {
+ for(KeyValue kv : result_.raw()) {
result.columns.put(
ByteBuffer.wrap(KeyValue.makeColumn(kv.getFamily(),
kv.getQualifier())),
Modified: hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TestSerialization.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TestSerialization.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TestSerialization.java (original)
+++ hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TestSerialization.java Fri Oct 28 23:37:20 2011
@@ -355,10 +355,10 @@ public class TestSerialization {
Result deResult = (Result)Writables.getWritable(rb, new Result());
assertTrue("results are not equivalent, first key mismatch",
- result.sorted()[0].equals(deResult.sorted()[0]));
+ result.raw()[0].equals(deResult.raw()[0]));
assertTrue("results are not equivalent, second key mismatch",
- result.sorted()[1].equals(deResult.sorted()[1]));
+ result.raw()[1].equals(deResult.raw()[1]));
// Test empty Result
Result r = new Result();
@@ -394,9 +394,9 @@ public class TestSerialization {
// Call sorted() first
deResult = (Result)Writables.getWritable(rb, new Result());
assertTrue("results are not equivalent, first key mismatch",
- result.sorted()[0].equals(deResult.sorted()[0]));
+ result.raw()[0].equals(deResult.raw()[0]));
assertTrue("results are not equivalent, second key mismatch",
- result.sorted()[1].equals(deResult.sorted()[1]));
+ result.raw()[1].equals(deResult.raw()[1]));
// Call raw() first
deResult = (Result)Writables.getWritable(rb, new Result());
@@ -443,8 +443,8 @@ public class TestSerialization {
assertTrue(results.length == deResults.length);
for(int i=0;i<results.length;i++) {
- KeyValue [] keysA = results[i].sorted();
- KeyValue [] keysB = deResults[i].sorted();
+ KeyValue [] keysA = results[i].raw();
+ KeyValue [] keysB = deResults[i].raw();
assertTrue(keysA.length == keysB.length);
for(int j=0;j<keysA.length;j++) {
assertTrue("Expected equivalent keys but found:\n" +
Modified: hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java (original)
+++ hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/TimestampTestBase.java Fri Oct 28 23:37:20 2011
@@ -106,7 +106,7 @@ public class TimestampTestBase extends H
get.setMaxVersions(3);
Result result = incommon.get(get);
assertEquals(1, result.size());
- long time = Bytes.toLong(result.sorted()[0].getValue());
+ long time = Bytes.toLong(result.raw()[0].getValue());
assertEquals(time, currentTime);
}
@@ -135,7 +135,7 @@ public class TimestampTestBase extends H
get.addColumn(FAMILY_NAME, QUALIFIER_NAME);
get.setMaxVersions(tss.length);
Result result = incommon.get(get);
- KeyValue [] kvs = result.sorted();
+ KeyValue [] kvs = result.raw();
assertEquals(kvs.length, tss.length);
for(int i=0;i<kvs.length;i++) {
t = Bytes.toLong(kvs[i].getValue());
@@ -151,7 +151,7 @@ public class TimestampTestBase extends H
get.setTimeRange(0, maxStamp);
get.setMaxVersions(kvs.length - 1);
result = incommon.get(get);
- kvs = result.sorted();
+ kvs = result.raw();
assertEquals(kvs.length, tss.length - 1);
for(int i=1;i<kvs.length;i++) {
t = Bytes.toLong(kvs[i-1].getValue());
Modified: hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java (original)
+++ hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java Fri Oct 28 23:37:20 2011
@@ -1861,15 +1861,15 @@ public class TestFromClientSide {
result = scanner.next();
assertTrue("Expected 1 key but received " + result.size(),
result.size() == 1);
- assertTrue(Bytes.equals(result.sorted()[0].getRow(), ROWS[3]));
- assertTrue(Bytes.equals(result.sorted()[0].getValue(), VALUES[0]));
+ assertTrue(Bytes.equals(result.raw()[0].getRow(), ROWS[3]));
+ assertTrue(Bytes.equals(result.raw()[0].getValue(), VALUES[0]));
result = scanner.next();
assertTrue("Expected 2 keys but received " + result.size(),
result.size() == 2);
- assertTrue(Bytes.equals(result.sorted()[0].getRow(), ROWS[4]));
- assertTrue(Bytes.equals(result.sorted()[1].getRow(), ROWS[4]));
- assertTrue(Bytes.equals(result.sorted()[0].getValue(), VALUES[1]));
- assertTrue(Bytes.equals(result.sorted()[1].getValue(), VALUES[2]));
+ assertTrue(Bytes.equals(result.raw()[0].getRow(), ROWS[4]));
+ assertTrue(Bytes.equals(result.raw()[1].getRow(), ROWS[4]));
+ assertTrue(Bytes.equals(result.raw()[0].getValue(), VALUES[1]));
+ assertTrue(Bytes.equals(result.raw()[1].getValue(), VALUES[2]));
scanner.close();
// Add test of bulk deleting.
@@ -1995,7 +1995,7 @@ public class TestFromClientSide {
Get get = new Get(ROWS[numRows-1]);
Result result = ht.get(get);
assertNumKeys(result, numColsPerRow);
- KeyValue [] keys = result.sorted();
+ KeyValue [] keys = result.raw();
for(int i=0;i<result.size();i++) {
assertKey(keys[i], ROWS[numRows-1], FAMILY, QUALIFIERS[i], QUALIFIERS[i]);
}
@@ -2006,7 +2006,7 @@ public class TestFromClientSide {
int rowCount = 0;
while((result = scanner.next()) != null) {
assertNumKeys(result, numColsPerRow);
- KeyValue [] kvs = result.sorted();
+ KeyValue [] kvs = result.raw();
for(int i=0;i<numColsPerRow;i++) {
assertKey(kvs[i], ROWS[rowCount], FAMILY, QUALIFIERS[i], QUALIFIERS[i]);
}
@@ -2024,7 +2024,7 @@ public class TestFromClientSide {
get = new Get(ROWS[numRows-1]);
result = ht.get(get);
assertNumKeys(result, numColsPerRow);
- keys = result.sorted();
+ keys = result.raw();
for(int i=0;i<result.size();i++) {
assertKey(keys[i], ROWS[numRows-1], FAMILY, QUALIFIERS[i], QUALIFIERS[i]);
}
@@ -2035,7 +2035,7 @@ public class TestFromClientSide {
rowCount = 0;
while((result = scanner.next()) != null) {
assertNumKeys(result, numColsPerRow);
- KeyValue [] kvs = result.sorted();
+ KeyValue [] kvs = result.raw();
for(int i=0;i<numColsPerRow;i++) {
assertKey(kvs[i], ROWS[rowCount], FAMILY, QUALIFIERS[i], QUALIFIERS[i]);
}
@@ -2857,7 +2857,7 @@ public class TestFromClientSide {
assertTrue("Expected " + idxs.length + " keys but result contains "
+ result.size(), result.size() == idxs.length);
- KeyValue [] keys = result.sorted();
+ KeyValue [] keys = result.raw();
for(int i=0;i<keys.length;i++) {
byte [] family = families[idxs[i][0]];
@@ -2887,7 +2887,7 @@ public class TestFromClientSide {
int expectedResults = end - start + 1;
assertEquals(expectedResults, result.size());
- KeyValue [] keys = result.sorted();
+ KeyValue [] keys = result.raw();
for (int i=0; i<keys.length; i++) {
byte [] value = values[end-i];
@@ -2921,7 +2921,7 @@ public class TestFromClientSide {
equals(row, result.getRow()));
assertTrue("Expected two keys but result contains " + result.size(),
result.size() == 2);
- KeyValue [] kv = result.sorted();
+ KeyValue [] kv = result.raw();
KeyValue kvA = kv[0];
assertTrue("(A) Expected family [" + Bytes.toString(familyA) + "] " +
"Got family [" + Bytes.toString(kvA.getFamily()) + "]",
@@ -2952,7 +2952,7 @@ public class TestFromClientSide {
equals(row, result.getRow()));
assertTrue("Expected a single key but result contains " + result.size(),
result.size() == 1);
- KeyValue kv = result.sorted()[0];
+ KeyValue kv = result.raw()[0];
assertTrue("Expected family [" + Bytes.toString(family) + "] " +
"Got family [" + Bytes.toString(kv.getFamily()) + "]",
equals(family, kv.getFamily()));
@@ -2972,7 +2972,7 @@ public class TestFromClientSide {
equals(row, result.getRow()));
assertTrue("Expected a single key but result contains " + result.size(),
result.size() == 1);
- KeyValue kv = result.sorted()[0];
+ KeyValue kv = result.raw()[0];
assertTrue("Expected family [" + Bytes.toString(family) + "] " +
"Got family [" + Bytes.toString(kv.getFamily()) + "]",
equals(family, kv.getFamily()));
@@ -3511,7 +3511,7 @@ public class TestFromClientSide {
scan.addColumn(CONTENTS_FAMILY, null);
ResultScanner scanner = table.getScanner(scan);
for (Result r : scanner) {
- for(KeyValue key : r.sorted()) {
+ for(KeyValue key : r.raw()) {
System.out.println(Bytes.toString(r.getRow()) + ": " + key.toString());
}
}
@@ -3691,7 +3691,7 @@ public class TestFromClientSide {
int index = 0;
Result r = null;
while ((r = s.next()) != null) {
- for(KeyValue key : r.sorted()) {
+ for(KeyValue key : r.raw()) {
times[index++] = key.getTimestamp();
}
}
@@ -3725,7 +3725,7 @@ public class TestFromClientSide {
int index = 0;
Result r = null;
while ((r = s.next()) != null) {
- for(KeyValue key : r.sorted()) {
+ for(KeyValue key : r.raw()) {
times[index++] = key.getTimestamp();
}
}
@@ -3792,7 +3792,7 @@ public class TestFromClientSide {
try {
for (Result r : s) {
put = new Put(r.getRow());
- for (KeyValue kv : r.sorted()) {
+ for (KeyValue kv : r.raw()) {
put.add(kv);
}
b.put(put);
Modified: hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java (original)
+++ hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java Fri Oct 28 23:37:20 2011
@@ -2749,7 +2749,7 @@ public class TestHRegion extends HBaseTe
// TODO this was removed, now what dangit?!
// search looking for the qualifier in question?
long timestamp = 0;
- for (KeyValue kv : result.sorted()) {
+ for (KeyValue kv : result.raw()) {
if (Bytes.equals(kv.getFamily(), families[0])
&& Bytes.equals(kv.getQualifier(), qualifiers[0])) {
timestamp = kv.getTimestamp();
@@ -3074,7 +3074,7 @@ public class TestHRegion extends HBaseTe
get.addColumn(family, qf);
}
Result result = newReg.get(get, null);
- KeyValue [] raw = result.sorted();
+ KeyValue [] raw = result.raw();
assertEquals(families.length, result.size());
for(int j=0; j<families.length; j++) {
assertEquals(0, Bytes.compareTo(row, raw[j].getRow()));
Modified: hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java?rev=1190672&r1=1190671&r2=1190672&view=diff
==============================================================================
--- hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java (original)
+++ hbase/branches/0.92/src/test/java/org/apache/hadoop/hbase/util/TestMergeTool.java Fri Oct 28 23:37:20 2011
@@ -230,7 +230,7 @@ public class TestMergeTool extends HBase
get.addFamily(FAMILY);
Result result = merged.get(get, null);
assertEquals(1, result.size());
- byte [] bytes = result.sorted()[0].getValue();
+ byte [] bytes = result.raw()[0].getValue();
assertNotNull(Bytes.toStringBinary(rows[i][j]), bytes);
assertTrue(Bytes.equals(bytes, rows[i][j]));
}
@@ -249,7 +249,7 @@ public class TestMergeTool extends HBase
Get get = new Get(rows[i][j]);
get.addFamily(FAMILY);
Result result = regions[i].get(get, null);
- byte [] bytes = result.sorted()[0].getValue();
+ byte [] bytes = result.raw()[0].getValue();
assertNotNull(bytes);
assertTrue(Bytes.equals(bytes, rows[i][j]));
}