You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by om...@apache.org on 2009/01/08 22:56:44 UTC
svn commit: r732840 - in /hadoop/core/branches/branch-0.20: ./
src/core/org/apache/hadoop/io/compress/
Author: omalley
Date: Thu Jan 8 13:56:43 2009
New Revision: 732840
URL: http://svn.apache.org/viewvc?rev=732840&view=rev
Log:
HADOOP-4950. Make the CompressorStream, DecompressorStream,
BlockCompressorStream, and BlockDecompressorStream public to facilitate
non-Hadoop codecs. (omalley)
Modified:
hadoop/core/branches/branch-0.20/ (props changed)
hadoop/core/branches/branch-0.20/CHANGES.txt (contents, props changed)
hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockCompressorStream.java
hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockDecompressorStream.java
hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/CompressorStream.java
hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/DecompressorStream.java
Propchange: hadoop/core/branches/branch-0.20/
------------------------------------------------------------------------------
--- svn:mergeinfo (original)
+++ svn:mergeinfo Thu Jan 8 13:56:43 2009
@@ -1,2 +1,2 @@
/hadoop/core/branches/branch-0.19:713112
-/hadoop/core/trunk:727001,727117,727191,727212,727217,727228,727255,727869,728187,729052,729987,732385,732572,732777
+/hadoop/core/trunk:727001,727117,727191,727212,727217,727228,727255,727869,728187,729052,729987,732385,732572,732777,732838
Modified: hadoop/core/branches/branch-0.20/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.20/CHANGES.txt?rev=732840&r1=732839&r2=732840&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.20/CHANGES.txt (original)
+++ hadoop/core/branches/branch-0.20/CHANGES.txt Thu Jan 8 13:56:43 2009
@@ -298,6 +298,10 @@
HADOOP-4916. Make user/location of Chukwa installation configurable by an
external properties file. (Eric Yang via cdouglas)
+ HADOOP-4950. Make the CompressorStream, DecompressorStream,
+ BlockCompressorStream, and BlockDecompressorStream public to facilitate
+ non-Hadoop codecs. (omalley)
+
OPTIMIZATIONS
HADOOP-3293. Fixes FileInputFormat to do provide locations for splits
Propchange: hadoop/core/branches/branch-0.20/CHANGES.txt
------------------------------------------------------------------------------
--- svn:mergeinfo (original)
+++ svn:mergeinfo Thu Jan 8 13:56:43 2009
@@ -1,3 +1,3 @@
/hadoop/core/branches/branch-0.18/CHANGES.txt:727226
/hadoop/core/branches/branch-0.19/CHANGES.txt:713112
-/hadoop/core/trunk/CHANGES.txt:727001,727117,727191,727212,727228,727255,727869,728187,729052,729987,732385,732572,732777
+/hadoop/core/trunk/CHANGES.txt:727001,727117,727191,727212,727228,727255,727869,728187,729052,729987,732385,732572,732777,732838
Modified: hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockCompressorStream.java
URL: http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockCompressorStream.java?rev=732840&r1=732839&r2=732840&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockCompressorStream.java (original)
+++ hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockCompressorStream.java Thu Jan 8 13:56:43 2009
@@ -31,7 +31,7 @@
* {@link org.apache.hadoop.io.compress.Compressor} requires buffering to
* effect meaningful compression, it is responsible for it.
*/
-class BlockCompressorStream extends CompressorStream {
+public class BlockCompressorStream extends CompressorStream {
// The 'maximum' size of input data to be compressed, to account
// for the overhead of the compression algorithm.
@@ -137,7 +137,7 @@
}
}
- void compress() throws IOException {
+ protected void compress() throws IOException {
int len = compressor.compress(buffer, 0, buffer.length);
if (len > 0) {
// Write out the compressed chunk
Modified: hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockDecompressorStream.java
URL: http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockDecompressorStream.java?rev=732840&r1=732839&r2=732840&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockDecompressorStream.java (original)
+++ hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/BlockDecompressorStream.java Thu Jan 8 13:56:43 2009
@@ -28,7 +28,7 @@
* 'stream-based' compression algorithms.
*
*/
-class BlockDecompressorStream extends DecompressorStream {
+public class BlockDecompressorStream extends DecompressorStream {
private int originalBlockSize = 0;
private int noUncompressedBytes = 0;
@@ -58,7 +58,7 @@
super(in);
}
- int decompress(byte[] b, int off, int len) throws IOException {
+ protected int decompress(byte[] b, int off, int len) throws IOException {
// Check if we are the beginning of a block
if (noUncompressedBytes == originalBlockSize) {
// Get original data size
@@ -89,7 +89,7 @@
return n;
}
- void getCompressedData() throws IOException {
+ protected void getCompressedData() throws IOException {
checkStream();
// Get the size of the compressed chunk
Modified: hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/CompressorStream.java
URL: http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/CompressorStream.java?rev=732840&r1=732839&r2=732840&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/CompressorStream.java (original)
+++ hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/CompressorStream.java Thu Jan 8 13:56:43 2009
@@ -24,10 +24,10 @@
import org.apache.hadoop.io.compress.CompressionOutputStream;
import org.apache.hadoop.io.compress.Compressor;
-class CompressorStream extends CompressionOutputStream {
- Compressor compressor;
- byte[] buffer;
- boolean closed = false;
+public class CompressorStream extends CompressionOutputStream {
+ protected Compressor compressor;
+ protected byte[] buffer;
+ protected boolean closed = false;
public CompressorStream(OutputStream out, Compressor compressor, int bufferSize) {
super(out);
@@ -72,7 +72,7 @@
}
}
- void compress() throws IOException {
+ protected void compress() throws IOException {
int len = compressor.compress(buffer, 0, buffer.length);
if (len > 0) {
out.write(buffer, 0, len);
@@ -100,7 +100,7 @@
}
}
- byte[] oneByte = new byte[1];
+ private byte[] oneByte = new byte[1];
public void write(int b) throws IOException {
oneByte[0] = (byte)(b & 0xff);
write(oneByte, 0, oneByte.length);
Modified: hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/DecompressorStream.java
URL: http://svn.apache.org/viewvc/hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/DecompressorStream.java?rev=732840&r1=732839&r2=732840&view=diff
==============================================================================
--- hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/DecompressorStream.java (original)
+++ hadoop/core/branches/branch-0.20/src/core/org/apache/hadoop/io/compress/DecompressorStream.java Thu Jan 8 13:56:43 2009
@@ -24,11 +24,11 @@
import org.apache.hadoop.io.compress.Decompressor;
-class DecompressorStream extends CompressionInputStream {
- Decompressor decompressor = null;
- byte[] buffer;
- boolean eof = false;
- boolean closed = false;
+public class DecompressorStream extends CompressionInputStream {
+ protected Decompressor decompressor = null;
+ protected byte[] buffer;
+ protected boolean eof = false;
+ protected boolean closed = false;
public DecompressorStream(InputStream in, Decompressor decompressor, int bufferSize) {
super(in);
@@ -56,7 +56,7 @@
super(in);
}
- byte[] oneByte = new byte[1];
+ private byte[] oneByte = new byte[1];
public int read() throws IOException {
checkStream();
return (read(oneByte, 0, oneByte.length) == -1) ? -1 : (oneByte[0] & 0xff);
@@ -74,7 +74,7 @@
return decompress(b, off, len);
}
- int decompress(byte[] b, int off, int len) throws IOException {
+ protected int decompress(byte[] b, int off, int len) throws IOException {
int n = 0;
while ((n = decompressor.decompress(b, off, len)) == 0) {
@@ -90,7 +90,7 @@
return n;
}
- void getCompressedData() throws IOException {
+ protected void getCompressedData() throws IOException {
checkStream();
int n = in.read(buffer, 0, buffer.length);
@@ -101,7 +101,7 @@
decompressor.setInput(buffer, 0, n);
}
- void checkStream() throws IOException {
+ protected void checkStream() throws IOException {
if (closed) {
throw new IOException("Stream closed");
}
@@ -111,7 +111,7 @@
decompressor.reset();
}
- byte[] skipBytes = new byte[512];
+ private byte[] skipBytes = new byte[512];
public long skip(long n) throws IOException {
// Sanity checks
if (n < 0) {