You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Chris Douglas (JIRA)" <ji...@apache.org> on 2009/03/17 00:35:50 UTC
[jira] Updated: (HADOOP-5281) GzipCodec fails second time it is
used in a process
[ https://issues.apache.org/jira/browse/HADOOP-5281?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Chris Douglas updated HADOOP-5281:
----------------------------------
Description:
The attached code (GZt.java) raises:
{noformat}
java.io.IOException: incorrect header check
at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.inflateBytesDirect(Native Method)
at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.decompress(ZlibDecompressor.java:221)
at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:80)
at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:74)
at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:62)
at java.io.DataInputStream.readByte(DataInputStream.java:248)
at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:325)
at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:346)
at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1853)
at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1876)
at org.apache.hadoop.io.MapFile$Reader.readIndex(MapFile.java:319)
at org.apache.hadoop.io.MapFile$Reader.seekInternal(MapFile.java:435)
at org.apache.hadoop.io.MapFile$Reader.seekInternal(MapFile.java:417)
at org.apache.hadoop.io.MapFile$Reader.seek(MapFile.java:404)
at org.apache.hadoop.io.MapFile$Reader.get(MapFile.java:523)
{noformat}
was:
The following code raises:
java.io.IOException: incorrect header check
at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.inflateBytesDirect(Native Method)
at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.decompress(ZlibDecompressor.java:221)
at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:80)
at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:74)
at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:62)
at java.io.DataInputStream.readByte(DataInputStream.java:248)
at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:325)
at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:346)
at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1853)
at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1876)
at org.apache.hadoop.io.MapFile$Reader.readIndex(MapFile.java:319)
at org.apache.hadoop.io.MapFile$Reader.seekInternal(MapFile.java:435)
at org.apache.hadoop.io.MapFile$Reader.seekInternal(MapFile.java:417)
at org.apache.hadoop.io.MapFile$Reader.seek(MapFile.java:404)
at org.apache.hadoop.io.MapFile$Reader.get(MapFile.java:523)
{quote}
import java.io.IOException;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.MapFile;
import org.apache.hadoop.io.SequenceFile.CompressionType;
import org.apache.hadoop.io.compress.GzipCodec;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner;
public class GZt extends Configured implements Tool {
private FileSystem fileSystem;
MapFile.Writer newMap(Path p) throws IOException {
return new MapFile.Writer(getConf(), fileSystem, p.toString(),
IntWritable.class, IntWritable.class, CompressionType.BLOCK,
new GzipCodec(), null);
}
@Override
public int run(String[] args) throws Exception {
Path p = new Path(args[0]);
fileSystem = FileSystem.get(getConf());
fileSystem.mkdirs(p);
for (int i = 0; i < 2; i++) {
MapFile.Writer w = newMap(new Path(p, "b" + i));
for (int j = 0; j < 100; j++) {
w.append(new IntWritable(j), new IntWritable(j));
}
w.close();
}
MapFile.Reader rdr = new MapFile.Reader(fileSystem, new Path(p, "b1")
.toString(), getConf());
rdr.get(new IntWritable(1), new IntWritable());
return 0;
}
public static void main(String[] args) throws Exception {
int res = ToolRunner.run(new Configuration(), new GZt(), args);
System.exit(res);
}
}
{quote}
Edited description
> GzipCodec fails second time it is used in a process
> ---------------------------------------------------
>
> Key: HADOOP-5281
> URL: https://issues.apache.org/jira/browse/HADOOP-5281
> Project: Hadoop Core
> Issue Type: Bug
> Affects Versions: 0.19.0
> Environment: 0.19.0 / Linux / amd64
> Reporter: Ben Maurer
> Priority: Blocker
> Attachments: GZt.java
>
>
> The attached code (GZt.java) raises:
> {noformat}
> java.io.IOException: incorrect header check
> at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.inflateBytesDirect(Native Method)
> at org.apache.hadoop.io.compress.zlib.ZlibDecompressor.decompress(ZlibDecompressor.java:221)
> at org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:80)
> at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:74)
> at org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:62)
> at java.io.DataInputStream.readByte(DataInputStream.java:248)
> at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:325)
> at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:346)
> at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1853)
> at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1876)
> at org.apache.hadoop.io.MapFile$Reader.readIndex(MapFile.java:319)
> at org.apache.hadoop.io.MapFile$Reader.seekInternal(MapFile.java:435)
> at org.apache.hadoop.io.MapFile$Reader.seekInternal(MapFile.java:417)
> at org.apache.hadoop.io.MapFile$Reader.seek(MapFile.java:404)
> at org.apache.hadoop.io.MapFile$Reader.get(MapFile.java:523)
> {noformat}
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.