You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Hairong Kuang (JIRA)" <ji...@apache.org> on 2007/03/16 02:23:09 UTC

[jira] Updated: (HADOOP-1123) LocalFileSystem gets a NullPointerException when tries to recover from ChecksumError

     [ https://issues.apache.org/jira/browse/HADOOP-1123?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Hairong Kuang updated HADOOP-1123:
----------------------------------

    Attachment: NPEReportChecksum.patch

a patch for review.

> LocalFileSystem gets a NullPointerException when tries to recover from ChecksumError
> ------------------------------------------------------------------------------------
>
>                 Key: HADOOP-1123
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1123
>             Project: Hadoop
>          Issue Type: Bug
>          Components: fs
>    Affects Versions: 0.12.0
>            Reporter: Hairong Kuang
>         Assigned To: Hairong Kuang
>             Fix For: 0.13.0
>
>         Attachments: NPEReportChecksum.patch
>
>
> NullPointerException occurs when run a large sort
> java.lang.NullPointerException
> 	at org.apache.hadoop.fs.FSDataInputStream$Buffer.seek(FSDataInputStream.java:74)
> 	at org.apache.hadoop.fs.FSDataInputStream.seek(FSDataInputStream.java:121)
> 	at org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.readBuffer(ChecksumFileSystem.java:221)
> 	at org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.read(ChecksumFileSystem.java:167)
> 	at org.apache.hadoop.fs.FSDataInputStream$PositionCache.read(FSDataInputStream.java:41)
> 	at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
> 	at java.io.BufferedInputStream.read(BufferedInputStream.java:237)
> 	at org.apache.hadoop.fs.FSDataInputStream$Buffer.read(FSDataInputStream.java:93)
> 	at java.io.DataInputStream.readInt(DataInputStream.java:370)
> 	at org.apache.hadoop.io.SequenceFile$Reader.nextRawKey(SequenceFile.java:1616)
> 	at org.apache.hadoop.io.SequenceFile$Sorter$SegmentDescriptor.nextRawKey(SequenceFile.java:2567)
> 	at org.apache.hadoop.io.SequenceFile$Sorter$MergeQueue.next(SequenceFile.java:2353)
> 	at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.getNext(ReduceTask.java:180)
> 	at org.apache.hadoop.mapred.ReduceTask$ValuesIterator.next(ReduceTask.java:149)
> 	at org.apache.hadoop.mapred.lib.IdentityReducer.reduce(IdentityReducer.java:41)
> 	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:313)
> 	at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:1445)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.