You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Hairong Kuang (JIRA)" <ji...@apache.org> on 2010/12/01 23:22:12 UTC

[jira] Resolved: (HDFS-1524) Image loader should make sure to read every byte in image file

     [ https://issues.apache.org/jira/browse/HDFS-1524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Hairong Kuang resolved HDFS-1524.
---------------------------------

      Resolution: Fixed
    Hadoop Flags: [Reviewed]

I've just committed this.

> Image loader should make sure to read every byte in image file
> --------------------------------------------------------------
>
>                 Key: HDFS-1524
>                 URL: https://issues.apache.org/jira/browse/HDFS-1524
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: name-node
>    Affects Versions: 0.22.0, 0.23.0
>            Reporter: Hairong Kuang
>            Assignee: Hairong Kuang
>            Priority: Blocker
>             Fix For: 0.22.0, 0.23.0
>
>         Attachments: eofImageFile.patch
>
>
> When I work on HDFS-1070, I come across a very strange bug. Occasionally when loading a compressed image, NameNode throws an exception complaining that the image file is corrupt. However, the result returned by the md5sum utility matches the checksum value stored in the VERSION file. 
> It turns out the image loader leaves 4 bytes unread after loading all the real data of an image. Those unread bytes may be some compression-related meta-info. The image loader should make sure to read to the end of file in order for an correct checksum.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.