You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "Andrew Purtell (JIRA)" <ji...@apache.org> on 2016/01/12 01:48:39 UTC

[jira] [Updated] (HBASE-14738) Backport HBASE-11927 (Use Native Hadoop Library for HFile checksum) to 0.98

     [ https://issues.apache.org/jira/browse/HBASE-14738?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Andrew Purtell updated HBASE-14738:
-----------------------------------
    Fix Version/s:     (was: 0.98.17)

Unscheduling. I'm not seeing gains worth the risk of changing this code in 0.98. Let me leave this issue open with patch parked here in case we want to revisit this later.

> Backport HBASE-11927 (Use Native Hadoop Library for HFile checksum) to 0.98
> ---------------------------------------------------------------------------
>
>                 Key: HBASE-14738
>                 URL: https://issues.apache.org/jira/browse/HBASE-14738
>             Project: HBase
>          Issue Type: Task
>            Reporter: Andrew Purtell
>            Assignee: Andrew Purtell
>         Attachments: HBASE-14738-0.98.patch
>
>
> Profiling 0.98.15 I see 20-30% of CPU time spent in Hadoop's PureJavaCrc32. Not surprising given previous results described on HBASE-11927. Backport.
> There are two issues with the backport:
> # The patch on 11927 changes the default CRC type from CRC32 to CRC32C. Although the changes are backwards compatible -files with either CRC type will be handled correctly in a transparent manner - we should probably leave the default alone in 0.98 and advise users on a site configuration change to use CRC32C if desired, for potential hardware acceleration.
> # Need a shim for differences between Hadoop's DataChecksum type.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)