You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "Todd Lipcon (JIRA)" <ji...@apache.org> on 2011/03/16 01:11:29 UTC

[jira] Commented: (HBASE-3649) Separate compression setting for flush files

    [ https://issues.apache.org/jira/browse/HBASE-3649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13007291#comment-13007291 ] 

Todd Lipcon commented on HBASE-3649:
------------------------------------

Not sure I follow here... turning off compression for the flush files would just make them bigger, but we'd still have the same amount....

I think the better solution would be "merging flushes"?

> Separate compression setting for flush files
> --------------------------------------------
>
>                 Key: HBASE-3649
>                 URL: https://issues.apache.org/jira/browse/HBASE-3649
>             Project: HBase
>          Issue Type: Improvement
>            Reporter: Andrew Purtell
>            Assignee: Andrew Purtell
>             Fix For: 0.90.2, 0.92.0
>
>
> In this thread on user@hbase: http://search-hadoop.com/m/WUnLM6ojHm1 J-D conjectures that compressing flush files leads to a suboptimal situation where "the puts are sometimes blocked on the memstores which are blocked by the flusher thread which is blocked because there's too many files to compact because the compactor is given too many small files to compact and has to compact the same data a bunch of times."
> We have a separate compression setting already for major compaction vs store files written during minor compaction, for background/archival apps. Add a separate compression setting for flush files, default to none, to avoid the above condition.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira