You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "Andrew Purtell (JIRA)" <ji...@apache.org> on 2018/04/16 20:39:00 UTC

[jira] [Commented] (HBASE-20429) Support for mixed or write-heavy workloads on non-HDFS filesystems

    [ https://issues.apache.org/jira/browse/HBASE-20429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16439990#comment-16439990 ] 

Andrew Purtell commented on HBASE-20429:
----------------------------------------

Let me just state this to get it out of the way. As you can imagine, reading between the lines, the motivation to look at this where I work is the good probability our storage stack is either going to utilize Amazon's S3 service "where applicable" or a compatible API analogue. Please don't take this to infer anything about business relationships, or not. Really, I would personally have no idea one way or the other. 

> Support for mixed or write-heavy workloads on non-HDFS filesystems
> ------------------------------------------------------------------
>
>                 Key: HBASE-20429
>                 URL: https://issues.apache.org/jira/browse/HBASE-20429
>             Project: HBase
>          Issue Type: Umbrella
>            Reporter: Andrew Purtell
>            Priority: Major
>
> We can support reasonably well use cases on non-HDFS filesystems, like S3, where an external writer has loaded (and continues to load) HFiles via the bulk load mechanism, and then we serve out a read only workload at the HBase API.
> Mixed workloads or write-heavy workloads won't fare as well. In fact, data loss seems certain. It will depend in the specific filesystem, but all of the S3 backed Hadoop filesystems suffer from a couple of obvious problems, notably a lack of atomic rename. 
> This umbrella will serve to collect some related ideas for consideration.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)