You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hbase.apache.org by "stack (JIRA)" <ji...@apache.org> on 2009/11/24 20:45:39 UTC

[jira] Updated: (HBASE-1901) "General" partitioner for "hbase-48" bulk (behind the api, write hfiles direct) uploader

     [ https://issues.apache.org/jira/browse/HBASE-1901?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

stack updated HBASE-1901:
-------------------------

       Resolution: Fixed
    Fix Version/s: 0.21.0
           Status: Resolved  (was: Patch Available)

Committed to TRUNK.

> "General" partitioner for "hbase-48" bulk (behind the api, write hfiles direct) uploader
> ----------------------------------------------------------------------------------------
>
>                 Key: HBASE-1901
>                 URL: https://issues.apache.org/jira/browse/HBASE-1901
>             Project: Hadoop HBase
>          Issue Type: Wish
>            Reporter: stack
>             Fix For: 0.21.0
>
>         Attachments: 1901.patch
>
>
> For users to bulk upload by writing hfiles directly to the filesystem, they currently need to write a partitioner that is intimate with how their key schema works.  This issue is about providing a general partitioner, one that could never be as fair as a custom-written partitioner but that might just work for many cases.  The idea is that a user would supply the first and last keys in their dataset to upload.  We'd then do bigdecimal on the range between start and end rowids dividing it by the number of reducers to come up with key ranges per reducer.
> (I thought jgray had done some BigDecimal work dividing keys already but I can't find it)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.