You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "churro morales (JIRA)" <ji...@apache.org> on 2014/07/25 01:37:38 UTC
[jira] [Commented] (HBASE-11409) Add more flexibility for input
directory structure to LoadIncrementalHFiles
[ https://issues.apache.org/jira/browse/HBASE-11409?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14073842#comment-14073842 ]
churro morales commented on HBASE-11409:
----------------------------------------
new patch fixes bug
> Add more flexibility for input directory structure to LoadIncrementalHFiles
> ---------------------------------------------------------------------------
>
> Key: HBASE-11409
> URL: https://issues.apache.org/jira/browse/HBASE-11409
> Project: HBase
> Issue Type: Bug
> Affects Versions: 0.94.20
> Reporter: churro morales
> Attachments: HBASE-11409-0.94.patch, HBASE-11409.0.94.v1.patch
>
>
> Use case:
> We were trying to combine two very large tables into a single table. Thus we ran jobs in one datacenter that populated certain column families and another datacenter which populated other column families. Took a snapshot and exported them to their respective datacenters. Wanted to simply take the hdfs restored snapshot and use LoadIncremental to merge the data.
> It would be nice to add support where we could run LoadIncremental on a directory where the depth of store files is something other than two (current behavior).
> With snapshots it would be nice if you could pass a restored hdfs snapshot's directory and have the tool run.
> I am attaching a patch where I parameterize the bulkLoad timeout as well as the default store file depth.
--
This message was sent by Atlassian JIRA
(v6.2#6252)