You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "Anoop Sam John (Jira)" <ji...@apache.org> on 2020/07/04 07:43:00 UTC

[jira] [Created] (HBASE-24679) HBase on Cloud Blob FS : Provide config to skip HFile archival while table deletion

Anoop Sam John created HBASE-24679:
--------------------------------------

             Summary: HBase on Cloud Blob FS : Provide config to skip HFile archival while table deletion 
                 Key: HBASE-24679
                 URL: https://issues.apache.org/jira/browse/HBASE-24679
             Project: HBase
          Issue Type: Improvement
            Reporter: Anoop Sam John
            Assignee: Anoop Sam John
             Fix For: 3.0.0-alpha-1, 2.4.0


When we delete a table as part of delete of table from FS, we do below things
1. Rename to table directory to come under /hbase/.tmp. This is an atomic rename op
2. Go through each of HFiles under every region:cf and archive that one by one. (Rename the file from .tmp path to go to /hbase/archive)
3. Delete the table dir under .tmp dir

In case of HDFS this is not a big deal as every rename op is just a meta op (Though the HFiles archival is a costly only as there will be so many calls to NN based the table's regions# and total storesfiles#)  But on Cloud blob based FS impl, this is a concerning op. Every rename will be a copy blob op. And we are doing it twice per each of the HFiles in this table !

The proposal here is to provide a config option (default to false) to skip this archival step.
We can provide another config to even avoid the .tmp rename? The atomicity of the Table delete can be achieved by HM side procedure and proc WAL. In table delete the 1st step is to delete the table form META anyways




--
This message was sent by Atlassian Jira
(v8.3.4#803005)