You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hbase.apache.org by "Sergey Soldatov (JIRA)" <ji...@apache.org> on 2016/10/17 06:55:58 UTC

[jira] [Created] (HBASE-16857) RateLimiter may fails during parallel scan execution

Sergey Soldatov created HBASE-16857:
---------------------------------------

             Summary: RateLimiter may fails during parallel scan execution
                 Key: HBASE-16857
                 URL: https://issues.apache.org/jira/browse/HBASE-16857
             Project: HBase
          Issue Type: Bug
    Affects Versions: 1.1.2
         Environment: hbase.quota.enabled=true 
hbase.quota.refresh.period=5000 
            Reporter: Sergey Soldatov
            Assignee: Sergey Soldatov


Steps to reproduce using phoenix (that's the easiest way to run a lot of parallel scans):
1. Create table:
{code}
create table "abc" (id bigint not null primary key, name varchar) salt_buckets=50; 
{code}
2. set quota from hbase shell:
{code}
set_quota TYPE => THROTTLE, TABLE => 'abc', LIMIT => '10G/sec' 
{code}
3. in phoenix run 
{code}
 select * from "abc"; 
{code}
It will fail with ThrottlingException.
Sometimes it requires to run it several times to reproduce.
That happens because the logic in DefaultOperationQuota. First we run limiter.checkQuota where we may change available to Long.MAX_VALUE, after that we run limiter.grabQuota where we reduce available for 1000 (is it scan overhead or what?) and in close() we adding this 1000 back. 
When number of parallel scans are executing there is a chance that one of the threads run limiter.checkQuota right before the second run close(). We get overflow and as the result available value becomes negative, so during the next check we just fail. 
This behavior was introduced in HBASE-13686.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)