You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@cassandra.apache.org by "ADARSH KUMAR (JIRA)" <ji...@apache.org> on 2016/05/30 09:04:12 UTC
[jira] [Created] (CASSANDRA-11920) Not able to set
bloom_filter_fp_chance as .00001
ADARSH KUMAR created CASSANDRA-11920:
----------------------------------------
Summary: Not able to set bloom_filter_fp_chance as .00001
Key: CASSANDRA-11920
URL: https://issues.apache.org/jira/browse/CASSANDRA-11920
Project: Cassandra
Issue Type: Bug
Components: Lifecycle, Local Write-Read Paths
Reporter: ADARSH KUMAR
Hi,
I was doing some bench-marking on bloom_filter_fp_chance values. Everything worked fine for values .01(default for STCS), .001, .0001. But when I set bloom_filter_fp_chance = .00001 i observed following behaviour:
1). Reads and writes looked normal from cqlsh.
2). SSttables are never created.
3). It just creates two files (*-Data.db and *-index.db) of size 0kb.
4). nodetool flush does not work and produce following exception:
java.lang.UnsupportedOperationException: Unable to satisfy 1.0E-5 with 20 buckets per element
at org.apache.cassandra.utils.BloomCalculations.computeBloomSpec(BloomCalculations.java:150) .....
I checked BloomCalculations class and following lines are responsible for this exception:
if (maxFalsePosProb < probs[maxBucketsPerElement][maxK]) {
throw new UnsupportedOperationException(String.format("Unable to satisfy %s with %s buckets per element",
maxFalsePosProb, maxBucketsPerElement));
}
From the code it looks like a hard coaded validation (unless we can change the nuber of buckets).
So, if this validation is hard coaded then why it is even allowed to set such value of bloom_fileter_fp_chance, that can prevent ssTable generation?
Please correct this issue.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)