You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by gi...@apache.org on 2018/11/27 14:52:39 UTC
[07/26] hbase-site git commit: Published site at
6f15cecaed2f1f76bfe1880b7c578ed369daa5d5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dccdd274/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheConfig.html
----------------------------------------------------------------------
diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheConfig.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheConfig.html
index f8e34f2..ea22449 100644
--- a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheConfig.html
+++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheConfig.html
@@ -217,184 +217,188 @@
<span class="sourceLineNo">209</span> @Test<a name="line.209"></a>
<span class="sourceLineNo">210</span> public void testDisableCacheDataBlock() throws IOException {<a name="line.210"></a>
<span class="sourceLineNo">211</span> Configuration conf = HBaseConfiguration.create();<a name="line.211"></a>
-<span class="sourceLineNo">212</span> CacheConfig cacheConfig = new CacheConfig(conf);<a name="line.212"></a>
-<span class="sourceLineNo">213</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.213"></a>
-<span class="sourceLineNo">214</span> assertFalse(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.214"></a>
-<span class="sourceLineNo">215</span> assertFalse(cacheConfig.shouldCacheDataCompressed());<a name="line.215"></a>
-<span class="sourceLineNo">216</span> assertFalse(cacheConfig.shouldCacheDataOnWrite());<a name="line.216"></a>
-<span class="sourceLineNo">217</span> assertTrue(cacheConfig.shouldCacheDataOnRead());<a name="line.217"></a>
-<span class="sourceLineNo">218</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.218"></a>
-<span class="sourceLineNo">219</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.219"></a>
-<span class="sourceLineNo">220</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.220"></a>
-<span class="sourceLineNo">221</span> assertFalse(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.221"></a>
-<span class="sourceLineNo">222</span> assertFalse(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.222"></a>
-<span class="sourceLineNo">223</span><a name="line.223"></a>
-<span class="sourceLineNo">224</span> conf.setBoolean(CacheConfig.CACHE_BLOCKS_ON_WRITE_KEY, true);<a name="line.224"></a>
-<span class="sourceLineNo">225</span> conf.setBoolean(CacheConfig.CACHE_DATA_BLOCKS_COMPRESSED_KEY, true);<a name="line.225"></a>
-<span class="sourceLineNo">226</span> conf.setBoolean(CacheConfig.CACHE_BLOOM_BLOCKS_ON_WRITE_KEY, true);<a name="line.226"></a>
-<span class="sourceLineNo">227</span> conf.setBoolean(CacheConfig.CACHE_INDEX_BLOCKS_ON_WRITE_KEY, true);<a name="line.227"></a>
-<span class="sourceLineNo">228</span><a name="line.228"></a>
-<span class="sourceLineNo">229</span> cacheConfig = new CacheConfig(conf);<a name="line.229"></a>
-<span class="sourceLineNo">230</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.230"></a>
-<span class="sourceLineNo">231</span> assertTrue(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.231"></a>
-<span class="sourceLineNo">232</span> assertTrue(cacheConfig.shouldCacheDataCompressed());<a name="line.232"></a>
-<span class="sourceLineNo">233</span> assertTrue(cacheConfig.shouldCacheDataOnWrite());<a name="line.233"></a>
-<span class="sourceLineNo">234</span> assertTrue(cacheConfig.shouldCacheDataOnRead());<a name="line.234"></a>
-<span class="sourceLineNo">235</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.235"></a>
-<span class="sourceLineNo">236</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.236"></a>
-<span class="sourceLineNo">237</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.237"></a>
-<span class="sourceLineNo">238</span> assertTrue(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.238"></a>
-<span class="sourceLineNo">239</span> assertTrue(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.239"></a>
-<span class="sourceLineNo">240</span><a name="line.240"></a>
-<span class="sourceLineNo">241</span> conf.setBoolean(CacheConfig.CACHE_DATA_ON_READ_KEY, false);<a name="line.241"></a>
-<span class="sourceLineNo">242</span> conf.setBoolean(CacheConfig.CACHE_BLOCKS_ON_WRITE_KEY, false);<a name="line.242"></a>
-<span class="sourceLineNo">243</span><a name="line.243"></a>
-<span class="sourceLineNo">244</span> cacheConfig = new CacheConfig(conf);<a name="line.244"></a>
-<span class="sourceLineNo">245</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.245"></a>
-<span class="sourceLineNo">246</span> assertFalse(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.246"></a>
-<span class="sourceLineNo">247</span> assertFalse(cacheConfig.shouldCacheDataCompressed());<a name="line.247"></a>
-<span class="sourceLineNo">248</span> assertFalse(cacheConfig.shouldCacheDataOnWrite());<a name="line.248"></a>
-<span class="sourceLineNo">249</span> assertFalse(cacheConfig.shouldCacheDataOnRead());<a name="line.249"></a>
-<span class="sourceLineNo">250</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.250"></a>
-<span class="sourceLineNo">251</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.251"></a>
-<span class="sourceLineNo">252</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.252"></a>
-<span class="sourceLineNo">253</span> assertTrue(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.253"></a>
-<span class="sourceLineNo">254</span> assertTrue(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.254"></a>
-<span class="sourceLineNo">255</span><a name="line.255"></a>
-<span class="sourceLineNo">256</span> conf.setBoolean(CacheConfig.CACHE_DATA_ON_READ_KEY, true);<a name="line.256"></a>
-<span class="sourceLineNo">257</span> conf.setBoolean(CacheConfig.CACHE_BLOCKS_ON_WRITE_KEY, false);<a name="line.257"></a>
-<span class="sourceLineNo">258</span><a name="line.258"></a>
-<span class="sourceLineNo">259</span> HColumnDescriptor family = new HColumnDescriptor("testDisableCacheDataBlock");<a name="line.259"></a>
-<span class="sourceLineNo">260</span> family.setBlockCacheEnabled(false);<a name="line.260"></a>
-<span class="sourceLineNo">261</span><a name="line.261"></a>
-<span class="sourceLineNo">262</span> cacheConfig = new CacheConfig(conf, family);<a name="line.262"></a>
-<span class="sourceLineNo">263</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.263"></a>
-<span class="sourceLineNo">264</span> assertFalse(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.264"></a>
-<span class="sourceLineNo">265</span> assertFalse(cacheConfig.shouldCacheDataCompressed());<a name="line.265"></a>
-<span class="sourceLineNo">266</span> assertFalse(cacheConfig.shouldCacheDataOnWrite());<a name="line.266"></a>
-<span class="sourceLineNo">267</span> assertFalse(cacheConfig.shouldCacheDataOnRead());<a name="line.267"></a>
-<span class="sourceLineNo">268</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.268"></a>
-<span class="sourceLineNo">269</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.269"></a>
-<span class="sourceLineNo">270</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.270"></a>
-<span class="sourceLineNo">271</span> assertTrue(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.271"></a>
-<span class="sourceLineNo">272</span> assertTrue(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.272"></a>
-<span class="sourceLineNo">273</span> }<a name="line.273"></a>
-<span class="sourceLineNo">274</span><a name="line.274"></a>
-<span class="sourceLineNo">275</span> @Test<a name="line.275"></a>
-<span class="sourceLineNo">276</span> public void testCacheConfigDefaultLRUBlockCache() {<a name="line.276"></a>
-<span class="sourceLineNo">277</span> CacheConfig cc = new CacheConfig(this.conf);<a name="line.277"></a>
-<span class="sourceLineNo">278</span> assertTrue(cc.isBlockCacheEnabled());<a name="line.278"></a>
-<span class="sourceLineNo">279</span> assertTrue(CacheConfig.DEFAULT_IN_MEMORY == cc.isInMemory());<a name="line.279"></a>
-<span class="sourceLineNo">280</span> basicBlockCacheOps(cc, false, true);<a name="line.280"></a>
-<span class="sourceLineNo">281</span> assertTrue(cc.getBlockCache() instanceof LruBlockCache);<a name="line.281"></a>
-<span class="sourceLineNo">282</span> }<a name="line.282"></a>
-<span class="sourceLineNo">283</span><a name="line.283"></a>
-<span class="sourceLineNo">284</span> /**<a name="line.284"></a>
-<span class="sourceLineNo">285</span> * Assert that the caches are deployed with CombinedBlockCache and of the appropriate sizes.<a name="line.285"></a>
-<span class="sourceLineNo">286</span> */<a name="line.286"></a>
-<span class="sourceLineNo">287</span> @Test<a name="line.287"></a>
-<span class="sourceLineNo">288</span> public void testOffHeapBucketCacheConfig() {<a name="line.288"></a>
-<span class="sourceLineNo">289</span> this.conf.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap");<a name="line.289"></a>
-<span class="sourceLineNo">290</span> doBucketCacheConfigTest();<a name="line.290"></a>
-<span class="sourceLineNo">291</span> }<a name="line.291"></a>
-<span class="sourceLineNo">292</span><a name="line.292"></a>
-<span class="sourceLineNo">293</span> @Test<a name="line.293"></a>
-<span class="sourceLineNo">294</span> public void testFileBucketCacheConfig() throws IOException {<a name="line.294"></a>
-<span class="sourceLineNo">295</span> HBaseTestingUtility htu = new HBaseTestingUtility(this.conf);<a name="line.295"></a>
-<span class="sourceLineNo">296</span> try {<a name="line.296"></a>
-<span class="sourceLineNo">297</span> Path p = new Path(htu.getDataTestDir(), "bc.txt");<a name="line.297"></a>
-<span class="sourceLineNo">298</span> FileSystem fs = FileSystem.get(this.conf);<a name="line.298"></a>
-<span class="sourceLineNo">299</span> fs.create(p).close();<a name="line.299"></a>
-<span class="sourceLineNo">300</span> this.conf.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "file:" + p);<a name="line.300"></a>
-<span class="sourceLineNo">301</span> doBucketCacheConfigTest();<a name="line.301"></a>
-<span class="sourceLineNo">302</span> } finally {<a name="line.302"></a>
-<span class="sourceLineNo">303</span> htu.cleanupTestDir();<a name="line.303"></a>
-<span class="sourceLineNo">304</span> }<a name="line.304"></a>
-<span class="sourceLineNo">305</span> }<a name="line.305"></a>
-<span class="sourceLineNo">306</span><a name="line.306"></a>
-<span class="sourceLineNo">307</span> private void doBucketCacheConfigTest() {<a name="line.307"></a>
-<span class="sourceLineNo">308</span> final int bcSize = 100;<a name="line.308"></a>
-<span class="sourceLineNo">309</span> this.conf.setInt(HConstants.BUCKET_CACHE_SIZE_KEY, bcSize);<a name="line.309"></a>
-<span class="sourceLineNo">310</span> CacheConfig cc = new CacheConfig(this.conf);<a name="line.310"></a>
-<span class="sourceLineNo">311</span> basicBlockCacheOps(cc, false, false);<a name="line.311"></a>
-<span class="sourceLineNo">312</span> assertTrue(cc.getBlockCache() instanceof CombinedBlockCache);<a name="line.312"></a>
-<span class="sourceLineNo">313</span> // TODO: Assert sizes allocated are right and proportions.<a name="line.313"></a>
-<span class="sourceLineNo">314</span> CombinedBlockCache cbc = (CombinedBlockCache)cc.getBlockCache();<a name="line.314"></a>
-<span class="sourceLineNo">315</span> BlockCache [] bcs = cbc.getBlockCaches();<a name="line.315"></a>
-<span class="sourceLineNo">316</span> assertTrue(bcs[0] instanceof LruBlockCache);<a name="line.316"></a>
-<span class="sourceLineNo">317</span> LruBlockCache lbc = (LruBlockCache)bcs[0];<a name="line.317"></a>
-<span class="sourceLineNo">318</span> assertEquals(MemorySizeUtil.getOnHeapCacheSize(this.conf), lbc.getMaxSize());<a name="line.318"></a>
-<span class="sourceLineNo">319</span> assertTrue(bcs[1] instanceof BucketCache);<a name="line.319"></a>
-<span class="sourceLineNo">320</span> BucketCache bc = (BucketCache)bcs[1];<a name="line.320"></a>
-<span class="sourceLineNo">321</span> // getMaxSize comes back in bytes but we specified size in MB<a name="line.321"></a>
-<span class="sourceLineNo">322</span> assertEquals(bcSize, bc.getMaxSize() / (1024 * 1024));<a name="line.322"></a>
-<span class="sourceLineNo">323</span> }<a name="line.323"></a>
-<span class="sourceLineNo">324</span><a name="line.324"></a>
-<span class="sourceLineNo">325</span> /**<a name="line.325"></a>
-<span class="sourceLineNo">326</span> * Assert that when BUCKET_CACHE_COMBINED_KEY is false, the non-default, that we deploy<a name="line.326"></a>
-<span class="sourceLineNo">327</span> * LruBlockCache as L1 with a BucketCache for L2.<a name="line.327"></a>
-<span class="sourceLineNo">328</span> */<a name="line.328"></a>
-<span class="sourceLineNo">329</span> @Test<a name="line.329"></a>
-<span class="sourceLineNo">330</span> public void testBucketCacheConfigL1L2Setup() {<a name="line.330"></a>
-<span class="sourceLineNo">331</span> this.conf.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap");<a name="line.331"></a>
-<span class="sourceLineNo">332</span> // Make lru size is smaller than bcSize for sure. Need this to be true so when eviction<a name="line.332"></a>
-<span class="sourceLineNo">333</span> // from L1 happens, it does not fail because L2 can't take the eviction because block too big.<a name="line.333"></a>
-<span class="sourceLineNo">334</span> this.conf.setFloat(HConstants.HFILE_BLOCK_CACHE_SIZE_KEY, 0.001f);<a name="line.334"></a>
-<span class="sourceLineNo">335</span> MemoryUsage mu = ManagementFactory.getMemoryMXBean().getHeapMemoryUsage();<a name="line.335"></a>
-<span class="sourceLineNo">336</span> long lruExpectedSize = MemorySizeUtil.getOnHeapCacheSize(this.conf);<a name="line.336"></a>
-<span class="sourceLineNo">337</span> final int bcSize = 100;<a name="line.337"></a>
-<span class="sourceLineNo">338</span> long bcExpectedSize = 100 * 1024 * 1024; // MB.<a name="line.338"></a>
-<span class="sourceLineNo">339</span> assertTrue(lruExpectedSize < bcExpectedSize);<a name="line.339"></a>
-<span class="sourceLineNo">340</span> this.conf.setInt(HConstants.BUCKET_CACHE_SIZE_KEY, bcSize);<a name="line.340"></a>
-<span class="sourceLineNo">341</span> CacheConfig cc = new CacheConfig(this.conf);<a name="line.341"></a>
-<span class="sourceLineNo">342</span> basicBlockCacheOps(cc, false, false);<a name="line.342"></a>
-<span class="sourceLineNo">343</span> assertTrue(cc.getBlockCache() instanceof CombinedBlockCache);<a name="line.343"></a>
-<span class="sourceLineNo">344</span> // TODO: Assert sizes allocated are right and proportions.<a name="line.344"></a>
-<span class="sourceLineNo">345</span> CombinedBlockCache cbc = (CombinedBlockCache)cc.getBlockCache();<a name="line.345"></a>
-<span class="sourceLineNo">346</span> LruBlockCache lbc = cbc.onHeapCache;<a name="line.346"></a>
-<span class="sourceLineNo">347</span> assertEquals(lruExpectedSize, lbc.getMaxSize());<a name="line.347"></a>
-<span class="sourceLineNo">348</span> BlockCache bc = cbc.l2Cache;<a name="line.348"></a>
-<span class="sourceLineNo">349</span> // getMaxSize comes back in bytes but we specified size in MB<a name="line.349"></a>
-<span class="sourceLineNo">350</span> assertEquals(bcExpectedSize, ((BucketCache) bc).getMaxSize());<a name="line.350"></a>
-<span class="sourceLineNo">351</span> // Test the L1+L2 deploy works as we'd expect with blocks evicted from L1 going to L2.<a name="line.351"></a>
-<span class="sourceLineNo">352</span> long initialL1BlockCount = lbc.getBlockCount();<a name="line.352"></a>
-<span class="sourceLineNo">353</span> long initialL2BlockCount = bc.getBlockCount();<a name="line.353"></a>
-<span class="sourceLineNo">354</span> Cacheable c = new DataCacheEntry();<a name="line.354"></a>
-<span class="sourceLineNo">355</span> BlockCacheKey bck = new BlockCacheKey("bck", 0);<a name="line.355"></a>
-<span class="sourceLineNo">356</span> lbc.cacheBlock(bck, c, false);<a name="line.356"></a>
-<span class="sourceLineNo">357</span> assertEquals(initialL1BlockCount + 1, lbc.getBlockCount());<a name="line.357"></a>
-<span class="sourceLineNo">358</span> assertEquals(initialL2BlockCount, bc.getBlockCount());<a name="line.358"></a>
-<span class="sourceLineNo">359</span> // Force evictions by putting in a block too big.<a name="line.359"></a>
-<span class="sourceLineNo">360</span> final long justTooBigSize = lbc.acceptableSize() + 1;<a name="line.360"></a>
-<span class="sourceLineNo">361</span> lbc.cacheBlock(new BlockCacheKey("bck2", 0), new DataCacheEntry() {<a name="line.361"></a>
-<span class="sourceLineNo">362</span> @Override<a name="line.362"></a>
-<span class="sourceLineNo">363</span> public long heapSize() {<a name="line.363"></a>
-<span class="sourceLineNo">364</span> return justTooBigSize;<a name="line.364"></a>
-<span class="sourceLineNo">365</span> }<a name="line.365"></a>
-<span class="sourceLineNo">366</span><a name="line.366"></a>
-<span class="sourceLineNo">367</span> @Override<a name="line.367"></a>
-<span class="sourceLineNo">368</span> public int getSerializedLength() {<a name="line.368"></a>
-<span class="sourceLineNo">369</span> return (int)heapSize();<a name="line.369"></a>
-<span class="sourceLineNo">370</span> }<a name="line.370"></a>
-<span class="sourceLineNo">371</span> });<a name="line.371"></a>
-<span class="sourceLineNo">372</span> // The eviction thread in lrublockcache needs to run.<a name="line.372"></a>
-<span class="sourceLineNo">373</span> while (initialL1BlockCount != lbc.getBlockCount()) Threads.sleep(10);<a name="line.373"></a>
-<span class="sourceLineNo">374</span> assertEquals(initialL1BlockCount, lbc.getBlockCount());<a name="line.374"></a>
-<span class="sourceLineNo">375</span> }<a name="line.375"></a>
-<span class="sourceLineNo">376</span><a name="line.376"></a>
-<span class="sourceLineNo">377</span> @Test<a name="line.377"></a>
-<span class="sourceLineNo">378</span> public void testL2CacheWithInvalidBucketSize() {<a name="line.378"></a>
-<span class="sourceLineNo">379</span> Configuration c = new Configuration(this.conf);<a name="line.379"></a>
-<span class="sourceLineNo">380</span> c.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap");<a name="line.380"></a>
-<span class="sourceLineNo">381</span> c.set(CacheConfig.BUCKET_CACHE_BUCKETS_KEY, "256,512,1024,2048,4000,4096");<a name="line.381"></a>
-<span class="sourceLineNo">382</span> c.setFloat(HConstants.BUCKET_CACHE_SIZE_KEY, 1024);<a name="line.382"></a>
-<span class="sourceLineNo">383</span> try {<a name="line.383"></a>
-<span class="sourceLineNo">384</span> CacheConfig.getBucketCache(c);<a name="line.384"></a>
-<span class="sourceLineNo">385</span> fail("Should throw IllegalArgumentException when passing illegal value for bucket size");<a name="line.385"></a>
-<span class="sourceLineNo">386</span> } catch (IllegalArgumentException e) {<a name="line.386"></a>
-<span class="sourceLineNo">387</span> }<a name="line.387"></a>
-<span class="sourceLineNo">388</span> }<a name="line.388"></a>
-<span class="sourceLineNo">389</span>}<a name="line.389"></a>
+<span class="sourceLineNo">212</span> CacheConfig.instantiateBlockCache(conf);<a name="line.212"></a>
+<span class="sourceLineNo">213</span> CacheConfig cacheConfig = new CacheConfig(conf);<a name="line.213"></a>
+<span class="sourceLineNo">214</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.214"></a>
+<span class="sourceLineNo">215</span> assertFalse(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.215"></a>
+<span class="sourceLineNo">216</span> assertFalse(cacheConfig.shouldCacheDataCompressed());<a name="line.216"></a>
+<span class="sourceLineNo">217</span> assertFalse(cacheConfig.shouldCacheDataOnWrite());<a name="line.217"></a>
+<span class="sourceLineNo">218</span> assertTrue(cacheConfig.shouldCacheDataOnRead());<a name="line.218"></a>
+<span class="sourceLineNo">219</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.219"></a>
+<span class="sourceLineNo">220</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.220"></a>
+<span class="sourceLineNo">221</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.221"></a>
+<span class="sourceLineNo">222</span> assertFalse(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.222"></a>
+<span class="sourceLineNo">223</span> assertFalse(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.223"></a>
+<span class="sourceLineNo">224</span><a name="line.224"></a>
+<span class="sourceLineNo">225</span> conf.setBoolean(CacheConfig.CACHE_BLOCKS_ON_WRITE_KEY, true);<a name="line.225"></a>
+<span class="sourceLineNo">226</span> conf.setBoolean(CacheConfig.CACHE_DATA_BLOCKS_COMPRESSED_KEY, true);<a name="line.226"></a>
+<span class="sourceLineNo">227</span> conf.setBoolean(CacheConfig.CACHE_BLOOM_BLOCKS_ON_WRITE_KEY, true);<a name="line.227"></a>
+<span class="sourceLineNo">228</span> conf.setBoolean(CacheConfig.CACHE_INDEX_BLOCKS_ON_WRITE_KEY, true);<a name="line.228"></a>
+<span class="sourceLineNo">229</span><a name="line.229"></a>
+<span class="sourceLineNo">230</span> cacheConfig = new CacheConfig(conf);<a name="line.230"></a>
+<span class="sourceLineNo">231</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.231"></a>
+<span class="sourceLineNo">232</span> assertTrue(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.232"></a>
+<span class="sourceLineNo">233</span> assertTrue(cacheConfig.shouldCacheDataCompressed());<a name="line.233"></a>
+<span class="sourceLineNo">234</span> assertTrue(cacheConfig.shouldCacheDataOnWrite());<a name="line.234"></a>
+<span class="sourceLineNo">235</span> assertTrue(cacheConfig.shouldCacheDataOnRead());<a name="line.235"></a>
+<span class="sourceLineNo">236</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.236"></a>
+<span class="sourceLineNo">237</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.237"></a>
+<span class="sourceLineNo">238</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.238"></a>
+<span class="sourceLineNo">239</span> assertTrue(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.239"></a>
+<span class="sourceLineNo">240</span> assertTrue(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.240"></a>
+<span class="sourceLineNo">241</span><a name="line.241"></a>
+<span class="sourceLineNo">242</span> conf.setBoolean(CacheConfig.CACHE_DATA_ON_READ_KEY, false);<a name="line.242"></a>
+<span class="sourceLineNo">243</span> conf.setBoolean(CacheConfig.CACHE_BLOCKS_ON_WRITE_KEY, false);<a name="line.243"></a>
+<span class="sourceLineNo">244</span><a name="line.244"></a>
+<span class="sourceLineNo">245</span> cacheConfig = new CacheConfig(conf);<a name="line.245"></a>
+<span class="sourceLineNo">246</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.246"></a>
+<span class="sourceLineNo">247</span> assertFalse(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.247"></a>
+<span class="sourceLineNo">248</span> assertFalse(cacheConfig.shouldCacheDataCompressed());<a name="line.248"></a>
+<span class="sourceLineNo">249</span> assertFalse(cacheConfig.shouldCacheDataOnWrite());<a name="line.249"></a>
+<span class="sourceLineNo">250</span> assertFalse(cacheConfig.shouldCacheDataOnRead());<a name="line.250"></a>
+<span class="sourceLineNo">251</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.251"></a>
+<span class="sourceLineNo">252</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.252"></a>
+<span class="sourceLineNo">253</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.253"></a>
+<span class="sourceLineNo">254</span> assertTrue(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.254"></a>
+<span class="sourceLineNo">255</span> assertTrue(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.255"></a>
+<span class="sourceLineNo">256</span><a name="line.256"></a>
+<span class="sourceLineNo">257</span> conf.setBoolean(CacheConfig.CACHE_DATA_ON_READ_KEY, true);<a name="line.257"></a>
+<span class="sourceLineNo">258</span> conf.setBoolean(CacheConfig.CACHE_BLOCKS_ON_WRITE_KEY, false);<a name="line.258"></a>
+<span class="sourceLineNo">259</span><a name="line.259"></a>
+<span class="sourceLineNo">260</span> HColumnDescriptor family = new HColumnDescriptor("testDisableCacheDataBlock");<a name="line.260"></a>
+<span class="sourceLineNo">261</span> family.setBlockCacheEnabled(false);<a name="line.261"></a>
+<span class="sourceLineNo">262</span><a name="line.262"></a>
+<span class="sourceLineNo">263</span> cacheConfig = new CacheConfig(conf, family);<a name="line.263"></a>
+<span class="sourceLineNo">264</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.DATA));<a name="line.264"></a>
+<span class="sourceLineNo">265</span> assertFalse(cacheConfig.shouldCacheCompressed(BlockCategory.DATA));<a name="line.265"></a>
+<span class="sourceLineNo">266</span> assertFalse(cacheConfig.shouldCacheDataCompressed());<a name="line.266"></a>
+<span class="sourceLineNo">267</span> assertFalse(cacheConfig.shouldCacheDataOnWrite());<a name="line.267"></a>
+<span class="sourceLineNo">268</span> assertFalse(cacheConfig.shouldCacheDataOnRead());<a name="line.268"></a>
+<span class="sourceLineNo">269</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.INDEX));<a name="line.269"></a>
+<span class="sourceLineNo">270</span> assertFalse(cacheConfig.shouldCacheBlockOnRead(BlockCategory.META));<a name="line.270"></a>
+<span class="sourceLineNo">271</span> assertTrue(cacheConfig.shouldCacheBlockOnRead(BlockCategory.BLOOM));<a name="line.271"></a>
+<span class="sourceLineNo">272</span> assertTrue(cacheConfig.shouldCacheBloomsOnWrite());<a name="line.272"></a>
+<span class="sourceLineNo">273</span> assertTrue(cacheConfig.shouldCacheIndexesOnWrite());<a name="line.273"></a>
+<span class="sourceLineNo">274</span> }<a name="line.274"></a>
+<span class="sourceLineNo">275</span><a name="line.275"></a>
+<span class="sourceLineNo">276</span> @Test<a name="line.276"></a>
+<span class="sourceLineNo">277</span> public void testCacheConfigDefaultLRUBlockCache() {<a name="line.277"></a>
+<span class="sourceLineNo">278</span> CacheConfig.instantiateBlockCache(this.conf);<a name="line.278"></a>
+<span class="sourceLineNo">279</span> CacheConfig cc = new CacheConfig(this.conf);<a name="line.279"></a>
+<span class="sourceLineNo">280</span> assertTrue(cc.isBlockCacheEnabled());<a name="line.280"></a>
+<span class="sourceLineNo">281</span> assertTrue(CacheConfig.DEFAULT_IN_MEMORY == cc.isInMemory());<a name="line.281"></a>
+<span class="sourceLineNo">282</span> basicBlockCacheOps(cc, false, true);<a name="line.282"></a>
+<span class="sourceLineNo">283</span> assertTrue(cc.getBlockCache() instanceof LruBlockCache);<a name="line.283"></a>
+<span class="sourceLineNo">284</span> }<a name="line.284"></a>
+<span class="sourceLineNo">285</span><a name="line.285"></a>
+<span class="sourceLineNo">286</span> /**<a name="line.286"></a>
+<span class="sourceLineNo">287</span> * Assert that the caches are deployed with CombinedBlockCache and of the appropriate sizes.<a name="line.287"></a>
+<span class="sourceLineNo">288</span> */<a name="line.288"></a>
+<span class="sourceLineNo">289</span> @Test<a name="line.289"></a>
+<span class="sourceLineNo">290</span> public void testOffHeapBucketCacheConfig() {<a name="line.290"></a>
+<span class="sourceLineNo">291</span> this.conf.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap");<a name="line.291"></a>
+<span class="sourceLineNo">292</span> doBucketCacheConfigTest();<a name="line.292"></a>
+<span class="sourceLineNo">293</span> }<a name="line.293"></a>
+<span class="sourceLineNo">294</span><a name="line.294"></a>
+<span class="sourceLineNo">295</span> @Test<a name="line.295"></a>
+<span class="sourceLineNo">296</span> public void testFileBucketCacheConfig() throws IOException {<a name="line.296"></a>
+<span class="sourceLineNo">297</span> HBaseTestingUtility htu = new HBaseTestingUtility(this.conf);<a name="line.297"></a>
+<span class="sourceLineNo">298</span> try {<a name="line.298"></a>
+<span class="sourceLineNo">299</span> Path p = new Path(htu.getDataTestDir(), "bc.txt");<a name="line.299"></a>
+<span class="sourceLineNo">300</span> FileSystem fs = FileSystem.get(this.conf);<a name="line.300"></a>
+<span class="sourceLineNo">301</span> fs.create(p).close();<a name="line.301"></a>
+<span class="sourceLineNo">302</span> this.conf.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "file:" + p);<a name="line.302"></a>
+<span class="sourceLineNo">303</span> doBucketCacheConfigTest();<a name="line.303"></a>
+<span class="sourceLineNo">304</span> } finally {<a name="line.304"></a>
+<span class="sourceLineNo">305</span> htu.cleanupTestDir();<a name="line.305"></a>
+<span class="sourceLineNo">306</span> }<a name="line.306"></a>
+<span class="sourceLineNo">307</span> }<a name="line.307"></a>
+<span class="sourceLineNo">308</span><a name="line.308"></a>
+<span class="sourceLineNo">309</span> private void doBucketCacheConfigTest() {<a name="line.309"></a>
+<span class="sourceLineNo">310</span> final int bcSize = 100;<a name="line.310"></a>
+<span class="sourceLineNo">311</span> this.conf.setInt(HConstants.BUCKET_CACHE_SIZE_KEY, bcSize);<a name="line.311"></a>
+<span class="sourceLineNo">312</span> CacheConfig.instantiateBlockCache(this.conf);<a name="line.312"></a>
+<span class="sourceLineNo">313</span> CacheConfig cc = new CacheConfig(this.conf);<a name="line.313"></a>
+<span class="sourceLineNo">314</span> basicBlockCacheOps(cc, false, false);<a name="line.314"></a>
+<span class="sourceLineNo">315</span> assertTrue(cc.getBlockCache() instanceof CombinedBlockCache);<a name="line.315"></a>
+<span class="sourceLineNo">316</span> // TODO: Assert sizes allocated are right and proportions.<a name="line.316"></a>
+<span class="sourceLineNo">317</span> CombinedBlockCache cbc = (CombinedBlockCache)cc.getBlockCache();<a name="line.317"></a>
+<span class="sourceLineNo">318</span> BlockCache [] bcs = cbc.getBlockCaches();<a name="line.318"></a>
+<span class="sourceLineNo">319</span> assertTrue(bcs[0] instanceof LruBlockCache);<a name="line.319"></a>
+<span class="sourceLineNo">320</span> LruBlockCache lbc = (LruBlockCache)bcs[0];<a name="line.320"></a>
+<span class="sourceLineNo">321</span> assertEquals(MemorySizeUtil.getOnHeapCacheSize(this.conf), lbc.getMaxSize());<a name="line.321"></a>
+<span class="sourceLineNo">322</span> assertTrue(bcs[1] instanceof BucketCache);<a name="line.322"></a>
+<span class="sourceLineNo">323</span> BucketCache bc = (BucketCache)bcs[1];<a name="line.323"></a>
+<span class="sourceLineNo">324</span> // getMaxSize comes back in bytes but we specified size in MB<a name="line.324"></a>
+<span class="sourceLineNo">325</span> assertEquals(bcSize, bc.getMaxSize() / (1024 * 1024));<a name="line.325"></a>
+<span class="sourceLineNo">326</span> }<a name="line.326"></a>
+<span class="sourceLineNo">327</span><a name="line.327"></a>
+<span class="sourceLineNo">328</span> /**<a name="line.328"></a>
+<span class="sourceLineNo">329</span> * Assert that when BUCKET_CACHE_COMBINED_KEY is false, the non-default, that we deploy<a name="line.329"></a>
+<span class="sourceLineNo">330</span> * LruBlockCache as L1 with a BucketCache for L2.<a name="line.330"></a>
+<span class="sourceLineNo">331</span> */<a name="line.331"></a>
+<span class="sourceLineNo">332</span> @Test<a name="line.332"></a>
+<span class="sourceLineNo">333</span> public void testBucketCacheConfigL1L2Setup() {<a name="line.333"></a>
+<span class="sourceLineNo">334</span> this.conf.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap");<a name="line.334"></a>
+<span class="sourceLineNo">335</span> // Make lru size is smaller than bcSize for sure. Need this to be true so when eviction<a name="line.335"></a>
+<span class="sourceLineNo">336</span> // from L1 happens, it does not fail because L2 can't take the eviction because block too big.<a name="line.336"></a>
+<span class="sourceLineNo">337</span> this.conf.setFloat(HConstants.HFILE_BLOCK_CACHE_SIZE_KEY, 0.001f);<a name="line.337"></a>
+<span class="sourceLineNo">338</span> MemoryUsage mu = ManagementFactory.getMemoryMXBean().getHeapMemoryUsage();<a name="line.338"></a>
+<span class="sourceLineNo">339</span> long lruExpectedSize = MemorySizeUtil.getOnHeapCacheSize(this.conf);<a name="line.339"></a>
+<span class="sourceLineNo">340</span> final int bcSize = 100;<a name="line.340"></a>
+<span class="sourceLineNo">341</span> long bcExpectedSize = 100 * 1024 * 1024; // MB.<a name="line.341"></a>
+<span class="sourceLineNo">342</span> assertTrue(lruExpectedSize < bcExpectedSize);<a name="line.342"></a>
+<span class="sourceLineNo">343</span> this.conf.setInt(HConstants.BUCKET_CACHE_SIZE_KEY, bcSize);<a name="line.343"></a>
+<span class="sourceLineNo">344</span> CacheConfig.instantiateBlockCache(this.conf);<a name="line.344"></a>
+<span class="sourceLineNo">345</span> CacheConfig cc = new CacheConfig(this.conf);<a name="line.345"></a>
+<span class="sourceLineNo">346</span> basicBlockCacheOps(cc, false, false);<a name="line.346"></a>
+<span class="sourceLineNo">347</span> assertTrue(cc.getBlockCache() instanceof CombinedBlockCache);<a name="line.347"></a>
+<span class="sourceLineNo">348</span> // TODO: Assert sizes allocated are right and proportions.<a name="line.348"></a>
+<span class="sourceLineNo">349</span> CombinedBlockCache cbc = (CombinedBlockCache)cc.getBlockCache();<a name="line.349"></a>
+<span class="sourceLineNo">350</span> LruBlockCache lbc = cbc.onHeapCache;<a name="line.350"></a>
+<span class="sourceLineNo">351</span> assertEquals(lruExpectedSize, lbc.getMaxSize());<a name="line.351"></a>
+<span class="sourceLineNo">352</span> BlockCache bc = cbc.l2Cache;<a name="line.352"></a>
+<span class="sourceLineNo">353</span> // getMaxSize comes back in bytes but we specified size in MB<a name="line.353"></a>
+<span class="sourceLineNo">354</span> assertEquals(bcExpectedSize, ((BucketCache) bc).getMaxSize());<a name="line.354"></a>
+<span class="sourceLineNo">355</span> // Test the L1+L2 deploy works as we'd expect with blocks evicted from L1 going to L2.<a name="line.355"></a>
+<span class="sourceLineNo">356</span> long initialL1BlockCount = lbc.getBlockCount();<a name="line.356"></a>
+<span class="sourceLineNo">357</span> long initialL2BlockCount = bc.getBlockCount();<a name="line.357"></a>
+<span class="sourceLineNo">358</span> Cacheable c = new DataCacheEntry();<a name="line.358"></a>
+<span class="sourceLineNo">359</span> BlockCacheKey bck = new BlockCacheKey("bck", 0);<a name="line.359"></a>
+<span class="sourceLineNo">360</span> lbc.cacheBlock(bck, c, false);<a name="line.360"></a>
+<span class="sourceLineNo">361</span> assertEquals(initialL1BlockCount + 1, lbc.getBlockCount());<a name="line.361"></a>
+<span class="sourceLineNo">362</span> assertEquals(initialL2BlockCount, bc.getBlockCount());<a name="line.362"></a>
+<span class="sourceLineNo">363</span> // Force evictions by putting in a block too big.<a name="line.363"></a>
+<span class="sourceLineNo">364</span> final long justTooBigSize = lbc.acceptableSize() + 1;<a name="line.364"></a>
+<span class="sourceLineNo">365</span> lbc.cacheBlock(new BlockCacheKey("bck2", 0), new DataCacheEntry() {<a name="line.365"></a>
+<span class="sourceLineNo">366</span> @Override<a name="line.366"></a>
+<span class="sourceLineNo">367</span> public long heapSize() {<a name="line.367"></a>
+<span class="sourceLineNo">368</span> return justTooBigSize;<a name="line.368"></a>
+<span class="sourceLineNo">369</span> }<a name="line.369"></a>
+<span class="sourceLineNo">370</span><a name="line.370"></a>
+<span class="sourceLineNo">371</span> @Override<a name="line.371"></a>
+<span class="sourceLineNo">372</span> public int getSerializedLength() {<a name="line.372"></a>
+<span class="sourceLineNo">373</span> return (int)heapSize();<a name="line.373"></a>
+<span class="sourceLineNo">374</span> }<a name="line.374"></a>
+<span class="sourceLineNo">375</span> });<a name="line.375"></a>
+<span class="sourceLineNo">376</span> // The eviction thread in lrublockcache needs to run.<a name="line.376"></a>
+<span class="sourceLineNo">377</span> while (initialL1BlockCount != lbc.getBlockCount()) Threads.sleep(10);<a name="line.377"></a>
+<span class="sourceLineNo">378</span> assertEquals(initialL1BlockCount, lbc.getBlockCount());<a name="line.378"></a>
+<span class="sourceLineNo">379</span> }<a name="line.379"></a>
+<span class="sourceLineNo">380</span><a name="line.380"></a>
+<span class="sourceLineNo">381</span> @Test<a name="line.381"></a>
+<span class="sourceLineNo">382</span> public void testL2CacheWithInvalidBucketSize() {<a name="line.382"></a>
+<span class="sourceLineNo">383</span> Configuration c = new Configuration(this.conf);<a name="line.383"></a>
+<span class="sourceLineNo">384</span> c.set(HConstants.BUCKET_CACHE_IOENGINE_KEY, "offheap");<a name="line.384"></a>
+<span class="sourceLineNo">385</span> c.set(CacheConfig.BUCKET_CACHE_BUCKETS_KEY, "256,512,1024,2048,4000,4096");<a name="line.385"></a>
+<span class="sourceLineNo">386</span> c.setFloat(HConstants.BUCKET_CACHE_SIZE_KEY, 1024);<a name="line.386"></a>
+<span class="sourceLineNo">387</span> try {<a name="line.387"></a>
+<span class="sourceLineNo">388</span> CacheConfig.getBucketCache(c);<a name="line.388"></a>
+<span class="sourceLineNo">389</span> fail("Should throw IllegalArgumentException when passing illegal value for bucket size");<a name="line.389"></a>
+<span class="sourceLineNo">390</span> } catch (IllegalArgumentException e) {<a name="line.390"></a>
+<span class="sourceLineNo">391</span> }<a name="line.391"></a>
+<span class="sourceLineNo">392</span> }<a name="line.392"></a>
+<span class="sourceLineNo">393</span>}<a name="line.393"></a>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dccdd274/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.CacheOnWriteType.html
----------------------------------------------------------------------
diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.CacheOnWriteType.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.CacheOnWriteType.html
index 4cd4f1b..21a78bc 100644
--- a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.CacheOnWriteType.html
+++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.CacheOnWriteType.html
@@ -168,75 +168,75 @@
<span class="sourceLineNo">160</span> Configuration conf = TEST_UTIL.getConfiguration();<a name="line.160"></a>
<span class="sourceLineNo">161</span> List<BlockCache> blockcaches = new ArrayList<>();<a name="line.161"></a>
<span class="sourceLineNo">162</span> // default<a name="line.162"></a>
-<span class="sourceLineNo">163</span> blockcaches.add(new CacheConfig(conf).getBlockCache());<a name="line.163"></a>
-<span class="sourceLineNo">164</span><a name="line.164"></a>
-<span class="sourceLineNo">165</span> //set LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME to 2.0f due to HBASE-16287<a name="line.165"></a>
-<span class="sourceLineNo">166</span> TEST_UTIL.getConfiguration().setFloat(LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME, 2.0f);<a name="line.166"></a>
-<span class="sourceLineNo">167</span> // memory<a name="line.167"></a>
-<span class="sourceLineNo">168</span> BlockCache lru = new LruBlockCache(128 * 1024 * 1024, 64 * 1024, TEST_UTIL.getConfiguration());<a name="line.168"></a>
-<span class="sourceLineNo">169</span> blockcaches.add(lru);<a name="line.169"></a>
-<span class="sourceLineNo">170</span><a name="line.170"></a>
-<span class="sourceLineNo">171</span> // bucket cache<a name="line.171"></a>
-<span class="sourceLineNo">172</span> FileSystem.get(conf).mkdirs(TEST_UTIL.getDataTestDir());<a name="line.172"></a>
-<span class="sourceLineNo">173</span> int[] bucketSizes =<a name="line.173"></a>
-<span class="sourceLineNo">174</span> { INDEX_BLOCK_SIZE, DATA_BLOCK_SIZE, BLOOM_BLOCK_SIZE, 64 * 1024, 128 * 1024 };<a name="line.174"></a>
-<span class="sourceLineNo">175</span> BlockCache bucketcache =<a name="line.175"></a>
-<span class="sourceLineNo">176</span> new BucketCache("offheap", 128 * 1024 * 1024, 64 * 1024, bucketSizes, 5, 64 * 100, null);<a name="line.176"></a>
-<span class="sourceLineNo">177</span> blockcaches.add(bucketcache);<a name="line.177"></a>
-<span class="sourceLineNo">178</span> return blockcaches;<a name="line.178"></a>
-<span class="sourceLineNo">179</span> }<a name="line.179"></a>
-<span class="sourceLineNo">180</span><a name="line.180"></a>
-<span class="sourceLineNo">181</span> @Parameters<a name="line.181"></a>
-<span class="sourceLineNo">182</span> public static Collection<Object[]> getParameters() throws IOException {<a name="line.182"></a>
-<span class="sourceLineNo">183</span> List<Object[]> params = new ArrayList<>();<a name="line.183"></a>
-<span class="sourceLineNo">184</span> for (BlockCache blockCache : getBlockCaches()) {<a name="line.184"></a>
-<span class="sourceLineNo">185</span> for (CacheOnWriteType cowType : CacheOnWriteType.values()) {<a name="line.185"></a>
-<span class="sourceLineNo">186</span> for (Compression.Algorithm compress : HBaseCommonTestingUtility.COMPRESSION_ALGORITHMS) {<a name="line.186"></a>
-<span class="sourceLineNo">187</span> for (boolean cacheCompressedData : new boolean[] { false, true }) {<a name="line.187"></a>
-<span class="sourceLineNo">188</span> params.add(new Object[] { cowType, compress, cacheCompressedData, blockCache });<a name="line.188"></a>
-<span class="sourceLineNo">189</span> }<a name="line.189"></a>
-<span class="sourceLineNo">190</span> }<a name="line.190"></a>
-<span class="sourceLineNo">191</span> }<a name="line.191"></a>
-<span class="sourceLineNo">192</span> }<a name="line.192"></a>
-<span class="sourceLineNo">193</span> return params;<a name="line.193"></a>
-<span class="sourceLineNo">194</span> }<a name="line.194"></a>
-<span class="sourceLineNo">195</span><a name="line.195"></a>
-<span class="sourceLineNo">196</span> private void clearBlockCache(BlockCache blockCache) throws InterruptedException {<a name="line.196"></a>
-<span class="sourceLineNo">197</span> if (blockCache instanceof LruBlockCache) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span> ((LruBlockCache) blockCache).clearCache();<a name="line.198"></a>
-<span class="sourceLineNo">199</span> } else {<a name="line.199"></a>
-<span class="sourceLineNo">200</span> // BucketCache may not return all cached blocks(blocks in write queue), so check it here.<a name="line.200"></a>
-<span class="sourceLineNo">201</span> for (int clearCount = 0; blockCache.getBlockCount() > 0; clearCount++) {<a name="line.201"></a>
-<span class="sourceLineNo">202</span> if (clearCount > 0) {<a name="line.202"></a>
-<span class="sourceLineNo">203</span> LOG.warn("clear block cache " + blockCache + " " + clearCount + " times, "<a name="line.203"></a>
-<span class="sourceLineNo">204</span> + blockCache.getBlockCount() + " blocks remaining");<a name="line.204"></a>
-<span class="sourceLineNo">205</span> Thread.sleep(10);<a name="line.205"></a>
-<span class="sourceLineNo">206</span> }<a name="line.206"></a>
-<span class="sourceLineNo">207</span> for (CachedBlock block : Lists.newArrayList(blockCache)) {<a name="line.207"></a>
-<span class="sourceLineNo">208</span> BlockCacheKey key = new BlockCacheKey(block.getFilename(), block.getOffset());<a name="line.208"></a>
-<span class="sourceLineNo">209</span> // CombinedBucketCache may need evict two times.<a name="line.209"></a>
-<span class="sourceLineNo">210</span> for (int evictCount = 0; blockCache.evictBlock(key); evictCount++) {<a name="line.210"></a>
-<span class="sourceLineNo">211</span> if (evictCount > 1) {<a name="line.211"></a>
-<span class="sourceLineNo">212</span> LOG.warn("evict block " + block + " in " + blockCache + " " + evictCount<a name="line.212"></a>
-<span class="sourceLineNo">213</span> + " times, maybe a bug here");<a name="line.213"></a>
-<span class="sourceLineNo">214</span> }<a name="line.214"></a>
-<span class="sourceLineNo">215</span> }<a name="line.215"></a>
-<span class="sourceLineNo">216</span> }<a name="line.216"></a>
-<span class="sourceLineNo">217</span> }<a name="line.217"></a>
-<span class="sourceLineNo">218</span> }<a name="line.218"></a>
-<span class="sourceLineNo">219</span> }<a name="line.219"></a>
-<span class="sourceLineNo">220</span><a name="line.220"></a>
-<span class="sourceLineNo">221</span> @Before<a name="line.221"></a>
-<span class="sourceLineNo">222</span> public void setUp() throws IOException {<a name="line.222"></a>
-<span class="sourceLineNo">223</span> conf = TEST_UTIL.getConfiguration();<a name="line.223"></a>
-<span class="sourceLineNo">224</span> this.conf.set("dfs.datanode.data.dir.perm", "700");<a name="line.224"></a>
-<span class="sourceLineNo">225</span> conf.setInt(HFileBlockIndex.MAX_CHUNK_SIZE_KEY, INDEX_BLOCK_SIZE);<a name="line.225"></a>
-<span class="sourceLineNo">226</span> conf.setInt(BloomFilterFactory.IO_STOREFILE_BLOOM_BLOCK_SIZE,<a name="line.226"></a>
-<span class="sourceLineNo">227</span> BLOOM_BLOCK_SIZE);<a name="line.227"></a>
-<span class="sourceLineNo">228</span> conf.setBoolean(CacheConfig.CACHE_DATA_BLOCKS_COMPRESSED_KEY, cacheCompressedData);<a name="line.228"></a>
-<span class="sourceLineNo">229</span> cowType.modifyConf(conf);<a name="line.229"></a>
-<span class="sourceLineNo">230</span> fs = HFileSystem.get(conf);<a name="line.230"></a>
-<span class="sourceLineNo">231</span> CacheConfig.GLOBAL_BLOCK_CACHE_INSTANCE = blockCache;<a name="line.231"></a>
+<span class="sourceLineNo">163</span> CacheConfig.instantiateBlockCache(conf);<a name="line.163"></a>
+<span class="sourceLineNo">164</span> blockcaches.add(new CacheConfig(conf).getBlockCache());<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span> //set LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME to 2.0f due to HBASE-16287<a name="line.166"></a>
+<span class="sourceLineNo">167</span> TEST_UTIL.getConfiguration().setFloat(LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME, 2.0f);<a name="line.167"></a>
+<span class="sourceLineNo">168</span> // memory<a name="line.168"></a>
+<span class="sourceLineNo">169</span> BlockCache lru = new LruBlockCache(128 * 1024 * 1024, 64 * 1024, TEST_UTIL.getConfiguration());<a name="line.169"></a>
+<span class="sourceLineNo">170</span> blockcaches.add(lru);<a name="line.170"></a>
+<span class="sourceLineNo">171</span><a name="line.171"></a>
+<span class="sourceLineNo">172</span> // bucket cache<a name="line.172"></a>
+<span class="sourceLineNo">173</span> FileSystem.get(conf).mkdirs(TEST_UTIL.getDataTestDir());<a name="line.173"></a>
+<span class="sourceLineNo">174</span> int[] bucketSizes =<a name="line.174"></a>
+<span class="sourceLineNo">175</span> { INDEX_BLOCK_SIZE, DATA_BLOCK_SIZE, BLOOM_BLOCK_SIZE, 64 * 1024, 128 * 1024 };<a name="line.175"></a>
+<span class="sourceLineNo">176</span> BlockCache bucketcache =<a name="line.176"></a>
+<span class="sourceLineNo">177</span> new BucketCache("offheap", 128 * 1024 * 1024, 64 * 1024, bucketSizes, 5, 64 * 100, null);<a name="line.177"></a>
+<span class="sourceLineNo">178</span> blockcaches.add(bucketcache);<a name="line.178"></a>
+<span class="sourceLineNo">179</span> return blockcaches;<a name="line.179"></a>
+<span class="sourceLineNo">180</span> }<a name="line.180"></a>
+<span class="sourceLineNo">181</span><a name="line.181"></a>
+<span class="sourceLineNo">182</span> @Parameters<a name="line.182"></a>
+<span class="sourceLineNo">183</span> public static Collection<Object[]> getParameters() throws IOException {<a name="line.183"></a>
+<span class="sourceLineNo">184</span> List<Object[]> params = new ArrayList<>();<a name="line.184"></a>
+<span class="sourceLineNo">185</span> for (BlockCache blockCache : getBlockCaches()) {<a name="line.185"></a>
+<span class="sourceLineNo">186</span> for (CacheOnWriteType cowType : CacheOnWriteType.values()) {<a name="line.186"></a>
+<span class="sourceLineNo">187</span> for (Compression.Algorithm compress : HBaseCommonTestingUtility.COMPRESSION_ALGORITHMS) {<a name="line.187"></a>
+<span class="sourceLineNo">188</span> for (boolean cacheCompressedData : new boolean[] { false, true }) {<a name="line.188"></a>
+<span class="sourceLineNo">189</span> params.add(new Object[] { cowType, compress, cacheCompressedData, blockCache });<a name="line.189"></a>
+<span class="sourceLineNo">190</span> }<a name="line.190"></a>
+<span class="sourceLineNo">191</span> }<a name="line.191"></a>
+<span class="sourceLineNo">192</span> }<a name="line.192"></a>
+<span class="sourceLineNo">193</span> }<a name="line.193"></a>
+<span class="sourceLineNo">194</span> return params;<a name="line.194"></a>
+<span class="sourceLineNo">195</span> }<a name="line.195"></a>
+<span class="sourceLineNo">196</span><a name="line.196"></a>
+<span class="sourceLineNo">197</span> private void clearBlockCache(BlockCache blockCache) throws InterruptedException {<a name="line.197"></a>
+<span class="sourceLineNo">198</span> if (blockCache instanceof LruBlockCache) {<a name="line.198"></a>
+<span class="sourceLineNo">199</span> ((LruBlockCache) blockCache).clearCache();<a name="line.199"></a>
+<span class="sourceLineNo">200</span> } else {<a name="line.200"></a>
+<span class="sourceLineNo">201</span> // BucketCache may not return all cached blocks(blocks in write queue), so check it here.<a name="line.201"></a>
+<span class="sourceLineNo">202</span> for (int clearCount = 0; blockCache.getBlockCount() > 0; clearCount++) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span> if (clearCount > 0) {<a name="line.203"></a>
+<span class="sourceLineNo">204</span> LOG.warn("clear block cache " + blockCache + " " + clearCount + " times, "<a name="line.204"></a>
+<span class="sourceLineNo">205</span> + blockCache.getBlockCount() + " blocks remaining");<a name="line.205"></a>
+<span class="sourceLineNo">206</span> Thread.sleep(10);<a name="line.206"></a>
+<span class="sourceLineNo">207</span> }<a name="line.207"></a>
+<span class="sourceLineNo">208</span> for (CachedBlock block : Lists.newArrayList(blockCache)) {<a name="line.208"></a>
+<span class="sourceLineNo">209</span> BlockCacheKey key = new BlockCacheKey(block.getFilename(), block.getOffset());<a name="line.209"></a>
+<span class="sourceLineNo">210</span> // CombinedBucketCache may need evict two times.<a name="line.210"></a>
+<span class="sourceLineNo">211</span> for (int evictCount = 0; blockCache.evictBlock(key); evictCount++) {<a name="line.211"></a>
+<span class="sourceLineNo">212</span> if (evictCount > 1) {<a name="line.212"></a>
+<span class="sourceLineNo">213</span> LOG.warn("evict block " + block + " in " + blockCache + " " + evictCount<a name="line.213"></a>
+<span class="sourceLineNo">214</span> + " times, maybe a bug here");<a name="line.214"></a>
+<span class="sourceLineNo">215</span> }<a name="line.215"></a>
+<span class="sourceLineNo">216</span> }<a name="line.216"></a>
+<span class="sourceLineNo">217</span> }<a name="line.217"></a>
+<span class="sourceLineNo">218</span> }<a name="line.218"></a>
+<span class="sourceLineNo">219</span> }<a name="line.219"></a>
+<span class="sourceLineNo">220</span> }<a name="line.220"></a>
+<span class="sourceLineNo">221</span><a name="line.221"></a>
+<span class="sourceLineNo">222</span> @Before<a name="line.222"></a>
+<span class="sourceLineNo">223</span> public void setUp() throws IOException {<a name="line.223"></a>
+<span class="sourceLineNo">224</span> conf = TEST_UTIL.getConfiguration();<a name="line.224"></a>
+<span class="sourceLineNo">225</span> this.conf.set("dfs.datanode.data.dir.perm", "700");<a name="line.225"></a>
+<span class="sourceLineNo">226</span> conf.setInt(HFileBlockIndex.MAX_CHUNK_SIZE_KEY, INDEX_BLOCK_SIZE);<a name="line.226"></a>
+<span class="sourceLineNo">227</span> conf.setInt(BloomFilterFactory.IO_STOREFILE_BLOOM_BLOCK_SIZE,<a name="line.227"></a>
+<span class="sourceLineNo">228</span> BLOOM_BLOCK_SIZE);<a name="line.228"></a>
+<span class="sourceLineNo">229</span> conf.setBoolean(CacheConfig.CACHE_DATA_BLOCKS_COMPRESSED_KEY, cacheCompressedData);<a name="line.229"></a>
+<span class="sourceLineNo">230</span> cowType.modifyConf(conf);<a name="line.230"></a>
+<span class="sourceLineNo">231</span> fs = HFileSystem.get(conf);<a name="line.231"></a>
<span class="sourceLineNo">232</span> cacheConf =<a name="line.232"></a>
<span class="sourceLineNo">233</span> new CacheConfig(blockCache, true, true, cowType.shouldBeCached(BlockType.DATA),<a name="line.233"></a>
<span class="sourceLineNo">234</span> cowType.shouldBeCached(BlockType.LEAF_INDEX),<a name="line.234"></a>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dccdd274/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.html
----------------------------------------------------------------------
diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.html
index 4cd4f1b..21a78bc 100644
--- a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.html
+++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestCacheOnWrite.html
@@ -168,75 +168,75 @@
<span class="sourceLineNo">160</span> Configuration conf = TEST_UTIL.getConfiguration();<a name="line.160"></a>
<span class="sourceLineNo">161</span> List<BlockCache> blockcaches = new ArrayList<>();<a name="line.161"></a>
<span class="sourceLineNo">162</span> // default<a name="line.162"></a>
-<span class="sourceLineNo">163</span> blockcaches.add(new CacheConfig(conf).getBlockCache());<a name="line.163"></a>
-<span class="sourceLineNo">164</span><a name="line.164"></a>
-<span class="sourceLineNo">165</span> //set LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME to 2.0f due to HBASE-16287<a name="line.165"></a>
-<span class="sourceLineNo">166</span> TEST_UTIL.getConfiguration().setFloat(LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME, 2.0f);<a name="line.166"></a>
-<span class="sourceLineNo">167</span> // memory<a name="line.167"></a>
-<span class="sourceLineNo">168</span> BlockCache lru = new LruBlockCache(128 * 1024 * 1024, 64 * 1024, TEST_UTIL.getConfiguration());<a name="line.168"></a>
-<span class="sourceLineNo">169</span> blockcaches.add(lru);<a name="line.169"></a>
-<span class="sourceLineNo">170</span><a name="line.170"></a>
-<span class="sourceLineNo">171</span> // bucket cache<a name="line.171"></a>
-<span class="sourceLineNo">172</span> FileSystem.get(conf).mkdirs(TEST_UTIL.getDataTestDir());<a name="line.172"></a>
-<span class="sourceLineNo">173</span> int[] bucketSizes =<a name="line.173"></a>
-<span class="sourceLineNo">174</span> { INDEX_BLOCK_SIZE, DATA_BLOCK_SIZE, BLOOM_BLOCK_SIZE, 64 * 1024, 128 * 1024 };<a name="line.174"></a>
-<span class="sourceLineNo">175</span> BlockCache bucketcache =<a name="line.175"></a>
-<span class="sourceLineNo">176</span> new BucketCache("offheap", 128 * 1024 * 1024, 64 * 1024, bucketSizes, 5, 64 * 100, null);<a name="line.176"></a>
-<span class="sourceLineNo">177</span> blockcaches.add(bucketcache);<a name="line.177"></a>
-<span class="sourceLineNo">178</span> return blockcaches;<a name="line.178"></a>
-<span class="sourceLineNo">179</span> }<a name="line.179"></a>
-<span class="sourceLineNo">180</span><a name="line.180"></a>
-<span class="sourceLineNo">181</span> @Parameters<a name="line.181"></a>
-<span class="sourceLineNo">182</span> public static Collection<Object[]> getParameters() throws IOException {<a name="line.182"></a>
-<span class="sourceLineNo">183</span> List<Object[]> params = new ArrayList<>();<a name="line.183"></a>
-<span class="sourceLineNo">184</span> for (BlockCache blockCache : getBlockCaches()) {<a name="line.184"></a>
-<span class="sourceLineNo">185</span> for (CacheOnWriteType cowType : CacheOnWriteType.values()) {<a name="line.185"></a>
-<span class="sourceLineNo">186</span> for (Compression.Algorithm compress : HBaseCommonTestingUtility.COMPRESSION_ALGORITHMS) {<a name="line.186"></a>
-<span class="sourceLineNo">187</span> for (boolean cacheCompressedData : new boolean[] { false, true }) {<a name="line.187"></a>
-<span class="sourceLineNo">188</span> params.add(new Object[] { cowType, compress, cacheCompressedData, blockCache });<a name="line.188"></a>
-<span class="sourceLineNo">189</span> }<a name="line.189"></a>
-<span class="sourceLineNo">190</span> }<a name="line.190"></a>
-<span class="sourceLineNo">191</span> }<a name="line.191"></a>
-<span class="sourceLineNo">192</span> }<a name="line.192"></a>
-<span class="sourceLineNo">193</span> return params;<a name="line.193"></a>
-<span class="sourceLineNo">194</span> }<a name="line.194"></a>
-<span class="sourceLineNo">195</span><a name="line.195"></a>
-<span class="sourceLineNo">196</span> private void clearBlockCache(BlockCache blockCache) throws InterruptedException {<a name="line.196"></a>
-<span class="sourceLineNo">197</span> if (blockCache instanceof LruBlockCache) {<a name="line.197"></a>
-<span class="sourceLineNo">198</span> ((LruBlockCache) blockCache).clearCache();<a name="line.198"></a>
-<span class="sourceLineNo">199</span> } else {<a name="line.199"></a>
-<span class="sourceLineNo">200</span> // BucketCache may not return all cached blocks(blocks in write queue), so check it here.<a name="line.200"></a>
-<span class="sourceLineNo">201</span> for (int clearCount = 0; blockCache.getBlockCount() > 0; clearCount++) {<a name="line.201"></a>
-<span class="sourceLineNo">202</span> if (clearCount > 0) {<a name="line.202"></a>
-<span class="sourceLineNo">203</span> LOG.warn("clear block cache " + blockCache + " " + clearCount + " times, "<a name="line.203"></a>
-<span class="sourceLineNo">204</span> + blockCache.getBlockCount() + " blocks remaining");<a name="line.204"></a>
-<span class="sourceLineNo">205</span> Thread.sleep(10);<a name="line.205"></a>
-<span class="sourceLineNo">206</span> }<a name="line.206"></a>
-<span class="sourceLineNo">207</span> for (CachedBlock block : Lists.newArrayList(blockCache)) {<a name="line.207"></a>
-<span class="sourceLineNo">208</span> BlockCacheKey key = new BlockCacheKey(block.getFilename(), block.getOffset());<a name="line.208"></a>
-<span class="sourceLineNo">209</span> // CombinedBucketCache may need evict two times.<a name="line.209"></a>
-<span class="sourceLineNo">210</span> for (int evictCount = 0; blockCache.evictBlock(key); evictCount++) {<a name="line.210"></a>
-<span class="sourceLineNo">211</span> if (evictCount > 1) {<a name="line.211"></a>
-<span class="sourceLineNo">212</span> LOG.warn("evict block " + block + " in " + blockCache + " " + evictCount<a name="line.212"></a>
-<span class="sourceLineNo">213</span> + " times, maybe a bug here");<a name="line.213"></a>
-<span class="sourceLineNo">214</span> }<a name="line.214"></a>
-<span class="sourceLineNo">215</span> }<a name="line.215"></a>
-<span class="sourceLineNo">216</span> }<a name="line.216"></a>
-<span class="sourceLineNo">217</span> }<a name="line.217"></a>
-<span class="sourceLineNo">218</span> }<a name="line.218"></a>
-<span class="sourceLineNo">219</span> }<a name="line.219"></a>
-<span class="sourceLineNo">220</span><a name="line.220"></a>
-<span class="sourceLineNo">221</span> @Before<a name="line.221"></a>
-<span class="sourceLineNo">222</span> public void setUp() throws IOException {<a name="line.222"></a>
-<span class="sourceLineNo">223</span> conf = TEST_UTIL.getConfiguration();<a name="line.223"></a>
-<span class="sourceLineNo">224</span> this.conf.set("dfs.datanode.data.dir.perm", "700");<a name="line.224"></a>
-<span class="sourceLineNo">225</span> conf.setInt(HFileBlockIndex.MAX_CHUNK_SIZE_KEY, INDEX_BLOCK_SIZE);<a name="line.225"></a>
-<span class="sourceLineNo">226</span> conf.setInt(BloomFilterFactory.IO_STOREFILE_BLOOM_BLOCK_SIZE,<a name="line.226"></a>
-<span class="sourceLineNo">227</span> BLOOM_BLOCK_SIZE);<a name="line.227"></a>
-<span class="sourceLineNo">228</span> conf.setBoolean(CacheConfig.CACHE_DATA_BLOCKS_COMPRESSED_KEY, cacheCompressedData);<a name="line.228"></a>
-<span class="sourceLineNo">229</span> cowType.modifyConf(conf);<a name="line.229"></a>
-<span class="sourceLineNo">230</span> fs = HFileSystem.get(conf);<a name="line.230"></a>
-<span class="sourceLineNo">231</span> CacheConfig.GLOBAL_BLOCK_CACHE_INSTANCE = blockCache;<a name="line.231"></a>
+<span class="sourceLineNo">163</span> CacheConfig.instantiateBlockCache(conf);<a name="line.163"></a>
+<span class="sourceLineNo">164</span> blockcaches.add(new CacheConfig(conf).getBlockCache());<a name="line.164"></a>
+<span class="sourceLineNo">165</span><a name="line.165"></a>
+<span class="sourceLineNo">166</span> //set LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME to 2.0f due to HBASE-16287<a name="line.166"></a>
+<span class="sourceLineNo">167</span> TEST_UTIL.getConfiguration().setFloat(LruBlockCache.LRU_HARD_CAPACITY_LIMIT_FACTOR_CONFIG_NAME, 2.0f);<a name="line.167"></a>
+<span class="sourceLineNo">168</span> // memory<a name="line.168"></a>
+<span class="sourceLineNo">169</span> BlockCache lru = new LruBlockCache(128 * 1024 * 1024, 64 * 1024, TEST_UTIL.getConfiguration());<a name="line.169"></a>
+<span class="sourceLineNo">170</span> blockcaches.add(lru);<a name="line.170"></a>
+<span class="sourceLineNo">171</span><a name="line.171"></a>
+<span class="sourceLineNo">172</span> // bucket cache<a name="line.172"></a>
+<span class="sourceLineNo">173</span> FileSystem.get(conf).mkdirs(TEST_UTIL.getDataTestDir());<a name="line.173"></a>
+<span class="sourceLineNo">174</span> int[] bucketSizes =<a name="line.174"></a>
+<span class="sourceLineNo">175</span> { INDEX_BLOCK_SIZE, DATA_BLOCK_SIZE, BLOOM_BLOCK_SIZE, 64 * 1024, 128 * 1024 };<a name="line.175"></a>
+<span class="sourceLineNo">176</span> BlockCache bucketcache =<a name="line.176"></a>
+<span class="sourceLineNo">177</span> new BucketCache("offheap", 128 * 1024 * 1024, 64 * 1024, bucketSizes, 5, 64 * 100, null);<a name="line.177"></a>
+<span class="sourceLineNo">178</span> blockcaches.add(bucketcache);<a name="line.178"></a>
+<span class="sourceLineNo">179</span> return blockcaches;<a name="line.179"></a>
+<span class="sourceLineNo">180</span> }<a name="line.180"></a>
+<span class="sourceLineNo">181</span><a name="line.181"></a>
+<span class="sourceLineNo">182</span> @Parameters<a name="line.182"></a>
+<span class="sourceLineNo">183</span> public static Collection<Object[]> getParameters() throws IOException {<a name="line.183"></a>
+<span class="sourceLineNo">184</span> List<Object[]> params = new ArrayList<>();<a name="line.184"></a>
+<span class="sourceLineNo">185</span> for (BlockCache blockCache : getBlockCaches()) {<a name="line.185"></a>
+<span class="sourceLineNo">186</span> for (CacheOnWriteType cowType : CacheOnWriteType.values()) {<a name="line.186"></a>
+<span class="sourceLineNo">187</span> for (Compression.Algorithm compress : HBaseCommonTestingUtility.COMPRESSION_ALGORITHMS) {<a name="line.187"></a>
+<span class="sourceLineNo">188</span> for (boolean cacheCompressedData : new boolean[] { false, true }) {<a name="line.188"></a>
+<span class="sourceLineNo">189</span> params.add(new Object[] { cowType, compress, cacheCompressedData, blockCache });<a name="line.189"></a>
+<span class="sourceLineNo">190</span> }<a name="line.190"></a>
+<span class="sourceLineNo">191</span> }<a name="line.191"></a>
+<span class="sourceLineNo">192</span> }<a name="line.192"></a>
+<span class="sourceLineNo">193</span> }<a name="line.193"></a>
+<span class="sourceLineNo">194</span> return params;<a name="line.194"></a>
+<span class="sourceLineNo">195</span> }<a name="line.195"></a>
+<span class="sourceLineNo">196</span><a name="line.196"></a>
+<span class="sourceLineNo">197</span> private void clearBlockCache(BlockCache blockCache) throws InterruptedException {<a name="line.197"></a>
+<span class="sourceLineNo">198</span> if (blockCache instanceof LruBlockCache) {<a name="line.198"></a>
+<span class="sourceLineNo">199</span> ((LruBlockCache) blockCache).clearCache();<a name="line.199"></a>
+<span class="sourceLineNo">200</span> } else {<a name="line.200"></a>
+<span class="sourceLineNo">201</span> // BucketCache may not return all cached blocks(blocks in write queue), so check it here.<a name="line.201"></a>
+<span class="sourceLineNo">202</span> for (int clearCount = 0; blockCache.getBlockCount() > 0; clearCount++) {<a name="line.202"></a>
+<span class="sourceLineNo">203</span> if (clearCount > 0) {<a name="line.203"></a>
+<span class="sourceLineNo">204</span> LOG.warn("clear block cache " + blockCache + " " + clearCount + " times, "<a name="line.204"></a>
+<span class="sourceLineNo">205</span> + blockCache.getBlockCount() + " blocks remaining");<a name="line.205"></a>
+<span class="sourceLineNo">206</span> Thread.sleep(10);<a name="line.206"></a>
+<span class="sourceLineNo">207</span> }<a name="line.207"></a>
+<span class="sourceLineNo">208</span> for (CachedBlock block : Lists.newArrayList(blockCache)) {<a name="line.208"></a>
+<span class="sourceLineNo">209</span> BlockCacheKey key = new BlockCacheKey(block.getFilename(), block.getOffset());<a name="line.209"></a>
+<span class="sourceLineNo">210</span> // CombinedBucketCache may need evict two times.<a name="line.210"></a>
+<span class="sourceLineNo">211</span> for (int evictCount = 0; blockCache.evictBlock(key); evictCount++) {<a name="line.211"></a>
+<span class="sourceLineNo">212</span> if (evictCount > 1) {<a name="line.212"></a>
+<span class="sourceLineNo">213</span> LOG.warn("evict block " + block + " in " + blockCache + " " + evictCount<a name="line.213"></a>
+<span class="sourceLineNo">214</span> + " times, maybe a bug here");<a name="line.214"></a>
+<span class="sourceLineNo">215</span> }<a name="line.215"></a>
+<span class="sourceLineNo">216</span> }<a name="line.216"></a>
+<span class="sourceLineNo">217</span> }<a name="line.217"></a>
+<span class="sourceLineNo">218</span> }<a name="line.218"></a>
+<span class="sourceLineNo">219</span> }<a name="line.219"></a>
+<span class="sourceLineNo">220</span> }<a name="line.220"></a>
+<span class="sourceLineNo">221</span><a name="line.221"></a>
+<span class="sourceLineNo">222</span> @Before<a name="line.222"></a>
+<span class="sourceLineNo">223</span> public void setUp() throws IOException {<a name="line.223"></a>
+<span class="sourceLineNo">224</span> conf = TEST_UTIL.getConfiguration();<a name="line.224"></a>
+<span class="sourceLineNo">225</span> this.conf.set("dfs.datanode.data.dir.perm", "700");<a name="line.225"></a>
+<span class="sourceLineNo">226</span> conf.setInt(HFileBlockIndex.MAX_CHUNK_SIZE_KEY, INDEX_BLOCK_SIZE);<a name="line.226"></a>
+<span class="sourceLineNo">227</span> conf.setInt(BloomFilterFactory.IO_STOREFILE_BLOOM_BLOCK_SIZE,<a name="line.227"></a>
+<span class="sourceLineNo">228</span> BLOOM_BLOCK_SIZE);<a name="line.228"></a>
+<span class="sourceLineNo">229</span> conf.setBoolean(CacheConfig.CACHE_DATA_BLOCKS_COMPRESSED_KEY, cacheCompressedData);<a name="line.229"></a>
+<span class="sourceLineNo">230</span> cowType.modifyConf(conf);<a name="line.230"></a>
+<span class="sourceLineNo">231</span> fs = HFileSystem.get(conf);<a name="line.231"></a>
<span class="sourceLineNo">232</span> cacheConf =<a name="line.232"></a>
<span class="sourceLineNo">233</span> new CacheConfig(blockCache, true, true, cowType.shouldBeCached(BlockType.DATA),<a name="line.233"></a>
<span class="sourceLineNo">234</span> cowType.shouldBeCached(BlockType.LEAF_INDEX),<a name="line.234"></a>
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dccdd274/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.html
----------------------------------------------------------------------
diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.html
index b16e1c7..169b747 100644
--- a/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.html
+++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.html
@@ -114,50 +114,51 @@
<span class="sourceLineNo">106</span> // Make sure we make a new one each time.<a name="line.106"></a>
<span class="sourceLineNo">107</span> CacheConfig.clearGlobalInstances();<a name="line.107"></a>
<span class="sourceLineNo">108</span> HFile.DATABLOCK_READ_COUNT.reset();<a name="line.108"></a>
-<span class="sourceLineNo">109</span> }<a name="line.109"></a>
-<span class="sourceLineNo">110</span><a name="line.110"></a>
-<span class="sourceLineNo">111</span> @Test<a name="line.111"></a>
-<span class="sourceLineNo">112</span> public void testCacheBlocks() throws IOException {<a name="line.112"></a>
-<span class="sourceLineNo">113</span> // Set index block size to be the same as normal block size.<a name="line.113"></a>
-<span class="sourceLineNo">114</span> TEST_UTIL.getConfiguration().setInt(HFileBlockIndex.MAX_CHUNK_SIZE_KEY, BLOCK_SIZE);<a name="line.114"></a>
-<span class="sourceLineNo">115</span> HColumnDescriptor hcd = new HColumnDescriptor(Bytes.toBytes(CF)).setMaxVersions(MAX_VERSIONS).<a name="line.115"></a>
-<span class="sourceLineNo">116</span> setCompressionType(COMPRESSION_ALGORITHM).<a name="line.116"></a>
-<span class="sourceLineNo">117</span> setBloomFilterType(BLOOM_TYPE);<a name="line.117"></a>
-<span class="sourceLineNo">118</span> hcd.setBlocksize(BLOCK_SIZE);<a name="line.118"></a>
-<span class="sourceLineNo">119</span> hcd.setBlockCacheEnabled(cfCacheEnabled);<a name="line.119"></a>
-<span class="sourceLineNo">120</span> HRegion region = TEST_UTIL.createTestRegion(TABLE, hcd);<a name="line.120"></a>
-<span class="sourceLineNo">121</span> BlockCache cache = region.getStore(hcd.getName()).getCacheConfig().getBlockCache();<a name="line.121"></a>
-<span class="sourceLineNo">122</span> CacheStats stats = cache.getStats();<a name="line.122"></a>
-<span class="sourceLineNo">123</span> writeTestData(region);<a name="line.123"></a>
-<span class="sourceLineNo">124</span> assertEquals(0, stats.getHitCount());<a name="line.124"></a>
-<span class="sourceLineNo">125</span> assertEquals(0, HFile.DATABLOCK_READ_COUNT.sum());<a name="line.125"></a>
-<span class="sourceLineNo">126</span> // Do a single get, take count of caches. If we are NOT caching DATA blocks, the miss<a name="line.126"></a>
-<span class="sourceLineNo">127</span> // count should go up. Otherwise, all should be cached and the miss count should not rise.<a name="line.127"></a>
-<span class="sourceLineNo">128</span> region.get(new Get(Bytes.toBytes("row" + 0)));<a name="line.128"></a>
-<span class="sourceLineNo">129</span> assertTrue(stats.getHitCount() > 0);<a name="line.129"></a>
-<span class="sourceLineNo">130</span> assertTrue(HFile.DATABLOCK_READ_COUNT.sum() > 0);<a name="line.130"></a>
-<span class="sourceLineNo">131</span> long missCount = stats.getMissCount();<a name="line.131"></a>
-<span class="sourceLineNo">132</span> region.get(new Get(Bytes.toBytes("row" + 0)));<a name="line.132"></a>
-<span class="sourceLineNo">133</span> if (this.cfCacheEnabled) assertEquals(missCount, stats.getMissCount());<a name="line.133"></a>
-<span class="sourceLineNo">134</span> else assertTrue(stats.getMissCount() > missCount);<a name="line.134"></a>
-<span class="sourceLineNo">135</span> }<a name="line.135"></a>
-<span class="sourceLineNo">136</span><a name="line.136"></a>
-<span class="sourceLineNo">137</span> private void writeTestData(HRegion region) throws IOException {<a name="line.137"></a>
-<span class="sourceLineNo">138</span> for (int i = 0; i < NUM_ROWS; ++i) {<a name="line.138"></a>
-<span class="sourceLineNo">139</span> Put put = new Put(Bytes.toBytes("row" + i));<a name="line.139"></a>
-<span class="sourceLineNo">140</span> for (int j = 0; j < NUM_COLS_PER_ROW; ++j) {<a name="line.140"></a>
-<span class="sourceLineNo">141</span> for (long ts = 1; ts < NUM_TIMESTAMPS_PER_COL; ++ts) {<a name="line.141"></a>
-<span class="sourceLineNo">142</span> put.addColumn(CF_BYTES, Bytes.toBytes("col" + j), ts,<a name="line.142"></a>
-<span class="sourceLineNo">143</span> Bytes.toBytes("value" + i + "_" + j + "_" + ts));<a name="line.143"></a>
-<span class="sourceLineNo">144</span> }<a name="line.144"></a>
-<span class="sourceLineNo">145</span> }<a name="line.145"></a>
-<span class="sourceLineNo">146</span> region.put(put);<a name="line.146"></a>
-<span class="sourceLineNo">147</span> if ((i + 1) % ROWS_PER_HFILE == 0) {<a name="line.147"></a>
-<span class="sourceLineNo">148</span> region.flush(true);<a name="line.148"></a>
-<span class="sourceLineNo">149</span> }<a name="line.149"></a>
-<span class="sourceLineNo">150</span> }<a name="line.150"></a>
-<span class="sourceLineNo">151</span> }<a name="line.151"></a>
-<span class="sourceLineNo">152</span>}<a name="line.152"></a>
+<span class="sourceLineNo">109</span> CacheConfig.instantiateBlockCache(TEST_UTIL.getConfiguration());<a name="line.109"></a>
+<span class="sourceLineNo">110</span> }<a name="line.110"></a>
+<span class="sourceLineNo">111</span><a name="line.111"></a>
+<span class="sourceLineNo">112</span> @Test<a name="line.112"></a>
+<span class="sourceLineNo">113</span> public void testCacheBlocks() throws IOException {<a name="line.113"></a>
+<span class="sourceLineNo">114</span> // Set index block size to be the same as normal block size.<a name="line.114"></a>
+<span class="sourceLineNo">115</span> TEST_UTIL.getConfiguration().setInt(HFileBlockIndex.MAX_CHUNK_SIZE_KEY, BLOCK_SIZE);<a name="line.115"></a>
+<span class="sourceLineNo">116</span> HColumnDescriptor hcd = new HColumnDescriptor(Bytes.toBytes(CF)).setMaxVersions(MAX_VERSIONS).<a name="line.116"></a>
+<span class="sourceLineNo">117</span> setCompressionType(COMPRESSION_ALGORITHM).<a name="line.117"></a>
+<span class="sourceLineNo">118</span> setBloomFilterType(BLOOM_TYPE);<a name="line.118"></a>
+<span class="sourceLineNo">119</span> hcd.setBlocksize(BLOCK_SIZE);<a name="line.119"></a>
+<span class="sourceLineNo">120</span> hcd.setBlockCacheEnabled(cfCacheEnabled);<a name="line.120"></a>
+<span class="sourceLineNo">121</span> HRegion region = TEST_UTIL.createTestRegion(TABLE, hcd);<a name="line.121"></a>
+<span class="sourceLineNo">122</span> BlockCache cache = region.getStore(hcd.getName()).getCacheConfig().getBlockCache();<a name="line.122"></a>
+<span class="sourceLineNo">123</span> CacheStats stats = cache.getStats();<a name="line.123"></a>
+<span class="sourceLineNo">124</span> writeTestData(region);<a name="line.124"></a>
+<span class="sourceLineNo">125</span> assertEquals(0, stats.getHitCount());<a name="line.125"></a>
+<span class="sourceLineNo">126</span> assertEquals(0, HFile.DATABLOCK_READ_COUNT.sum());<a name="line.126"></a>
+<span class="sourceLineNo">127</span> // Do a single get, take count of caches. If we are NOT caching DATA blocks, the miss<a name="line.127"></a>
+<span class="sourceLineNo">128</span> // count should go up. Otherwise, all should be cached and the miss count should not rise.<a name="line.128"></a>
+<span class="sourceLineNo">129</span> region.get(new Get(Bytes.toBytes("row" + 0)));<a name="line.129"></a>
+<span class="sourceLineNo">130</span> assertTrue(stats.getHitCount() > 0);<a name="line.130"></a>
+<span class="sourceLineNo">131</span> assertTrue(HFile.DATABLOCK_READ_COUNT.sum() > 0);<a name="line.131"></a>
+<span class="sourceLineNo">132</span> long missCount = stats.getMissCount();<a name="line.132"></a>
+<span class="sourceLineNo">133</span> region.get(new Get(Bytes.toBytes("row" + 0)));<a name="line.133"></a>
+<span class="sourceLineNo">134</span> if (this.cfCacheEnabled) assertEquals(missCount, stats.getMissCount());<a name="line.134"></a>
+<span class="sourceLineNo">135</span> else assertTrue(stats.getMissCount() > missCount);<a name="line.135"></a>
+<span class="sourceLineNo">136</span> }<a name="line.136"></a>
+<span class="sourceLineNo">137</span><a name="line.137"></a>
+<span class="sourceLineNo">138</span> private void writeTestData(HRegion region) throws IOException {<a name="line.138"></a>
+<span class="sourceLineNo">139</span> for (int i = 0; i < NUM_ROWS; ++i) {<a name="line.139"></a>
+<span class="sourceLineNo">140</span> Put put = new Put(Bytes.toBytes("row" + i));<a name="line.140"></a>
+<span class="sourceLineNo">141</span> for (int j = 0; j < NUM_COLS_PER_ROW; ++j) {<a name="line.141"></a>
+<span class="sourceLineNo">142</span> for (long ts = 1; ts < NUM_TIMESTAMPS_PER_COL; ++ts) {<a name="line.142"></a>
+<span class="sourceLineNo">143</span> put.addColumn(CF_BYTES, Bytes.toBytes("col" + j), ts,<a name="line.143"></a>
+<span class="sourceLineNo">144</span> Bytes.toBytes("value" + i + "_" + j + "_" + ts));<a name="line.144"></a>
+<span class="sourceLineNo">145</span> }<a name="line.145"></a>
+<span class="sourceLineNo">146</span> }<a name="line.146"></a>
+<span class="sourceLineNo">147</span> region.put(put);<a name="line.147"></a>
+<span class="sourceLineNo">148</span> if ((i + 1) % ROWS_PER_HFILE == 0) {<a name="line.148"></a>
+<span class="sourceLineNo">149</span> region.flush(true);<a name="line.149"></a>
+<span class="sourceLineNo">150</span> }<a name="line.150"></a>
+<span class="sourceLineNo">151</span> }<a name="line.151"></a>
+<span class="sourceLineNo">152</span> }<a name="line.152"></a>
+<span class="sourceLineNo">153</span>}<a name="line.153"></a>