You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@cassandra.apache.org by Oleg Krayushkin <al...@gmail.com> on 2016/10/20 14:29:23 UTC
strange node load decrease after nodetool repair -pr
Hi. After I've run token-ranged repair from node at 12.5.13.125 with
nodetool repair -full -st ${start_tokens[i]} -et ${end_tokens[i]}
on every token range, I got this node load:
-- Address Load Tokens Owns Rack
UN 12.5.13.141 23.94 GB 256 32.3% rack1
DN 12.5.13.125 34.71 GB 256 31.8% rack1
UN 12.5.13.46 29.01 GB 512 58.1% rack1
UN 12.5.13.228 41.17 GB 512 58.5% rack1
UN 12.5.13.34 45.93 GB 512 59.8% rack1
UN 12.5.13.82 42.05 GB 512 59.4% rack1
Then I've run partitioner-range repair from the same node with
nodetool repair -full -pr
And unexpectedly I got such a different load:
-- Address Load Tokens Owns Rack
UN 12.5.13.141 22.93 GB 256 32.3% rack1
UN 12.5.13.125 30.94 GB 256 31.8% rack1
UN 12.5.13.46 27.38 GB 512 58.1% rack1
UN 12.5.13.228 39.51 GB 512 58.5% rack1
UN 12.5.13.34 41.58 GB 512 59.8% rack1
UN 12.5.13.82 33.9 GB 512 59.4% rack1
What are posible reasons of such load decrease after last repair? Maybe
some compaction, that were not done after token-ranged repairs? But at
12.5.13.82 gone about 8GB!
Additional info:
- There were no writes to db during these periods.
- All repair operations completed without errors, exceptions or fails.
- Before the first repair I've done sstablescrub on every node -- maybe
this gives a clue?
- cassandra version is 3.0.8
--
Oleg Krayushkin