You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@cassandra.apache.org by nowarry <no...@gmail.com> on 2015/07/06 22:45:44 UTC
Wrong peers
Hey guys,
I'm using Ruby driver( http://datastax.github.io/ruby-driver/ ) for backup scripts. I tried to discover all peers and got wrong peers that are different with nodetool status.
=================
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
-- Address Load Tokens Owns Host ID Rack
UN 10.40.231.53 1.18 TB 256 ? b2d877d7-f031-4190-8569-976bb0ce034f RACK01
UN 10.40.231.11 1.24 TB 256 ? e15cda1c-65cc-40cb-b85c-c4bd665d02d7 RACK01
cqlsh> use system;
cqlsh:system> select peer from system.peers;
peer
--------------
10.40.231.31
10.40.231.53
(2 rows)
What to do with these old peers, whether they can be removed without consequences since they are not in production cluster? And how to keep up to date the peers?
--------------------------------------
Anton Koshevoy
Re: Wrong peers
Posted by Carlos Rolo <ro...@pythian.com>.
There is a bug in Jira related to this, it is not a driver issue, is a
Cassandra issue. It is solved on 2.0.14 I think. I will post the ticket
once I find it.
Regards,
Carlos Juzarte Rolo
Cassandra Consultant
Pythian - Love your data
rolo@pythian | Twitter: cjrolo | Linkedin: *linkedin.com/in/carlosjuzarterolo
<http://linkedin.com/in/carlosjuzarterolo>*
Mobile: +31 6 159 61 814 | Tel: +1 613 565 8696 x1649
www.pythian.com
On Mon, Jul 6, 2015 at 10:50 PM, Jeff Williams <je...@wherethebitsroam.com>
wrote:
> Anton,
>
> I have also seen this issue with decommissioned nodes remaining in the
> system.peers table.
>
> On the bright side, they can be safely removed from the system.peers table
> without issue. You will have to check every node in the cluster since this
> is a local setting per node.
>
> Jeff
>
> On 6 July 2015 at 22:45, nowarry <no...@gmail.com> wrote:
>
>> Hey guys,
>>
>> I'm using Ruby driver( http://datastax.github.io/ruby-driver/ ) for
>> backup scripts. I tried to discover all peers and got wrong peers that are
>> different with nodetool status.
>>
>> =================
>> Status=Up/Down
>> |/ State=Normal/Leaving/Joining/Moving
>> -- Address Load Tokens Owns Host ID
>> Rack
>> UN 10.40.231.53 1.18 TB 256 ?
>> b2d877d7-f031-4190-8569-976bb0ce034f RACK01
>> UN 10.40.231.11 1.24 TB 256 ?
>> e15cda1c-65cc-40cb-b85c-c4bd665d02d7 RACK01
>>
>> cqlsh> use system;
>> cqlsh:system> select peer from system.peers;
>>
>> peer
>> --------------
>> 10.40.231.31
>> 10.40.231.53
>>
>> (2 rows)
>>
>> What to do with these old peers, whether they can be removed without
>> consequences since they are not in production cluster? And how to keep up
>> to date the peers?
>>
>> --------------------------------------
>> Anton Koshevoy
>>
>>
>
--
--
Re: Wrong peers
Posted by Jeff Williams <je...@wherethebitsroam.com>.
Anton,
I have also seen this issue with decommissioned nodes remaining in the
system.peers table.
On the bright side, they can be safely removed from the system.peers table
without issue. You will have to check every node in the cluster since this
is a local setting per node.
Jeff
On 6 July 2015 at 22:45, nowarry <no...@gmail.com> wrote:
> Hey guys,
>
> I'm using Ruby driver( http://datastax.github.io/ruby-driver/ ) for
> backup scripts. I tried to discover all peers and got wrong peers that are
> different with nodetool status.
>
> =================
> Status=Up/Down
> |/ State=Normal/Leaving/Joining/Moving
> -- Address Load Tokens Owns Host ID
> Rack
> UN 10.40.231.53 1.18 TB 256 ?
> b2d877d7-f031-4190-8569-976bb0ce034f RACK01
> UN 10.40.231.11 1.24 TB 256 ?
> e15cda1c-65cc-40cb-b85c-c4bd665d02d7 RACK01
>
> cqlsh> use system;
> cqlsh:system> select peer from system.peers;
>
> peer
> --------------
> 10.40.231.31
> 10.40.231.53
>
> (2 rows)
>
> What to do with these old peers, whether they can be removed without
> consequences since they are not in production cluster? And how to keep up
> to date the peers?
>
> --------------------------------------
> Anton Koshevoy
>
>