You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by "Marouane RAJI (JIRA)" <ji...@apache.org> on 2019/08/13 09:32:00 UTC

[jira] [Created] (KAFKA-8796) A broker joining the cluster should be able to replicate without impacting the cluster

Marouane RAJI created KAFKA-8796:
------------------------------------

             Summary: A broker joining the cluster should be able to replicate without impacting the cluster
                 Key: KAFKA-8796
                 URL: https://issues.apache.org/jira/browse/KAFKA-8796
             Project: Kafka
          Issue Type: Bug
    Affects Versions: 1.1.0
            Reporter: Marouane RAJI
         Attachments: image-2019-08-13-10-26-19-282.png, image-2019-08-13-10-28-42-337.png

Hi, 

We run a cluster of 50 brokers, 1.4M msgs/sec at max, on AWS. We were using m4.2xlarge. We are now moving to m5.2xlarge. Everytime we replace a broker from scratch (EBSs are linked to ec2 instance..), the byte sent on the replaced broker increase significantly and that seem to impact the cluster, increasing the produce time and fetch time..

This is our configuration per broker :

 

 
{code:java}
broker.id=11
############################# Socket Server Settings #############################
# The port the socket server listens on
port=9092

advertised.host.name=ec2-xx-xx-xx-xx.eu-west-1.compute.amazonaws.com
# The number of threads handling network requests
num.network.threads=32
# The number of threads doing disk I/O
num.io.threads=16socket server socket.receive.buffer.bytes=1048576 

socket.request.max.bytes=104857600 # The max time a connection can be idle connections.max.idle.ms = 60000 

num.partitions=2 

default.replication.factor=2 

auto.leader.rebalance.enable=true 

delete.topic.enable=true 

compression.type=producer 

log.message.format.version=0.9.0.1


message.max.bytes=8000000 
# The minimum age of a log file to be eligible for deletion log.retention.hours=48 

log.retention.bytes=3000000000 

log.segment.bytes=268435456 

log.retention.check.interval.ms=60000  

log.cleaner.enable=true 

log.cleaner.dedupe.buffer.size=268435456

replica.fetch.max.bytes=8388608 

replica.fetch.wait.max.ms=500 

replica.lag.time.max.ms=10000 

num.replica.fetchers = 3 

# Auto creation of topics on the server 
auto.create.topics.enable=true 

controlled.shutdown.enable=true 

inter.broker.protocol.version=0.10.2 

unclean.leader.election.enabled=True
{code}
 

This is what we notice on replication :

I high increase in byte received on the replaced broker

 

!image-2019-08-13-10-26-19-282.png!

!image-2019-08-13-10-28-42-337.png!

You can't see it the graph above but the increase in produce time stayed high for 20minutes..

We didn't see anything out of the ordinary in the logs.

Please let us know if there is anything wrong in our config or if it is a potential issue that needs fixing with kafka. 

Thanks.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)