You are viewing a plain text version of this content. The canonical link for it is here.
Posted to jira@kafka.apache.org by "Evan Williams (Jira)" <ji...@apache.org> on 2020/02/03 16:30:00 UTC

[jira] [Comment Edited] (KAFKA-8796) A broker joining the cluster should be able to replicate without impacting the cluster

    [ https://issues.apache.org/jira/browse/KAFKA-8796?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17029073#comment-17029073 ] 

Evan Williams edited comment on KAFKA-8796 at 2/3/20 4:29 PM:
--------------------------------------------------------------

[~dajac] Must you enforce throttling at both the topic and broker levels ? Or can you just enforce it at the broker level, which would be much simpler to automate ?

Ideally the following can be set on the brokers, to limit replication and hopefully lower CPU.


{{leader.replication.throttled.rate}}
{{follower.replication.throttled.rate}}


was (Author: blodsbror):
[~dajac] Must you enforce throttling at both the topic and broker levels ? Or can you just enforce it at the broker level, which would be much simpler to automate ?

> A broker joining the cluster should be able to replicate without impacting the cluster
> --------------------------------------------------------------------------------------
>
>                 Key: KAFKA-8796
>                 URL: https://issues.apache.org/jira/browse/KAFKA-8796
>             Project: Kafka
>          Issue Type: Bug
>    Affects Versions: 1.1.0
>            Reporter: Marouane RAJI
>            Priority: Major
>         Attachments: image-2019-08-13-10-26-19-282.png, image-2019-08-13-10-28-42-337.png
>
>
> Hi, 
> We run a cluster of 50 brokers, 1.4M msgs/sec at max, on AWS. We were using m4.2xlarge. We are now moving to m5.2xlarge. Everytime we replace a broker from scratch (EBSs are linked to ec2 instance..), the byte sent on the replaced broker increase significantly and that seem to impact the cluster, increasing the produce time and fetch time..
> This is our configuration per broker :
>  
>  
> {code:java}
> broker.id=11
> ############################# Socket Server Settings #############################
> # The port the socket server listens on
> port=9092
> advertised.host.name=ec2-xx-xx-xx-xx.eu-west-1.compute.amazonaws.com
> # The number of threads handling network requests
> num.network.threads=32
> # The number of threads doing disk I/O
> num.io.threads=16socket server socket.receive.buffer.bytes=1048576 
> socket.request.max.bytes=104857600 # The max time a connection can be idle connections.max.idle.ms = 60000 
> num.partitions=2 
> default.replication.factor=2 
> auto.leader.rebalance.enable=true 
> delete.topic.enable=true 
> compression.type=producer 
> log.message.format.version=0.9.0.1
> message.max.bytes=8000000 
> # The minimum age of a log file to be eligible for deletion log.retention.hours=48 
> log.retention.bytes=3000000000 
> log.segment.bytes=268435456 
> log.retention.check.interval.ms=60000  
> log.cleaner.enable=true 
> log.cleaner.dedupe.buffer.size=268435456
> replica.fetch.max.bytes=8388608 
> replica.fetch.wait.max.ms=500 
> replica.lag.time.max.ms=10000 
> num.replica.fetchers = 3 
> # Auto creation of topics on the server 
> auto.create.topics.enable=true 
> controlled.shutdown.enable=true 
> inter.broker.protocol.version=0.10.2 
> unclean.leader.election.enabled=True
> {code}
>  
> This is what we notice on replication :
> I high increase in byte received on the replaced broker
>  
> !image-2019-08-13-10-26-19-282.png!
> !image-2019-08-13-10-28-42-337.png!
> You can't see it the graph above but the increase in produce time stayed high for 20minutes..
> We didn't see anything out of the ordinary in the logs.
> Please let us know if there is anything wrong in our config or if it is a potential issue that needs fixing with kafka. 
> Thanks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)