You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@kafka.apache.org by Jason Gustafson <ja...@confluent.io> on 2016/02/18 19:07:23 UTC

Re: 0.9 client AbstractCoordinator - Attempt to join group failed due to obsolete coordinator information

Hi Gary,

The coordinator is a special broker which is chosen for each consumer group
to manage its state. It facilitates group membership, partition assignment
and offset commits. If the coordinator is shutdown, then Kafka will choose
another broker to assume the role. The log message might be a little
unclear, but basically it is saying that the broker the consumer thought
was its coordinator no longer is (maybe because it's shutting down). Note
this is not an error: the consumer will rediscover the new coordinator and
continue happily along.

Thanks,
Jason

On Tue, Feb 16, 2016 at 7:12 PM, Gary Struthers <ag...@earthlink.net>
wrote:

> Hi,
>
> My local Java client consumer and producer fail with log messages I don’t
> understand. What does "obsolete coordinator information” mean?
>
> 2016-02-16 18:49:01,795 INFO  o.a.kafka.common.utils.AppInfoParser  -
> Kafka version : 0.9.0.0
> 2016-02-16 18:49:01,795 INFO  o.a.kafka.common.utils.AppInfoParser  -
> Kafka commitId : fc7243c2af4b2b4a
> 2016-02-16 18:49:02,756 INFO  o.a.k.c.c.i.AbstractCoordinator  - Marking
> the coordinator 2147483647 dead.
> 2016-02-16 18:49:02,759 INFO  o.a.k.c.c.i.AbstractCoordinator  - Attempt
> to join group dendrites-group failed due to obsolete coordinator
> information, retrying.
> 2016-02-16 18:50:01,881 INFO  o.a.k.clients.producer.KafkaProducer  -
> Closing the Kafka producer with timeoutMillis = 100 ms.
>
> Here are the logged configs
> 2016-02-16 18:49:01,093 INFO  o.a.k.c.producer.ProducerConfig  -
> ProducerConfig values:
>         compression.type = none
>         metric.reporters = []
>         metadata.max.age.ms = 300000
>         metadata.fetch.timeout.ms = 60000
>         reconnect.backoff.ms = 50
>         sasl.kerberos.ticket.renew.window.factor = 0.8
>         bootstrap.servers = [localhost:9092]
>         retry.backoff.ms = 100
>         sasl.kerberos.kinit.cmd = /usr/bin/kinit
>         buffer.memory = 33554432
>         timeout.ms = 30000
>         key.serializer = class
> org.apache.kafka.common.serialization.StringSerializer
>         sasl.kerberos.service.name = null
>         sasl.kerberos.ticket.renew.jitter = 0.05
>         ssl.keystore.type = JKS
>         ssl.trustmanager.algorithm = PKIX
>         block.on.buffer.full = false
>         ssl.key.password = null
>         max.block.ms = 60000
>         sasl.kerberos.min.time.before.relogin = 60000
>         connections.max.idle.ms = 540000
>         ssl.truststore.password = null
>         max.in.flight.requests.per.connection = 5
>         metrics.num.samples = 2
>         client.id =
>         ssl.endpoint.identification.algorithm = null
>         ssl.protocol = TLS
>         request.timeout.ms = 30000
>         ssl.provider = null
>         ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
>         acks = all
>         batch.size = 16384
>         ssl.keystore.location = null
>         receive.buffer.bytes = 32768
>         ssl.cipher.suites = null
>         ssl.truststore.type = JKS
>         security.protocol = PLAINTEXT
>         retries = 0
>         max.request.size = 1048576
>         value.serializer = class
> org.apache.kafka.common.serialization.ByteArraySerializer
>         ssl.truststore.location = null
>         ssl.keystore.password = null
>         ssl.keymanager.algorithm = SunX509
>         metrics.sample.window.ms = 30000
>         partitioner.class = class
> org.apache.kafka.clients.producer.internals.DefaultPartitioner
>         send.buffer.bytes = 131072
>         linger.ms = 1
>
> 2016-02-16 18:49:01,217 INFO  o.a.k.c.consumer.ConsumerConfig  -
> ConsumerConfig values:
>         metric.reporters = []
>         metadata.max.age.ms = 300000
>         value.deserializer = class
> org.apache.kafka.common.serialization.ByteArrayDeserializer
>         group.id = dendrites-group
>         partition.assignment.strategy =
> [org.apache.kafka.clients.consumer.RangeAssignor]
>         reconnect.backoff.ms = 50
>         sasl.kerberos.ticket.renew.window.factor = 0.8
>         max.partition.fetch.bytes = 1048576
>         bootstrap.servers = [localhost:9092]
>         retry.backoff.ms = 100
>         sasl.kerberos.kinit.cmd = /usr/bin/kinit
>         sasl.kerberos.service.name = null
>         sasl.kerberos.ticket.renew.jitter = 0.05
>         ssl.keystore.type = JKS
>         ssl.trustmanager.algorithm = PKIX
>         enable.auto.commit = false
>         ssl.key.password = null
>         fetch.max.wait.ms = 500
>         sasl.kerberos.min.time.before.relogin = 60000
>         connections.max.idle.ms = 540000
>         ssl.truststore.password = null
>         session.timeout.ms = 30000
>         metrics.num.samples = 2
>         client.id =
>         ssl.endpoint.identification.algorithm = null
>         key.deserializer = class
> org.apache.kafka.common.serialization.StringDeserializer
>         ssl.protocol = TLS
>         check.crcs = true
>         request.timeout.ms = 40000
>         ssl.provider = null
>         ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
>         ssl.keystore.location = null
>         heartbeat.interval.ms = 3000
>         auto.commit.interval.ms = 1000
>         receive.buffer.bytes = 32768
>         ssl.cipher.suites = null
>         ssl.truststore.type = JKS
>         security.protocol = PLAINTEXT
>         ssl.truststore.location = null
>         ssl.keystore.password = null
>         ssl.keymanager.algorithm = SunX509
>         metrics.sample.window.ms = 30000
>         fetch.min.bytes = 1024
>         send.buffer.bytes = 131072
>         auto.offset.reset = latest
>
> thanks,
> Gary