You are viewing a plain text version of this content. The canonical link for it is here.
Posted to jira@kafka.apache.org by "Andrey (JIRA)" <ji...@apache.org> on 2017/11/08 15:41:00 UTC
[jira] [Created] (KAFKA-6189) Loosing messages while getting
OFFSET_OUT_OF_RANGE eror in consumer
Andrey created KAFKA-6189:
-----------------------------
Summary: Loosing messages while getting OFFSET_OUT_OF_RANGE eror in consumer
Key: KAFKA-6189
URL: https://issues.apache.org/jira/browse/KAFKA-6189
Project: Kafka
Issue Type: Bug
Components: clients
Affects Versions: 0.11.0.0
Reporter: Andrey
Attachments: kafkaLossingMessages.png
Steps to reproduce:
* Setup test:
** producer sends messages constantly. If cluster not available, then it will retry
** consumer polling
** topic has 3 partitions and replication factor 3.
** min.insync.replicas=2
** producer has "acks=all"
** consumer has default "auto.offset.reset=latest"
** consumer manually commitSync offsets after handling messages.
** kafka cluster has 3 brokers
* Kill broker 0
* In consumer's logs:
{code}
2017-11-08 11:36:33,967 INFO org.apache.kafka.clients.consumer.internals.Fetcher - Fetch offset 10706682 is out of range for partition mytopic-2, resetting offset [kafka-consumer]
2017-11-08 11:36:33,968 INFO org.apache.kafka.clients.consumer.internals.Fetcher - Fetch offset 8024431 is out of range for partition mytopic-1, resetting offset [kafka-consumer]
2017-11-08 11:36:34,045 INFO org.apache.kafka.clients.consumer.internals.Fetcher - Fetch offset 8029505 is out of range for partition mytopic-0, resetting offset [kafka-consumer]
{code}
After that, consumer lost several messages on each partition.
Expected:
* return upper bound of range
* consumer should resume from that offset instead of "auto.offset.reset".
Workaround:
* put "auto.offset.reset=earliest"
* get a lot of duplicate messages, instead of lost
Looks like this is what happening during the recovery from broker failure:
!kafkaLossingMessages.png|thumbnail!
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)