You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by "lokesh Birla (JIRA)" <ji...@apache.org> on 2014/12/04 22:47:12 UTC
[jira] [Commented] (KAFKA-727) broker can still expose uncommitted
data to a consumer
[ https://issues.apache.org/jira/browse/KAFKA-727?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14234678#comment-14234678 ]
lokesh Birla commented on KAFKA-727:
------------------------------------
Hi,
Is this really fixed? I still see this issue when I am using 4 topics, 3 partitions and 3 replication factor. I am using kafka_2.9.2-0.8.1.1.
Currently I am using 3 node broker and 1 zookeeper. I did not see this issue when I used 1,2 or 3 topics.
2014-08-18 06:43:58,356] ERROR [KafkaApi-1] Error when processing fetch request for partition [mmetopic4,2] offset 1940029 from consumer with correlation id 21 (kafka.server.Kaf
kaApis)
java.lang.IllegalArgumentException: Attempt to read with a maximum offset (1818353) less than the start offset (1940029).
at kafka.log.LogSegment.read(LogSegment.scala:136)
at kafka.log.Log.read(Log.scala:386)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:530)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:476)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:471)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:233)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:233)
at scala.collection.immutable.Map$Map1.foreach(Map.scala:119)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:233)
at scala.collection.immutable.Map$Map1.map(Map.scala:107)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:471)
at kafka.server.KafkaApis$FetchRequestPurgatory.expire(KafkaApis.scala:783)
at kafka.server.KafkaApis$FetchRequestPurgatory.expire(KafkaApis.scala:765)
at kafka.server.RequestPurgatory$ExpiredRequestReaper.run(RequestPurgatory.scala:216)
at java.lang.Thread.run(Thread.java:745)
THanks for your help.
> broker can still expose uncommitted data to a consumer
> ------------------------------------------------------
>
> Key: KAFKA-727
> URL: https://issues.apache.org/jira/browse/KAFKA-727
> Project: Kafka
> Issue Type: Bug
> Components: core
> Affects Versions: 0.8.0
> Reporter: Jun Rao
> Assignee: Jay Kreps
> Priority: Blocker
> Labels: p1
> Attachments: KAFKA-727-v1.patch
>
>
> Even after kafka-698 is fixed, we still see consumer clients occasionally see uncommitted data. The following is how this can happen.
> 1. In Log.read(), we pass in startOffset < HW and maxOffset = HW.
> 2. Then we call LogSegment.read(), in which we call translateOffset on the maxOffset. The offset doesn't exist and translateOffset returns null.
> 3. Continue in LogSegment.read(), we then call messageSet.sizeInBytes() to fetch and return the data.
> What can happen is that between step 2 and step 3, a new message is appended to the log and is not committed yet. Now, we have exposed uncommitted data to the client.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)