You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@storm.apache.org by "Jiasheng Wang (JIRA)" <ji...@apache.org> on 2016/05/25 04:15:13 UTC

[jira] [Commented] (STORM-1863) Throw exception if messages fetched by storm-kafka is emtpy

    [ https://issues.apache.org/jira/browse/STORM-1863?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15299428#comment-15299428 ] 

Jiasheng Wang commented on STORM-1863:
--------------------------------------

PR: https://github.com/apache/storm/pull/1443

> Throw exception if messages fetched by storm-kafka is emtpy
> -----------------------------------------------------------
>
>                 Key: STORM-1863
>                 URL: https://issues.apache.org/jira/browse/STORM-1863
>             Project: Apache Storm
>          Issue Type: Improvement
>          Components: storm-kafka
>            Reporter: Jiasheng Wang
>
> In kafka ConsumerIterator, there is some codes like this:
> // if we just updated the current chunk and it is empty that means the fetch size is too small! if(currentDataChunk.messages.validBytes == 0) throw new MessageSizeTooLargeException("Found a message larger than the maximum fetch size of this consumer on topic " + "%s partition %d at fetch offset %d. Increase the fetch size, or decrease the maximum message size the broker will allow." .format(currentDataChunk.topicInfo.topic, currentDataChunk.topicInfo.partitionId, currentDataChunk.fetchOffset))
> When "fetch.message.max.bytes" config is smaller than the actual message size in topic, ConsumerIterator will throw an exception to notify user.
> But in storm-kafka, there is no such logic. And as a result, if KafkaConfig.fetchSizeBytes is smaller than actual message size, the topology will fetch no data but still be running.
> To prevent this situation, we need throw MessageSizeTooLargeException as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)