You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@ignite.apache.org by "Nikolay Izhikov (Jira)" <ji...@apache.org> on 2023/04/26 10:48:00 UTC
[jira] [Updated] (IGNITE-19369) Metadata topic offset must be stored only after commit
[ https://issues.apache.org/jira/browse/IGNITE-19369?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Nikolay Izhikov updated IGNITE-19369:
-------------------------------------
Priority: Blocker (was: Minor)
> Metadata topic offset must be stored only after commit
> ------------------------------------------------------
>
> Key: IGNITE-19369
> URL: https://issues.apache.org/jira/browse/IGNITE-19369
> Project: Ignite
> Issue Type: Improvement
> Components: extensions
> Reporter: Nikolay Izhikov
> Assignee: Nikolay Izhikov
> Priority: Blocker
> Labels: IEP-59, ise
>
> Currently, when CDC through Kafka is used there are possible delays in replication between clusters when {{KafkaToIgniteCdcStreamerApplier}} tries to update binary metadata and marshaller mappings.
> Delays caused by calls of {{KafkaConsumer#poll}} in {{KafkaToIgniteMetadataUpdater#updateMetadata}} , when meta topic is empty:
> # When first {{KafkaToIgniteCdcStreamerApplier}} meets {{META_UPDATE_MARKER}} it calls {{KafkaToIgniteMetadataUpdater#updateMetadata}} which in turn calls {{KafkaConsumer#poll}}, which returns immediately [1] when data is present in metadata topic. If there are few binary types and mappings to update, first {{KafkaToIgniteCdcStreamerApplier}} will consume all entries from metadata topic.
> # {{KafkaToIgniteCdcStreamerApplier}} consequently call {{KafkaToIgniteMetadataUpdater#updateMetadata}} for each partition with meta update marker. All further consequent calls will wait for {{kafkaReqTimeout}}.
> # Also there is a bottleneck, when multiple applier threads tries to update metadata and call synchronized method {{KafkaToIgniteMetadataUpdater#updateMetadata}}, because {{KafkaToIgniteMetadataUpdater}} is shared between applier threads.
> # Because {{META_UPDATE_MARKER}} is sent twice to each Kafka partition of event topic from every node: firstly, in case of type mappings updates, secondly, in case of binary types update there are possible delays up to {{clusterSize x (topicPartitions x 2 - 1) x kafkaReqTimeout}}.
> # Data updates are blocked for Kafka partitions with unprocessed update markers.
> # For example for default timeout and 16 Kafka partitions _last partition will be consumed after 1.5 minutes_ in case of two one-node clusters.
> Links:
> # [https://kafka.apache.org/27/javadoc/org/apache/kafka/clients/consumer/KafkaConsumer.html#poll-java.time.Duration-]
--
This message was sent by Atlassian Jira
(v8.20.10#820010)