You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@camel.apache.org by da...@apache.org on 2023/02/15 09:31:09 UTC
[camel] 01/02: Regen
This is an automated email from the ASF dual-hosted git repository.
davsclaus pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/camel.git
commit b25dc28e8667b3515cc4553ef945593554dcf0fb
Author: Claus Ibsen <cl...@gmail.com>
AuthorDate: Wed Feb 15 10:30:28 2023 +0100
Regen
---
.../dsl/DebeziumDb2EndpointBuilderFactory.java | 619 ++++++------------
.../dsl/DebeziumMongodbEndpointBuilderFactory.java | 196 +++---
.../dsl/DebeziumMySqlEndpointBuilderFactory.java | 675 ++++++-------------
.../dsl/DebeziumOracleEndpointBuilderFactory.java | 719 ++++++++-------------
.../DebeziumPostgresEndpointBuilderFactory.java | 360 ++++-------
.../DebeziumSqlserverEndpointBuilderFactory.java | 683 ++++++-------------
6 files changed, 1102 insertions(+), 2150 deletions(-)
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumDb2EndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumDb2EndpointBuilderFactory.java
index df5789c0953..50bad611985 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumDb2EndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumDb2EndpointBuilderFactory.java
@@ -328,22 +328,6 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("offsetStorageTopic", offsetStorageTopic);
return this;
}
- /**
- * Regular expressions matching columns to exclude from change events
- * (deprecated, use column.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param columnBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder columnBlacklist(
- String columnBlacklist) {
- doSetProperty("columnBlacklist", columnBlacklist);
- return this;
- }
/**
* Regular expressions matching columns to exclude from change events.
*
@@ -392,22 +376,6 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("columnPropagateSourceType", columnPropagateSourceType);
return this;
}
- /**
- * Regular expressions matching columns to include in change events
- * (deprecated, use column.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param columnWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder columnWhitelist(
- String columnWhitelist) {
- doSetProperty("columnWhitelist", columnWhitelist);
- return this;
- }
/**
* Optional list of custom converters that would be used instead of
* default ones. The converters are defined using '.type' config option
@@ -439,299 +407,6 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("databaseDbname", databaseDbname);
return this;
}
- /**
- * The name of the DatabaseHistory class that should be used to store
- * and recover database schema changes. The configuration properties for
- * the history are prefixed with the 'database.history.' string.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: io.debezium.relational.history.KafkaDatabaseHistory
- * Group: db2
- *
- * @param databaseHistory the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistory(
- String databaseHistory) {
- doSetProperty("databaseHistory", databaseHistory);
- return this;
- }
- /**
- * The path to the file that will be used to record the database
- * history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param databaseHistoryFileFilename the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryFileFilename(
- String databaseHistoryFileFilename) {
- doSetProperty("databaseHistoryFileFilename", databaseHistoryFileFilename);
- return this;
- }
- /**
- * A list of host/port pairs that the connector will use for
- * establishing the initial connection to the Kafka cluster for
- * retrieving database schema history previously stored by the
- * connector. This should point to the same Kafka cluster used by the
- * Kafka Connect process.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param databaseHistoryKafkaBootstrapServers the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaBootstrapServers(
- String databaseHistoryKafkaBootstrapServers) {
- doSetProperty("databaseHistoryKafkaBootstrapServers", databaseHistoryKafkaBootstrapServers);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option is a: <code>long</code> type.
- *
- * Default: 3s
- * Group: db2
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- long databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option will be converted to a <code>long</code> type.
- *
- * Default: 3s
- * Group: db2
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- String databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100
- * Group: db2
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- int databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100
- * Group: db2
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- String databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100ms
- * Group: db2
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- int databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100ms
- * Group: db2
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- String databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The name of the topic for the database schema history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param databaseHistoryKafkaTopic the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryKafkaTopic(
- String databaseHistoryKafkaTopic) {
- doSetProperty("databaseHistoryKafkaTopic", databaseHistoryKafkaTopic);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: db2
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistorySkipUnparseableDdl(
- boolean databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: db2
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistorySkipUnparseableDdl(
- String databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: db2
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- boolean databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: db2
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- String databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: db2
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- boolean databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: db2
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- String databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
/**
* Resolvable hostname or IP address of the database server.
*
@@ -794,25 +469,6 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("databasePort", databasePort);
return this;
}
- /**
- * Unique name that identifies the database server and all recorded
- * offsets, and that is used as a prefix for all schemas and topics.
- * Each distinct installation should have a separate namespace and be
- * monitored by at most one Debezium connector.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Required: true
- * Group: db2
- *
- * @param databaseServerName the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder databaseServerName(
- String databaseServerName) {
- doSetProperty("databaseServerName", databaseServerName);
- return this;
- }
/**
* Name of the database user to be used when connecting to the database.
*
@@ -847,7 +503,7 @@ public interface DebeziumDb2EndpointBuilderFactory {
}
/**
* Specify how DECIMAL and NUMERIC columns should be represented in
- * change events, including:'precise' (the default) uses
+ * change events, including: 'precise' (the default) uses
* java.math.BigDecimal to represent values, which are encoded in the
* change events using a binary representation and Kafka Connect's
* 'org.apache.kafka.connect.data.Decimal' type; 'string' uses string to
@@ -870,11 +526,11 @@ public interface DebeziumDb2EndpointBuilderFactory {
}
/**
* Specify how failures during processing of events (i.e. when
- * encountering a corrupted event) should be handled, including:'fail'
+ * encountering a corrupted event) should be handled, including: 'fail'
* (the default) an exception indicating the problematic event and its
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its position will be logged and the event will
- * be skipped;'ignore' the problematic event will be skipped.
+ * be skipped; 'ignore' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -947,8 +603,9 @@ public interface DebeziumDb2EndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option is a: <code>boolean</code> type.
*
@@ -968,8 +625,9 @@ public interface DebeziumDb2EndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1120,12 +778,12 @@ public interface DebeziumDb2EndpointBuilderFactory {
/**
* A semicolon-separated list of expressions that match fully-qualified
* tables and column(s) to be used as message key. Each expression must
- * match the pattern ':',where the table names could be defined as
+ * match the pattern ':', where the table names could be defined as
* (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on the
- * specific connector,and the key columns are a comma-separated list of
+ * specific connector, and the key columns are a comma-separated list of
* columns representing the custom key. For any table without an
* explicit key configuration the table's primary key column(s) will be
- * used as message key.Example:
+ * used as message key. Example:
* dbserver1.inventory.orderlines:orderId,orderLineId;dbserver1.inventory.orders:id.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1207,11 +865,12 @@ public interface DebeziumDb2EndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size. The
+ * default value is '10000'.
*
* The option is a: <code>int</code> type.
*
- * Default: 0
+ * Default: 10000
* Group: db2
*
* @param queryFetchSize the value to set
@@ -1223,11 +882,12 @@ public interface DebeziumDb2EndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size. The
+ * default value is '10000'.
*
* The option will be converted to a <code>int</code> type.
*
- * Default: 0
+ * Default: 10000
* Group: db2
*
* @param queryFetchSize the value to set
@@ -1304,15 +964,131 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("sanitizeFieldNames", sanitizeFieldNames);
return this;
}
+ /**
+ * The name of the SchemaHistory class that should be used to store and
+ * recover database schema changes. The configuration properties for the
+ * history are prefixed with the 'schema.history.internal.' string.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.storage.kafka.history.KafkaSchemaHistory
+ * Group: db2
+ *
+ * @param schemaHistoryInternal the value to set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder schemaHistoryInternal(
+ String schemaHistoryInternal) {
+ doSetProperty("schemaHistoryInternal", schemaHistoryInternal);
+ return this;
+ }
+ /**
+ * The path to the file that will be used to record the database schema
+ * history.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Group: db2
+ *
+ * @param schemaHistoryInternalFileFilename the value to set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder schemaHistoryInternalFileFilename(
+ String schemaHistoryInternalFileFilename) {
+ doSetProperty("schemaHistoryInternalFileFilename", schemaHistoryInternalFileFilename);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: db2
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ boolean schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: db2
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ String schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: db2
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ boolean schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: db2
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ String schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
/**
* Specify how schema names should be adjusted for compatibility with
- * the message converter used by the connector, including:'avro'
+ * the message converter used by the connector, including: 'avro'
* replaces the characters that cannot be used in the Avro type name
- * with underscore (default)'none' does not apply any adjustment.
+ * with underscore; 'none' does not apply any adjustment (default).
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: avro
+ * Default: none
* Group: db2
*
* @param schemaNameAdjustmentMode the value to set
@@ -1343,10 +1119,11 @@ public interface DebeziumDb2EndpointBuilderFactory {
* The comma-separated list of operations to skip during streaming,
* defined as: 'c' for inserts/create; 'u' for updates; 'd' for deletes,
* 't' for truncates, and 'none' to indicate nothing skipped. By
- * default, no operations will be skipped.
+ * default, only truncate operations will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
+ * Default: t
* Group: db2
*
* @param skippedOperations the value to set
@@ -1423,7 +1200,7 @@ public interface DebeziumDb2EndpointBuilderFactory {
return this;
}
/**
- * this setting must be set to specify a list of tables/collections
+ * This setting must be set to specify a list of tables/collections
* whose snapshot must be taken on creating or restarting the connector.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1497,7 +1274,7 @@ public interface DebeziumDb2EndpointBuilderFactory {
/**
* This property contains a comma-separated list of fully-qualified
* tables (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on
- * thespecific connectors. Select statements for the individual tables
+ * the specific connectors. Select statements for the individual tables
* are specified in further configuration properties, one for each
* table, identified by the id
* 'snapshot.select.statement.overrides.DB_NAME.TABLE_NAME' or
@@ -1520,39 +1297,6 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("snapshotSelectStatementOverrides", snapshotSelectStatementOverrides);
return this;
}
- /**
- * A version of the format of the publicly visible source part in the
- * message.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: v2
- * Group: db2
- *
- * @param sourceStructVersion the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder sourceStructVersion(
- String sourceStructVersion) {
- doSetProperty("sourceStructVersion", sourceStructVersion);
- return this;
- }
- /**
- * A comma-separated list of regular expressions that match the
- * fully-qualified names of tables to be excluded from monitoring
- * (deprecated, use table.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param tableBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder tableBlacklist(String tableBlacklist) {
- doSetProperty("tableBlacklist", tableBlacklist);
- return this;
- }
/**
* A comma-separated list of regular expressions that match the
* fully-qualified names of tables to be excluded from monitoring.
@@ -1617,30 +1361,15 @@ public interface DebeziumDb2EndpointBuilderFactory {
doSetProperty("tableIncludeList", tableIncludeList);
return this;
}
- /**
- * The tables for which changes are to be captured (deprecated, use
- * table.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: db2
- *
- * @param tableWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumDb2EndpointBuilder tableWhitelist(String tableWhitelist) {
- doSetProperty("tableWhitelist", tableWhitelist);
- return this;
- }
/**
* Time, date, and timestamps can be represented with different kinds of
- * precisions, including:'adaptive' (the default) bases the precision of
- * time, date, and timestamp values on the database column's precision;
- * 'adaptive_time_microseconds' like 'adaptive' mode, but TIME fields
- * always use microseconds precision;'connect' always represents time,
- * date, and timestamp values using Kafka Connect's built-in
- * representations for Time, Date, and Timestamp, which uses millisecond
- * precision regardless of the database columns' precision .
+ * precisions, including: 'adaptive' (the default) bases the precision
+ * of time, date, and timestamp values on the database column's
+ * precision; 'adaptive_time_microseconds' like 'adaptive' mode, but
+ * TIME fields always use microseconds precision; 'connect' always
+ * represents time, date, and timestamp values using Kafka Connect's
+ * built-in representations for Time, Date, and Timestamp, which uses
+ * millisecond precision regardless of the database columns' precision.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1657,10 +1386,10 @@ public interface DebeziumDb2EndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option is a: <code>boolean</code> type.
*
@@ -1677,10 +1406,10 @@ public interface DebeziumDb2EndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1697,21 +1426,41 @@ public interface DebeziumDb2EndpointBuilderFactory {
return this;
}
/**
- * The name of the transaction metadata topic. The placeholder
- * ${database.server.name} can be used for referring to the connector's
- * logical name; defaults to ${database.server.name}.transaction.
+ * The name of the TopicNamingStrategy class that should be used to
+ * determine the topic name for data change, schema change, transaction,
+ * heartbeat event etc.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.schema.SchemaTopicNamingStrategy
+ * Group: db2
+ *
+ * @param topicNamingStrategy the value to set
+ * @return the dsl builder
+ */
+ default DebeziumDb2EndpointBuilder topicNamingStrategy(
+ String topicNamingStrategy) {
+ doSetProperty("topicNamingStrategy", topicNamingStrategy);
+ return this;
+ }
+ /**
+ * Topic prefix that identifies and provides a namespace for the
+ * particular database server/cluster is capturing changes. The topic
+ * prefix should be unique across all other connectors, since it is used
+ * as a prefix for all Kafka topic names that receive events emitted by
+ * this connector. Only alphanumeric characters, hyphens, dots and
+ * underscores must be accepted.
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: ${database.server.name}.transaction
+ * Required: true
* Group: db2
*
- * @param transactionTopic the value to set
+ * @param topicPrefix the value to set
* @return the dsl builder
*/
- default DebeziumDb2EndpointBuilder transactionTopic(
- String transactionTopic) {
- doSetProperty("transactionTopic", transactionTopic);
+ default DebeziumDb2EndpointBuilder topicPrefix(String topicPrefix) {
+ doSetProperty("topicPrefix", topicPrefix);
return this;
}
}
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMongodbEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMongodbEndpointBuilderFactory.java
index ab356443194..2d96c3b6e88 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMongodbEndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMongodbEndpointBuilderFactory.java
@@ -331,11 +331,10 @@ public interface DebeziumMongodbEndpointBuilderFactory {
}
/**
* The method used to capture changes from MongoDB server. Options
- * include: 'oplog' to capture changes from the oplog; 'change_streams'
- * to capture changes via MongoDB Change Streams, update events do not
- * contain full documents; 'change_streams_update_full' (the default) to
- * capture changes via MongoDB Change Streams, update events contain
- * full documents.
+ * include: 'change_streams' to capture changes via MongoDB Change
+ * Streams, update events do not contain full documents;
+ * 'change_streams_update_full' (the default) to capture changes via
+ * MongoDB Change Streams, update events contain full documents.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -557,22 +556,6 @@ public interface DebeziumMongodbEndpointBuilderFactory {
doSetProperty("databaseExcludeList", databaseExcludeList);
return this;
}
- /**
- * The path to the file that will be used to record the database
- * history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mongodb
- *
- * @param databaseHistoryFileFilename the value to set
- * @return the dsl builder
- */
- default DebeziumMongodbEndpointBuilder databaseHistoryFileFilename(
- String databaseHistoryFileFilename) {
- doSetProperty("databaseHistoryFileFilename", databaseHistoryFileFilename);
- return this;
- }
/**
* A comma-separated list of regular expressions that match the database
* names for which changes are to be captured.
@@ -591,11 +574,11 @@ public interface DebeziumMongodbEndpointBuilderFactory {
}
/**
* Specify how failures during processing of events (i.e. when
- * encountering a corrupted event) should be handled, including:'fail'
+ * encountering a corrupted event) should be handled, including: 'fail'
* (the default) an exception indicating the problematic event and its
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its position will be logged and the event will
- * be skipped;'ignore' the problematic event will be skipped.
+ * be skipped; 'ignore' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -812,6 +795,21 @@ public interface DebeziumMongodbEndpointBuilderFactory {
doSetProperty("mongodbAuthsource", mongodbAuthsource);
return this;
}
+ /**
+ * Database connection string.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Group: mongodb
+ *
+ * @param mongodbConnectionString the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMongodbEndpointBuilder mongodbConnectionString(
+ String mongodbConnectionString) {
+ doSetProperty("mongodbConnectionString", mongodbConnectionString);
+ return this;
+ }
/**
* The connection timeout, given in milliseconds. Defaults to 10 seconds
* (10,000 ms).
@@ -846,6 +844,40 @@ public interface DebeziumMongodbEndpointBuilderFactory {
doSetProperty("mongodbConnectTimeoutMs", mongodbConnectTimeoutMs);
return this;
}
+ /**
+ * The frequency that the cluster monitor attempts to reach each server.
+ * Defaults to 10 seconds (10,000 ms).
+ *
+ * The option is a: <code>int</code> type.
+ *
+ * Default: 10s
+ * Group: mongodb
+ *
+ * @param mongodbHeartbeatFrequencyMs the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMongodbEndpointBuilder mongodbHeartbeatFrequencyMs(
+ int mongodbHeartbeatFrequencyMs) {
+ doSetProperty("mongodbHeartbeatFrequencyMs", mongodbHeartbeatFrequencyMs);
+ return this;
+ }
+ /**
+ * The frequency that the cluster monitor attempts to reach each server.
+ * Defaults to 10 seconds (10,000 ms).
+ *
+ * The option will be converted to a <code>int</code> type.
+ *
+ * Default: 10s
+ * Group: mongodb
+ *
+ * @param mongodbHeartbeatFrequencyMs the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMongodbEndpointBuilder mongodbHeartbeatFrequencyMs(
+ String mongodbHeartbeatFrequencyMs) {
+ doSetProperty("mongodbHeartbeatFrequencyMs", mongodbHeartbeatFrequencyMs);
+ return this;
+ }
/**
* The hostname and port pairs (in the form 'host' or 'host:port') of
* the MongoDB server(s) in the replica set.
@@ -900,24 +932,6 @@ public interface DebeziumMongodbEndpointBuilderFactory {
doSetProperty("mongodbMembersAutoDiscover", mongodbMembersAutoDiscover);
return this;
}
- /**
- * Unique name that identifies the MongoDB replica set or cluster and
- * all recorded offsets, and that is used as a prefix for all schemas
- * and topics. Each distinct MongoDB installation should have a separate
- * namespace and monitored by at most one Debezium connector.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Required: true
- * Group: mongodb
- *
- * @param mongodbName the value to set
- * @return the dsl builder
- */
- default DebeziumMongodbEndpointBuilder mongodbName(String mongodbName) {
- doSetProperty("mongodbName", mongodbName);
- return this;
- }
/**
* Password to be used when connecting to MongoDB, if necessary.
*
@@ -1185,7 +1199,7 @@ public interface DebeziumMongodbEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option is a: <code>int</code> type.
*
@@ -1201,7 +1215,7 @@ public interface DebeziumMongodbEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option will be converted to a <code>int</code> type.
*
@@ -1283,15 +1297,31 @@ public interface DebeziumMongodbEndpointBuilderFactory {
doSetProperty("sanitizeFieldNames", sanitizeFieldNames);
return this;
}
+ /**
+ * The path to the file that will be used to record the database schema
+ * history.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Group: mongodb
+ *
+ * @param schemaHistoryInternalFileFilename the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMongodbEndpointBuilder schemaHistoryInternalFileFilename(
+ String schemaHistoryInternalFileFilename) {
+ doSetProperty("schemaHistoryInternalFileFilename", schemaHistoryInternalFileFilename);
+ return this;
+ }
/**
* Specify how schema names should be adjusted for compatibility with
- * the message converter used by the connector, including:'avro'
+ * the message converter used by the connector, including: 'avro'
* replaces the characters that cannot be used in the Avro type name
- * with underscore (default)'none' does not apply any adjustment.
+ * with underscore; 'none' does not apply any adjustment (default).
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: avro
+ * Default: none
* Group: mongodb
*
* @param schemaNameAdjustmentMode the value to set
@@ -1322,10 +1352,11 @@ public interface DebeziumMongodbEndpointBuilderFactory {
* The comma-separated list of operations to skip during streaming,
* defined as: 'c' for inserts/create; 'u' for updates; 'd' for deletes,
* 't' for truncates, and 'none' to indicate nothing skipped. By
- * default, no operations will be skipped.
+ * default, only truncate operations will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
+ * Default: t
* Group: mongodb
*
* @param skippedOperations the value to set
@@ -1421,7 +1452,7 @@ public interface DebeziumMongodbEndpointBuilderFactory {
return this;
}
/**
- * this setting must be set to specify a list of tables/collections
+ * This setting must be set to specify a list of tables/collections
* whose snapshot must be taken on creating or restarting the connector.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1488,29 +1519,12 @@ public interface DebeziumMongodbEndpointBuilderFactory {
doSetProperty("snapshotMode", snapshotMode);
return this;
}
- /**
- * A version of the format of the publicly visible source part in the
- * message.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: v2
- * Group: mongodb
- *
- * @param sourceStructVersion the value to set
- * @return the dsl builder
- */
- default DebeziumMongodbEndpointBuilder sourceStructVersion(
- String sourceStructVersion) {
- doSetProperty("sourceStructVersion", sourceStructVersion);
- return this;
- }
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option is a: <code>boolean</code> type.
*
@@ -1527,10 +1541,10 @@ public interface DebeziumMongodbEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1547,21 +1561,41 @@ public interface DebeziumMongodbEndpointBuilderFactory {
return this;
}
/**
- * The name of the transaction metadata topic. The placeholder
- * ${database.server.name} can be used for referring to the connector's
- * logical name; defaults to ${database.server.name}.transaction.
+ * The name of the TopicNamingStrategy class that should be used to
+ * determine the topic name for data change, schema change, transaction,
+ * heartbeat event etc.
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: ${database.server.name}.transaction
+ * Default: io.debezium.schema.SchemaTopicNamingStrategy
+ * Group: mongodb
+ *
+ * @param topicNamingStrategy the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMongodbEndpointBuilder topicNamingStrategy(
+ String topicNamingStrategy) {
+ doSetProperty("topicNamingStrategy", topicNamingStrategy);
+ return this;
+ }
+ /**
+ * Topic prefix that identifies and provides a namespace for the
+ * particular database server/cluster is capturing changes. The topic
+ * prefix should be unique across all other connectors, since it is used
+ * as a prefix for all Kafka topic names that receive events emitted by
+ * this connector. Only alphanumeric characters, hyphens, dots and
+ * underscores must be accepted.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Required: true
* Group: mongodb
*
- * @param transactionTopic the value to set
+ * @param topicPrefix the value to set
* @return the dsl builder
*/
- default DebeziumMongodbEndpointBuilder transactionTopic(
- String transactionTopic) {
- doSetProperty("transactionTopic", transactionTopic);
+ default DebeziumMongodbEndpointBuilder topicPrefix(String topicPrefix) {
+ doSetProperty("topicPrefix", topicPrefix);
return this;
}
}
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMySqlEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMySqlEndpointBuilderFactory.java
index b89c3a65bcb..25a0bbb614b 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMySqlEndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumMySqlEndpointBuilderFactory.java
@@ -330,7 +330,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Specify how BIGINT UNSIGNED columns should be represented in change
- * events, including:'precise' uses java.math.BigDecimal to represent
+ * events, including: 'precise' uses java.math.BigDecimal to represent
* values, which are encoded in the change events using a binary
* representation and Kafka Connect's
* 'org.apache.kafka.connect.data.Decimal' type; 'long' (the default)
@@ -352,9 +352,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Specify how binary (blob, binary, etc.) columns should be represented
- * in change events, including:'bytes' represents binary data as byte
- * array (default)'base64' represents binary data as base64-encoded
- * string'hex' represents binary data as hex-encoded (base16) string.
+ * in change events, including: 'bytes' represents binary data as byte
+ * array (default); 'base64' represents binary data as base64-encoded
+ * string; 'base64-url-safe' represents binary data as
+ * base64-url-safe-encoded string; 'hex' represents binary data as
+ * hex-encoded (base16) string.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -407,22 +409,6 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("binlogBufferSize", binlogBufferSize);
return this;
}
- /**
- * Regular expressions matching columns to exclude from change events
- * (deprecated, use column.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param columnBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder columnBlacklist(
- String columnBlacklist) {
- doSetProperty("columnBlacklist", columnBlacklist);
- return this;
- }
/**
* Regular expressions matching columns to exclude from change events.
*
@@ -471,22 +457,6 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("columnPropagateSourceType", columnPropagateSourceType);
return this;
}
- /**
- * Regular expressions matching columns to include in change events
- * (deprecated, use column.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param columnWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder columnWhitelist(
- String columnWhitelist) {
- doSetProperty("columnWhitelist", columnWhitelist);
- return this;
- }
/**
* Whether a separate thread should be used to ensure the connection is
* kept alive.
@@ -624,299 +594,6 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("databaseExcludeList", databaseExcludeList);
return this;
}
- /**
- * The name of the DatabaseHistory class that should be used to store
- * and recover database schema changes. The configuration properties for
- * the history are prefixed with the 'database.history.' string.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: io.debezium.relational.history.FileDatabaseHistory
- * Group: mysql
- *
- * @param databaseHistory the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistory(
- String databaseHistory) {
- doSetProperty("databaseHistory", databaseHistory);
- return this;
- }
- /**
- * The path to the file that will be used to record the database
- * history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param databaseHistoryFileFilename the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryFileFilename(
- String databaseHistoryFileFilename) {
- doSetProperty("databaseHistoryFileFilename", databaseHistoryFileFilename);
- return this;
- }
- /**
- * A list of host/port pairs that the connector will use for
- * establishing the initial connection to the Kafka cluster for
- * retrieving database schema history previously stored by the
- * connector. This should point to the same Kafka cluster used by the
- * Kafka Connect process.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param databaseHistoryKafkaBootstrapServers the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaBootstrapServers(
- String databaseHistoryKafkaBootstrapServers) {
- doSetProperty("databaseHistoryKafkaBootstrapServers", databaseHistoryKafkaBootstrapServers);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option is a: <code>long</code> type.
- *
- * Default: 3s
- * Group: mysql
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- long databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option will be converted to a <code>long</code> type.
- *
- * Default: 3s
- * Group: mysql
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- String databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100
- * Group: mysql
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- int databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100
- * Group: mysql
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- String databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100ms
- * Group: mysql
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- int databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100ms
- * Group: mysql
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- String databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The name of the topic for the database schema history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param databaseHistoryKafkaTopic the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryKafkaTopic(
- String databaseHistoryKafkaTopic) {
- doSetProperty("databaseHistoryKafkaTopic", databaseHistoryKafkaTopic);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: mysql
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistorySkipUnparseableDdl(
- boolean databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: mysql
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistorySkipUnparseableDdl(
- String databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: mysql
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- boolean databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: mysql
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- String databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: mysql
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- boolean databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: mysql
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- String databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
/**
* Resolvable hostname or IP address of the database server.
*
@@ -952,7 +629,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* JDBC connection (not binlog reading connection) to the database is
* established. Note that the connector may establish JDBC connections
* at its own discretion, so this should typically be used for
- * configuration of session parameters only,but not for executing DML
+ * configuration of session parameters only, but not for executing DML
* statements. Use doubled semicolon (';;') to use a semicolon as a
* character and not as a delimiter.
*
@@ -1035,8 +712,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* A numeric ID of this database client, which must be unique across all
* currently-running database processes in the cluster. This connector
* joins the MySQL database cluster as another server (with this unique
- * ID) so it can read the binlog. By default, a random number is
- * generated between 5400 and 6400.
+ * ID) so it can read the binlog.
*
* The option is a: <code>long</code> type.
*
@@ -1054,8 +730,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* A numeric ID of this database client, which must be unique across all
* currently-running database processes in the cluster. This connector
* joins the MySQL database cluster as another server (with this unique
- * ID) so it can read the binlog. By default, a random number is
- * generated between 5400 and 6400.
+ * ID) so it can read the binlog.
*
* The option will be converted to a <code>long</code> type.
*
@@ -1107,25 +782,6 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("databaseServerIdOffset", databaseServerIdOffset);
return this;
}
- /**
- * Unique name that identifies the database server and all recorded
- * offsets, and that is used as a prefix for all schemas and topics.
- * Each distinct installation should have a separate namespace and be
- * monitored by at most one Debezium connector.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Required: true
- * Group: mysql
- *
- * @param databaseServerName the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder databaseServerName(
- String databaseServerName) {
- doSetProperty("databaseServerName", databaseServerName);
- return this;
- }
/**
* The location of the key store file. This is optional and can be used
* for two-way authentication between the client and the MySQL Server.
@@ -1159,8 +815,8 @@ public interface DebeziumMySqlEndpointBuilderFactory {
return this;
}
/**
- * Whether to use an encrypted connection to MySQL. Options
- * include'disabled' (the default) to use an unencrypted connection;
+ * Whether to use an encrypted connection to MySQL. Options include:
+ * 'disabled' (the default) to use an unencrypted connection;
* 'preferred' to establish a secure (encrypted) connection if the
* server supports secure connections, but fall back to an unencrypted
* connection otherwise; 'required' to use a secure (encrypted)
@@ -1250,7 +906,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Specify how DECIMAL and NUMERIC columns should be represented in
- * change events, including:'precise' (the default) uses
+ * change events, including: 'precise' (the default) uses
* java.math.BigDecimal to represent values, which are encoded in the
* change events using a binary representation and Kafka Connect's
* 'org.apache.kafka.connect.data.Decimal' type; 'string' uses string to
@@ -1312,12 +968,12 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Specify how failures during deserialization of binlog events (i.e.
- * when encountering a corrupted event) should be handled,
- * including:'fail' (the default) an exception indicating the
- * problematic event and its binlog position is raised, causing the
- * connector to be stopped; 'warn' the problematic event and its binlog
- * position will be logged and the event will be skipped;'ignore' the
- * problematic event will be skipped.
+ * when encountering a corrupted event) should be handled, including:
+ * 'fail' (the default) an exception indicating the problematic event
+ * and its binlog position is raised, causing the connector to be
+ * stopped; 'warn' the problematic event and its binlog position will be
+ * logged and the event will be skipped; 'ignore' the problematic event
+ * will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1334,11 +990,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Specify how failures during processing of events (i.e. when
- * encountering a corrupted event) should be handled, including:'fail'
+ * encountering a corrupted event) should be handled, including: 'fail'
* (the default) an exception indicating the problematic event and its
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its position will be logged and the event will
- * be skipped;'ignore' the problematic event will be skipped.
+ * be skipped; 'ignore' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1498,7 +1154,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* configured with the binlog_rows_query_log_events option set to ON.
* Query will not be present for events generated from snapshot.
* WARNING: Enabling this option may expose tables or fields explicitly
- * blacklisted or masked by including the original SQL statement in the
+ * excluded or masked by including the original SQL statement in the
* change event. For this reason the default value is 'false'.
*
* The option is a: <code>boolean</code> type.
@@ -1519,7 +1175,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* configured with the binlog_rows_query_log_events option set to ON.
* Query will not be present for events generated from snapshot.
* WARNING: Enabling this option may expose tables or fields explicitly
- * blacklisted or masked by including the original SQL statement in the
+ * excluded or masked by including the original SQL statement in the
* change event. For this reason the default value is 'false'.
*
* The option will be converted to a <code>boolean</code>
@@ -1540,8 +1196,9 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option is a: <code>boolean</code> type.
*
@@ -1561,8 +1218,9 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1580,11 +1238,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option is a: <code>boolean</code> type.
*
@@ -1601,11 +1259,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1624,11 +1282,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
/**
* Specify how binlog events that belong to a table missing from
* internal schema representation (i.e. internal representation is not
- * consistent with database) should be handled, including:'fail' (the
+ * consistent with database) should be handled, including: 'fail' (the
* default) an exception indicating the problematic event and its binlog
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its binlog position will be logged and the
- * event will be skipped;'skip' the problematic event will be skipped.
+ * event will be skipped; 'skip' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1825,12 +1483,12 @@ public interface DebeziumMySqlEndpointBuilderFactory {
/**
* A semicolon-separated list of expressions that match fully-qualified
* tables and column(s) to be used as message key. Each expression must
- * match the pattern ':',where the table names could be defined as
+ * match the pattern ':', where the table names could be defined as
* (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on the
- * specific connector,and the key columns are a comma-separated list of
+ * specific connector, and the key columns are a comma-separated list of
* columns representing the custom key. For any table without an
* explicit key configuration the table's primary key column(s) will be
- * used as message key.Example:
+ * used as message key. Example:
* dbserver1.inventory.orderlines:orderId,orderLineId;dbserver1.inventory.orders:id.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1951,7 +1609,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option is a: <code>int</code> type.
*
@@ -1967,7 +1625,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option will be converted to a <code>int</code> type.
*
@@ -2049,15 +1707,131 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("sanitizeFieldNames", sanitizeFieldNames);
return this;
}
+ /**
+ * The name of the SchemaHistory class that should be used to store and
+ * recover database schema changes. The configuration properties for the
+ * history are prefixed with the 'schema.history.internal.' string.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.storage.kafka.history.KafkaSchemaHistory
+ * Group: mysql
+ *
+ * @param schemaHistoryInternal the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder schemaHistoryInternal(
+ String schemaHistoryInternal) {
+ doSetProperty("schemaHistoryInternal", schemaHistoryInternal);
+ return this;
+ }
+ /**
+ * The path to the file that will be used to record the database schema
+ * history.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Group: mysql
+ *
+ * @param schemaHistoryInternalFileFilename the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder schemaHistoryInternalFileFilename(
+ String schemaHistoryInternalFileFilename) {
+ doSetProperty("schemaHistoryInternalFileFilename", schemaHistoryInternalFileFilename);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: mysql
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ boolean schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: mysql
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ String schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: mysql
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ boolean schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: mysql
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ String schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
/**
* Specify how schema names should be adjusted for compatibility with
- * the message converter used by the connector, including:'avro'
+ * the message converter used by the connector, including: 'avro'
* replaces the characters that cannot be used in the Avro type name
- * with underscore (default)'none' does not apply any adjustment.
+ * with underscore; 'none' does not apply any adjustment (default).
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: avro
+ * Default: none
* Group: mysql
*
* @param schemaNameAdjustmentMode the value to set
@@ -2088,10 +1862,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
* The comma-separated list of operations to skip during streaming,
* defined as: 'c' for inserts/create; 'u' for updates; 'd' for deletes,
* 't' for truncates, and 'none' to indicate nothing skipped. By
- * default, no operations will be skipped.
+ * default, only truncate operations will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
+ * Default: t
* Group: mysql
*
* @param skippedOperations the value to set
@@ -2169,7 +1944,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
return this;
}
/**
- * this setting must be set to specify a list of tables/collections
+ * This setting must be set to specify a list of tables/collections
* whose snapshot must be taken on creating or restarting the connector.
*
* The option is a: <code>java.lang.String</code> type.
@@ -2314,10 +2089,10 @@ public interface DebeziumMySqlEndpointBuilderFactory {
/**
* BETA FEATURE: On connector restart, the connector will check if there
* have been any new tables added to the configuration, and snapshot
- * them. There is presently only two options:'off': Default behavior. Do
- * not snapshot new tables.'parallel': The snapshot of the new tables
- * will occur in parallel to the continued binlog reading of the old
- * tables. When the snapshot completes, an independent binlog reader
+ * them. There is presently only two options: 'off': Default behavior.
+ * Do not snapshot new tables. 'parallel': The snapshot of the new
+ * tables will occur in parallel to the continued binlog reading of the
+ * old tables. When the snapshot completes, an independent binlog reader
* will begin reading the events for the new tables until it catches up
* to present time. At this point, both old and new binlog readers will
* be momentarily halted and new binlog reader will start that will read
@@ -2341,7 +2116,7 @@ public interface DebeziumMySqlEndpointBuilderFactory {
/**
* This property contains a comma-separated list of fully-qualified
* tables (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on
- * thespecific connectors. Select statements for the individual tables
+ * the specific connectors. Select statements for the individual tables
* are specified in further configuration properties, one for each
* table, identified by the id
* 'snapshot.select.statement.overrides.DB_NAME.TABLE_NAME' or
@@ -2364,40 +2139,6 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("snapshotSelectStatementOverrides", snapshotSelectStatementOverrides);
return this;
}
- /**
- * A version of the format of the publicly visible source part in the
- * message.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: v2
- * Group: mysql
- *
- * @param sourceStructVersion the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder sourceStructVersion(
- String sourceStructVersion) {
- doSetProperty("sourceStructVersion", sourceStructVersion);
- return this;
- }
- /**
- * A comma-separated list of regular expressions that match the
- * fully-qualified names of tables to be excluded from monitoring
- * (deprecated, use table.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param tableBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder tableBlacklist(
- String tableBlacklist) {
- doSetProperty("tableBlacklist", tableBlacklist);
- return this;
- }
/**
* A comma-separated list of regular expressions that match the
* fully-qualified names of tables to be excluded from monitoring.
@@ -2462,27 +2203,11 @@ public interface DebeziumMySqlEndpointBuilderFactory {
doSetProperty("tableIncludeList", tableIncludeList);
return this;
}
- /**
- * The tables for which changes are to be captured (deprecated, use
- * table.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: mysql
- *
- * @param tableWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumMySqlEndpointBuilder tableWhitelist(
- String tableWhitelist) {
- doSetProperty("tableWhitelist", tableWhitelist);
- return this;
- }
/**
* Time, date and timestamps can be represented with different kinds of
- * precisions, including:'adaptive_time_microseconds': the precision of
+ * precisions, including: 'adaptive_time_microseconds': the precision of
* date and timestamp values is based the database column's precision;
- * but time fields always use microseconds precision;'connect': always
+ * but time fields always use microseconds precision; 'connect': always
* represents time, date and timestamp values using Kafka Connect's
* built-in representations for Time, Date, and Timestamp, which uses
* millisecond precision regardless of the database columns' precision.
@@ -2502,10 +2227,10 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option is a: <code>boolean</code> type.
*
@@ -2522,10 +2247,10 @@ public interface DebeziumMySqlEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -2542,21 +2267,41 @@ public interface DebeziumMySqlEndpointBuilderFactory {
return this;
}
/**
- * The name of the transaction metadata topic. The placeholder
- * ${database.server.name} can be used for referring to the connector's
- * logical name; defaults to ${database.server.name}.transaction.
+ * The name of the TopicNamingStrategy class that should be used to
+ * determine the topic name for data change, schema change, transaction,
+ * heartbeat event etc.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.schema.SchemaTopicNamingStrategy
+ * Group: mysql
+ *
+ * @param topicNamingStrategy the value to set
+ * @return the dsl builder
+ */
+ default DebeziumMySqlEndpointBuilder topicNamingStrategy(
+ String topicNamingStrategy) {
+ doSetProperty("topicNamingStrategy", topicNamingStrategy);
+ return this;
+ }
+ /**
+ * Topic prefix that identifies and provides a namespace for the
+ * particular database server/cluster is capturing changes. The topic
+ * prefix should be unique across all other connectors, since it is used
+ * as a prefix for all Kafka topic names that receive events emitted by
+ * this connector. Only alphanumeric characters, hyphens, dots and
+ * underscores must be accepted.
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: ${database.server.name}.transaction
+ * Required: true
* Group: mysql
*
- * @param transactionTopic the value to set
+ * @param topicPrefix the value to set
* @return the dsl builder
*/
- default DebeziumMySqlEndpointBuilder transactionTopic(
- String transactionTopic) {
- doSetProperty("transactionTopic", transactionTopic);
+ default DebeziumMySqlEndpointBuilder topicPrefix(String topicPrefix) {
+ doSetProperty("topicPrefix", topicPrefix);
return this;
}
}
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumOracleEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumOracleEndpointBuilderFactory.java
index 1c9f702a261..84b305ea300 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumOracleEndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumOracleEndpointBuilderFactory.java
@@ -330,9 +330,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Specify how binary (blob, binary, etc.) columns should be represented
- * in change events, including:'bytes' represents binary data as byte
- * array (default)'base64' represents binary data as base64-encoded
- * string'hex' represents binary data as hex-encoded (base16) string.
+ * in change events, including: 'bytes' represents binary data as byte
+ * array (default); 'base64' represents binary data as base64-encoded
+ * string; 'base64-url-safe' represents binary data as
+ * base64-url-safe-encoded string; 'hex' represents binary data as
+ * hex-encoded (base16) string.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -347,22 +349,6 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("binaryHandlingMode", binaryHandlingMode);
return this;
}
- /**
- * Regular expressions matching columns to exclude from change events
- * (deprecated, use column.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param columnBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder columnBlacklist(
- String columnBlacklist) {
- doSetProperty("columnBlacklist", columnBlacklist);
- return this;
- }
/**
* Regular expressions matching columns to exclude from change events.
*
@@ -411,22 +397,6 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("columnPropagateSourceType", columnPropagateSourceType);
return this;
}
- /**
- * Regular expressions matching columns to include in change events
- * (deprecated, use column.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param columnWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder columnWhitelist(
- String columnWhitelist) {
- doSetProperty("columnWhitelist", columnWhitelist);
- return this;
- }
/**
* Optional list of custom converters that would be used instead of
* default ones. The converters are defined using '.type' config option
@@ -477,299 +447,6 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("databaseDbname", databaseDbname);
return this;
}
- /**
- * The name of the DatabaseHistory class that should be used to store
- * and recover database schema changes. The configuration properties for
- * the history are prefixed with the 'database.history.' string.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: io.debezium.relational.history.KafkaDatabaseHistory
- * Group: oracle
- *
- * @param databaseHistory the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistory(
- String databaseHistory) {
- doSetProperty("databaseHistory", databaseHistory);
- return this;
- }
- /**
- * The path to the file that will be used to record the database
- * history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param databaseHistoryFileFilename the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryFileFilename(
- String databaseHistoryFileFilename) {
- doSetProperty("databaseHistoryFileFilename", databaseHistoryFileFilename);
- return this;
- }
- /**
- * A list of host/port pairs that the connector will use for
- * establishing the initial connection to the Kafka cluster for
- * retrieving database schema history previously stored by the
- * connector. This should point to the same Kafka cluster used by the
- * Kafka Connect process.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param databaseHistoryKafkaBootstrapServers the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaBootstrapServers(
- String databaseHistoryKafkaBootstrapServers) {
- doSetProperty("databaseHistoryKafkaBootstrapServers", databaseHistoryKafkaBootstrapServers);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option is a: <code>long</code> type.
- *
- * Default: 3s
- * Group: oracle
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- long databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option will be converted to a <code>long</code> type.
- *
- * Default: 3s
- * Group: oracle
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- String databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100
- * Group: oracle
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- int databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100
- * Group: oracle
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- String databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100ms
- * Group: oracle
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- int databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100ms
- * Group: oracle
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- String databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The name of the topic for the database schema history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param databaseHistoryKafkaTopic the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryKafkaTopic(
- String databaseHistoryKafkaTopic) {
- doSetProperty("databaseHistoryKafkaTopic", databaseHistoryKafkaTopic);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: oracle
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistorySkipUnparseableDdl(
- boolean databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: oracle
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistorySkipUnparseableDdl(
- String databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: oracle
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- boolean databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: oracle
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- String databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: oracle
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- boolean databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: oracle
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- String databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
/**
* Resolvable hostname or IP address of the database server.
*
@@ -863,25 +540,6 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("databasePort", databasePort);
return this;
}
- /**
- * Unique name that identifies the database server and all recorded
- * offsets, and that is used as a prefix for all schemas and topics.
- * Each distinct installation should have a separate namespace and be
- * monitored by at most one Debezium connector.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Required: true
- * Group: oracle
- *
- * @param databaseServerName the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder databaseServerName(
- String databaseServerName) {
- doSetProperty("databaseServerName", databaseServerName);
- return this;
- }
/**
* Complete JDBC URL as an alternative to specifying hostname, port and
* database provided as a way to support alternative connection
@@ -932,7 +590,7 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Specify how DECIMAL and NUMERIC columns should be represented in
- * change events, including:'precise' (the default) uses
+ * change events, including: 'precise' (the default) uses
* java.math.BigDecimal to represent values, which are encoded in the
* change events using a binary representation and Kafka Connect's
* 'org.apache.kafka.connect.data.Decimal' type; 'string' uses string to
@@ -955,11 +613,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Specify how failures during processing of events (i.e. when
- * encountering a corrupted event) should be handled, including:'fail'
+ * encountering a corrupted event) should be handled, including: 'fail'
* (the default) an exception indicating the problematic event and its
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its position will be logged and the event will
- * be skipped;'ignore' the problematic event will be skipped.
+ * be skipped; 'ignore' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1047,8 +705,9 @@ public interface DebeziumOracleEndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option is a: <code>boolean</code> type.
*
@@ -1068,8 +727,9 @@ public interface DebeziumOracleEndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1087,11 +747,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option is a: <code>boolean</code> type.
*
@@ -1108,11 +768,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1130,8 +790,8 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Specify how INTERVAL columns should be represented in change events,
- * including:'string' represents values as an exact ISO formatted
- * string'numeric' (default) represents values using the inexact
+ * including: 'string' represents values as an exact ISO formatted
+ * string; 'numeric' (default) represents values using the inexact
* conversion into microseconds.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1148,9 +808,9 @@ public interface DebeziumOracleEndpointBuilderFactory {
return this;
}
/**
- * When set to false, the default, LOB fields will not be captured nor
- * emitted. When set to true, the connector will capture LOB fields and
- * emit changes for those fields like any other column type.
+ * When set to 'false', the default, LOB fields will not be captured nor
+ * emitted. When set to 'true', the connector will capture LOB fields
+ * and emit changes for those fields like any other column type.
*
* The option is a: <code>boolean</code> type.
*
@@ -1165,9 +825,9 @@ public interface DebeziumOracleEndpointBuilderFactory {
return this;
}
/**
- * When set to false, the default, LOB fields will not be captured nor
- * emitted. When set to true, the connector will capture LOB fields and
- * emit changes for those fields like any other column type.
+ * When set to 'false', the default, LOB fields will not be captured nor
+ * emitted. When set to 'true', the connector will capture LOB fields
+ * and emit changes for those fields like any other column type.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1234,11 +894,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
return this;
}
/**
- * When set to false, the default, the connector will mine both archive
- * log and redo logs to emit change events. When set to true, the
- * connector will only mine archive logs. There are circumstances where
- * its advantageous to only mine archive logs and accept latency in
- * event emission due to frequent revolving redo logs.
+ * When set to 'false', the default, the connector will mine both
+ * archive log and redo logs to emit change events. When set to 'true',
+ * the connector will only mine archive logs. There are circumstances
+ * where its advantageous to only mine archive logs and accept latency
+ * in event emission due to frequent revolving redo logs.
*
* The option is a: <code>boolean</code> type.
*
@@ -1254,11 +914,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
return this;
}
/**
- * When set to false, the default, the connector will mine both archive
- * log and redo logs to emit change events. When set to true, the
- * connector will only mine archive logs. There are circumstances where
- * its advantageous to only mine archive logs and accept latency in
- * event emission due to frequent revolving redo logs.
+ * When set to 'false', the default, the connector will mine both
+ * archive log and redo logs to emit change events. When set to 'true',
+ * the connector will only mine archive logs. There are circumstances
+ * where its advantageous to only mine archive logs and accept latency
+ * in event emission due to frequent revolving redo logs.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1515,6 +1175,46 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("logMiningBufferInfinispanCacheTransactions", logMiningBufferInfinispanCacheTransactions);
return this;
}
+ /**
+ * The number of events a transaction can include before the transaction
+ * is discarded. This is useful for managing buffer memory and/or space
+ * when dealing with very large transactions. Defaults to 0, meaning
+ * that no threshold is applied and transactions can have unlimited
+ * events.
+ *
+ * The option is a: <code>long</code> type.
+ *
+ * Default: 0
+ * Group: oracle
+ *
+ * @param logMiningBufferTransactionEventsThreshold the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder logMiningBufferTransactionEventsThreshold(
+ long logMiningBufferTransactionEventsThreshold) {
+ doSetProperty("logMiningBufferTransactionEventsThreshold", logMiningBufferTransactionEventsThreshold);
+ return this;
+ }
+ /**
+ * The number of events a transaction can include before the transaction
+ * is discarded. This is useful for managing buffer memory and/or space
+ * when dealing with very large transactions. Defaults to 0, meaning
+ * that no threshold is applied and transactions can have unlimited
+ * events.
+ *
+ * The option will be converted to a <code>long</code> type.
+ *
+ * Default: 0
+ * Group: oracle
+ *
+ * @param logMiningBufferTransactionEventsThreshold the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder logMiningBufferTransactionEventsThreshold(
+ String logMiningBufferTransactionEventsThreshold) {
+ doSetProperty("logMiningBufferTransactionEventsThreshold", logMiningBufferTransactionEventsThreshold);
+ return this;
+ }
/**
* The buffer type controls how the connector manages buffering
* transaction data. memory - Uses the JVM process' heap to buffer all
@@ -1966,12 +1666,12 @@ public interface DebeziumOracleEndpointBuilderFactory {
/**
* A semicolon-separated list of expressions that match fully-qualified
* tables and column(s) to be used as message key. Each expression must
- * match the pattern ':',where the table names could be defined as
+ * match the pattern ':', where the table names could be defined as
* (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on the
- * specific connector,and the key columns are a comma-separated list of
+ * specific connector, and the key columns are a comma-separated list of
* columns representing the custom key. For any table without an
* explicit key configuration the table's primary key column(s) will be
- * used as message key.Example:
+ * used as message key. Example:
* dbserver1.inventory.orderlines:orderId,orderLineId;dbserver1.inventory.orders:id.
*
* The option is a: <code>java.lang.String</code> type.
@@ -2054,7 +1754,7 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option is a: <code>int</code> type.
*
@@ -2070,7 +1770,7 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option will be converted to a <code>int</code> type.
*
@@ -2166,15 +1866,131 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("sanitizeFieldNames", sanitizeFieldNames);
return this;
}
+ /**
+ * The name of the SchemaHistory class that should be used to store and
+ * recover database schema changes. The configuration properties for the
+ * history are prefixed with the 'schema.history.internal.' string.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.storage.kafka.history.KafkaSchemaHistory
+ * Group: oracle
+ *
+ * @param schemaHistoryInternal the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder schemaHistoryInternal(
+ String schemaHistoryInternal) {
+ doSetProperty("schemaHistoryInternal", schemaHistoryInternal);
+ return this;
+ }
+ /**
+ * The path to the file that will be used to record the database schema
+ * history.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Group: oracle
+ *
+ * @param schemaHistoryInternalFileFilename the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder schemaHistoryInternalFileFilename(
+ String schemaHistoryInternalFileFilename) {
+ doSetProperty("schemaHistoryInternalFileFilename", schemaHistoryInternalFileFilename);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: oracle
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ boolean schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: oracle
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ String schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: oracle
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ boolean schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: oracle
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ String schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
/**
* Specify how schema names should be adjusted for compatibility with
- * the message converter used by the connector, including:'avro'
+ * the message converter used by the connector, including: 'avro'
* replaces the characters that cannot be used in the Avro type name
- * with underscore (default)'none' does not apply any adjustment.
+ * with underscore; 'none' does not apply any adjustment (default).
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: avro
+ * Default: none
* Group: oracle
*
* @param schemaNameAdjustmentMode the value to set
@@ -2205,10 +2021,11 @@ public interface DebeziumOracleEndpointBuilderFactory {
* The comma-separated list of operations to skip during streaming,
* defined as: 'c' for inserts/create; 'u' for updates; 'd' for deletes,
* 't' for truncates, and 'none' to indicate nothing skipped. By
- * default, no operations will be skipped.
+ * default, only truncate operations will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
+ * Default: t
* Group: oracle
*
* @param skippedOperations the value to set
@@ -2301,7 +2118,7 @@ public interface DebeziumOracleEndpointBuilderFactory {
return this;
}
/**
- * this setting must be set to specify a list of tables/collections
+ * This setting must be set to specify a list of tables/collections
* whose snapshot must be taken on creating or restarting the connector.
*
* The option is a: <code>java.lang.String</code> type.
@@ -2434,7 +2251,7 @@ public interface DebeziumOracleEndpointBuilderFactory {
/**
* This property contains a comma-separated list of fully-qualified
* tables (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on
- * thespecific connectors. Select statements for the individual tables
+ * the specific connectors. Select statements for the individual tables
* are specified in further configuration properties, one for each
* table, identified by the id
* 'snapshot.select.statement.overrides.DB_NAME.TABLE_NAME' or
@@ -2457,40 +2274,6 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("snapshotSelectStatementOverrides", snapshotSelectStatementOverrides);
return this;
}
- /**
- * A version of the format of the publicly visible source part in the
- * message.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: v2
- * Group: oracle
- *
- * @param sourceStructVersion the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder sourceStructVersion(
- String sourceStructVersion) {
- doSetProperty("sourceStructVersion", sourceStructVersion);
- return this;
- }
- /**
- * A comma-separated list of regular expressions that match the
- * fully-qualified names of tables to be excluded from monitoring
- * (deprecated, use table.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param tableBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder tableBlacklist(
- String tableBlacklist) {
- doSetProperty("tableBlacklist", tableBlacklist);
- return this;
- }
/**
* A comma-separated list of regular expressions that match the
* fully-qualified names of tables to be excluded from monitoring.
@@ -2522,31 +2305,15 @@ public interface DebeziumOracleEndpointBuilderFactory {
doSetProperty("tableIncludeList", tableIncludeList);
return this;
}
- /**
- * The tables for which changes are to be captured (deprecated, use
- * table.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: oracle
- *
- * @param tableWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumOracleEndpointBuilder tableWhitelist(
- String tableWhitelist) {
- doSetProperty("tableWhitelist", tableWhitelist);
- return this;
- }
/**
* Time, date, and timestamps can be represented with different kinds of
- * precisions, including:'adaptive' (the default) bases the precision of
- * time, date, and timestamp values on the database column's precision;
- * 'adaptive_time_microseconds' like 'adaptive' mode, but TIME fields
- * always use microseconds precision;'connect' always represents time,
- * date, and timestamp values using Kafka Connect's built-in
- * representations for Time, Date, and Timestamp, which uses millisecond
- * precision regardless of the database columns' precision .
+ * precisions, including: 'adaptive' (the default) bases the precision
+ * of time, date, and timestamp values on the database column's
+ * precision; 'adaptive_time_microseconds' like 'adaptive' mode, but
+ * TIME fields always use microseconds precision; 'connect' always
+ * represents time, date, and timestamp values using Kafka Connect's
+ * built-in representations for Time, Date, and Timestamp, which uses
+ * millisecond precision regardless of the database columns' precision.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -2563,10 +2330,10 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option is a: <code>boolean</code> type.
*
@@ -2583,10 +2350,10 @@ public interface DebeziumOracleEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -2603,21 +2370,41 @@ public interface DebeziumOracleEndpointBuilderFactory {
return this;
}
/**
- * The name of the transaction metadata topic. The placeholder
- * ${database.server.name} can be used for referring to the connector's
- * logical name; defaults to ${database.server.name}.transaction.
+ * The name of the TopicNamingStrategy class that should be used to
+ * determine the topic name for data change, schema change, transaction,
+ * heartbeat event etc.
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: ${database.server.name}.transaction
+ * Default: io.debezium.schema.SchemaTopicNamingStrategy
+ * Group: oracle
+ *
+ * @param topicNamingStrategy the value to set
+ * @return the dsl builder
+ */
+ default DebeziumOracleEndpointBuilder topicNamingStrategy(
+ String topicNamingStrategy) {
+ doSetProperty("topicNamingStrategy", topicNamingStrategy);
+ return this;
+ }
+ /**
+ * Topic prefix that identifies and provides a namespace for the
+ * particular database server/cluster is capturing changes. The topic
+ * prefix should be unique across all other connectors, since it is used
+ * as a prefix for all Kafka topic names that receive events emitted by
+ * this connector. Only alphanumeric characters, hyphens, dots and
+ * underscores must be accepted.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Required: true
* Group: oracle
*
- * @param transactionTopic the value to set
+ * @param topicPrefix the value to set
* @return the dsl builder
*/
- default DebeziumOracleEndpointBuilder transactionTopic(
- String transactionTopic) {
- doSetProperty("transactionTopic", transactionTopic);
+ default DebeziumOracleEndpointBuilder topicPrefix(String topicPrefix) {
+ doSetProperty("topicPrefix", topicPrefix);
return this;
}
/**
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumPostgresEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumPostgresEndpointBuilderFactory.java
index 1a65cacfa06..9ab021a5e37 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumPostgresEndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumPostgresEndpointBuilderFactory.java
@@ -331,9 +331,11 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify how binary (blob, binary, etc.) columns should be represented
- * in change events, including:'bytes' represents binary data as byte
- * array (default)'base64' represents binary data as base64-encoded
- * string'hex' represents binary data as hex-encoded (base16) string.
+ * in change events, including: 'bytes' represents binary data as byte
+ * array (default); 'base64' represents binary data as base64-encoded
+ * string; 'base64-url-safe' represents binary data as
+ * base64-url-safe-encoded string; 'hex' represents binary data as
+ * hex-encoded (base16) string.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -348,22 +350,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("binaryHandlingMode", binaryHandlingMode);
return this;
}
- /**
- * Regular expressions matching columns to exclude from change events
- * (deprecated, use column.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: postgres
- *
- * @param columnBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder columnBlacklist(
- String columnBlacklist) {
- doSetProperty("columnBlacklist", columnBlacklist);
- return this;
- }
/**
* Regular expressions matching columns to exclude from change events.
*
@@ -412,22 +398,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("columnPropagateSourceType", columnPropagateSourceType);
return this;
}
- /**
- * Regular expressions matching columns to include in change events
- * (deprecated, use column.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: postgres
- *
- * @param columnWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder columnWhitelist(
- String columnWhitelist) {
- doSetProperty("columnWhitelist", columnWhitelist);
- return this;
- }
/**
* Optional list of custom converters that would be used instead of
* default ones. The converters are defined using '.type' config option
@@ -460,22 +430,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("databaseDbname", databaseDbname);
return this;
}
- /**
- * The path to the file that will be used to record the database
- * history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: postgres
- *
- * @param databaseHistoryFileFilename the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder databaseHistoryFileFilename(
- String databaseHistoryFileFilename) {
- doSetProperty("databaseHistoryFileFilename", databaseHistoryFileFilename);
- return this;
- }
/**
* Resolvable hostname or IP address of the database server.
*
@@ -495,7 +449,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
* A semicolon separated list of SQL statements to be executed when a
* JDBC connection to the database is established. Note that the
* connector may establish JDBC connections at its own discretion, so
- * this should typically be used for configurationof session parameters
+ * this should typically be used for configuration of session parameters
* only, but not for executing DML statements. Use doubled semicolon
* (';;') to use a semicolon as a character and not as a delimiter.
*
@@ -558,25 +512,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("databasePort", databasePort);
return this;
}
- /**
- * Unique name that identifies the database server and all recorded
- * offsets, and that is used as a prefix for all schemas and topics.
- * Each distinct installation should have a separate namespace and be
- * monitored by at most one Debezium connector.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Required: true
- * Group: postgres
- *
- * @param databaseServerName the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder databaseServerName(
- String databaseServerName) {
- doSetProperty("databaseServerName", databaseServerName);
- return this;
- }
/**
* File containing the SSL Certificate for the client. See the Postgres
* SSL docs for further information.
@@ -627,15 +562,15 @@ public interface DebeziumPostgresEndpointBuilderFactory {
return this;
}
/**
- * Whether to use an encrypted connection to Postgres. Options
- * include'disable' (the default) to use an unencrypted connection;
- * 'require' to use a secure (encrypted) connection, and fail if one
- * cannot be established; 'verify-ca' like 'required' but additionally
- * verify the server TLS certificate against the configured Certificate
- * Authority (CA) certificates, or fail if no valid matching CA
- * certificates are found; or'verify-full' like 'verify-ca' but
- * additionally verify that the server certificate matches the host to
- * which the connection is attempted.
+ * Whether to use an encrypted connection to Postgres. Options include:
+ * 'disable' (the default) to use an unencrypted connection; 'require'
+ * to use a secure (encrypted) connection, and fail if one cannot be
+ * established; 'verify-ca' like 'required' but additionally verify the
+ * server TLS certificate against the configured Certificate Authority
+ * (CA) certificates, or fail if no valid matching CA certificates are
+ * found; or 'verify-full' like 'verify-ca' but additionally verify that
+ * the server certificate matches the host to which the connection is
+ * attempted.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -751,7 +686,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify how DECIMAL and NUMERIC columns should be represented in
- * change events, including:'precise' (the default) uses
+ * change events, including: 'precise' (the default) uses
* java.math.BigDecimal to represent values, which are encoded in the
* change events using a binary representation and Kafka Connect's
* 'org.apache.kafka.connect.data.Decimal' type; 'string' uses string to
@@ -774,11 +709,11 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify how failures during processing of events (i.e. when
- * encountering a corrupted event) should be handled, including:'fail'
+ * encountering a corrupted event) should be handled, including: 'fail'
* (the default) an exception indicating the problematic event and its
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its position will be logged and the event will
- * be skipped;'ignore' the problematic event will be skipped.
+ * be skipped; 'ignore' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -793,6 +728,43 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("eventProcessingFailureHandlingMode", eventProcessingFailureHandlingMode);
return this;
}
+ /**
+ * Boolean to determine if Debezium should flush LSN in the source
+ * postgres database. If set to false, user will have to flush the LSN
+ * manually outside Debezium.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: true
+ * Group: postgres
+ *
+ * @param flushLsnSource the value to set
+ * @return the dsl builder
+ */
+ default DebeziumPostgresEndpointBuilder flushLsnSource(
+ boolean flushLsnSource) {
+ doSetProperty("flushLsnSource", flushLsnSource);
+ return this;
+ }
+ /**
+ * Boolean to determine if Debezium should flush LSN in the source
+ * postgres database. If set to false, user will have to flush the LSN
+ * manually outside Debezium.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: true
+ * Group: postgres
+ *
+ * @param flushLsnSource the value to set
+ * @return the dsl builder
+ */
+ default DebeziumPostgresEndpointBuilder flushLsnSource(
+ String flushLsnSource) {
+ doSetProperty("flushLsnSource", flushLsnSource);
+ return this;
+ }
/**
* The query executed with every heartbeat.
*
@@ -863,8 +835,8 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify how HSTORE columns should be represented in change events,
- * including:'json' represents values as string-ified JSON
- * (default)'map' represents values as a key/value map.
+ * including: 'json' represents values as string-ified JSON (default);
+ * 'map' represents values as a key/value map.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -881,11 +853,11 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option is a: <code>boolean</code> type.
*
@@ -902,11 +874,11 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -924,7 +896,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify whether the fields of data type not supported by Debezium
- * should be processed:'false' (the default) omits the fields; 'true'
+ * should be processed: 'false' (the default) omits the fields; 'true'
* converts the field into an implementation dependent binary
* representation.
*
@@ -943,7 +915,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify whether the fields of data type not supported by Debezium
- * should be processed:'false' (the default) omits the fields; 'true'
+ * should be processed: 'false' (the default) omits the fields; 'true'
* converts the field into an implementation dependent binary
* representation.
*
@@ -995,8 +967,8 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify how INTERVAL columns should be represented in change events,
- * including:'string' represents values as an exact ISO formatted
- * string'numeric' (default) represents values using the inexact
+ * including: 'string' represents values as an exact ISO formatted
+ * string; 'numeric' (default) represents values using the inexact
* conversion into microseconds.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1115,12 +1087,12 @@ public interface DebeziumPostgresEndpointBuilderFactory {
/**
* A semicolon-separated list of expressions that match fully-qualified
* tables and column(s) to be used as message key. Each expression must
- * match the pattern ':',where the table names could be defined as
+ * match the pattern ':', where the table names could be defined as
* (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on the
- * specific connector,and the key columns are a comma-separated list of
+ * specific connector, and the key columns are a comma-separated list of
* columns representing the custom key. For any table without an
* explicit key configuration the table's primary key column(s) will be
- * used as message key.Example:
+ * used as message key. Example:
* dbserver1.inventory.orderlines:orderId,orderLineId;dbserver1.inventory.orders:id.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1283,7 +1255,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* The name of the Postgres 10 publication used for streaming changes
- * from a plugin.Defaults to 'dbz_publication'.
+ * from a plugin. Defaults to 'dbz_publication'.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1300,7 +1272,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option is a: <code>int</code> type.
*
@@ -1317,7 +1289,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option will be converted to a <code>int</code> type.
*
@@ -1400,34 +1372,34 @@ public interface DebeziumPostgresEndpointBuilderFactory {
return this;
}
/**
- * The schemas for which events must not be captured (deprecated, use
- * schema.exclude.list instead).
+ * The schemas for which events must not be captured.
*
* The option is a: <code>java.lang.String</code> type.
*
* Group: postgres
*
- * @param schemaBlacklist the value to set
+ * @param schemaExcludeList the value to set
* @return the dsl builder
*/
- default DebeziumPostgresEndpointBuilder schemaBlacklist(
- String schemaBlacklist) {
- doSetProperty("schemaBlacklist", schemaBlacklist);
+ default DebeziumPostgresEndpointBuilder schemaExcludeList(
+ String schemaExcludeList) {
+ doSetProperty("schemaExcludeList", schemaExcludeList);
return this;
}
/**
- * The schemas for which events must not be captured.
+ * The path to the file that will be used to record the database schema
+ * history.
*
* The option is a: <code>java.lang.String</code> type.
*
* Group: postgres
*
- * @param schemaExcludeList the value to set
+ * @param schemaHistoryInternalFileFilename the value to set
* @return the dsl builder
*/
- default DebeziumPostgresEndpointBuilder schemaExcludeList(
- String schemaExcludeList) {
- doSetProperty("schemaExcludeList", schemaExcludeList);
+ default DebeziumPostgresEndpointBuilder schemaHistoryInternalFileFilename(
+ String schemaHistoryInternalFileFilename) {
+ doSetProperty("schemaHistoryInternalFileFilename", schemaHistoryInternalFileFilename);
return this;
}
/**
@@ -1447,13 +1419,13 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Specify how schema names should be adjusted for compatibility with
- * the message converter used by the connector, including:'avro'
+ * the message converter used by the connector, including: 'avro'
* replaces the characters that cannot be used in the Avro type name
- * with underscore (default)'none' does not apply any adjustment.
+ * with underscore; 'none' does not apply any adjustment (default).
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: avro
+ * Default: none
* Group: postgres
*
* @param schemaNameAdjustmentMode the value to set
@@ -1491,22 +1463,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("schemaRefreshMode", schemaRefreshMode);
return this;
}
- /**
- * The schemas for which events should be captured (deprecated, use
- * schema.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: postgres
- *
- * @param schemaWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder schemaWhitelist(
- String schemaWhitelist) {
- doSetProperty("schemaWhitelist", schemaWhitelist);
- return this;
- }
/**
* The name of the data collection that is used to send signals/commands
* to Debezium. Signaling is disabled when not set.
@@ -1527,7 +1483,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
* The comma-separated list of operations to skip during streaming,
* defined as: 'c' for inserts/create; 'u' for updates; 'd' for deletes,
* 't' for truncates, and 'none' to indicate nothing skipped. By
- * default, no operations will be skipped.
+ * default, only truncate operations will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1544,8 +1500,8 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Whether or not to drop the logical replication slot when the
- * connector finishes orderlyBy default the replication is kept so that
- * on restart progress can resume from the last recorded location.
+ * connector finishes orderly. By default the replication is kept so
+ * that on restart progress can resume from the last recorded location.
*
* The option is a: <code>boolean</code> type.
*
@@ -1562,8 +1518,8 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Whether or not to drop the logical replication slot when the
- * connector finishes orderlyBy default the replication is kept so that
- * on restart progress can resume from the last recorded location.
+ * connector finishes orderly. By default the replication is kept so
+ * that on restart progress can resume from the last recorded location.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1615,7 +1571,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* The name of the Postgres logical decoding slot created for streaming
- * changes from a plugin.Defaults to 'debezium.
+ * changes from a plugin. Defaults to 'debezium.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1685,7 +1641,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
/**
* When 'snapshot.mode' is set as custom, this setting must be set to
* specify a fully qualified class name to load (via the default class
- * loader).This class must implement the 'Snapshotter' interface and is
+ * loader). This class must implement the 'Snapshotter' interface and is
* called on each app boot to determine whether to do a snapshot and how
* to build queries.
*
@@ -1768,7 +1724,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
return this;
}
/**
- * this setting must be set to specify a list of tables/collections
+ * This setting must be set to specify a list of tables/collections
* whose snapshot must be taken on creating or restarting the connector.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1882,7 +1838,7 @@ public interface DebeziumPostgresEndpointBuilderFactory {
/**
* This property contains a comma-separated list of fully-qualified
* tables (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on
- * thespecific connectors. Select statements for the individual tables
+ * the specific connectors. Select statements for the individual tables
* are specified in further configuration properties, one for each
* table, identified by the id
* 'snapshot.select.statement.overrides.DB_NAME.TABLE_NAME' or
@@ -1905,23 +1861,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("snapshotSelectStatementOverrides", snapshotSelectStatementOverrides);
return this;
}
- /**
- * A version of the format of the publicly visible source part in the
- * message.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: v2
- * Group: postgres
- *
- * @param sourceStructVersion the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder sourceStructVersion(
- String sourceStructVersion) {
- doSetProperty("sourceStructVersion", sourceStructVersion);
- return this;
- }
/**
* Frequency for sending replication connection status updates to the
* server, given in milliseconds. Defaults to 10 seconds (10,000 ms).
@@ -1956,23 +1895,6 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("statusUpdateIntervalMs", statusUpdateIntervalMs);
return this;
}
- /**
- * A comma-separated list of regular expressions that match the
- * fully-qualified names of tables to be excluded from monitoring
- * (deprecated, use table.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: postgres
- *
- * @param tableBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder tableBlacklist(
- String tableBlacklist) {
- doSetProperty("tableBlacklist", tableBlacklist);
- return this;
- }
/**
* A comma-separated list of regular expressions that match the
* fully-qualified names of tables to be excluded from monitoring.
@@ -2037,31 +1959,15 @@ public interface DebeziumPostgresEndpointBuilderFactory {
doSetProperty("tableIncludeList", tableIncludeList);
return this;
}
- /**
- * The tables for which changes are to be captured (deprecated, use
- * table.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: postgres
- *
- * @param tableWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumPostgresEndpointBuilder tableWhitelist(
- String tableWhitelist) {
- doSetProperty("tableWhitelist", tableWhitelist);
- return this;
- }
/**
* Time, date, and timestamps can be represented with different kinds of
- * precisions, including:'adaptive' (the default) bases the precision of
- * time, date, and timestamp values on the database column's precision;
- * 'adaptive_time_microseconds' like 'adaptive' mode, but TIME fields
- * always use microseconds precision;'connect' always represents time,
- * date, and timestamp values using Kafka Connect's built-in
- * representations for Time, Date, and Timestamp, which uses millisecond
- * precision regardless of the database columns' precision .
+ * precisions, including: 'adaptive' (the default) bases the precision
+ * of time, date, and timestamp values on the database column's
+ * precision; 'adaptive_time_microseconds' like 'adaptive' mode, but
+ * TIME fields always use microseconds precision; 'connect' always
+ * represents time, date, and timestamp values using Kafka Connect's
+ * built-in representations for Time, Date, and Timestamp, which uses
+ * millisecond precision regardless of the database columns' precision.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -2078,10 +1984,10 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option is a: <code>boolean</code> type.
*
@@ -2098,10 +2004,10 @@ public interface DebeziumPostgresEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -2118,21 +2024,41 @@ public interface DebeziumPostgresEndpointBuilderFactory {
return this;
}
/**
- * The name of the transaction metadata topic. The placeholder
- * ${database.server.name} can be used for referring to the connector's
- * logical name; defaults to ${database.server.name}.transaction.
+ * The name of the TopicNamingStrategy class that should be used to
+ * determine the topic name for data change, schema change, transaction,
+ * heartbeat event etc.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.schema.SchemaTopicNamingStrategy
+ * Group: postgres
+ *
+ * @param topicNamingStrategy the value to set
+ * @return the dsl builder
+ */
+ default DebeziumPostgresEndpointBuilder topicNamingStrategy(
+ String topicNamingStrategy) {
+ doSetProperty("topicNamingStrategy", topicNamingStrategy);
+ return this;
+ }
+ /**
+ * Topic prefix that identifies and provides a namespace for the
+ * particular database server/cluster is capturing changes. The topic
+ * prefix should be unique across all other connectors, since it is used
+ * as a prefix for all Kafka topic names that receive events emitted by
+ * this connector. Only alphanumeric characters, hyphens, dots and
+ * underscores must be accepted.
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: ${database.server.name}.transaction
+ * Required: true
* Group: postgres
*
- * @param transactionTopic the value to set
+ * @param topicPrefix the value to set
* @return the dsl builder
*/
- default DebeziumPostgresEndpointBuilder transactionTopic(
- String transactionTopic) {
- doSetProperty("transactionTopic", transactionTopic);
+ default DebeziumPostgresEndpointBuilder topicPrefix(String topicPrefix) {
+ doSetProperty("topicPrefix", topicPrefix);
return this;
}
/**
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumSqlserverEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumSqlserverEndpointBuilderFactory.java
index ed725b84cad..a6691772dcb 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumSqlserverEndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/DebeziumSqlserverEndpointBuilderFactory.java
@@ -331,9 +331,11 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Specify how binary (blob, binary, etc.) columns should be represented
- * in change events, including:'bytes' represents binary data as byte
- * array (default)'base64' represents binary data as base64-encoded
- * string'hex' represents binary data as hex-encoded (base16) string.
+ * in change events, including: 'bytes' represents binary data as byte
+ * array (default); 'base64' represents binary data as base64-encoded
+ * string; 'base64-url-safe' represents binary data as
+ * base64-url-safe-encoded string; 'hex' represents binary data as
+ * hex-encoded (base16) string.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -348,22 +350,6 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("binaryHandlingMode", binaryHandlingMode);
return this;
}
- /**
- * Regular expressions matching columns to exclude from change events
- * (deprecated, use column.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param columnBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder columnBlacklist(
- String columnBlacklist) {
- doSetProperty("columnBlacklist", columnBlacklist);
- return this;
- }
/**
* Regular expressions matching columns to exclude from change events.
*
@@ -412,22 +398,6 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("columnPropagateSourceType", columnPropagateSourceType);
return this;
}
- /**
- * Regular expressions matching columns to include in change events
- * (deprecated, use column.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param columnWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder columnWhitelist(
- String columnWhitelist) {
- doSetProperty("columnWhitelist", columnWhitelist);
- return this;
- }
/**
* Optional list of custom converters that would be used instead of
* default ones. The converters are defined using '.type' config option
@@ -444,315 +414,6 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("converters", converters);
return this;
}
- /**
- * The name of the database from which the connector should capture
- * changes.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param databaseDbname the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseDbname(
- String databaseDbname) {
- doSetProperty("databaseDbname", databaseDbname);
- return this;
- }
- /**
- * The name of the DatabaseHistory class that should be used to store
- * and recover database schema changes. The configuration properties for
- * the history are prefixed with the 'database.history.' string.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: io.debezium.relational.history.FileDatabaseHistory
- * Group: sqlserver
- *
- * @param databaseHistory the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistory(
- String databaseHistory) {
- doSetProperty("databaseHistory", databaseHistory);
- return this;
- }
- /**
- * The path to the file that will be used to record the database
- * history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param databaseHistoryFileFilename the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryFileFilename(
- String databaseHistoryFileFilename) {
- doSetProperty("databaseHistoryFileFilename", databaseHistoryFileFilename);
- return this;
- }
- /**
- * A list of host/port pairs that the connector will use for
- * establishing the initial connection to the Kafka cluster for
- * retrieving database schema history previously stored by the
- * connector. This should point to the same Kafka cluster used by the
- * Kafka Connect process.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaBootstrapServers the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaBootstrapServers(
- String databaseHistoryKafkaBootstrapServers) {
- doSetProperty("databaseHistoryKafkaBootstrapServers", databaseHistoryKafkaBootstrapServers);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option is a: <code>long</code> type.
- *
- * Default: 3s
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- long databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while fetching cluster information
- * using Kafka admin client.
- *
- * The option will be converted to a <code>long</code> type.
- *
- * Default: 3s
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaQueryTimeoutMs the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaQueryTimeoutMs(
- String databaseHistoryKafkaQueryTimeoutMs) {
- doSetProperty("databaseHistoryKafkaQueryTimeoutMs", databaseHistoryKafkaQueryTimeoutMs);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- int databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of attempts in a row that no data are returned from Kafka
- * before recover completes. The maximum amount of time to wait after
- * receiving no data is (recovery.attempts) x
- * (recovery.poll.interval.ms).
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaRecoveryAttempts the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaRecoveryAttempts(
- String databaseHistoryKafkaRecoveryAttempts) {
- doSetProperty("databaseHistoryKafkaRecoveryAttempts", databaseHistoryKafkaRecoveryAttempts);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option is a: <code>int</code> type.
- *
- * Default: 100ms
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- int databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The number of milliseconds to wait while polling for persisted data
- * during recovery.
- *
- * The option will be converted to a <code>int</code> type.
- *
- * Default: 100ms
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaRecoveryPollIntervalMs the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaRecoveryPollIntervalMs(
- String databaseHistoryKafkaRecoveryPollIntervalMs) {
- doSetProperty("databaseHistoryKafkaRecoveryPollIntervalMs", databaseHistoryKafkaRecoveryPollIntervalMs);
- return this;
- }
- /**
- * The name of the topic for the database schema history.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param databaseHistoryKafkaTopic the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryKafkaTopic(
- String databaseHistoryKafkaTopic) {
- doSetProperty("databaseHistoryKafkaTopic", databaseHistoryKafkaTopic);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: sqlserver
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistorySkipUnparseableDdl(
- boolean databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls the action Debezium will take when it meets a DDL statement
- * in binlog, that it cannot parse.By default the connector will stop
- * operating but by changing the setting it can ignore the statements
- * which it cannot parse. If skipping is enabled then Debezium can miss
- * metadata changes.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: sqlserver
- *
- * @param databaseHistorySkipUnparseableDdl the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistorySkipUnparseableDdl(
- String databaseHistorySkipUnparseableDdl) {
- doSetProperty("databaseHistorySkipUnparseableDdl", databaseHistorySkipUnparseableDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: sqlserver
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- boolean databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a captured table will be stored.
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: sqlserver
- *
- * @param databaseHistoryStoreOnlyCapturedTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryStoreOnlyCapturedTablesDdl(
- String databaseHistoryStoreOnlyCapturedTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyCapturedTablesDdl", databaseHistoryStoreOnlyCapturedTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option is a: <code>boolean</code> type.
- *
- * Default: false
- * Group: sqlserver
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- boolean databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
- /**
- * Controls what DDL will Debezium store in database history. By default
- * (false) Debezium will store all incoming DDL statements. If set to
- * true, then only DDL that manipulates a monitored table will be stored
- * (deprecated, use database.history.store.only.captured.tables.ddl
- * instead).
- *
- * The option will be converted to a <code>boolean</code>
- * type.
- *
- * Default: false
- * Group: sqlserver
- *
- * @param databaseHistoryStoreOnlyMonitoredTablesDdl the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseHistoryStoreOnlyMonitoredTablesDdl(
- String databaseHistoryStoreOnlyMonitoredTablesDdl) {
- doSetProperty("databaseHistoryStoreOnlyMonitoredTablesDdl", databaseHistoryStoreOnlyMonitoredTablesDdl);
- return this;
- }
/**
* Resolvable hostname or IP address of the database server.
*
@@ -847,25 +508,6 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("databasePort", databasePort);
return this;
}
- /**
- * Unique name that identifies the database server and all recorded
- * offsets, and that is used as a prefix for all schemas and topics.
- * Each distinct installation should have a separate namespace and be
- * monitored by at most one Debezium connector.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Required: true
- * Group: sqlserver
- *
- * @param databaseServerName the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder databaseServerName(
- String databaseServerName) {
- doSetProperty("databaseServerName", databaseServerName);
- return this;
- }
/**
* Name of the database user to be used when connecting to the database.
*
@@ -901,7 +543,7 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Specify how DECIMAL and NUMERIC columns should be represented in
- * change events, including:'precise' (the default) uses
+ * change events, including: 'precise' (the default) uses
* java.math.BigDecimal to represent values, which are encoded in the
* change events using a binary representation and Kafka Connect's
* 'org.apache.kafka.connect.data.Decimal' type; 'string' uses string to
@@ -924,11 +566,11 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Specify how failures during processing of events (i.e. when
- * encountering a corrupted event) should be handled, including:'fail'
+ * encountering a corrupted event) should be handled, including: 'fail'
* (the default) an exception indicating the problematic event and its
* position is raised, causing the connector to be stopped; 'warn' the
* problematic event and its position will be logged and the event will
- * be skipped;'ignore' the problematic event will be skipped.
+ * be skipped; 'ignore' the problematic event will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1016,8 +658,9 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option is a: <code>boolean</code> type.
*
@@ -1037,8 +680,9 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
* to a Kafka topic with the same name as the database server ID. Each
* schema change will be recorded using a key that contains the database
* name and whose value include logical description of the new schema
- * and optionally the DDL statement(s).The default is 'true'. This is
- * independent of how the connector internally records database history.
+ * and optionally the DDL statement(s). The default is 'true'. This is
+ * independent of how the connector internally records database schema
+ * history.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1056,11 +700,11 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option is a: <code>boolean</code> type.
*
@@ -1077,11 +721,11 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Whether the connector parse table and column's comment to metadata
- * object.Note: Enable this option will bring the implications on memory
- * usage. The number and size of ColumnImpl objects is what largely
- * impacts how much memory is consumed by the Debezium connectors, and
- * adding a String to each of them can potentially be quite heavy. The
- * default is 'false'.
+ * object. Note: Enable this option will bring the implications on
+ * memory usage. The number and size of ColumnImpl objects is what
+ * largely impacts how much memory is consumed by the Debezium
+ * connectors, and adding a String to each of them can potentially be
+ * quite heavy. The default is 'false'.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -1354,12 +998,12 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
/**
* A semicolon-separated list of expressions that match fully-qualified
* tables and column(s) to be used as message key. Each expression must
- * match the pattern ':',where the table names could be defined as
+ * match the pattern ':', where the table names could be defined as
* (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on the
- * specific connector,and the key columns are a comma-separated list of
+ * specific connector, and the key columns are a comma-separated list of
* columns representing the custom key. For any table without an
* explicit key configuration the table's primary key column(s) will be
- * used as message key.Example:
+ * used as message key. Example:
* dbserver1.inventory.orderlines:orderId,orderLineId;dbserver1.inventory.orders:id.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1443,7 +1087,7 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option is a: <code>int</code> type.
*
@@ -1460,7 +1104,7 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* The maximum number of records that should be loaded into memory while
- * streaming. A value of 0 uses the default JDBC fetch size.
+ * streaming. A value of '0' uses the default JDBC fetch size.
*
* The option will be converted to a <code>int</code> type.
*
@@ -1542,15 +1186,131 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("sanitizeFieldNames", sanitizeFieldNames);
return this;
}
+ /**
+ * The name of the SchemaHistory class that should be used to store and
+ * recover database schema changes. The configuration properties for the
+ * history are prefixed with the 'schema.history.internal.' string.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Default: io.debezium.storage.kafka.history.KafkaSchemaHistory
+ * Group: sqlserver
+ *
+ * @param schemaHistoryInternal the value to set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder schemaHistoryInternal(
+ String schemaHistoryInternal) {
+ doSetProperty("schemaHistoryInternal", schemaHistoryInternal);
+ return this;
+ }
+ /**
+ * The path to the file that will be used to record the database schema
+ * history.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Group: sqlserver
+ *
+ * @param schemaHistoryInternalFileFilename the value to set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder schemaHistoryInternalFileFilename(
+ String schemaHistoryInternalFileFilename) {
+ doSetProperty("schemaHistoryInternalFileFilename", schemaHistoryInternalFileFilename);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: sqlserver
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ boolean schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls the action Debezium will take when it meets a DDL statement
+ * in binlog, that it cannot parse.By default the connector will stop
+ * operating but by changing the setting it can ignore the statements
+ * which it cannot parse. If skipping is enabled then Debezium can miss
+ * metadata changes.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: sqlserver
+ *
+ * @param schemaHistoryInternalSkipUnparseableDdl the value to set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder schemaHistoryInternalSkipUnparseableDdl(
+ String schemaHistoryInternalSkipUnparseableDdl) {
+ doSetProperty("schemaHistoryInternalSkipUnparseableDdl", schemaHistoryInternalSkipUnparseableDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option is a: <code>boolean</code> type.
+ *
+ * Default: false
+ * Group: sqlserver
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ boolean schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
+ /**
+ * Controls what DDL will Debezium store in database schema history. By
+ * default (false) Debezium will store all incoming DDL statements. If
+ * set to true, then only DDL that manipulates a captured table will be
+ * stored.
+ *
+ * The option will be converted to a <code>boolean</code>
+ * type.
+ *
+ * Default: false
+ * Group: sqlserver
+ *
+ * @param schemaHistoryInternalStoreOnlyCapturedTablesDdl the value to
+ * set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder schemaHistoryInternalStoreOnlyCapturedTablesDdl(
+ String schemaHistoryInternalStoreOnlyCapturedTablesDdl) {
+ doSetProperty("schemaHistoryInternalStoreOnlyCapturedTablesDdl", schemaHistoryInternalStoreOnlyCapturedTablesDdl);
+ return this;
+ }
/**
* Specify how schema names should be adjusted for compatibility with
- * the message converter used by the connector, including:'avro'
+ * the message converter used by the connector, including: 'avro'
* replaces the characters that cannot be used in the Avro type name
- * with underscore (default)'none' does not apply any adjustment.
+ * with underscore; 'none' does not apply any adjustment (default).
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: avro
+ * Default: none
* Group: sqlserver
*
* @param schemaNameAdjustmentMode the value to set
@@ -1581,10 +1341,11 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
* The comma-separated list of operations to skip during streaming,
* defined as: 'c' for inserts/create; 'u' for updates; 'd' for deletes,
* 't' for truncates, and 'none' to indicate nothing skipped. By
- * default, no operations will be skipped.
+ * default, only truncate operations will be skipped.
*
* The option is a: <code>java.lang.String</code> type.
*
+ * Default: t
* Group: sqlserver
*
* @param skippedOperations the value to set
@@ -1662,7 +1423,7 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
return this;
}
/**
- * this setting must be set to specify a list of tables/collections
+ * This setting must be set to specify a list of tables/collections
* whose snapshot must be taken on creating or restarting the connector.
*
* The option is a: <code>java.lang.String</code> type.
@@ -1679,12 +1440,12 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Controls which transaction isolation level is used and how long the
- * connector locks the monitored tables. The default is
+ * connector locks the captured tables. The default is
* 'repeatable_read', which means that repeatable read isolation level
* is used. In addition, exclusive locks are taken only during schema
* snapshot. Using a value of 'exclusive' ensures that the connector
* holds the exclusive lock (and thus prevents any reads and updates)
- * for all monitored tables during the entire snapshot duration. When
+ * for all captured tables during the entire snapshot duration. When
* 'snapshot' is specified, connector runs the initial snapshot in
* SNAPSHOT isolation level, which guarantees snapshot consistency. In
* addition, neither table nor row-level locks are held. When
@@ -1802,7 +1563,7 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
/**
* This property contains a comma-separated list of fully-qualified
* tables (DB_NAME.TABLE_NAME) or (SCHEMA_NAME.TABLE_NAME), depending on
- * thespecific connectors. Select statements for the individual tables
+ * the specific connectors. Select statements for the individual tables
* are specified in further configuration properties, one for each
* table, identified by the id
* 'snapshot.select.statement.overrides.DB_NAME.TABLE_NAME' or
@@ -1825,60 +1586,6 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("snapshotSelectStatementOverrides", snapshotSelectStatementOverrides);
return this;
}
- /**
- * A version of the format of the publicly visible source part in the
- * message.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: v2
- * Group: sqlserver
- *
- * @param sourceStructVersion the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder sourceStructVersion(
- String sourceStructVersion) {
- doSetProperty("sourceStructVersion", sourceStructVersion);
- return this;
- }
- /**
- * Configures the criteria of the attached timestamp within the source
- * record (ts_ms).Options include:'commit', (default) the source
- * timestamp is set to the instant where the record was committed in the
- * database'processing', (deprecated) the source timestamp is set to the
- * instant where the record was processed by Debezium.
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Default: commit
- * Group: sqlserver
- *
- * @param sourceTimestampMode the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder sourceTimestampMode(
- String sourceTimestampMode) {
- doSetProperty("sourceTimestampMode", sourceTimestampMode);
- return this;
- }
- /**
- * A comma-separated list of regular expressions that match the
- * fully-qualified names of tables to be excluded from monitoring
- * (deprecated, use table.exclude.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param tableBlacklist the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder tableBlacklist(
- String tableBlacklist) {
- doSetProperty("tableBlacklist", tableBlacklist);
- return this;
- }
/**
* A comma-separated list of regular expressions that match the
* fully-qualified names of tables to be excluded from monitoring.
@@ -1943,31 +1650,15 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
doSetProperty("tableIncludeList", tableIncludeList);
return this;
}
- /**
- * The tables for which changes are to be captured (deprecated, use
- * table.include.list instead).
- *
- * The option is a: <code>java.lang.String</code> type.
- *
- * Group: sqlserver
- *
- * @param tableWhitelist the value to set
- * @return the dsl builder
- */
- default DebeziumSqlserverEndpointBuilder tableWhitelist(
- String tableWhitelist) {
- doSetProperty("tableWhitelist", tableWhitelist);
- return this;
- }
/**
* Time, date, and timestamps can be represented with different kinds of
- * precisions, including:'adaptive' (the default) bases the precision of
- * time, date, and timestamp values on the database column's precision;
- * 'adaptive_time_microseconds' like 'adaptive' mode, but TIME fields
- * always use microseconds precision;'connect' always represents time,
- * date, and timestamp values using Kafka Connect's built-in
- * representations for Time, Date, and Timestamp, which uses millisecond
- * precision regardless of the database columns' precision .
+ * precisions, including: 'adaptive' (the default) bases the precision
+ * of time, date, and timestamp values on the database column's
+ * precision; 'adaptive_time_microseconds' like 'adaptive' mode, but
+ * TIME fields always use microseconds precision; 'connect' always
+ * represents time, date, and timestamp values using Kafka Connect's
+ * built-in representations for Time, Date, and Timestamp, which uses
+ * millisecond precision regardless of the database columns' precision.
*
* The option is a: <code>java.lang.String</code> type.
*
@@ -1984,10 +1675,10 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option is a: <code>boolean</code> type.
*
@@ -2004,10 +1695,10 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
}
/**
* Whether delete operations should be represented by a delete event and
- * a subsquenttombstone event (true) or only by a delete event (false).
- * Emitting the tombstone event (the default behavior) allows Kafka to
- * completely delete all events pertaining to the given key once the
- * source record got deleted.
+ * a subsequent tombstone event (true) or only by a delete event
+ * (false). Emitting the tombstone event (the default behavior) allows
+ * Kafka to completely delete all events pertaining to the given key
+ * once the source record got deleted.
*
* The option will be converted to a <code>boolean</code>
* type.
@@ -2024,21 +1715,41 @@ public interface DebeziumSqlserverEndpointBuilderFactory {
return this;
}
/**
- * The name of the transaction metadata topic. The placeholder
- * ${database.server.name} can be used for referring to the connector's
- * logical name; defaults to ${database.server.name}.transaction.
+ * The name of the TopicNamingStrategy class that should be used to
+ * determine the topic name for data change, schema change, transaction,
+ * heartbeat event etc.
*
* The option is a: <code>java.lang.String</code> type.
*
- * Default: ${database.server.name}.transaction
+ * Default: io.debezium.schema.SchemaTopicNamingStrategy
+ * Group: sqlserver
+ *
+ * @param topicNamingStrategy the value to set
+ * @return the dsl builder
+ */
+ default DebeziumSqlserverEndpointBuilder topicNamingStrategy(
+ String topicNamingStrategy) {
+ doSetProperty("topicNamingStrategy", topicNamingStrategy);
+ return this;
+ }
+ /**
+ * Topic prefix that identifies and provides a namespace for the
+ * particular database server/cluster is capturing changes. The topic
+ * prefix should be unique across all other connectors, since it is used
+ * as a prefix for all Kafka topic names that receive events emitted by
+ * this connector. Only alphanumeric characters, hyphens, dots and
+ * underscores must be accepted.
+ *
+ * The option is a: <code>java.lang.String</code> type.
+ *
+ * Required: true
* Group: sqlserver
*
- * @param transactionTopic the value to set
+ * @param topicPrefix the value to set
* @return the dsl builder
*/
- default DebeziumSqlserverEndpointBuilder transactionTopic(
- String transactionTopic) {
- doSetProperty("transactionTopic", transactionTopic);
+ default DebeziumSqlserverEndpointBuilder topicPrefix(String topicPrefix) {
+ doSetProperty("topicPrefix", topicPrefix);
return this;
}
}