You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@camel.apache.org by ac...@apache.org on 2023/07/11 08:02:52 UTC

[camel] branch optaplanner created (now 1f81d30ee3a)

This is an automated email from the ASF dual-hosted git repository.

acosentino pushed a change to branch optaplanner
in repository https://gitbox.apache.org/repos/asf/camel.git


      at 1f81d30ee3a Regen

This branch includes the following new commits:

     new 1db9e2b01d8 Upgrade Optaplanner to version 9.41.0.Final
     new 1f81d30ee3a Regen

The 2 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.



[camel] 01/02: Upgrade Optaplanner to version 9.41.0.Final

Posted by ac...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

acosentino pushed a commit to branch optaplanner
in repository https://gitbox.apache.org/repos/asf/camel.git

commit 1db9e2b01d84204f1432c5d8bc746cbea2f03512
Author: Andrea Cosentino <an...@gmail.com>
AuthorDate: Tue Jul 11 09:18:15 2023 +0200

    Upgrade Optaplanner to version 9.41.0.Final
    
    Signed-off-by: Andrea Cosentino <an...@gmail.com>
---
 parent/pom.xml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/parent/pom.xml b/parent/pom.xml
index 894871b50a5..31c0721a29b 100644
--- a/parent/pom.xml
+++ b/parent/pom.xml
@@ -385,7 +385,7 @@
         <openstack4j-version>3.11</openstack4j-version>
         <opentelemetry-version>1.26.0</opentelemetry-version>
         <opentelemetry-alpha-version>${opentelemetry-version}-alpha</opentelemetry-alpha-version>
-        <optaplanner-version>9.37.0.Final</optaplanner-version>
+        <optaplanner-version>9.41.0.Final</optaplanner-version>
         <os-maven-plugin-version>1.7.0</os-maven-plugin-version>
         <paho-version>1.2.5</paho-version>
         <parquet-common-version>1.13.1</parquet-common-version>


[camel] 02/02: Regen

Posted by ac...@apache.org.
This is an automated email from the ASF dual-hosted git repository.

acosentino pushed a commit to branch optaplanner
in repository https://gitbox.apache.org/repos/asf/camel.git

commit 1f81d30ee3aba3866cddc1833a039cdeb987873a
Author: Andrea Cosentino <an...@gmail.com>
AuthorDate: Tue Jul 11 09:57:16 2023 +0200

    Regen
    
    Signed-off-by: Andrea Cosentino <an...@gmail.com>
---
 ...neCacheProducerMultiOperationSameCacheTest.java |  3 -
 ...adCacheProducerMultiOperationSameCacheTest.java |  3 -
 .../camel/component/kafka/KafkaComponent.java      | 10 ++--
 .../camel/component/kafka/KafkaConfiguration.java  | 26 ++++-----
 .../others/examples/json/platform-http-main.json   |  1 +
 docs/components/modules/others/nav.adoc            |  1 +
 .../modules/others/pages/platform-http-main.adoc   |  1 +
 .../dsl/KafkaComponentBuilderFactory.java          | 38 ++++++-------
 .../endpoint/dsl/KafkaEndpointBuilderFactory.java  | 64 +++++++++++-----------
 9 files changed, 72 insertions(+), 75 deletions(-)

diff --git a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
index abcebc693ab..fec5aeefcc1 100644
--- a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
+++ b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/cache/CaffeineCacheProducerMultiOperationSameCacheTest.java
@@ -16,9 +16,6 @@
  */
 package org.apache.camel.component.caffeine.cache;
 
-import java.util.HashMap;
-import java.util.Map;
-
 import org.apache.camel.builder.RouteBuilder;
 import org.apache.camel.component.caffeine.CaffeineConstants;
 import org.apache.camel.component.mock.MockEndpoint;
diff --git a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
index 68eec560449..3a955336dc2 100644
--- a/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
+++ b/components/camel-caffeine/src/test/java/org/apache/camel/component/caffeine/loadcache/CaffeineLoadCacheProducerMultiOperationSameCacheTest.java
@@ -16,9 +16,6 @@
  */
 package org.apache.camel.component.caffeine.loadcache;
 
-import java.util.HashMap;
-import java.util.Map;
-
 import org.apache.camel.builder.RouteBuilder;
 import org.apache.camel.component.caffeine.CaffeineConstants;
 import org.apache.camel.component.mock.MockEndpoint;
diff --git a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
index 2f2618939f7..7b8b5322a2e 100644
--- a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
+++ b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaComponent.java
@@ -175,8 +175,8 @@ public class KafkaComponent extends DefaultComponent implements SSLContextParame
      * invalid kafka configurations. Unfortunately, kafka-client does not separate this kind of errors.
      *
      * Camel will by default retry forever, and therefore never give up. If you want to give up after many attempts then
-     * set this option and Camel will then when giving up terminate the consumer. To try again, you can manually restart the consumer
-     * by stopping, and starting the route.
+     * set this option and Camel will then when giving up terminate the consumer. To try again, you can manually restart
+     * the consumer by stopping, and starting the route.
      */
     public void setCreateConsumerBackoffMaxAttempts(int createConsumerBackoffMaxAttempts) {
         this.createConsumerBackoffMaxAttempts = createConsumerBackoffMaxAttempts;
@@ -204,9 +204,9 @@ public class KafkaComponent extends DefaultComponent implements SSLContextParame
      * Error during subscribing the consumer to the kafka topic could be temporary errors due to network issues, and
      * could potentially be recoverable.
      *
-     * Camel will by default retry forever, and therefore never give up. If you want to give up after many attempts, then
-     * set this option and Camel will then when giving up terminate the consumer. You can manually restart the consumer
-     * by stopping and starting the route, to try again.
+     * Camel will by default retry forever, and therefore never give up. If you want to give up after many attempts,
+     * then set this option and Camel will then when giving up terminate the consumer. You can manually restart the
+     * consumer by stopping and starting the route, to try again.
      */
     public void setSubscribeConsumerBackoffMaxAttempts(int subscribeConsumerBackoffMaxAttempts) {
         this.subscribeConsumerBackoffMaxAttempts = subscribeConsumerBackoffMaxAttempts;
diff --git a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
index a695f2028fb..06aad479337 100755
--- a/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
+++ b/components/camel-kafka/src/main/java/org/apache/camel/component/kafka/KafkaConfiguration.java
@@ -824,8 +824,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware
      * <tt>false</tt> then the consumer continues to the next message and processes it. If the option is <tt>true</tt>
      * then the consumer breaks out, and will seek back to offset of the message that caused a failure, and then
      * re-attempt to process this message. However this can lead to endless processing of the same message if its bound
-     * to fail every time, eg a poison message. Therefore it is recommended to deal with that for example by using Camel's
-     * error handler.
+     * to fail every time, eg a poison message. Therefore it is recommended to deal with that for example by using
+     * Camel's error handler.
      */
     public void setBreakOnFirstError(boolean breakOnFirstError) {
         this.breakOnFirstError = breakOnFirstError;
@@ -889,8 +889,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware
 
     /**
      * Before each retry, the producer refreshes the metadata of relevant topics to see if a new leader has been
-     * elected. Since the leader election takes a bit of time, this property specifies the amount of time that the producer
-     * waits before refreshing the metadata.
+     * elected. Since the leader election takes a bit of time, this property specifies the amount of time that the
+     * producer waits before refreshing the metadata.
      */
     public void setRetryBackoffMs(Integer retryBackoffMs) {
         this.retryBackoffMs = retryBackoffMs;
@@ -1220,8 +1220,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware
     }
 
     /**
-     * The location of the key store file. This is optional for the client and can be used for two-way authentication for
-     * the client.
+     * The location of the key store file. This is optional for the client and can be used for two-way authentication
+     * for the client.
      */
     public void setSslKeystoreLocation(String sslKeystoreLocation) {
         this.sslKeystoreLocation = sslKeystoreLocation;
@@ -1232,8 +1232,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware
     }
 
     /**
-     * The store password for the key store file. This is optional for the client and only needed if sslKeystoreLocation' is
-     * configured. Key store password is not supported for PEM format.
+     * The store password for the key store file. This is optional for the client and only needed if
+     * sslKeystoreLocation' is configured. Key store password is not supported for PEM format.
      */
     public void setSslKeystorePassword(String sslKeystorePassword) {
         this.sslKeystorePassword = sslKeystorePassword;
@@ -1399,9 +1399,9 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware
      * request. Normally this occurs only under load when records arrive faster than they can be sent out. However in
      * some circumstances the client may want to reduce the number of requests even under moderate load. This setting
      * accomplishes this by adding a small amount of artificial delay that is, rather than immediately sending out a
-     * record the producer will wait for up to the given delay to allow other records to be sent so that they can
-     * be batched together. This can be thought of as analogous to Nagle's algorithm in TCP. This setting gives the
-     * upper bound on the delay for batching: once we get batch.size worth of records for a partition it will be sent
+     * record the producer will wait for up to the given delay to allow other records to be sent so that they can be
+     * batched together. This can be thought of as analogous to Nagle's algorithm in TCP. This setting gives the upper
+     * bound on the delay for batching: once we get batch.size worth of records for a partition it will be sent
      * immediately regardless of this setting, however if we have fewer than this many bytes accumulated for this
      * partition we will 'linger' for the specified time waiting for more records to show up. This setting defaults to 0
      * (i.e. no delay). Setting linger.ms=5, for example, would have the effect of reducing the number of requests sent
@@ -1663,8 +1663,8 @@ public class KafkaConfiguration implements Cloneable, HeaderFilterStrategyAware
     }
 
     /**
-     * Set if KafkaConsumer will read from the beginning or the end on startup: SeekPolicy.BEGINNING: read from the beginning.
-     * SeekPolicy.END: read from the end.
+     * Set if KafkaConsumer will read from the beginning or the end on startup: SeekPolicy.BEGINNING: read from the
+     * beginning. SeekPolicy.END: read from the end.
      */
     public void setSeekTo(SeekPolicy seekTo) {
         this.seekTo = seekTo;
diff --git a/docs/components/modules/others/examples/json/platform-http-main.json b/docs/components/modules/others/examples/json/platform-http-main.json
new file mode 120000
index 00000000000..cabf7d49547
--- /dev/null
+++ b/docs/components/modules/others/examples/json/platform-http-main.json
@@ -0,0 +1 @@
+../../../../../../components/camel-platform-http-main/src/generated/resources/platform-http-main.json
\ No newline at end of file
diff --git a/docs/components/modules/others/nav.adoc b/docs/components/modules/others/nav.adoc
index 6f03eacacad..fd975c765b7 100644
--- a/docs/components/modules/others/nav.adoc
+++ b/docs/components/modules/others/nav.adoc
@@ -36,6 +36,7 @@
 ** xref:microprofile-health.adoc[Microprofile Health]
 ** xref:openapi-java.adoc[Openapi Java]
 ** xref:opentelemetry.adoc[OpenTelemetry]
+** xref:platform-http-main.adoc[Platform Http Main]
 ** xref:platform-http-vertx.adoc[Platform Http Vertx]
 ** xref:reactive-executor-tomcat.adoc[Reactive Executor Tomcat]
 ** xref:reactive-executor-vertx.adoc[Reactive Executor Vert.x]
diff --git a/docs/components/modules/others/pages/platform-http-main.adoc b/docs/components/modules/others/pages/platform-http-main.adoc
new file mode 120000
index 00000000000..2ac31070d90
--- /dev/null
+++ b/docs/components/modules/others/pages/platform-http-main.adoc
@@ -0,0 +1 @@
+../../../../../components/camel-platform-http-main/src/main/docs/platform-http-main.adoc
\ No newline at end of file
diff --git a/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java b/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
index feaa422d09a..d919a48dfea 100644
--- a/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
+++ b/dsl/camel-componentdsl/src/generated/java/org/apache/camel/builder/component/dsl/KafkaComponentBuilderFactory.java
@@ -254,7 +254,7 @@ public interface KafkaComponentBuilderFactory {
          * message that caused a failure, and then re-attempt to process this
          * message. However this can lead to endless processing of the same
          * message if its bound to fail every time, eg a poison message.
-         * Therefore its recommended to deal with that for example by using
+         * Therefore it is recommended to deal with that for example by using
          * Camel's error handler.
          * 
          * The option is a: &lt;code&gt;boolean&lt;/code&gt; type.
@@ -501,7 +501,7 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * Deserializer class for key that implements the Deserializer
+         * Deserializer class for the key that implements the Deserializer
          * interface.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
@@ -654,9 +654,9 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * Set if KafkaConsumer will read from beginning or end on startup:
-         * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from
-         * end.
+         * Set if KafkaConsumer will read from the beginning or the end on
+         * startup: SeekPolicy.BEGINNING: read from the beginning.
+         * SeekPolicy.END: read from the end.
          * 
          * The option is a:
          * &lt;code&gt;org.apache.camel.component.kafka.SeekPolicy&lt;/code&gt;
@@ -765,12 +765,12 @@ public interface KafkaComponentBuilderFactory {
          * possible. However, one part of the validation is DNS resolution of
          * the bootstrap broker hostnames. This may be a temporary networking
          * problem, and could potentially be recoverable. While other errors are
-         * fatal such as some invalid kafka configurations. Unfortunately
+         * fatal, such as some invalid kafka configurations. Unfortunately,
          * kafka-client does not separate this kind of errors. Camel will by
          * default retry forever, and therefore never give up. If you want to
          * give up after many attempts then set this option and Camel will then
-         * when giving up terminate the consumer. You can manually restart the
-         * consumer by stopping and starting the route, to try again.
+         * when giving up terminate the consumer. To try again, you can manually
+         * restart the consumer by stopping, and starting the route.
          * 
          * The option is a: &lt;code&gt;int&lt;/code&gt; type.
          * 
@@ -874,7 +874,7 @@ public interface KafkaComponentBuilderFactory {
          * subscribing the consumer to the kafka topic could be temporary errors
          * due to network issues, and could potentially be recoverable. Camel
          * will by default retry forever, and therefore never give up. If you
-         * want to give up after many attempts then set this option and Camel
+         * want to give up after many attempts, then set this option and Camel
          * will then when giving up terminate the consumer. You can manually
          * restart the consumer by stopping and starting the route, to try
          * again.
@@ -1100,7 +1100,7 @@ public interface KafkaComponentBuilderFactory {
          * accomplishes this by adding a small amount of artificial delay that
          * is, rather than immediately sending out a record the producer will
          * wait for up to the given delay to allow other records to be sent so
-         * that the sends can be batched together. This can be thought of as
+         * that they can be batched together. This can be thought of as
          * analogous to Nagle's algorithm in TCP. This setting gives the upper
          * bound on the delay for batching: once we get batch.size worth of
          * records for a partition it will be sent immediately regardless of
@@ -1132,7 +1132,7 @@ public interface KafkaComponentBuilderFactory {
          * serializers or partitioner is not counted against this timeout). For
          * partitionsFor() this timeout bounds the time spent waiting for
          * metadata if it is unavailable. The transaction-related methods always
-         * block, but may timeout if the transaction coordinator could not be
+         * block, but may time out if the transaction coordinator could not be
          * discovered or did not respond within the timeout.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
@@ -1293,9 +1293,9 @@ public interface KafkaComponentBuilderFactory {
          * requests whenever multiple records are being sent to the same
          * partition. This helps performance on both the client and the server.
          * This configuration controls the default batch size in bytes. No
-         * attempt will be made to batch records larger than this size.Requests
+         * attempt will be made to batch records larger than this size. Requests
          * sent to brokers will contain multiple batches, one for each partition
-         * with data available to be sent.A small batch size will make batching
+         * with data available to be sent. A small batch size will make batching
          * less common and may reduce throughput (a batch size of zero will
          * disable batching entirely). A very large batch size may use memory a
          * bit more wastefully as we will always allocate a buffer of the
@@ -1472,9 +1472,9 @@ public interface KafkaComponentBuilderFactory {
         }
         /**
          * Before each retry, the producer refreshes the metadata of relevant
-         * topics to see if a new leader has been elected. Since leader election
-         * takes a bit of time, this property specifies the amount of time that
-         * the producer waits before refreshing the metadata.
+         * topics to see if a new leader has been elected. Since the leader
+         * election takes a bit of time, this property specifies the amount of
+         * time that the producer waits before refreshing the metadata.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
          * 
@@ -1958,8 +1958,8 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -1974,7 +1974,7 @@ public interface KafkaComponentBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 
diff --git a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
index 02ec4c69b27..2c777640af8 100644
--- a/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
+++ b/dsl/camel-endpointdsl/src/generated/java/org/apache/camel/builder/endpoint/dsl/KafkaEndpointBuilderFactory.java
@@ -376,7 +376,7 @@ public interface KafkaEndpointBuilderFactory {
          * message that caused a failure, and then re-attempt to process this
          * message. However this can lead to endless processing of the same
          * message if its bound to fail every time, eg a poison message.
-         * Therefore its recommended to deal with that for example by using
+         * Therefore it is recommended to deal with that for example by using
          * Camel's error handler.
          * 
          * The option is a: &lt;code&gt;boolean&lt;/code&gt; type.
@@ -400,7 +400,7 @@ public interface KafkaEndpointBuilderFactory {
          * message that caused a failure, and then re-attempt to process this
          * message. However this can lead to endless processing of the same
          * message if its bound to fail every time, eg a poison message.
-         * Therefore its recommended to deal with that for example by using
+         * Therefore it is recommended to deal with that for example by using
          * Camel's error handler.
          * 
          * The option will be converted to a &lt;code&gt;boolean&lt;/code&gt;
@@ -799,7 +799,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * Deserializer class for key that implements the Deserializer
+         * Deserializer class for the key that implements the Deserializer
          * interface.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
@@ -1073,9 +1073,9 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * Set if KafkaConsumer will read from beginning or end on startup:
-         * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from
-         * end.
+         * Set if KafkaConsumer will read from the beginning or the end on
+         * startup: SeekPolicy.BEGINNING: read from the beginning.
+         * SeekPolicy.END: read from the end.
          * 
          * The option is a:
          * &lt;code&gt;org.apache.camel.component.kafka.SeekPolicy&lt;/code&gt;
@@ -1092,9 +1092,9 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * Set if KafkaConsumer will read from beginning or end on startup:
-         * SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from
-         * end.
+         * Set if KafkaConsumer will read from the beginning or the end on
+         * startup: SeekPolicy.BEGINNING: read from the beginning.
+         * SeekPolicy.END: read from the end.
          * 
          * The option will be converted to a
          * &lt;code&gt;org.apache.camel.component.kafka.SeekPolicy&lt;/code&gt;
@@ -1633,8 +1633,8 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -1649,7 +1649,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 
@@ -2573,7 +2573,7 @@ public interface KafkaEndpointBuilderFactory {
          * accomplishes this by adding a small amount of artificial delay that
          * is, rather than immediately sending out a record the producer will
          * wait for up to the given delay to allow other records to be sent so
-         * that the sends can be batched together. This can be thought of as
+         * that they can be batched together. This can be thought of as
          * analogous to Nagle's algorithm in TCP. This setting gives the upper
          * bound on the delay for batching: once we get batch.size worth of
          * records for a partition it will be sent immediately regardless of
@@ -2605,7 +2605,7 @@ public interface KafkaEndpointBuilderFactory {
          * accomplishes this by adding a small amount of artificial delay that
          * is, rather than immediately sending out a record the producer will
          * wait for up to the given delay to allow other records to be sent so
-         * that the sends can be batched together. This can be thought of as
+         * that they can be batched together. This can be thought of as
          * analogous to Nagle's algorithm in TCP. This setting gives the upper
          * bound on the delay for batching: once we get batch.size worth of
          * records for a partition it will be sent immediately regardless of
@@ -2638,7 +2638,7 @@ public interface KafkaEndpointBuilderFactory {
          * serializers or partitioner is not counted against this timeout). For
          * partitionsFor() this timeout bounds the time spent waiting for
          * metadata if it is unavailable. The transaction-related methods always
-         * block, but may timeout if the transaction coordinator could not be
+         * block, but may time out if the transaction coordinator could not be
          * discovered or did not respond within the timeout.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
@@ -2662,7 +2662,7 @@ public interface KafkaEndpointBuilderFactory {
          * serializers or partitioner is not counted against this timeout). For
          * partitionsFor() this timeout bounds the time spent waiting for
          * metadata if it is unavailable. The transaction-related methods always
-         * block, but may timeout if the transaction coordinator could not be
+         * block, but may time out if the transaction coordinator could not be
          * discovered or did not respond within the timeout.
          * 
          * The option will be converted to a
@@ -2934,9 +2934,9 @@ public interface KafkaEndpointBuilderFactory {
          * requests whenever multiple records are being sent to the same
          * partition. This helps performance on both the client and the server.
          * This configuration controls the default batch size in bytes. No
-         * attempt will be made to batch records larger than this size.Requests
+         * attempt will be made to batch records larger than this size. Requests
          * sent to brokers will contain multiple batches, one for each partition
-         * with data available to be sent.A small batch size will make batching
+         * with data available to be sent. A small batch size will make batching
          * less common and may reduce throughput (a batch size of zero will
          * disable batching entirely). A very large batch size may use memory a
          * bit more wastefully as we will always allocate a buffer of the
@@ -2960,9 +2960,9 @@ public interface KafkaEndpointBuilderFactory {
          * requests whenever multiple records are being sent to the same
          * partition. This helps performance on both the client and the server.
          * This configuration controls the default batch size in bytes. No
-         * attempt will be made to batch records larger than this size.Requests
+         * attempt will be made to batch records larger than this size. Requests
          * sent to brokers will contain multiple batches, one for each partition
-         * with data available to be sent.A small batch size will make batching
+         * with data available to be sent. A small batch size will make batching
          * less common and may reduce throughput (a batch size of zero will
          * disable batching entirely). A very large batch size may use memory a
          * bit more wastefully as we will always allocate a buffer of the
@@ -3267,9 +3267,9 @@ public interface KafkaEndpointBuilderFactory {
         }
         /**
          * Before each retry, the producer refreshes the metadata of relevant
-         * topics to see if a new leader has been elected. Since leader election
-         * takes a bit of time, this property specifies the amount of time that
-         * the producer waits before refreshing the metadata.
+         * topics to see if a new leader has been elected. Since the leader
+         * election takes a bit of time, this property specifies the amount of
+         * time that the producer waits before refreshing the metadata.
          * 
          * The option is a: &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
          * 
@@ -3286,9 +3286,9 @@ public interface KafkaEndpointBuilderFactory {
         }
         /**
          * Before each retry, the producer refreshes the metadata of relevant
-         * topics to see if a new leader has been elected. Since leader election
-         * takes a bit of time, this property specifies the amount of time that
-         * the producer waits before refreshing the metadata.
+         * topics to see if a new leader has been elected. Since the leader
+         * election takes a bit of time, this property specifies the amount of
+         * time that the producer waits before refreshing the metadata.
          * 
          * The option will be converted to a
          * &lt;code&gt;java.lang.Integer&lt;/code&gt; type.
@@ -3864,8 +3864,8 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -3880,7 +3880,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          * 
@@ -4746,8 +4746,8 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The location of the key store file. This is optional for client and
-         * can be used for two-way authentication for client.
+         * The location of the key store file. This is optional for the client
+         * and can be used for two-way authentication for the client.
          * 
          * The option is a: &lt;code&gt;java.lang.String&lt;/code&gt; type.
          * 
@@ -4762,7 +4762,7 @@ public interface KafkaEndpointBuilderFactory {
             return this;
         }
         /**
-         * The store password for the key store file. This is optional for
+         * The store password for the key store file. This is optional for the
          * client and only needed if sslKeystoreLocation' is configured. Key
          * store password is not supported for PEM format.
          *