You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by gw...@apache.org on 2016/01/05 15:22:56 UTC

kafka git commit: MINOR: Update `config/producer.properties` to have new producer properties

Repository: kafka
Updated Branches:
  refs/heads/trunk b0b3e5aeb -> 23d607dc2


MINOR: Update `config/producer.properties` to have new producer properties

Also include some trivial clean-ups in `ProducerConfig`and `BaseProducer`.

Author: Ismael Juma <is...@juma.me.uk>

Reviewers: Gwen Shapira

Closes #710 from ijuma/use-new-producer-properties-in-config


Project: http://git-wip-us.apache.org/repos/asf/kafka/repo
Commit: http://git-wip-us.apache.org/repos/asf/kafka/commit/23d607dc
Tree: http://git-wip-us.apache.org/repos/asf/kafka/tree/23d607dc
Diff: http://git-wip-us.apache.org/repos/asf/kafka/diff/23d607dc

Branch: refs/heads/trunk
Commit: 23d607dc2174f8d2202d438fa65e03e951591fcd
Parents: b0b3e5a
Author: Ismael Juma <is...@juma.me.uk>
Authored: Tue Jan 5 06:22:28 2016 -0800
Committer: Gwen Shapira <cs...@gmail.com>
Committed: Tue Jan 5 06:22:28 2016 -0800

----------------------------------------------------------------------
 .../kafka/clients/producer/ProducerConfig.java  |  4 +-
 config/producer.properties                      | 40 ++++++++------------
 .../scala/kafka/producer/BaseProducer.scala     |  3 +-
 3 files changed, 19 insertions(+), 28 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/kafka/blob/23d607dc/clients/src/main/java/org/apache/kafka/clients/producer/ProducerConfig.java
----------------------------------------------------------------------
diff --git a/clients/src/main/java/org/apache/kafka/clients/producer/ProducerConfig.java b/clients/src/main/java/org/apache/kafka/clients/producer/ProducerConfig.java
index 98b119b..ae9aa08 100644
--- a/clients/src/main/java/org/apache/kafka/clients/producer/ProducerConfig.java
+++ b/clients/src/main/java/org/apache/kafka/clients/producer/ProducerConfig.java
@@ -130,7 +130,7 @@ public class ProducerConfig extends AbstractConfig {
 
     /** <code>max.request.size</code> */
     public static final String MAX_REQUEST_SIZE_CONFIG = "max.request.size";
-    private static final String MAX_REQUEST_SIZE_DOC = "The maximum size of a request. This is also effectively a cap on the maximum record size. Note that the server " + "has its own cap on record size which may be different from this. This setting will limit the number of record "
+    private static final String MAX_REQUEST_SIZE_DOC = "The maximum size of a request in bytes. This is also effectively a cap on the maximum record size. Note that the server " + "has its own cap on record size which may be different from this. This setting will limit the number of record "
                                                        + "batches the producer will send in a single request to avoid sending huge requests.";
 
     /** <code>reconnect.backoff.ms</code> */
@@ -143,7 +143,7 @@ public class ProducerConfig extends AbstractConfig {
     @Deprecated
     public static final String BLOCK_ON_BUFFER_FULL_CONFIG = "block.on.buffer.full";
     private static final String BLOCK_ON_BUFFER_FULL_DOC = "When our memory buffer is exhausted we must either stop accepting new records (block) or throw errors. By default " + "this setting is true and we block, however in some scenarios blocking is not desirable and it is better to "
-                                                           + "immediately give an error. Setting this to <code>false</code> will accomplish that: the producer will throw a BufferExhaustedException if a recrord is sent and the buffer space is full.";
+                                                           + "immediately give an error. Setting this to <code>false</code> will accomplish that: the producer will throw a BufferExhaustedException if a record is sent and the buffer space is full.";
 
     /** <code>retries</code> */
     public static final String RETRIES_CONFIG = "retries";

http://git-wip-us.apache.org/repos/asf/kafka/blob/23d607dc/config/producer.properties
----------------------------------------------------------------------
diff --git a/config/producer.properties b/config/producer.properties
index 47ae3e2..fcd745c 100644
--- a/config/producer.properties
+++ b/config/producer.properties
@@ -18,36 +18,28 @@
 
 # list of brokers used for bootstrapping knowledge about the rest of the cluster
 # format: host1:port1,host2:port2 ...
-metadata.broker.list=localhost:9092
+bootstrap.servers=localhost:9092
+
+# specify the compression codec for all data generated: none, gzip, snappy, lz4
+compression.type=none
 
 # name of the partitioner class for partitioning events; default partition spreads data randomly
 #partitioner.class=
 
-# specifies whether the messages are sent asynchronously (async) or synchronously (sync)
-producer.type=sync
-
-# specify the compression codec for all data generated: none, gzip, snappy, lz4.
-# the old config values work as well: 0, 1, 2, 3 for none, gzip, snappy, lz4, respectively
-compression.codec=none
-
-# message encoder
-serializer.class=kafka.serializer.DefaultEncoder
+# the maximum amount of time the client will wait for the response of a request
+#request.timeout.ms=
 
-# allow topic level compression
-#compressed.topics=
+# how long `KafkaProducer.send` and `KafkaProducer.partitionsFor` will block for
+#max.block.ms=
 
-############################# Async Producer #############################
-# maximum time, in milliseconds, for buffering data on the producer queue 
-#queue.buffering.max.ms=
+# the producer will wait for up to the given delay to allow other records to be sent so that the sends can be batched together
+#linger.ms=
 
-# the maximum size of the blocking queue for buffering on the producer 
-#queue.buffering.max.messages=
+# the maximum size of a request in bytes
+#max.request.size=
 
-# Timeout for event enqueue:
-# 0: events will be enqueued immediately or dropped if the queue is full
-# -ve: enqueue will block indefinitely if the queue is full
-# +ve: enqueue will block up to this many milliseconds if the queue is full
-#queue.enqueue.timeout.ms=
+# the default batch size in bytes when batching multiple records sent to a partition
+#batch.size=
 
-# the number of messages batched at the producer 
-#batch.num.messages=
+# the total bytes of memory the producer can use to buffer records waiting to be sent to the server
+#buffer.memory=

http://git-wip-us.apache.org/repos/asf/kafka/blob/23d607dc/core/src/main/scala/kafka/producer/BaseProducer.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/kafka/producer/BaseProducer.scala b/core/src/main/scala/kafka/producer/BaseProducer.scala
index 8e00713..9d0976f 100644
--- a/core/src/main/scala/kafka/producer/BaseProducer.scala
+++ b/core/src/main/scala/kafka/producer/BaseProducer.scala
@@ -27,7 +27,7 @@ trait BaseProducer {
 }
 
 class NewShinyProducer(producerProps: Properties) extends BaseProducer {
-  import org.apache.kafka.clients.producer.{KafkaProducer, ProducerConfig, ProducerRecord}
+  import org.apache.kafka.clients.producer.{KafkaProducer, ProducerRecord}
   import org.apache.kafka.clients.producer.internals.ErrorLoggingCallback
 
   // decide whether to send synchronously based on producer properties
@@ -51,7 +51,6 @@ class NewShinyProducer(producerProps: Properties) extends BaseProducer {
 }
 
 class OldProducer(producerProps: Properties) extends BaseProducer {
-  import kafka.producer.{KeyedMessage, ProducerConfig}
 
   // default to byte array partitioner
   if (producerProps.getProperty("partitioner.class") == null)