You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by sr...@apache.org on 2019/03/04 15:33:08 UTC

[spark] branch branch-2.4 updated: [SPARK-27046][DSTREAMS] Remove SPARK-19185 related references from documentation

This is an automated email from the ASF dual-hosted git repository.

srowen pushed a commit to branch branch-2.4
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-2.4 by this push:
     new ae462b1  [SPARK-27046][DSTREAMS] Remove SPARK-19185 related references from documentation
ae462b1 is described below

commit ae462b12cdb9a6406070d526ad73f686457d6bcd
Author: Gabor Somogyi <ga...@gmail.com>
AuthorDate: Mon Mar 4 09:31:46 2019 -0600

    [SPARK-27046][DSTREAMS] Remove SPARK-19185 related references from documentation
    
    ## What changes were proposed in this pull request?
    
    SPARK-19185 is resolved so the reference can be removed from the documentation.
    
    ## How was this patch tested?
    
    cd docs/
    SKIP_API=1 jekyll build
    Manual webpage check.
    
    Closes #23959 from gaborgsomogyi/SPARK-27046.
    
    Authored-by: Gabor Somogyi <ga...@gmail.com>
    Signed-off-by: Sean Owen <se...@databricks.com>
    (cherry picked from commit 5252d8b9872cbf200651b0bb7b8c6edd649ebb58)
    Signed-off-by: Sean Owen <se...@databricks.com>
---
 docs/streaming-kafka-0-10-integration.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/docs/streaming-kafka-0-10-integration.md b/docs/streaming-kafka-0-10-integration.md
index 386066a..d5e6371 100644
--- a/docs/streaming-kafka-0-10-integration.md
+++ b/docs/streaming-kafka-0-10-integration.md
@@ -93,7 +93,7 @@ In most cases, you should use `LocationStrategies.PreferConsistent` as shown abo
 
 The cache for consumers has a default maximum size of 64.  If you expect to be handling more than (64 * number of executors) Kafka partitions, you can change this setting via `spark.streaming.kafka.consumer.cache.maxCapacity`.
 
-If you would like to disable the caching for Kafka consumers, you can set `spark.streaming.kafka.consumer.cache.enabled` to `false`. Disabling the cache may be needed to workaround the problem described in SPARK-19185. This property may be removed in later versions of Spark, once SPARK-19185 is resolved.
+If you would like to disable the caching for Kafka consumers, you can set `spark.streaming.kafka.consumer.cache.enabled` to `false`.
 
 The cache is keyed by topicpartition and group.id, so use a **separate** `group.id` for each call to `createDirectStream`.
 


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org