You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by gu...@apache.org on 2021/08/27 01:28:19 UTC

[spark] branch branch-3.2 updated: [SPARK-35611][SS][FOLLOW-UP] Improve the user guide document

This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch branch-3.2
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.2 by this push:
     new f50f2d4  [SPARK-35611][SS][FOLLOW-UP] Improve the user guide document
f50f2d4 is described below

commit f50f2d474c31053caf7aeefec7ed499320d09dcc
Author: Yuanjian Li <yu...@databricks.com>
AuthorDate: Fri Aug 27 10:27:06 2021 +0900

    [SPARK-35611][SS][FOLLOW-UP] Improve the user guide document
    
    ### What changes were proposed in this pull request?
    Improve the user guide document.
    
    ### Why are the changes needed?
    Make the user guide clear.
    
    ### Does this PR introduce _any_ user-facing change?
    No.
    
    ### How was this patch tested?
    Doc change only.
    
    Closes #33854 from xuanyuanking/SPARK-35611-follow.
    
    Authored-by: Yuanjian Li <yu...@databricks.com>
    Signed-off-by: Hyukjin Kwon <gu...@apache.org>
    (cherry picked from commit dd3f0fa8c2318843f04910bb841329d1830b9838)
    Signed-off-by: Hyukjin Kwon <gu...@apache.org>
---
 docs/structured-streaming-kafka-integration.md | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git a/docs/structured-streaming-kafka-integration.md b/docs/structured-streaming-kafka-integration.md
index 0ec359f..6121f19 100644
--- a/docs/structured-streaming-kafka-integration.md
+++ b/docs/structured-streaming-kafka-integration.md
@@ -547,10 +547,10 @@ The following configurations are optional:
   <td>"error" or "latest"</td>
   <td>"error"</td>
   <td>streaming and batch</td>
-  <td>Defines the behavior when the starting offset by timestamp is specified (either global or per partition), and Kafka doesn't return the matched offset.<p/>
+  <td>The strategy will be used when the specified starting offset by timestamp (either global or per partition) doesn't match with the offset Kafka returned. Here's the strategy name and corresponding descriptions:<p/>
   <p/>
-  "error": fail the query.<p/>
-  "latest": set the offset to the latest, so that further new records in the partition are being read.<p/></td>
+  "error": fail the query and end users have to deal with workarounds requiring manual steps.<p/>
+  "latest": assigns the latest offset for these partitions, so that Spark can read newer records from these partitions in further micro-batches.<p/></td>
 </tr>
 </table>
 

---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org