You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Shixiong Zhu (JIRA)" <ji...@apache.org> on 2016/12/29 19:32:58 UTC
[jira] [Commented] (SPARK-18942) Support output operations for
kinesis
[ https://issues.apache.org/jira/browse/SPARK-18942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15785953#comment-15785953 ]
Shixiong Zhu commented on SPARK-18942:
--------------------------------------
Thanks for your prototype. Actually, you can just implement an RDD action, or DataFrame DataSource and put them as a Spark package like [spark-redshift|https://github.com/databricks/spark-redshift]. [Spark Packages|https://spark-packages.org/] is a better place for such third-party data sources.
> Support output operations for kinesis
> -------------------------------------
>
> Key: SPARK-18942
> URL: https://issues.apache.org/jira/browse/SPARK-18942
> Project: Spark
> Issue Type: New Feature
> Components: DStreams
> Affects Versions: 2.0.2
> Reporter: Takeshi Yamamuro
> Priority: Trivial
>
> Spark does not support output operations (e.g. DStream#saveAsTextFile) for Kinesis. So, officially supporting this is useful for some AWS users, I think. An usage of the output operations is assumed as follows;
> {code}
> // Import a class that includes an output function
> scala> import org.apache.spark.streaming.kinesis.KinesisDStreamFunctions._
> // Create a DStream
> scala> val stream: DStream[String] = ...
> // Define a handler to convert the DStream type for output
> scala> val msgHandler = (s: String) => s.getBytes("UTF-8")
> // Define the output operation
> scala> kinesisStream.count().saveAsKinesisStream(streamName, endpointUrl, msgHandler)
> {code}
> A prototype I made is here: https://github.com/apache/spark/compare/master...maropu:OutputOpForKinesis
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org