You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Brian London (JIRA)" <ji...@apache.org> on 2015/12/02 00:16:11 UTC

[jira] [Commented] (SPARK-10969) Spark Streaming Kinesis: Allow specifying separate credentials for Kinesis and DynamoDB

    [ https://issues.apache.org/jira/browse/SPARK-10969?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15034820#comment-15034820 ] 

Brian London commented on SPARK-10969:
--------------------------------------

Was this fixed with https://github.com/apache/spark/commit/63accc79625d8a03d0624717af5e1d81b18a6da3 ?

> Spark Streaming Kinesis: Allow specifying separate credentials for Kinesis and DynamoDB
> ---------------------------------------------------------------------------------------
>
>                 Key: SPARK-10969
>                 URL: https://issues.apache.org/jira/browse/SPARK-10969
>             Project: Spark
>          Issue Type: Improvement
>          Components: Streaming
>    Affects Versions: 1.5.1
>            Reporter: Christoph Pirkl
>            Priority: Critical
>
> {{KinesisUtils.createStream()}} allows specifying only one set of AWS credentials that will be used by Amazon KCL for accessing Kinesis, DynamoDB and CloudWatch.
> h5. Motivation
> In a scenario where one needs to read from a Kinesis Stream owned by a different AWS account the user usually has minimal rights (i.e. only read from the stream). In this case creating the DynamoDB table in KCL will fail.
> h5. Proposal
> My proposed solution would be to allow specifying multiple credentials in {{KinesisUtils.createStream()}} for Kinesis, DynamoDB and CloudWatch. The additional credentials could then be passed to the constructor of {{KinesisClientLibConfiguration}} or method {{KinesisClientLibConfiguration.withDynamoDBClientConfig()}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org