You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Gaurav Agarwal <ga...@gmail.com> on 2015/08/21 04:48:56 UTC
spark kafka partitioning
Hello
Regarding Spark Streaming and Kafka Partitioning
When i send message on kafka topic with 3 partitions and listens on
kafkareceiver with local value[4] . how will i come to know in Spark
Streaming that different Dstreams are created according to partitions of
kafka messages .
Thanks
Re: spark kafka partitioning
Posted by Gaurav Agarwal <ga...@gmail.com>.
when i send the message from kafka topic having three partitions.
Spark will listen the message when i say kafkautils.createStream or
createDirectstSream have local[4]
Now i want to see if spark will create partitions when it receive
message from kafka using dstream, how and where ,prwhich method of
spark api i have to see to find out
On 8/21/15, Gaurav Agarwal <ga...@gmail.com> wrote:
> Hello
>
> Regarding Spark Streaming and Kafka Partitioning
>
> When i send message on kafka topic with 3 partitions and listens on
> kafkareceiver with local value[4] . how will i come to know in Spark
> Streaming that different Dstreams are created according to partitions of
> kafka messages .
>
> Thanks
>
---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org
Re: spark kafka partitioning
Posted by Cody Koeninger <co...@koeninger.org>.
I'm not clear on your question, can you rephrase it? Also, are you talking
about createStream or createDirectStream?
On Thu, Aug 20, 2015 at 9:48 PM, Gaurav Agarwal <ga...@gmail.com>
wrote:
> Hello
>
> Regarding Spark Streaming and Kafka Partitioning
>
> When i send message on kafka topic with 3 partitions and listens on
> kafkareceiver with local value[4] . how will i come to know in Spark
> Streaming that different Dstreams are created according to partitions of
> kafka messages .
>
> Thanks
>
>
Re: spark kafka partitioning
Posted by ayan guha <gu...@gmail.com>.
If you have 1 topic, that means you have 1 DStream, which will have a
series of RDDs for each batch interval. In receiver-based integration,
there is no direct relationship b/w Kafka paritions with spark partitions.
in Direct approach, 1 partition will be created for each kafka partition.
On Fri, Aug 21, 2015 at 12:48 PM, Gaurav Agarwal <ga...@gmail.com>
wrote:
> Hello
>
> Regarding Spark Streaming and Kafka Partitioning
>
> When i send message on kafka topic with 3 partitions and listens on
> kafkareceiver with local value[4] . how will i come to know in Spark
> Streaming that different Dstreams are created according to partitions of
> kafka messages .
>
> Thanks
>
>
--
Best Regards,
Ayan Guha