You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Alex Sulimanov <as...@tremorvideodsp.com> on 2017/10/31 03:10:11 UTC

Question regarding cached partitions

Hi,

I started Spark Streaming job with 96 executors which reads from 96 Kafka partitions and applies mapWithState on the incoming DStream.

Why would it cache only 77 partitions? Do I have to allocate more memory? Currently each executor gets 10 GB and it is not clear why it can't cache all 96 partitions.
[cid:5DD89010-AB5E-43B3-88F6-486FDB58CD75]

Thank you!