You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by "Saumitra Shahapure (Vizury)" <sa...@vizury.com> on 2014/04/23 09:06:05 UTC

Sharing RDDs

Hello,

Is it possible in spark to reuse cached RDDs generated in earlier run?

Specifically, I am trying to have a setup where first scala script
generates cached RDDs. If another scala script tries to perform same
operations on same dataset, it should be able to get results from cache
generated in earlier run.

Is there any direct/indirect way to do this?

--
Regards,
Saumitra Shahapure