You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Patrick Wendell (JIRA)" <ji...@apache.org> on 2014/10/07 01:23:34 UTC
[jira] [Created] (SPARK-3824) Spark SQL should cache in
MEMORY_AND_DISK by default
Patrick Wendell created SPARK-3824:
--------------------------------------
Summary: Spark SQL should cache in MEMORY_AND_DISK by default
Key: SPARK-3824
URL: https://issues.apache.org/jira/browse/SPARK-3824
Project: Spark
Issue Type: Bug
Components: SQL
Reporter: Patrick Wendell
Assignee: Cheng Lian
Priority: Blocker
Spark SQL currently uses MEMORY_ONLY as the default format. Due to the use of column buffers however, there is a huge cost to having to recompute blocks, much more so than Spark core. Especially since now we are more conservative about caching blocks and sometimes won't cache blocks we think might exceed memory, it seems good to keep persisted blocks on disk by default.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org