You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "David Hodeffi (JIRA)" <ji...@apache.org> on 2016/10/25 16:06:58 UTC
[jira] [Created] (SPARK-18096) Spark on have - 'Update' save mode
David Hodeffi created SPARK-18096:
-------------------------------------
Summary: Spark on have - 'Update' save mode
Key: SPARK-18096
URL: https://issues.apache.org/jira/browse/SPARK-18096
Project: Spark
Issue Type: Improvement
Components: Spark Core
Affects Versions: 2.0.1
Reporter: David Hodeffi
when creating ETL with Spark on Hive, it is needed to update incrementally the destination table.
In case it is partitioned table it means that we don't need to update all partitions, but just the one who mutated.
right now there is only one way to update a Dataframe which is SaveMode.Overwrite , the problem is that when doing it incrementally you don't need to update all partitions but just those who changed/updated.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org