You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by GitBox <gi...@apache.org> on 2021/01/10 13:04:40 UTC

[GitHub] [hudi] lw309637554 commented on pull request #2196: [HUDI-1349]spark sql support overwrite use replace action

lw309637554 commented on pull request #2196:
URL: https://github.com/apache/hudi/pull/2196#issuecomment-757473127


   > @lw309637554 @n3nash @satishkotha
   > 
   > I have been trying to test master branch and the following change to make `Overwrite` be a INSERT_OVERWRITE_TABLE, adds complexity IMO - esp given interplay with newer features like table metadata.
   > 
   > What's the rationale behind this? Just to make the operation faster over the previous fs.delete() based approach? I mean, ultimately someone needs to do these deletes no.
   
   @vinothchandar  Hello,
   the original intention of this function is that, users can be able to recover accidentally deleted when using INSERT_OVERWRITE_TABLE, if the replace file have not been cleaned. The old fs.delete needs to retrieve data from the data source , it is complex.


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org