You are viewing a plain text version of this content. The canonical link for it is here.
Posted to github@arrow.apache.org by "alamb (via GitHub)" <gi...@apache.org> on 2023/04/23 15:33:36 UTC

[GitHub] [arrow-datafusion] alamb commented on a diff in pull request #6049: MemoryExec INSERT INTO refactor to use ExecutionPlan

alamb commented on code in PR #6049:
URL: https://github.com/apache/arrow-datafusion/pull/6049#discussion_r1174595757


##########
datafusion/core/src/datasource/datasource.rs:
##########
@@ -102,8 +102,8 @@ pub trait TableProvider: Sync + Send {
     async fn insert_into(
         &self,
         _state: &SessionState,
-        _input: &LogicalPlan,
-    ) -> Result<()> {
+        _input: Arc<dyn ExecutionPlan>,
+    ) -> Result<Arc<dyn ExecutionPlan>> {

Review Comment:
   I agree taking an `ExecutionPlan`  is probably better than LogicalPlan as it keeps planning separate from the execution
   
   I also thought about making it a `SendableRecordBatch` to avoid plans entirely, but that would prevent things like running multiple partitions in parallel if the datasource supports it. 



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: github-unsubscribe@arrow.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org