You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by GitBox <gi...@apache.org> on 2021/04/13 13:11:11 UTC

[GitHub] [spark] cloud-fan commented on pull request #32136: [WIP][SPARK-35022][CORE] Task Scheduling Plugin in Spark

cloud-fan commented on pull request #32136:
URL: https://github.com/apache/spark/pull/32136#issuecomment-818723560


   > to avoid Spark schedule streaming tasks which use state store (let me call them stateful tasks) to arbitrary executors.
   
   I don't think we can guarantee it. It's a best effort and tasks should be able to run on any executor, thought tasks can have preferred executors (locality). Otherwise, we need to revisit many design decisions like how to avoid infinite wait, how to auto-scale, etc.
   
   > current locality seems a hacky approach as we can just blindly assign stateful tasks to executors evenly.
   
   Can you elaborate? If it's a problem of delay scheduling let's fix it instead.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org