You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dongjoon Hyun (Jira)" <ji...@apache.org> on 2020/01/23 02:22:00 UTC

[jira] [Assigned] (SPARK-30556) SubqueryExec passes local properties to SubqueryExec.executionContext

     [ https://issues.apache.org/jira/browse/SPARK-30556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Dongjoon Hyun reassigned SPARK-30556:
-------------------------------------

    Assignee: Ajith S

> SubqueryExec passes local properties to SubqueryExec.executionContext
> ---------------------------------------------------------------------
>
>                 Key: SPARK-30556
>                 URL: https://issues.apache.org/jira/browse/SPARK-30556
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 2.4.4, 3.0.0
>            Reporter: Ajith S
>            Assignee: Ajith S
>            Priority: Major
>
> Local properties set via sparkContext are not available as TaskContext properties when executing  jobs and threadpools have idle threads which are reused
> Explanation:
> When SubqueryExec, the {{relationFuture}} is evaluated via a separate thread. The threads inherit the {{localProperties}} from sparkContext as they are the child threads.
> These threads are controlled via the executionContext (thread pools). Each Thread pool has a default {{keepAliveSeconds}} of 60 seconds for idle threads.
> Scenarios where the thread pool has threads which are idle and reused for a subsequent new query, the thread local properties will not be inherited from spark context (thread properties are inherited only on thread creation) hence end up having old or no properties set. This will cause taskset properties to be missing when properties are transferred by child thread via {{sparkContext.runJob/submitJob}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org