You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Marco Gaido (JIRA)" <ji...@apache.org> on 2018/01/17 13:13:00 UTC

[jira] [Commented] (SPARK-15401) Spark Thrift server creates empty directories in tmp directory on the driver

    [ https://issues.apache.org/jira/browse/SPARK-15401?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16328738#comment-16328738 ] 

Marco Gaido commented on SPARK-15401:
-------------------------------------

this should have been fixed in SPARK-22793.

> Spark Thrift server creates empty directories in tmp directory on the driver
> ----------------------------------------------------------------------------
>
>                 Key: SPARK-15401
>                 URL: https://issues.apache.org/jira/browse/SPARK-15401
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 1.6.1
>            Reporter: Christophe Préaud
>            Priority: Minor
>
> Each connection to the Spark thrift server (e.g. using beeline) creates two empty directories in the tmp directory on the driver which are never removed:
> cd <tmp directory>
> ls -ltd *_resources | wc -l && /opt/spark/bin/beeline -u jdbc:hive2://dc1-kdp-prod-hadoop-00.prod.dc1.kelkoo.net:10000 -n kookel -e '!quit' && ls -ltd *_resources | wc -l
> 9080
> Connecting to jdbc:hive2://dc1-kdp-prod-hadoop-00.prod.dc1.kelkoo.net:10000
> Connected to: Spark SQL (version 1.6.1)
> Driver: Spark Project Core (version 1.6.1)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> Closing: 0: jdbc:hive2://dc1-kdp-prod-hadoop-00.prod.dc1.kelkoo.net:10000
> Beeline version 1.6.1 by Apache Hive
> 9082
> Those directories accumulates over time and are not removed:
> ls -ld *_resources | wc -l
> 9064
> And they are indeed empty:
> find *_resources -type f | wc -l
> 0



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org