You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "thierry accart (Jira)" <ji...@apache.org> on 2019/12/09 13:43:00 UTC

[jira] [Created] (SPARK-30190) HistoryServerDiskManager will fail on appStoreDir in s3

thierry accart created SPARK-30190:
--------------------------------------

             Summary: HistoryServerDiskManager will fail on appStoreDir in s3
                 Key: SPARK-30190
                 URL: https://issues.apache.org/jira/browse/SPARK-30190
             Project: Spark
          Issue Type: Improvement
          Components: Spark Core
    Affects Versions: 2.4.4
            Reporter: thierry accart


Hi

While setting spark.eventLog.dir to s3a://... I realized that it *requires destination directory to preexists for S3* 

This is explained I think in HistoryServerDiskManager's appStoreDir: it tries check if directory exists or can be created

{{if (!appStoreDir.isDirectory() && !appStoreDir.mkdir()) \{throw new IllegalArgumentException(s"Failed to create app directory ($appStoreDir).")}}}

But in S3, a directory does not exists and cannot be created: directories don't exists by themselves, they are only materialized due to existence of objects.


Before proposing a patch, I wanted to know what are the prefered options : 
should we have a spark option to skip the appStoreDir test, or skip it only when a particular scheme is set, have a custom implementation of HistoryServerDiskManager ...? 

 

_Note for people facing the {{IllegalArgumentException:}} {{Failed to create app directory}} *you just have to put an empty file in bucket destination 'path'*._



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org