You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Cyanny (JIRA)" <ji...@apache.org> on 2017/08/29 08:37:02 UTC
[jira] [Created] (SPARK-21859) SparkFiles.get failed on driver in
yarn-cluster and yarn-client mode
Cyanny created SPARK-21859:
------------------------------
Summary: SparkFiles.get failed on driver in yarn-cluster and yarn-client mode
Key: SPARK-21859
URL: https://issues.apache.org/jira/browse/SPARK-21859
Project: Spark
Issue Type: Bug
Components: Spark Core
Affects Versions: 1.6.2, 2.2.1
Reporter: Cyanny
Fix For: 2.2.1
when use SparkFiles.get a file on driver in yarn-client or yarn-cluster, it will report file not found exception.
This exception only happens on driver, SparkFiles.get works fine on executor.
we can reproduce the bug as follows:
```scala
def testOnDriver(fileName: String) = {
val file = new File(SparkFiles.get(fileName))
if (!file.exists()) {
logging.info(s"$file not exist")
} else {
// print file content on driver
val content = Source.fromFile(file).getLines().mkString("\n")
logging.info(s"File content: ${content}")
}
}
// the output will be file not exist
```
```python
conf = SparkConf().setAppName("test files")
sc = SparkContext(appName="spark files test")
def test_on_driver(filename):
file = SparkFiles.get(filename)
print("file path: {}".format(file))
if os.path.exists(file):
with open(file) as f:
lines = f.readlines()
print(lines)
else:
print("file doesn't exist")
run_command("ls .")
```
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org