You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by we...@apache.org on 2017/07/04 01:16:48 UTC
spark git commit: [SPARK-21283][CORE] FileOutputStream should be
created as append mode
Repository: spark
Updated Branches:
refs/heads/master c79c10eba -> 6657e00de
[SPARK-21283][CORE] FileOutputStream should be created as append mode
## What changes were proposed in this pull request?
`FileAppender` is used to write `stderr` and `stdout` files in `ExecutorRunner`, But before writing `ErrorStream` into the the `stderr` file, the header information has been written into ,if FileOutputStream is not created as append mode, the header information will be lost
## How was this patch tested?
unit test case
Author: liuxian <li...@zte.com.cn>
Closes #18507 from 10110346/wip-lx-0703.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6657e00d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6657e00d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6657e00d
Branch: refs/heads/master
Commit: 6657e00de36b59011d3fe78e8613fb64e54c957a
Parents: c79c10e
Author: liuxian <li...@zte.com.cn>
Authored: Tue Jul 4 09:16:40 2017 +0800
Committer: Wenchen Fan <we...@databricks.com>
Committed: Tue Jul 4 09:16:40 2017 +0800
----------------------------------------------------------------------
.../main/scala/org/apache/spark/util/logging/FileAppender.scala | 2 +-
.../test/scala/org/apache/spark/util/FileAppenderSuite.scala | 5 ++++-
2 files changed, 5 insertions(+), 2 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/6657e00d/core/src/main/scala/org/apache/spark/util/logging/FileAppender.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/util/logging/FileAppender.scala b/core/src/main/scala/org/apache/spark/util/logging/FileAppender.scala
index fdb1495..8a0cc70 100644
--- a/core/src/main/scala/org/apache/spark/util/logging/FileAppender.scala
+++ b/core/src/main/scala/org/apache/spark/util/logging/FileAppender.scala
@@ -94,7 +94,7 @@ private[spark] class FileAppender(inputStream: InputStream, file: File, bufferSi
/** Open the file output stream */
protected def openFile() {
- outputStream = new FileOutputStream(file, false)
+ outputStream = new FileOutputStream(file, true)
logDebug(s"Opened file $file")
}
http://git-wip-us.apache.org/repos/asf/spark/blob/6657e00d/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala
----------------------------------------------------------------------
diff --git a/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala b/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala
index 7e2da8e..cd0ed5b 100644
--- a/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala
+++ b/core/src/test/scala/org/apache/spark/util/FileAppenderSuite.scala
@@ -52,10 +52,13 @@ class FileAppenderSuite extends SparkFunSuite with BeforeAndAfter with Logging {
test("basic file appender") {
val testString = (1 to 1000).mkString(", ")
val inputStream = new ByteArrayInputStream(testString.getBytes(StandardCharsets.UTF_8))
+ // The `header` should not be covered
+ val header = "Add header"
+ Files.write(header, testFile, StandardCharsets.UTF_8)
val appender = new FileAppender(inputStream, testFile)
inputStream.close()
appender.awaitTermination()
- assert(Files.toString(testFile, StandardCharsets.UTF_8) === testString)
+ assert(Files.toString(testFile, StandardCharsets.UTF_8) === header + testString)
}
test("rolling file appender - time-based rolling") {
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org