You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2014/01/15 00:00:19 UTC

git commit: Merge pull request #428 from pwendell/writeable-objects

Updated Branches:
  refs/heads/branch-0.9 329c9df13 -> 2f930d5ae


Merge pull request #428 from pwendell/writeable-objects

Don't clone records for text files

(cherry picked from commit 74b46acdc57293c103ab5dd5af931d0d0e32c0ed)
Signed-off-by: Reynold Xin <rx...@apache.org>


Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/2f930d5a
Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/2f930d5a
Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/2f930d5a

Branch: refs/heads/branch-0.9
Commit: 2f930d5ae10d603370b49b91719259f17dc25628
Parents: 329c9df
Author: Reynold Xin <rx...@apache.org>
Authored: Tue Jan 14 14:59:13 2014 -0800
Committer: Reynold Xin <rx...@apache.org>
Committed: Tue Jan 14 15:00:11 2014 -0800

----------------------------------------------------------------------
 core/src/main/scala/org/apache/spark/SparkContext.scala | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/2f930d5a/core/src/main/scala/org/apache/spark/SparkContext.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala b/core/src/main/scala/org/apache/spark/SparkContext.scala
index 55ac76b..ba3e91e 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -340,8 +340,8 @@ class SparkContext(
    * Hadoop-supported file system URI, and return it as an RDD of Strings.
    */
   def textFile(path: String, minSplits: Int = defaultMinSplits): RDD[String] = {
-    hadoopFile(path, classOf[TextInputFormat], classOf[LongWritable], classOf[Text], minSplits)
-      .map(pair => pair._2.toString)
+    hadoopFile(path, classOf[TextInputFormat], classOf[LongWritable], classOf[Text],
+      minSplits, cloneRecords = false).map(pair => pair._2.toString)
   }
 
   /**