You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by pw...@apache.org on 2014/06/08 01:22:36 UTC
git commit: SPARK-2056 Set RDD name to input path
Repository: spark
Updated Branches:
refs/heads/master 3ace10dc9 -> 7b877b270
SPARK-2056 Set RDD name to input path
Author: Neville Li <ne...@spotify.com>
Closes #992 from nevillelyh/master and squashes the following commits:
3011739 [Neville Li] [SPARK-2056] Set RDD name to input path
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/7b877b27
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/7b877b27
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/7b877b27
Branch: refs/heads/master
Commit: 7b877b27053bfb7092e250e01a3b887e1b50a109
Parents: 3ace10d
Author: Neville Li <ne...@spotify.com>
Authored: Sat Jun 7 16:22:26 2014 -0700
Committer: Patrick Wendell <pw...@gmail.com>
Committed: Sat Jun 7 16:22:26 2014 -0700
----------------------------------------------------------------------
core/src/main/scala/org/apache/spark/SparkContext.scala | 8 ++++----
1 file changed, 4 insertions(+), 4 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/7b877b27/core/src/main/scala/org/apache/spark/SparkContext.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala b/core/src/main/scala/org/apache/spark/SparkContext.scala
index d941aea..d721aba 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -455,7 +455,7 @@ class SparkContext(config: SparkConf) extends Logging {
*/
def textFile(path: String, minPartitions: Int = defaultMinPartitions): RDD[String] = {
hadoopFile(path, classOf[TextInputFormat], classOf[LongWritable], classOf[Text],
- minPartitions).map(pair => pair._2.toString)
+ minPartitions).map(pair => pair._2.toString).setName(path)
}
/**
@@ -496,7 +496,7 @@ class SparkContext(config: SparkConf) extends Logging {
classOf[String],
classOf[String],
updateConf,
- minPartitions)
+ minPartitions).setName(path)
}
/**
@@ -551,7 +551,7 @@ class SparkContext(config: SparkConf) extends Logging {
inputFormatClass,
keyClass,
valueClass,
- minPartitions)
+ minPartitions).setName(path)
}
/**
@@ -623,7 +623,7 @@ class SparkContext(config: SparkConf) extends Logging {
val job = new NewHadoopJob(conf)
NewFileInputFormat.addInputPath(job, new Path(path))
val updatedConf = job.getConfiguration
- new NewHadoopRDD(this, fClass, kClass, vClass, updatedConf)
+ new NewHadoopRDD(this, fClass, kClass, vClass, updatedConf).setName(path)
}
/**