You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by pw...@apache.org on 2014/06/17 21:03:48 UTC
git commit: SPARK-2146. Fix takeOrdered doc
Repository: spark
Updated Branches:
refs/heads/master b92d16b11 -> 2794990e9
SPARK-2146. Fix takeOrdered doc
Removes Python syntax in Scaladoc, corrects result in Scaladoc, and removes irrelevant cache() call in Python doc.
Author: Sandy Ryza <sa...@cloudera.com>
Closes #1086 from sryza/sandy-spark-2146 and squashes the following commits:
185ff18 [Sandy Ryza] Use Seq instead of Array
c996120 [Sandy Ryza] SPARK-2146. Fix takeOrdered doc
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2794990e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2794990e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2794990e
Branch: refs/heads/master
Commit: 2794990e9eb8712d76d3a0f0483063ddc295e639
Parents: b92d16b
Author: Sandy Ryza <sa...@cloudera.com>
Authored: Tue Jun 17 12:03:22 2014 -0700
Committer: Patrick Wendell <pw...@gmail.com>
Committed: Tue Jun 17 12:03:22 2014 -0700
----------------------------------------------------------------------
core/src/main/scala/org/apache/spark/rdd/RDD.scala | 16 ++++++++--------
python/pyspark/rdd.py | 2 +-
2 files changed, 9 insertions(+), 9 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/2794990e/core/src/main/scala/org/apache/spark/rdd/RDD.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/rdd/RDD.scala b/core/src/main/scala/org/apache/spark/rdd/RDD.scala
index cf915b8..1633b18 100644
--- a/core/src/main/scala/org/apache/spark/rdd/RDD.scala
+++ b/core/src/main/scala/org/apache/spark/rdd/RDD.scala
@@ -1074,11 +1074,11 @@ abstract class RDD[T: ClassTag](
* Returns the top K (largest) elements from this RDD as defined by the specified
* implicit Ordering[T]. This does the opposite of [[takeOrdered]]. For example:
* {{{
- * sc.parallelize([10, 4, 2, 12, 3]).top(1)
- * // returns [12]
+ * sc.parallelize(Seq(10, 4, 2, 12, 3)).top(1)
+ * // returns Array(12)
*
- * sc.parallelize([2, 3, 4, 5, 6]).top(2)
- * // returns [6, 5]
+ * sc.parallelize(Seq(2, 3, 4, 5, 6)).top(2)
+ * // returns Array(6, 5)
* }}}
*
* @param num the number of top elements to return
@@ -1092,11 +1092,11 @@ abstract class RDD[T: ClassTag](
* implicit Ordering[T] and maintains the ordering. This does the opposite of [[top]].
* For example:
* {{{
- * sc.parallelize([10, 4, 2, 12, 3]).takeOrdered(1)
- * // returns [12]
+ * sc.parallelize(Seq(10, 4, 2, 12, 3)).takeOrdered(1)
+ * // returns Array(2)
*
- * sc.parallelize([2, 3, 4, 5, 6]).takeOrdered(2)
- * // returns [2, 3]
+ * sc.parallelize(Seq(2, 3, 4, 5, 6)).takeOrdered(2)
+ * // returns Array(2, 3)
* }}}
*
* @param num the number of top elements to return
http://git-wip-us.apache.org/repos/asf/spark/blob/2794990e/python/pyspark/rdd.py
----------------------------------------------------------------------
diff --git a/python/pyspark/rdd.py b/python/pyspark/rdd.py
index 65f6315..a0b2c74 100644
--- a/python/pyspark/rdd.py
+++ b/python/pyspark/rdd.py
@@ -857,7 +857,7 @@ class RDD(object):
Note: It returns the list sorted in descending order.
>>> sc.parallelize([10, 4, 2, 12, 3]).top(1)
[12]
- >>> sc.parallelize([2, 3, 4, 5, 6], 2).cache().top(2)
+ >>> sc.parallelize([2, 3, 4, 5, 6], 2).top(2)
[6, 5]
"""
def topIterator(iterator):