You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by ma...@apache.org on 2014/12/01 22:12:35 UTC
spark git commit: [SQL] add @group tab in limit() and count()
Repository: spark
Updated Branches:
refs/heads/master 5db8dcaf4 -> bafee67eb
[SQL] add @group tab in limit() and count()
group tab is missing for scaladoc
Author: Jacky Li <ja...@gmail.com>
Closes #3458 from jackylk/patch-7 and squashes the following commits:
0121a70 [Jacky Li] add @group tab in limit() and count()
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/bafee67e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/bafee67e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/bafee67e
Branch: refs/heads/master
Commit: bafee67ebad01f7aea2cd393a70b57eb8345eeb0
Parents: 5db8dca
Author: Jacky Li <ja...@gmail.com>
Authored: Mon Dec 1 13:12:30 2014 -0800
Committer: Michael Armbrust <mi...@databricks.com>
Committed: Mon Dec 1 13:12:30 2014 -0800
----------------------------------------------------------------------
sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala | 4 ++++
1 file changed, 4 insertions(+)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/bafee67e/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
index 452baab..c6d4dab 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
@@ -225,6 +225,8 @@ class SchemaRDD(
* {{{
* schemaRDD.limit(10)
* }}}
+ *
+ * @group Query
*/
def limit(limitNum: Int): SchemaRDD =
new SchemaRDD(sqlContext, Limit(Literal(limitNum), logicalPlan))
@@ -355,6 +357,8 @@ class SchemaRDD(
* Return the number of elements in the RDD. Unlike the base RDD implementation of count, this
* implementation leverages the query optimizer to compute the count on the SchemaRDD, which
* supports features such as filter pushdown.
+ *
+ * @group Query
*/
@Experimental
override def count(): Long = aggregate(Count(Literal(1))).collect().head.getLong(0)
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org