You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by gu...@apache.org on 2018/03/05 14:26:31 UTC
spark git commit: [SPARK-23566][MINOR][DOC] Argument name mismatch
fixed
Repository: spark
Updated Branches:
refs/heads/master 42cf48e20 -> 5ff72ffcf
[SPARK-23566][MINOR][DOC] Argument name mismatch fixed
Argument name mismatch fixed.
## What changes were proposed in this pull request?
`col` changed to `new` in doc string to match the argument list.
Patch file added: https://issues.apache.org/jira/browse/SPARK-23566
Please review http://spark.apache.org/contributing.html before opening a pull request.
Author: Anirudh <an...@mail.com>
Closes #20716 from animenon/master.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/5ff72ffc
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/5ff72ffc
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/5ff72ffc
Branch: refs/heads/master
Commit: 5ff72ffcf495d2823f7f1186078d1cb261667c3d
Parents: 42cf48e
Author: Anirudh <an...@mail.com>
Authored: Mon Mar 5 23:17:16 2018 +0900
Committer: hyukjinkwon <gu...@gmail.com>
Committed: Mon Mar 5 23:17:16 2018 +0900
----------------------------------------------------------------------
python/pyspark/sql/dataframe.py | 20 +++++++++++++-------
1 file changed, 13 insertions(+), 7 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/5ff72ffc/python/pyspark/sql/dataframe.py
----------------------------------------------------------------------
diff --git a/python/pyspark/sql/dataframe.py b/python/pyspark/sql/dataframe.py
index f37777e..9d8e85c 100644
--- a/python/pyspark/sql/dataframe.py
+++ b/python/pyspark/sql/dataframe.py
@@ -588,6 +588,8 @@ class DataFrame(object):
"""
Returns a new :class:`DataFrame` that has exactly `numPartitions` partitions.
+ :param numPartitions: int, to specify the target number of partitions
+
Similar to coalesce defined on an :class:`RDD`, this operation results in a
narrow dependency, e.g. if you go from 1000 partitions to 100 partitions,
there will not be a shuffle, instead each of the 100 new partitions will
@@ -612,9 +614,10 @@ class DataFrame(object):
Returns a new :class:`DataFrame` partitioned by the given partitioning expressions. The
resulting DataFrame is hash partitioned.
- ``numPartitions`` can be an int to specify the target number of partitions or a Column.
- If it is a Column, it will be used as the first partitioning column. If not specified,
- the default number of partitions is used.
+ :param numPartitions:
+ can be an int to specify the target number of partitions or a Column.
+ If it is a Column, it will be used as the first partitioning column. If not specified,
+ the default number of partitions is used.
.. versionchanged:: 1.6
Added optional arguments to specify the partitioning columns. Also made numPartitions
@@ -673,9 +676,10 @@ class DataFrame(object):
Returns a new :class:`DataFrame` partitioned by the given partitioning expressions. The
resulting DataFrame is range partitioned.
- ``numPartitions`` can be an int to specify the target number of partitions or a Column.
- If it is a Column, it will be used as the first partitioning column. If not specified,
- the default number of partitions is used.
+ :param numPartitions:
+ can be an int to specify the target number of partitions or a Column.
+ If it is a Column, it will be used as the first partitioning column. If not specified,
+ the default number of partitions is used.
At least one partition-by expression must be specified.
When no explicit sort order is specified, "ascending nulls first" is assumed.
@@ -892,6 +896,8 @@ class DataFrame(object):
def alias(self, alias):
"""Returns a new :class:`DataFrame` with an alias set.
+ :param alias: string, an alias name to be set for the DataFrame.
+
>>> from pyspark.sql.functions import *
>>> df_as1 = df.alias("df_as1")
>>> df_as2 = df.alias("df_as2")
@@ -1900,7 +1906,7 @@ class DataFrame(object):
This is a no-op if schema doesn't contain the given column name.
:param existing: string, name of the existing column to rename.
- :param col: string, new name of the column.
+ :param new: string, new name of the column.
>>> df.withColumnRenamed('age', 'age2').collect()
[Row(age2=2, name=u'Alice'), Row(age2=5, name=u'Bob')]
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org