You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by GitBox <gi...@apache.org> on 2022/12/11 20:34:36 UTC
[GitHub] [spark] bjornjorgensen commented on a diff in pull request #39020: [MINOR][CORE][PYTHON][SQL][PS] Fix some typos
bjornjorgensen commented on code in PR #39020:
URL: https://github.com/apache/spark/pull/39020#discussion_r1045292765
##########
python/pyspark/pandas/frame.py:
##########
@@ -2598,15 +2598,15 @@ def transpose(self) -> "DataFrame":
.. note:: This method is based on an expensive operation due to the nature
of big data. Internally it needs to generate each row for each value, and
- then group twice - it is a huge operation. To prevent misusage, this method
+ then group twice - it is a huge operation. To prevent misuse, this method
has the 'compute.max_rows' default limit of input length, and raises a ValueError.
>>> from pyspark.pandas.config import option_context
>>> with option_context('compute.max_rows', 1000): # doctest: +NORMALIZE_WHITESPACE
... ps.DataFrame({'a': range(1001)}).transpose()
Traceback (most recent call last):
...
- ValueError: Current DataFrame has more then the given limit 1000 rows.
+ ValueError: Current DataFrame's length exceeds the given limit of 1000 rows.
Please set 'compute.max_rows' by using 'pyspark.pandas.config.set_option'
to retrieve to retrieve more than 1000 rows. Note that, before changing the
Review Comment:
to retrieve to retrieve -> to retrieve
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org