You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by me...@apache.org on 2015/09/21 23:24:23 UTC
spark git commit: [DOC] [PYSPARK] [MLLIB] Added newlines to
docstrings to fix parameter formatting
Repository: spark
Updated Branches:
refs/heads/master 362539f8d -> 7c4f852bf
[DOC] [PYSPARK] [MLLIB] Added newlines to docstrings to fix parameter formatting
Added newlines before `:param ...:` and `:return:` markup. Without these, parameter lists aren't formatted correctly in the API docs. I.e:
![screen shot 2015-09-21 at 21 49 26](https://cloud.githubusercontent.com/assets/11915197/10004686/de3c41d4-60aa-11e5-9c50-a46dcb51243f.png)
.. looks like this once newline is added:
![screen shot 2015-09-21 at 21 50 14](https://cloud.githubusercontent.com/assets/11915197/10004706/f86bfb08-60aa-11e5-8524-ae4436713502.png)
Author: noelsmith <ma...@noelsmith.com>
Closes #8851 from noel-smith/docstring-missing-newline-fix.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/7c4f852b
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/7c4f852b
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/7c4f852b
Branch: refs/heads/master
Commit: 7c4f852bfc39537840f56cd8121457a0dc1ad7c1
Parents: 362539f
Author: noelsmith <ma...@noelsmith.com>
Authored: Mon Sep 21 14:24:19 2015 -0700
Committer: Xiangrui Meng <me...@databricks.com>
Committed: Mon Sep 21 14:24:19 2015 -0700
----------------------------------------------------------------------
python/pyspark/ml/param/__init__.py | 4 ++++
python/pyspark/ml/pipeline.py | 1 +
python/pyspark/ml/tuning.py | 2 ++
python/pyspark/ml/wrapper.py | 2 ++
python/pyspark/mllib/evaluation.py | 2 +-
python/pyspark/mllib/linalg/__init__.py | 1 +
python/pyspark/streaming/context.py | 2 ++
python/pyspark/streaming/mqtt.py | 1 +
8 files changed, 14 insertions(+), 1 deletion(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/ml/param/__init__.py
----------------------------------------------------------------------
diff --git a/python/pyspark/ml/param/__init__.py b/python/pyspark/ml/param/__init__.py
index eeeac49..2e0c63cb 100644
--- a/python/pyspark/ml/param/__init__.py
+++ b/python/pyspark/ml/param/__init__.py
@@ -164,6 +164,7 @@ class Params(Identifiable):
a flat param map, where the latter value is used if there exist
conflicts, i.e., with ordering: default param values <
user-supplied values < extra.
+
:param extra: extra param values
:return: merged param map
"""
@@ -182,6 +183,7 @@ class Params(Identifiable):
embedded and extra parameters over and returns the copy.
Subclasses should override this method if the default approach
is not sufficient.
+
:param extra: Extra parameters to copy to the new instance
:return: Copy of this instance
"""
@@ -201,6 +203,7 @@ class Params(Identifiable):
def _resolveParam(self, param):
"""
Resolves a param and validates the ownership.
+
:param param: param name or the param instance, which must
belong to this Params instance
:return: resolved param instance
@@ -243,6 +246,7 @@ class Params(Identifiable):
"""
Copies param values from this instance to another instance for
params shared by them.
+
:param to: the target instance
:param extra: extra params to be copied
:return: the target instance with param values copied
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/ml/pipeline.py
----------------------------------------------------------------------
diff --git a/python/pyspark/ml/pipeline.py b/python/pyspark/ml/pipeline.py
index 13cf2b0..312a850 100644
--- a/python/pyspark/ml/pipeline.py
+++ b/python/pyspark/ml/pipeline.py
@@ -154,6 +154,7 @@ class Pipeline(Estimator):
def setStages(self, value):
"""
Set pipeline stages.
+
:param value: a list of transformers or estimators
:return: the pipeline instance
"""
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/ml/tuning.py
----------------------------------------------------------------------
diff --git a/python/pyspark/ml/tuning.py b/python/pyspark/ml/tuning.py
index ab5621f..705ee53 100644
--- a/python/pyspark/ml/tuning.py
+++ b/python/pyspark/ml/tuning.py
@@ -254,6 +254,7 @@ class CrossValidator(Estimator):
Creates a copy of this instance with a randomly generated uid
and some extra params. This copies creates a deep copy of
the embedded paramMap, and copies the embedded and extra parameters over.
+
:param extra: Extra parameters to copy to the new instance
:return: Copy of this instance
"""
@@ -290,6 +291,7 @@ class CrossValidatorModel(Model):
and some extra params. This copies the underlying bestModel,
creates a deep copy of the embedded paramMap, and
copies the embedded and extra parameters over.
+
:param extra: Extra parameters to copy to the new instance
:return: Copy of this instance
"""
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/ml/wrapper.py
----------------------------------------------------------------------
diff --git a/python/pyspark/ml/wrapper.py b/python/pyspark/ml/wrapper.py
index 8218c7c..4bcb4aa 100644
--- a/python/pyspark/ml/wrapper.py
+++ b/python/pyspark/ml/wrapper.py
@@ -119,6 +119,7 @@ class JavaEstimator(Estimator, JavaWrapper):
def _fit_java(self, dataset):
"""
Fits a Java model to the input dataset.
+
:param dataset: input dataset, which is an instance of
:py:class:`pyspark.sql.DataFrame`
:param params: additional params (overwriting embedded values)
@@ -173,6 +174,7 @@ class JavaModel(Model, JavaTransformer):
extra params. This implementation first calls Params.copy and
then make a copy of the companion Java model with extra params.
So both the Python wrapper and the Java model get copied.
+
:param extra: Extra parameters to copy to the new instance
:return: Copy of this instance
"""
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/mllib/evaluation.py
----------------------------------------------------------------------
diff --git a/python/pyspark/mllib/evaluation.py b/python/pyspark/mllib/evaluation.py
index 4398ca8..a90e5c5 100644
--- a/python/pyspark/mllib/evaluation.py
+++ b/python/pyspark/mllib/evaluation.py
@@ -147,7 +147,7 @@ class MulticlassMetrics(JavaModelWrapper):
"""
Evaluator for multiclass classification.
- :param predictionAndLabels an RDD of (prediction, label) pairs.
+ :param predictionAndLabels: an RDD of (prediction, label) pairs.
>>> predictionAndLabels = sc.parallelize([(0.0, 0.0), (0.0, 1.0), (0.0, 0.0),
... (1.0, 0.0), (1.0, 1.0), (1.0, 1.0), (1.0, 1.0), (2.0, 2.0), (2.0, 0.0)])
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/mllib/linalg/__init__.py
----------------------------------------------------------------------
diff --git a/python/pyspark/mllib/linalg/__init__.py b/python/pyspark/mllib/linalg/__init__.py
index f929e3e..ea42127 100644
--- a/python/pyspark/mllib/linalg/__init__.py
+++ b/python/pyspark/mllib/linalg/__init__.py
@@ -240,6 +240,7 @@ class Vector(object):
def toArray(self):
"""
Convert the vector into an numpy.ndarray
+
:return: numpy.ndarray
"""
raise NotImplementedError
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/streaming/context.py
----------------------------------------------------------------------
diff --git a/python/pyspark/streaming/context.py b/python/pyspark/streaming/context.py
index 4069d7a..a8c9ffc 100644
--- a/python/pyspark/streaming/context.py
+++ b/python/pyspark/streaming/context.py
@@ -240,6 +240,7 @@ class StreamingContext(object):
def awaitTermination(self, timeout=None):
"""
Wait for the execution to stop.
+
@param timeout: time to wait in seconds
"""
if timeout is None:
@@ -252,6 +253,7 @@ class StreamingContext(object):
Wait for the execution to stop. Return `true` if it's stopped; or
throw the reported error during the execution; or `false` if the
waiting time elapsed before returning from the method.
+
@param timeout: time to wait in seconds
"""
self._jssc.awaitTerminationOrTimeout(int(timeout * 1000))
http://git-wip-us.apache.org/repos/asf/spark/blob/7c4f852b/python/pyspark/streaming/mqtt.py
----------------------------------------------------------------------
diff --git a/python/pyspark/streaming/mqtt.py b/python/pyspark/streaming/mqtt.py
index f065989..fa83006 100644
--- a/python/pyspark/streaming/mqtt.py
+++ b/python/pyspark/streaming/mqtt.py
@@ -31,6 +31,7 @@ class MQTTUtils(object):
storageLevel=StorageLevel.MEMORY_AND_DISK_SER_2):
"""
Create an input stream that pulls messages from a Mqtt Broker.
+
:param ssc: StreamingContext object
:param brokerUrl: Url of remote mqtt publisher
:param topic: topic name to subscribe to
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org