You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Bryan Cutler (JIRA)" <ji...@apache.org> on 2018/01/29 17:24:00 UTC

[jira] [Resolved] (SPARK-23109) ML 2.3 QA: API: Python API coverage

     [ https://issues.apache.org/jira/browse/SPARK-23109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Bryan Cutler resolved SPARK-23109.
----------------------------------
    Resolution: Done

> ML 2.3 QA: API: Python API coverage
> -----------------------------------
>
>                 Key: SPARK-23109
>                 URL: https://issues.apache.org/jira/browse/SPARK-23109
>             Project: Spark
>          Issue Type: Sub-task
>          Components: Documentation, ML, PySpark
>    Affects Versions: 2.3.0
>            Reporter: Joseph K. Bradley
>            Assignee: Bryan Cutler
>            Priority: Blocker
>
> For new public APIs added to MLlib ({{spark.ml}} only), we need to check the generated HTML doc and compare the Scala & Python versions.
> * *GOAL*: Audit and create JIRAs to fix in the next release.
> * *NON-GOAL*: This JIRA is _not_ for fixing the API parity issues.
> We need to track:
> * Inconsistency: Do class/method/parameter names match?
> * Docs: Is the Python doc missing or just a stub?  We want the Python doc to be as complete as the Scala doc.
> * API breaking changes: These should be very rare but are occasionally either necessary (intentional) or accidental.  These must be recorded and added in the Migration Guide for this release.
> ** Note: If the API change is for an Alpha/Experimental/DeveloperApi component, please note that as well.
> * Missing classes/methods/parameters: We should create to-do JIRAs for functionality missing from Python, to be added in the next release cycle.  *Please use a _separate_ JIRA (linked below as "requires") for this list of to-do items.*



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org