You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Joseph K. Bradley (JIRA)" <ji...@apache.org> on 2015/05/15 22:55:59 UTC

[jira] [Created] (SPARK-7675) PySpark spark.ml Params type conversions

Joseph K. Bradley created SPARK-7675:
----------------------------------------

             Summary: PySpark spark.ml Params type conversions
                 Key: SPARK-7675
                 URL: https://issues.apache.org/jira/browse/SPARK-7675
             Project: Spark
          Issue Type: Improvement
          Components: ML, PySpark
            Reporter: Joseph K. Bradley
            Priority: Minor


Currently, PySpark wrappers for spark.ml Scala classes are brittle when accepting Param types.  E.g., Normalizer's "p" param cannot be set to "2" (an integer); it must be set to "2.0" (a float).  Fixing this is not trivial since there does not appear to be a natural place to insert the conversion before Python wrappers call Java's Params setter method.

A possible fix will be to include a method "_checkType" to PySpark's Param class which checks the type, prints an error if needed, and converts types when relevant (e.g., int to float, or scipy matrix to array).  The Java wrapper method which copies params to Scala can call this method when available.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org