You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sun Rui (JIRA)" <ji...@apache.org> on 2016/04/20 08:05:25 UTC
[jira] [Created] (SPARK-14746) Support transformations in R source
code for Dataset/DataFrame
Sun Rui created SPARK-14746:
-------------------------------
Summary: Support transformations in R source code for Dataset/DataFrame
Key: SPARK-14746
URL: https://issues.apache.org/jira/browse/SPARK-14746
Project: Spark
Issue Type: New Feature
Components: SparkR, SQL
Reporter: Sun Rui
there actually is a desired scenario mentioned several times in the Spark mailing list that users are writing Scala/Java Spark applications (not SparkR) but want to use R functions in some transformations. typically this can be achieved by calling Pipe() in RDD. However, there are limitations on pipe(). So we can support applying a R function in source code format to a Dataset/DataFrame (Thus SparkR is not needed for serializing an R function.)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org