You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Apache Spark (JIRA)" <ji...@apache.org> on 2016/01/19 21:27:40 UTC
[jira] [Assigned] (SPARK-12910) Support for specifying version of R
to use while creating sparkR libraries
[ https://issues.apache.org/jira/browse/SPARK-12910?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Apache Spark reassigned SPARK-12910:
------------------------------------
Assignee: (was: Apache Spark)
> Support for specifying version of R to use while creating sparkR libraries
> --------------------------------------------------------------------------
>
> Key: SPARK-12910
> URL: https://issues.apache.org/jira/browse/SPARK-12910
> Project: Spark
> Issue Type: Improvement
> Components: SparkR
> Environment: Linux
> Reporter: Shubhanshu Mishra
> Priority: Minor
> Labels: installation, sparkR
>
> When we use `$SPARK_HOME/R/install-dev.sh` it uses the default system R. However, a user might have locally installed their own version of R. There should be a way to specify which R version to use.
> I have fixed this in my code using the following patch:
> ```
> $ git diff HEAD
> diff --git a/R/README.md b/R/README.md
> index 005f56d..99182e5 100644
> --- a/R/README.md
> +++ b/R/README.md
> @@ -1,6 +1,15 @@
> # R on Spark
>
> SparkR is an R package that provides a light-weight frontend to use Spark from R.
> +### Installing sparkR
> +
> +Libraries of sparkR need to be created in `$SPARK_HOME/R/lib`. This can be done by running the script `$SPARK_HOME/R/install-dev.sh`.
> +By default the above script uses the system wide installation of R. However, this can be changed to any user installed location of R by giving the full path of the `$R_HOME` as the first argument to the install-dev.sh script.
> +Example:
> +```
> +# where /home/username/R is where R is installed and /home/username/R/bin contains the files R and RScript
> +./install-dev.sh /home/username/R
> +```
>
> ### SparkR development
>
> diff --git a/R/install-dev.sh b/R/install-dev.sh
> index 4972bb9..a8efa86 100755
> --- a/R/install-dev.sh
> +++ b/R/install-dev.sh
> @@ -35,12 +35,19 @@ LIB_DIR="$FWDIR/lib"
> mkdir -p $LIB_DIR
>
> pushd $FWDIR > /dev/null
> +if [ ! -z "$1" ]
> + then
> + R_HOME="$1/bin"
> + else
> + R_HOME="$(dirname $(which R))"
> +fi
> +echo "USING R_HOME = $R_HOME"
>
> # Generate Rd files if devtools is installed
> -Rscript -e ' if("devtools" %in% rownames(installed.packages())) { library(devtools); devtools::document(pkg="./pkg", roclets=c("rd")) }'
> +"$R_HOME/"Rscript -e ' if("devtools" %in% rownames(installed.packages())) { library(devtools); devtools::document(pkg="./pkg", roclets=c("rd")) }'
>
> # Install SparkR to $LIB_DIR
> -R CMD INSTALL --library=$LIB_DIR $FWDIR/pkg/
> +"$R_HOME/"R CMD INSTALL --library=$LIB_DIR $FWDIR/pkg/
>
> # Zip the SparkR package so that it can be distributed to worker nodes on YARN
> cd $LIB_DIR
> ```
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org