You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@toree.apache.org by "Gino Bustelo (JIRA)" <ji...@apache.org> on 2016/05/25 13:50:12 UTC

[jira] [Updated] (TOREE-315) SparkR interpreter hangs when running on Spark 1.6.1

     [ https://issues.apache.org/jira/browse/TOREE-315?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Gino Bustelo updated TOREE-315:
-------------------------------
    Fix Version/s: 0.1.0

> SparkR interpreter hangs when running on Spark 1.6.1
> ----------------------------------------------------
>
>                 Key: TOREE-315
>                 URL: https://issues.apache.org/jira/browse/TOREE-315
>             Project: TOREE
>          Issue Type: Bug
>    Affects Versions: 0.1.0
>         Environment: Spark 1.6.1
> Jupyter notebook 4.2.0
> R 3.2.2
> Toree 0.1.0.dev7
>            Reporter: Gurvinder
>            Priority: Minor
>             Fix For: 0.1.0
>
>
> Create notebook with scala interpreter and once notebook started with spark context, running SparkR magic as
> %%SparkR sc
> cause notebook to hang and no further execution of any cells. Need to restart the notebook. Issue seems to be caused by patching/updating install SparkR package in spark directory (/usr/local/spark/R/libs/SparkR , in my case). Here is the diff from default installed SparkR and updated one
> diff SparkR/DESCRIPTION SparkR.bak/DESCRIPTION                                                       
> 4c4                                                                                                  
> < Version: 1.6.1                                                                                     
> ---                                                                                                  
> > Version: 1.5.0                                                                                     
> 16,18c16,17                                                                                          
> <         'functions.R' 'mllib.R' 'serialize.R' 'sparkR.R' 'stats.R'                                 
> <         'types.R' 'utils.R'                                                                        
> < Built: R 3.1.1; ; 2016-02-27 04:45:19 UTC; unix                                                    
> ---                                                                                                  
> >         'functions.R' 'mllib.R' 'serialize.R' 'sparkR.R' 'utils.R'                                 
> > Built: R 3.2.1; ; 2015-10-01 18:38:50 UTC; unix                                                    
> Only in SparkR: INDEX                                                                                
> Common subdirectories: SparkR/Meta and SparkR.bak/Meta                                               
> diff SparkR/NAMESPACE SparkR.bak/NAMESPACE                                                           
> 12a13,22                                                                                             
> > # Needed exports for runner                                                                        
> > export("sparkR.connect")                                                                           
> > export("isInstanceOf")                                                                             
> > export("callJMethod")                                                                              
> > export("callJStatic")                                                                              
> > export("newJObject")                                                                               
> > export("removeJObject")                                                                            
> > export("isRemoveMethod")                                                                           
> > export("invokeJava")                                                                               
> >                                                                                                    
> 26,27d35                                                                                             
> <               "as.data.frame",                                                                     
> <               "attach",                                                                            
> 30,33d37                                                                                             
> <               "colnames",                                                                          
> <               "colnames<-",                                                                        
> <               "coltypes",                                                                          
> <               "coltypes<-",                                                                        
> 36,37d39                                                                                             
> <               "cov",                                                                               
> <               "corr",                                                                              
> 49d50                                                                                                
> <               "freqItems",
> Also once the default version is patched, running sparkR from terminal also give this error which worked earlier
> Launching java with spark-submit command /usr/local/spark/bin/spark-submit   "sparkr-shell" /tmp/RtmpbfR7Ea/backend_port27b11cd7f3f 
> 16/05/17 18:26:02 ERROR RBackendHandler: createSparkContext on org.apache.spark.api.r.RRDD failed
> Error in invokeJava(isStatic = TRUE, className, methodName, ...) :



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)