You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@zeppelin.apache.org by shane lee <sh...@gmail.com> on 2016/05/27 06:40:03 UTC

Problem starting SparkR Interpreter on external spark

Hello,

I have been trying to use the SparkR interpreter on Spark 1.6.1 and run
into a problem. The SparkR interpreter works well in stand alone
configuration with master set to local[*] and SPARK_HOME undefined.

With master set to yarn-client (with hadoop 2.6.4) and SPARK_HOME pointing
to the installed location of Spark 1.6.1, the spark and pyspark
interpreters work well. However the R/SparkR interpreters always fail with
"sparkr is not responding" error.

This is with the master branch of the code from GitHub and here is the full
output:

org.apache.zeppelin.interpreter.InterpreterException: sparkr is not
responding

R version 3.2.2 (2015-08-14) -- "Fire Safety"
Copyright (C) 2015 The R Foundation for Statistical Computing
Platform: x86_64-w64-mingw32/x64 (64-bit)

R is free software and comes with ABSOLUTELY NO WARRANTY.
You are welcome to redistribute it under certain conditions.
Type 'license()' or 'licence()' for distribution details.

  Natural language support but running in an English locale

R is a collaborative project with many contributors.
Type 'contributors()' for more information and
'citation()' on how to cite R or R packages in publications.

Type 'demo()' for some demos, 'help()' for on-line help, or
'help.start()' for an HTML browser interface to help.
Type 'q()' to quit R.

> #
> # Licensed to the Apache Software Foundation (ASF) under one
> # or more contributor license agreements.  See the NOTICE file
> # distributed with this work for additional information
> # regarding copyright ownership.  The ASF licenses this file
> # to you under the Apache License, Version 2.0 (the
> # "License"); you may not use this file except in compliance
> # with the License.  You may obtain a copy of the License at
> #
> #     http://www.apache.org/licenses/LICENSE-2.0
> #
> # Unless required by applicable law or agreed to in writing, software
> # distributed under the License is distributed on an "AS IS" BASIS,
> # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
> # See the License for the specific language governing permissions and
> # limitations under the License.
> #
>
> args <- commandArgs(trailingOnly = TRUE)
>
> hashCode <- as.integer(args[1])
> port <- as.integer(args[2])
> libPath <- args[3]
> rm(args)
> print(paste("hashCode ", toString(hashCode)))
[1] "hashCode  1850484525"
> print(paste("Port ", toString(port)))
[1] "Port  56762"
> print(paste("LibPath ", libPath))
[1] "LibPath  c:\\spark-1.6.1-bin-hadoop2.6\\bin\\../R/lib"
>
> .libPaths(c(file.path(libPath), .libPaths()))
> library(SparkR)

Attaching package: 'SparkR'

The following objects are masked from 'package:stats':

    cov, filter, lag, na.omit, predict, sd, var

The following objects are masked from 'package:base':

    colnames, colnames<-, intersect, rank, rbind, sample, subset,
    summary, table, transform

>
>
> SparkR:::connectBackend("localhost", port)
        description               class                mode
 text
"->localhost:56762"          "sockconn"                "wb"
 "binary"
             opened            can read           can write
           "opened"               "yes"               "yes"
>
> # scStartTime is needed by R/pkg/R/sparkR.R
> assign(".scStartTime", as.integer(Sys.time()), envir =
SparkR:::.sparkREnv)
>
> # getZeppelinR
> .zeppelinR = SparkR:::callJStatic("org.apache.zeppelin.spark.ZeppelinR",
"getZeppelinR", hashCode)
Error in invokeJava(isStatic = TRUE, className, methodName, ...) :
Calls: <Anonymous> -> invokeJava
Execution halted

at
org.apache.zeppelin.spark.ZeppelinR.waitForRScriptInitialized(ZeppelinR.java:312)
at org.apache.zeppelin.spark.ZeppelinR.request(ZeppelinR.java:250)
at org.apache.zeppelin.spark.ZeppelinR.eval(ZeppelinR.java:198)
at org.apache.zeppelin.spark.ZeppelinR.open(ZeppelinR.java:187)
at
org.apache.zeppelin.spark.SparkRInterpreter.open(SparkRInterpreter.java:104)
at
org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:345)
at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)