You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@zeppelin.apache.org by Meethu Mathew <me...@flytxt.com> on 2018/01/03 07:05:03 UTC

More than 2 notebooks in R failing with error sparkr intrepreter not responding

Hi,

I have met with a strange issue in running R notebooks in zeppelin(0.7.2).
Spark intrepreter is in per note Scoped mode and spark version is 1.6.2

Please find the steps below to reproduce the issue:
1. Create a notebook (Note1) and run any r code in a paragraph. I ran the
following code.

> %r
>
> rdf <- data.frame(c(1,2,3,4))
>
> colnames(rdf) <- c("myCol")
>
> sdf <- createDataFrame(sqlContext, rdf)
>
> withColumn(sdf, "newCol", sdf$myCol * 2.0)
>
>
2.  Create another notebook (Note2) and run any r code in a paragraph. I
ran the same code as above.

Till now everything works fine.

3. Create third notebook (Note3) and run any r code in a paragraph. I ran
the same code. This notebook fails with the error

> org.apache.zeppelin.interpreter.InterpreterException: sparkr is not
> responding


 What I understood from the analysis is that  the process created for
sparkr interpreter is not getting killed properly and this makes every
third model to throw an error while executing. The process will be killed
on restarting the sparkr interpreter and another 2 models could be executed
successfully. ie, For every third model run using the sparkr interpreter,
the error is thrown. We suspect this as a limitation with zeppelin.

Please help to solve this issue

Regards,
Meethu Mathew

Re: More than 2 notebooks in R failing with error sparkr intrepreter not responding

Posted by Meethu Mathew <me...@flytxt.com>.
Hi Jeff,

PFB the interpreter log.

INFO [2018-01-03 12:10:05,960] ({pool-2-thread-9}
Logging.scala[logInfo]:58) - Starting HTTP Server
 INFO [2018-01-03 12:10:05,961] ({pool-2-thread-9}
Server.java[doStart]:272) - jetty-8.y.z-SNAPSHOT
 INFO [2018-01-03 12:10:05,963] ({pool-2-thread-9}
AbstractConnector.java[doStart]:338) - Started SocketConnector@0.0.0.0:58989
 INFO [2018-01-03 12:10:05,963] ({pool-2-thread-9}
Logging.scala[logInfo]:58) - Successfully started service 'HTTP class
server' on port 58989.
 INFO [2018-01-03 12:10:06,094] ({dispatcher-event-loop-1}
Logging.scala[logInfo]:58) - Removed broadcast_1_piece0 on localhost:42453
in memory (size: 854.0 B, free: 511.1 MB)
 INFO [2018-01-03 12:10:07,049] ({pool-2-thread-9}
ZeppelinR.java[createRScript]:353) - File
/tmp/zeppelin_sparkr-5046601627391341672.R created
ERROR [2018-01-03 12:10:17,051] ({pool-2-thread-9} Job.java[run]:188) - Job
failed
*org.apache.zeppelin.interpreter.InterpreterException: sparkr is not
responding *


R version 3.4.1 (2017-06-30) -- "Single Candle"
Copyright (C) 2017 The R Foundation for Statistical Computing
Platform: x86_64-pc-linux-gnu (64-bit)
....
....
> args <- commandArgs(trailingOnly = TRUE)

> hashCode <- as.integer(args[1])

> port <- as.integer(args[2])

> libPath <- args[3]

> version <- as.integer(args[4])

> rm(args)

>
> print(paste("Port ", toString(port)))
[1]
 "Port  58063"

> print(paste("LibPath ", libPath))
[1]

 "LibPath  /home/meethu/spark-1.6.1-bin-hadoop2.6/R/lib"
>

> .libPaths(c(file.path(libPath), .libPaths()))

> library(SparkR)

Attaching package: ‘SparkR’

The following objects are masked from ‘package:stats’:

    cov, filter, lag, na.omit, predict, sd, var

The following objects are masked from ‘package:base’:

    colnames, colnames<-, endsWith, intersect, rank, rbind, sample,
    startsWith, subset, summary, table, transform

> SparkR:::connectBackend("localhost", port, 6000)
A connection with
description "->localhost:58063"
class
 "sockconn"
mode        "wb"
text        "binary"
opened      "opened"
can read    "yes"
can write   "yes"
>

> # scStartTime is needed by R/pkg/R/sparkR.R

> assign(".scStartTime", as.integer(Sys.time()), envir =
SparkR:::.sparkREnv)

> # getZeppelinR

> *.zeppelinR = SparkR:::callJStatic("org.apache.zeppelin.spark.ZeppelinR",
"getZeppelinR", hashCode)*

at
org.apache.zeppelin.spark.ZeppelinR.waitForRScriptInitialized(ZeppelinR.java:285)
at org.apache.zeppelin.spark.ZeppelinR.request(ZeppelinR.java:227)
at org.apache.zeppelin.spark.ZeppelinR.eval(ZeppelinR.java:176)
at org.apache.zeppelin.spark.ZeppelinR.open(ZeppelinR.java:165)
at
org.apache.zeppelin.spark.SparkRInterpreter.open(SparkRInterpreter.java:90)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:491)
at org.apache.zeppelin.scheduler.Job.run(Job.java:175)
at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
 INFO [2018-01-03 12:10:17,070] ({pool-2-thread-9}
SchedulerFactory.java[jobFinished]:137) - Job
remoteInterpretJob_1514961605951 finished by scheduler
org.apache.zeppelin.spark.SparkRInterpreter392022746
 INFO [2018-01-03 12:39:22,664] ({Spark Context Cleaner}
Logging.scala[logInfo]:58) - Cleaned accumulator 2


PFB the output of the command  *ps -ef | grep /usr/lib/R/bin/exec/R*

meethu    6647  6470  0 12:09 pts/1    00:00:00 /usr/lib/R/bin/exec/R
--no-save --no-restore -f /tmp/zeppelin_sparkr-1100854828050763213.R --args
214655664 58063 /home/meethu/spark-1.6.1-bin-hadoop2.6/R/lib 10601
meethu    6701  6470  0 12:09 pts/1    00:00:00 /usr/lib/R/bin/exec/R
--no-save --no-restore -f /tmp/zeppelin_sparkr-4152305170353311178.R --args
1642312173 58063 /home/meethu/spark-1.6.1-bin-hadoop2.6/R/lib 10601
meethu    6745  6470  0 12:10 pts/1    00:00:00 /usr/lib/R/bin/exec/R
--no-save --no-restore -f /tmp/zeppelin_sparkr-5046601627391341672.R --args
1158632477 58063 /home/meethu/spark-1.6.1-bin-hadoop2.6/R/lib 10601


Regards,
Meethu Mathew


On Wed, Jan 3, 2018 at 12:56 PM, Jeff Zhang <zj...@gmail.com> wrote:

>
> Could you check the interpreter log ?
>
> Meethu Mathew <me...@flytxt.com>于2018年1月3日周三 下午3:05写道:
>
>> Hi,
>>
>> I have met with a strange issue in running R notebooks in zeppelin(0.7.2).
>> Spark intrepreter is in per note Scoped mode and spark version is 1.6.2
>>
>> Please find the steps below to reproduce the issue:
>> 1. Create a notebook (Note1) and run any r code in a paragraph. I ran the
>> following code.
>>
>>> %r
>>>
>>> rdf <- data.frame(c(1,2,3,4))
>>>
>>> colnames(rdf) <- c("myCol")
>>>
>>> sdf <- createDataFrame(sqlContext, rdf)
>>>
>>> withColumn(sdf, "newCol", sdf$myCol * 2.0)
>>>
>>>
>> 2.  Create another notebook (Note2) and run any r code in a paragraph. I
>> ran the same code as above.
>>
>> Till now everything works fine.
>>
>> 3. Create third notebook (Note3) and run any r code in a paragraph. I
>> ran the same code. This notebook fails with the error
>>
>>> org.apache.zeppelin.interpreter.InterpreterException: sparkr is not
>>> responding
>>
>>
>>  What I understood from the analysis is that  the process created for
>> sparkr interpreter is not getting killed properly and this makes every
>> third model to throw an error while executing. The process will be killed
>> on restarting the sparkr interpreter and another 2 models could be executed
>> successfully. ie, For every third model run using the sparkr interpreter,
>> the error is thrown. We suspect this as a limitation with zeppelin.
>>
>> Please help to solve this issue
>>
>> Regards,
>>
>>
>> Meethu Mathew
>>
>>

Re: More than 2 notebooks in R failing with error sparkr intrepreter not responding

Posted by Jeff Zhang <zj...@gmail.com>.
Could you check the interpreter log ?

Meethu Mathew <me...@flytxt.com>于2018年1月3日周三 下午3:05写道:

> Hi,
>
> I have met with a strange issue in running R notebooks in zeppelin(0.7.2).
> Spark intrepreter is in per note Scoped mode and spark version is 1.6.2
>
> Please find the steps below to reproduce the issue:
> 1. Create a notebook (Note1) and run any r code in a paragraph. I ran the
> following code.
>
>> %r
>>
>> rdf <- data.frame(c(1,2,3,4))
>>
>> colnames(rdf) <- c("myCol")
>>
>> sdf <- createDataFrame(sqlContext, rdf)
>>
>> withColumn(sdf, "newCol", sdf$myCol * 2.0)
>>
>>
> 2.  Create another notebook (Note2) and run any r code in a paragraph. I
> ran the same code as above.
>
> Till now everything works fine.
>
> 3. Create third notebook (Note3) and run any r code in a paragraph. I ran
> the same code. This notebook fails with the error
>
>> org.apache.zeppelin.interpreter.InterpreterException: sparkr is not
>> responding
>
>
>  What I understood from the analysis is that  the process created for
> sparkr interpreter is not getting killed properly and this makes every
> third model to throw an error while executing. The process will be killed
> on restarting the sparkr interpreter and another 2 models could be executed
> successfully. ie, For every third model run using the sparkr interpreter,
> the error is thrown. We suspect this as a limitation with zeppelin.
>
> Please help to solve this issue
>
> Regards,
>
>
> Meethu Mathew
>
>