You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@zeppelin.apache.org by Theo Diefenthal <th...@scoop-software.de> on 2021/04/01 23:48:40 UTC

Connect Dockerized Zeppelin to dev Cloudera 6.3.1 cluster [Promise timeout when Registering interpreter process]

Hi there, 

I want to achieve the following usecase: Start Zeppelin 0.9.0 (in docker) on my local dev machine but let the Spark jobs in the notebook run on a remote cluster via YARN. 

For a few hours already, I try to setup that environment with my companies Cloudera CDH 6.3.1 development cluster. That cluster is unsecured (despite that it can only be reached when connected to VPN). With a lot of trial and error I finally achieved a successful connection from my dockerized Zeppelin to the cluster. This means that when I start running a spark cell in Zeppelin, I can see a new application in YARN on the cluster-side [named spark-shared_process] . However, eventually the execution of the cell will fail with the following stack trace in the yarn application [1]. I have no idea where this timeout could potentially come from and I'd be happy if you could help me out here. In the said VPN to the dev cluster, there are no connection restrictions like firewalls or stuff like that engaged. The cell I run is the first one in "3. Spark SQL (Scala)" Zeppelin quick start notebooks with title "Create Dataset/DataFrame via SparkSession". 

For reference, I also attach my docker-compose file [2] and my Dockerfile for building Zeppelin with Spark and Hadoop [3] (Note that I add hadoop conf files into the image because I'd like to distribute the image as ready-to-run for the other people in my project without needing them to copy over the hadoop conf files). After start of the container, I further change the interpreter settings by setting yarn-cluster in %spark interpreter settings and also set zeppelin.interpreter.connect.timeout to 600.000. 

Best regards 
Theo 

PS: HDFS in general seems to work well. [4] 
PPS: I also attach the docker container logs from an attempt [5] 



[1] 
INFO [2021-04-01 23:48:20,984] ({main} Logging.scala[logInfo]:54) - Registered signal handler for TERM 
INFO [2021-04-01 23:48:21,005] ({main} Logging.scala[logInfo]:54) - Registered signal handler for HUP 
INFO [2021-04-01 23:48:21,014] ({main} Logging.scala[logInfo]:54) - Registered signal handler for INT 
INFO [2021-04-01 23:48:22,158] ({main} Logging.scala[logInfo]:54) - Changing view acls to: yarn,sandbox 
INFO [2021-04-01 23:48:22,160] ({main} Logging.scala[logInfo]:54) - Changing modify acls to: yarn,sandbox 
INFO [2021-04-01 23:48:22,161] ({main} Logging.scala[logInfo]:54) - Changing view acls groups to: 
INFO [2021-04-01 23:48:22,162] ({main} Logging.scala[logInfo]:54) - Changing modify acls groups to: 
INFO [2021-04-01 23:48:22,168] ({main} Logging.scala[logInfo]:54) - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(yarn, sandbox); groups with view permissions: Set(); users with modify permissions: Set(yarn, sandbox); groups with modify permissions: Set() 
INFO [2021-04-01 23:48:25,388] ({main} Logging.scala[logInfo]:54) - Preparing Local resources 
WARN [2021-04-01 23:48:28,111] ({main} NativeCodeLoader.java[<clinit>]:62) - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 
INFO [2021-04-01 23:48:29,004] ({main} Logging.scala[logInfo]:54) - ApplicationAttemptId: appattempt_1617228950227_5781_000001 
INFO [2021-04-01 23:48:29,041] ({main} Logging.scala[logInfo]:54) - Starting the user application in a separate Thread 
INFO [2021-04-01 23:48:29,289] ({main} Logging.scala[logInfo]:54) - Waiting for spark context initialization... 
INFO [2021-04-01 23:48:30,007] ({RegisterThread} RemoteInterpreterServer.java[run]:595) - Start registration 
INFO [2021-04-01 23:48:30,009] ({RemoteInterpreterServer-Thread} RemoteInterpreterServer.java[run]:193) - Launching ThriftServer at 99.99.99.99:44802 
INFO [2021-04-01 23:48:31,276] ({RegisterThread} RemoteInterpreterServer.java[run]:609) - Registering interpreter process 
ERROR [2021-04-01 23:50:09,531] ({main} Logging.scala[logError]:91) - Uncaught exception: 
java.util.concurrent.TimeoutException: Futures timed out after [100000 milliseconds] 
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:223) 
at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:227) 
at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:220) 
at org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:469) 
at org.apache.spark.deploy.yarn.ApplicationMaster.org$apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:305) 
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply$mcV$sp(ApplicationMaster.scala:245) 
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245) 
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245) 
at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$3.run(ApplicationMaster.scala:780) 
at java.security.AccessController.doPrivileged(Native Method) 
at javax.security.auth.Subject.doAs(Subject.java:422) 
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) 
at org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:779) 
at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:244) 
at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:804) 
at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala) 
INFO [2021-04-01 23:50:09,547] ({main} Logging.scala[logInfo]:54) - Final app status: FAILED, exitCode: 13, (reason: Uncaught exception: java.util.concurrent.TimeoutException: Futures timed out after [100000 milliseconds] 


[2] 
version: '3.7' 
services: 
zeppelin: 
build: zeppelin-customized 
ports: 
- "9999:8080" 
environment: 
ZEPPELIN_PORT: 8080 
ZEPPELIN_JAVA_OPTS: >- 
-Dspark.driver.memory=1g 
-Dspark.executor.memory=2g 
HADOOP_USER_NAME: sandbox 
volumes: 
- zeppelindata:/zeppelin/data 
- zeppelinnotebooks:/zeppelin/notebook 
volumes: 
zeppelindata: 
zeppelinnotebooks: 


[3] 
FROM apache/zeppelin:0.9.0 
# default user is 1000 in zeppelin base.. 
USER root 
RUN mkdir /spark && chown 1000:1000 /spark && mkdir /hadoop && chown 1000:1000 /hadoop 
USER 1000 

# Add Spark 
RUN cd /spark \ 
&& wget https://artfiles.org/apache.org/spark/spark-2.4.7/spark-2.4.7-bin-hadoop2.7.tgz \ 
&& tar xf spark-2.4.7-bin-hadoop2.7.tgz \ 
&& rm spark-2.4.7-bin-hadoop2.7.tgz \ 
&& cd ~ 
ENV SPARK_HOME /spark/spark-2.4.7-bin-hadoop2.7 
ENV HADOOP_CONF_DIR /zeppelin/conf 

# Add Hadoop 
RUN cd /hadoop \ 
&& wget https://archive.apache.org/dist/hadoop/common/hadoop-3.0.0/hadoop-3.0.0.tar.gz \ 
&& tar xf hadoop-3.0.0.tar.gz \ 
&& rm hadoop-3.0.0.tar.gz \ 
&& cd ~ 
ENV HADOOP_HOME /hadoop/hadoop-3.0.0 
ENV HADOOP_INSTALL=$HADOOP_HOME 
ENV HADOOP_MAPRED_HOME=$HADOOP_HOME 
ENV HADOOP_COMMON_HOME=$HADOOP_HOME 
ENV HADOOP_HDFS_HOME=$HADOOP_HOME 
ENV YARN_HOME=$HADOOP_HOME 
ENV HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native 
ENV HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib/nativ" 
ENV PATH="${HADOOP_HOME}/bin:${HADOOP_HOME}/sbin:${PATH}" 
ENV USE_HADOOP=true 

# Copy over /etc/hadoop/conf from one of the cluster nodes... 
COPY cloudernode/conf/ /zeppelin/conf/ 


[4] 
%sh 
hdfs dfs -ls /user/sandbox 
=> prints out properly. 


[5] 
zeppelin_1 | WARN [2021-04-01 23:18:36,440] ({SchedulerFactory4} SparkInterpreterLauncher.java[buildEnvFromProperties]:221) - spark-defaults.conf doesn't exist: /spark/spark-2.4.7-bin-hadoop2.7/conf/spark-defaults.conf 
zeppelin_1 | INFO [2021-04-01 23:18:36,440] ({SchedulerFactory4} SparkInterpreterLauncher.java[buildEnvFromProperties]:224) - buildEnvFromProperties: {PATH=/hadoop/hadoop-3.0.0/bin:/hadoop/hadoop-3.0.0/sbin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin, ZEPPELIN_PORT=8080, HADOOP_CONF_DIR=/zeppelin/conf, ZEPPELIN_JAVA_OPTS=-Dspark.driver.memory=1g -Dspark.executor.memory=2g, ZEPPELIN_LOG_DIR=/opt/zeppelin/logs, MASTER=yarn, ZEPPELIN_WAR=/opt/zeppelin/zeppelin-web-0.9.0.war, ZEPPELIN_ENCODING=UTF-8, ZEPPELIN_SPARK_CONF= --conf spark.yarn.dist.archives=/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr --conf spark.yarn.isPython=true --conf spark.executor.instances=2 --conf spark.app.name=spark-shared_process --conf spark.webui.yarn.useProxy=false --conf spark.driver.cores=1 --conf spark.yarn.maxAppAttempts=1 --conf spark.executor.memory=2g --conf spark.master=yarn-cluster --conf spark.files=/opt/zeppelin/conf/log4j_yarn_cluster.properties --conf spark.driver.memory=1g --conf spark.jars=/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar,/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar --conf spark.executor.cores=1 --conf spark.yarn.submit.waitAppCompletion=false, JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64, JAVA_OPTS= -Dspark.driver.memory=1g -Dspark.executor.memory=2g -Dfile.encoding=UTF-8 -Xms1024m -Xmx1024m -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties -Dzeppelin.log.file=/opt/zeppelin/logs/zeppelin--d5ea32f1f431.log, INTERPRETER_GROUP_ID=spark-shared_process, Z_VERSION=0.9.0, LANG=en_US.UTF-8, JAVA_INTP_OPTS= -Dfile.encoding=UTF-8 -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties -Dlog4j.configurationFile=file:///opt/zeppelin/conf/log4j2.properties, PYSPARK_PYTHON=python, HADOOP_USER_NAME=sandbox, ZEPPELIN_SPARK_YARN_CLUSTER=true, Z_HOME=/opt/zeppelin, SPARK_HOME=/spark/spark-2.4.7-bin-hadoop2.7, ZEPPELIN_CONF_DIR=/opt/zeppelin/conf, YARN_HOME=/hadoop/hadoop-3.0.0, HADOOP_HDFS_HOME=/hadoop/hadoop-3.0.0, ZEPPELIN_RUNNER=/usr/lib/jvm/java-8-openjdk-amd64/bin/java, HADOOP_MAPRED_HOME=/hadoop/hadoop-3.0.0, PWD=/opt/zeppelin, HADOOP_COMMON_HOME=/hadoop/hadoop-3.0.0, HADOOP_INSTALL=/hadoop/hadoop-3.0.0, ZEPPELIN_HOME=/opt/zeppelin, LOG_TAG=[ZEPPELIN_0.9.0]:, ZEPPELIN_INTP_MEM=-Xms1024m -Xmx2048m, HADOOP_OPTS=-Djava.library.path=/hadoop/hadoop-3.0.0/lib/nativ, PYSPARK_DRIVER_PYTHON=python, ZEPPELIN_PID_DIR=/opt/zeppelin/run, ZEPPELIN_ANGULAR_WAR=/opt/zeppelin/zeppelin-web-angular-0.9.0.war, ZEPPELIN_MEM=-Xms1024m -Xmx1024m, HOSTNAME=d5ea32f1f431, LC_ALL=en_US.UTF-8, ZEPPELIN_IDENT_STRING=, PYSPARK_PIN_THREAD=true, HADOOP_HOME=/hadoop/hadoop-3.0.0, USE_HADOOP=true, HADOOP_COMMON_LIB_NATIVE_DIR=/hadoop/hadoop-3.0.0/lib/native, ZEPPELIN_ADDR=0.0.0.0, ZEPPELIN_INTERPRETER_REMOTE_RUNNER=bin/interpreter.sh, SHLVL=0, HOME=/opt/zeppelin} 
zeppelin_1 | INFO [2021-04-01 23:18:36,445] ({SchedulerFactory4} ProcessLauncher.java[transition]:109) - Process state is transitioned to LAUNCHED 
zeppelin_1 | INFO [2021-04-01 23:18:36,446] ({SchedulerFactory4} ProcessLauncher.java[launch]:96) - Process is launched: [/opt/zeppelin/bin/interpreter.sh, -d, /opt/zeppelin/interpreter/spark, -c, 172.2.0.2, -p, 46781, -r, :, -i, spark-shared_process, -l, /opt/zeppelin/local-repo/spark, -g, spark] 
zeppelin_1 | WARN [2021-04-01 23:20:51,930] ({Exec Default Executor} RemoteInterpreterManagedProcess.java[onProcessComplete]:255) - Process is exited with exit value 0 
zeppelin_1 | INFO [2021-04-01 23:20:51,933] ({Exec Default Executor} ProcessLauncher.java[transition]:109) - Process state is transitioned to COMPLETED 
zeppelin_1 | INFO [2021-04-01 23:24:06,162] ({qtp418304857-11} VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln 
zeppelin_1 | INFO [2021-04-01 23:24:15,933] ({qtp418304857-27} VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln 
zeppelin_1 | WARN [2021-04-01 23:28:36,539] ({SchedulerFactory4} NotebookServer.java[onStatusChange]:1928) - Job 20180530-101750_1491737301 is finished, status: ERROR, exception: null, result: %text org.apache.zeppelin.interpreter.InterpreterException: java.io.IOException: Fail to launch interpreter process: 
zeppelin_1 | Warning: Master yarn-cluster is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead. 
zeppelin_1 | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 
zeppelin_1 | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (16400 MB per container) 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM container, with 1408 MB memory including 384 MB overhead 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up container launch context for our AM 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch environment for our AM container 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for our AM container 
zeppelin_1 | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME. 
zeppelin_1 | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip 
zeppelin_1 | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar 
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar 
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar 
zeppelin_1 | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties 
zeppelin_1 | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip 
zeppelin_1 | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip 
zeppelin_1 | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip 
zeppelin_1 | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls to: zeppelin,sandbox 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls to: zeppelin,sandbox 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls groups to: 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls groups to: 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin, sandbox); groups with view permissions: Set(); users with modify permissions: Set(zeppelin, sandbox); groups with modify permissions: Set() 
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Submitting application application_1617315347811_0170 to ResourceManager 
zeppelin_1 | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted application application_1617315347811_0170 
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Application report for application_1617315347811_0170 (state: ACCEPTED) 
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: 
zeppelin_1 | client token: N/A 
zeppelin_1 | diagnostics: N/A 
zeppelin_1 | ApplicationMaster host: N/A 
zeppelin_1 | ApplicationMaster RPC port: -1 
zeppelin_1 | queue: root.users.sandbox 
zeppelin_1 | start time: 1617319251597 
zeppelin_1 | final status: UNDEFINED 
zeppelin_1 | tracking URL: http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/ 
zeppelin_1 | user: sandbox 
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown hook called 
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2 
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440 
zeppelin_1 | 
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:129) 
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getFormType(RemoteInterpreter.java:271) 
zeppelin_1 | at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:444) 
zeppelin_1 | at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:72) 
zeppelin_1 | at org.apache.zeppelin.scheduler.Job.run(Job.java:172) 
zeppelin_1 | at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:132) 
zeppelin_1 | at org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:182) 
zeppelin_1 | at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
zeppelin_1 | at java.util.concurrent.FutureTask.run(FutureTask.java:266) 
zeppelin_1 | at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) 
zeppelin_1 | at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) 
zeppelin_1 | at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
zeppelin_1 | at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
zeppelin_1 | at java.lang.Thread.run(Thread.java:748) 
zeppelin_1 | Caused by: java.io.IOException: Fail to launch interpreter process: 
zeppelin_1 | Warning: Master yarn-cluster is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead. 
zeppelin_1 | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 
zeppelin_1 | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (16400 MB per container) 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM container, with 1408 MB memory including 384 MB overhead 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up container launch context for our AM 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch environment for our AM container 
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for our AM container 
zeppelin_1 | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME. 
zeppelin_1 | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip 
zeppelin_1 | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar 
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar 
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar 
zeppelin_1 | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties 
zeppelin_1 | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip 
zeppelin_1 | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip 
zeppelin_1 | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip 
zeppelin_1 | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls to: zeppelin,sandbox 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls to: zeppelin,sandbox 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls groups to: 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls groups to: 
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin, sandbox); groups with view permissions: Set(); users with modify permissions: Set(zeppelin, sandbox); groups with modify permissions: Set() 
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Submitting application application_1617315347811_0170 to ResourceManager 
zeppelin_1 | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted application application_1617315347811_0170 
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Application report for application_1617315347811_0170 (state: ACCEPTED) 
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: 
zeppelin_1 | client token: N/A 
zeppelin_1 | diagnostics: N/A 
zeppelin_1 | ApplicationMaster host: N/A 
zeppelin_1 | ApplicationMaster RPC port: -1 
zeppelin_1 | queue: root.users.sandbox 
zeppelin_1 | start time: 1617319251597 
zeppelin_1 | final status: UNDEFINED 
zeppelin_1 | tracking URL: http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/ 
zeppelin_1 | user: sandbox 
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown hook called 
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440 
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2 
zeppelin_1 | 
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreterManagedProcess.start(RemoteInterpreterManagedProcess.java:126) 
zeppelin_1 | at org.apache.zeppelin.interpreter.ManagedInterpreterGroup.getOrCreateInterpreterProcess(ManagedInterpreterGroup.java:68) 
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getOrCreateInterpreterProcess(RemoteInterpreter.java:104) 
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.internal_create(RemoteInterpreter.java:154) 
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:126) 
zeppelin_1 | ... 13 more 
zeppelin_1 | 
zeppelin_1 | INFO [2021-04-01 23:28:36,542] ({SchedulerFactory4} VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln 


Re: Connect Dockerized Zeppelin to dev Cloudera 6.3.1 cluster [Promise timeout when Registering interpreter process]

Posted by Jeff Zhang <zj...@gmail.com>.
Most likely it is due to network issue, the connection between spark driver
(run in yarn container in yarn cluster mode) and zeppelin server is
bidirectional.  It looks like it is due to your spark driver unable to
connect to the zeppelin server.

Theo Diefenthal <th...@scoop-software.de> 于2021年4月2日周五 上午7:48写道:

> Hi there,
>
> I want to achieve the following usecase: Start Zeppelin 0.9.0 (in docker)
> on my local dev machine but let the Spark jobs in the notebook run on a
> remote cluster via YARN.
>
> For a few hours already, I try to setup that environment with my companies
> Cloudera CDH 6.3.1 development cluster. That cluster is unsecured (despite
> that it can only be reached when connected to VPN). With a lot of trial and
> error I finally achieved a successful connection from my dockerized
> Zeppelin to the cluster. This means that when I start running a spark cell
> in Zeppelin, I can see a new application in YARN on the cluster-side [named
> spark-shared_process]. However, eventually the execution of the cell will
> fail with the following stack trace in the yarn application [1]. I have no
> idea where this timeout could potentially come from and I'd be happy if you
> could help me out here. In the said VPN to the dev cluster, there are no
> connection restrictions like firewalls or stuff like that engaged. The cell
> I run is the first one in "3. Spark SQL (Scala)" Zeppelin quick start
> notebooks with title "Create Dataset/DataFrame via SparkSession".
>
> For reference, I also attach my docker-compose file [2] and my Dockerfile
> for building Zeppelin with Spark and Hadoop [3] (Note that I add hadoop
> conf files into the image because I'd like to distribute the image as
> ready-to-run for the other people in my project without needing them to
> copy over the hadoop conf files). After start of the container, I further
> change the interpreter settings by setting yarn-cluster in %spark
> interpreter settings and also set zeppelin.interpreter.connect.timeout to
> 600.000.
>
> Best regards
> Theo
>
> PS: HDFS in general seems to work well. [4]
> PPS: I also attach the docker container logs from an attempt [5]
>
>
>
> [1]
> INFO [2021-04-01 23:48:20,984] ({main} Logging.scala[logInfo]:54) -
> Registered signal handler for TERM
>  INFO [2021-04-01 23:48:21,005] ({main} Logging.scala[logInfo]:54) -
> Registered signal handler for HUP
>  INFO [2021-04-01 23:48:21,014] ({main} Logging.scala[logInfo]:54) -
> Registered signal handler for INT
>  INFO [2021-04-01 23:48:22,158] ({main} Logging.scala[logInfo]:54) -
> Changing view acls to: yarn,sandbox
>  INFO [2021-04-01 23:48:22,160] ({main} Logging.scala[logInfo]:54) -
> Changing modify acls to: yarn,sandbox
>  INFO [2021-04-01 23:48:22,161] ({main} Logging.scala[logInfo]:54) -
> Changing view acls groups to:
>  INFO [2021-04-01 23:48:22,162] ({main} Logging.scala[logInfo]:54) -
> Changing modify acls groups to:
>  INFO [2021-04-01 23:48:22,168] ({main} Logging.scala[logInfo]:54) -
> SecurityManager: authentication disabled; ui acls disabled; users  with
> view permissions: Set(yarn, sandbox); groups with view permissions: Set();
> users  with modify permissions: Set(yarn, sandbox); groups with modify
> permissions: Set()
>  INFO [2021-04-01 23:48:25,388] ({main} Logging.scala[logInfo]:54) -
> Preparing Local resources
>  WARN [2021-04-01 23:48:28,111] ({main}
> NativeCodeLoader.java[<clinit>]:62) - Unable to load native-hadoop library
> for your platform... using builtin-java classes where applicable
>  INFO [2021-04-01 23:48:29,004] ({main} Logging.scala[logInfo]:54) -
> ApplicationAttemptId: appattempt_1617228950227_5781_000001
>  INFO [2021-04-01 23:48:29,041] ({main} Logging.scala[logInfo]:54) -
> Starting the user application in a separate Thread
>  INFO [2021-04-01 23:48:29,289] ({main} Logging.scala[logInfo]:54) -
> Waiting for spark context initialization...
>  INFO [2021-04-01 23:48:30,007] ({RegisterThread}
> RemoteInterpreterServer.java[run]:595) - Start registration
>  INFO [2021-04-01 23:48:30,009] ({RemoteInterpreterServer-Thread}
> RemoteInterpreterServer.java[run]:193) - Launching ThriftServer at
> 99.99.99.99:44802
>  INFO [2021-04-01 23:48:31,276] ({RegisterThread}
> RemoteInterpreterServer.java[run]:609) - Registering interpreter process
> ERROR [2021-04-01 23:50:09,531] ({main} Logging.scala[logError]:91) -
> Uncaught exception:
> java.util.concurrent.TimeoutException: Futures timed out after [100000
> milliseconds]
>     at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:223)
>     at
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:227)
>     at
> org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:220)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:469)
>     at org.apache.spark.deploy.yarn.ApplicationMaster.org
> $apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:305)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply$mcV$sp(ApplicationMaster.scala:245)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$3.run(ApplicationMaster.scala:780)
>     at java.security.AccessController.doPrivileged(Native Method)
>     at javax.security.auth.Subject.doAs(Subject.java:422)
>     at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:779)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:244)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:804)
>     at
> org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
>  INFO [2021-04-01 23:50:09,547] ({main} Logging.scala[logInfo]:54) - Final
> app status: FAILED, exitCode: 13, (reason: Uncaught exception:
> java.util.concurrent.TimeoutException: Futures timed out after [100000
> milliseconds]
>
>
> [2]
> version: '3.7'
> services:
>   zeppelin:
>     build: zeppelin-customized
>     ports:
>       - "9999:8080"
>     environment:
>       ZEPPELIN_PORT: 8080
>       ZEPPELIN_JAVA_OPTS: >-
>          -Dspark.driver.memory=1g
>          -Dspark.executor.memory=2g
>       HADOOP_USER_NAME: sandbox
>     volumes:
>       - zeppelindata:/zeppelin/data
>       - zeppelinnotebooks:/zeppelin/notebook
> volumes:
>   zeppelindata:
>   zeppelinnotebooks:
>
>
> [3]
> FROM apache/zeppelin:0.9.0
> # default user is 1000 in zeppelin base..
> USER root
> RUN mkdir /spark && chown 1000:1000 /spark && mkdir /hadoop && chown
> 1000:1000 /hadoop
> USER 1000
>
> # Add Spark
> RUN cd /spark \
>     && wget
> https://artfiles.org/apache.org/spark/spark-2.4.7/spark-2.4.7-bin-hadoop2.7.tgz
> \
>     && tar xf spark-2.4.7-bin-hadoop2.7.tgz \
>     && rm spark-2.4.7-bin-hadoop2.7.tgz \
>     && cd ~
> ENV SPARK_HOME /spark/spark-2.4.7-bin-hadoop2.7
> ENV HADOOP_CONF_DIR /zeppelin/conf
>
> # Add Hadoop
> RUN cd /hadoop \
>     && wget
> https://archive.apache.org/dist/hadoop/common/hadoop-3.0.0/hadoop-3.0.0.tar.gz
> \
>     && tar xf hadoop-3.0.0.tar.gz \
>     && rm hadoop-3.0.0.tar.gz \
>     && cd ~
> ENV HADOOP_HOME /hadoop/hadoop-3.0.0
> ENV HADOOP_INSTALL=$HADOOP_HOME
> ENV HADOOP_MAPRED_HOME=$HADOOP_HOME
> ENV HADOOP_COMMON_HOME=$HADOOP_HOME
> ENV HADOOP_HDFS_HOME=$HADOOP_HOME
> ENV YARN_HOME=$HADOOP_HOME
> ENV HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
> ENV HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib/nativ"
> ENV PATH="${HADOOP_HOME}/bin:${HADOOP_HOME}/sbin:${PATH}"
> ENV USE_HADOOP=true
>
> # Copy over /etc/hadoop/conf from one of the cluster nodes...
> COPY cloudernode/conf/ /zeppelin/conf/
>
>
> [4]
> %sh
> hdfs dfs -ls /user/sandbox
> => prints out properly.
>
>
> [5]
> zeppelin_1  |  WARN [2021-04-01 23:18:36,440] ({SchedulerFactory4}
> SparkInterpreterLauncher.java[buildEnvFromProperties]:221) -
> spark-defaults.conf doesn't exist:
> /spark/spark-2.4.7-bin-hadoop2.7/conf/spark-defaults.conf
> zeppelin_1  |  INFO [2021-04-01 23:18:36,440] ({SchedulerFactory4}
> SparkInterpreterLauncher.java[buildEnvFromProperties]:224) -
> buildEnvFromProperties:
> {PATH=/hadoop/hadoop-3.0.0/bin:/hadoop/hadoop-3.0.0/sbin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin,
> ZEPPELIN_PORT=8080, HADOOP_CONF_DIR=/zeppelin/conf,
> ZEPPELIN_JAVA_OPTS=-Dspark.driver.memory=1g -Dspark.executor.memory=2g,
> ZEPPELIN_LOG_DIR=/opt/zeppelin/logs, MASTER=yarn,
> ZEPPELIN_WAR=/opt/zeppelin/zeppelin-web-0.9.0.war, ZEPPELIN_ENCODING=UTF-8,
> ZEPPELIN_SPARK_CONF= --conf
> spark.yarn.dist.archives=/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr
> --conf spark.yarn.isPython=true --conf spark.executor.instances=2 --conf
> spark.app.name=spark-shared_process --conf
> spark.webui.yarn.useProxy=false --conf spark.driver.cores=1 --conf
> spark.yarn.maxAppAttempts=1 --conf spark.executor.memory=2g --conf
> spark.master=yarn-cluster --conf
> spark.files=/opt/zeppelin/conf/log4j_yarn_cluster.properties --conf
> spark.driver.memory=1g --conf
> spark.jars=/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar,/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar
> --conf spark.executor.cores=1 --conf
> spark.yarn.submit.waitAppCompletion=false,
> JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64, JAVA_OPTS=
> -Dspark.driver.memory=1g -Dspark.executor.memory=2g -Dfile.encoding=UTF-8
> -Xms1024m -Xmx1024m
> -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties
> -Dzeppelin.log.file=/opt/zeppelin/logs/zeppelin--d5ea32f1f431.log,
> INTERPRETER_GROUP_ID=spark-shared_process, Z_VERSION=0.9.0,
> LANG=en_US.UTF-8, JAVA_INTP_OPTS= -Dfile.encoding=UTF-8
> -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties
> -Dlog4j.configurationFile=file:///opt/zeppelin/conf/log4j2.properties,
> PYSPARK_PYTHON=python, HADOOP_USER_NAME=sandbox,
> ZEPPELIN_SPARK_YARN_CLUSTER=true, Z_HOME=/opt/zeppelin,
> SPARK_HOME=/spark/spark-2.4.7-bin-hadoop2.7,
> ZEPPELIN_CONF_DIR=/opt/zeppelin/conf, YARN_HOME=/hadoop/hadoop-3.0.0,
> HADOOP_HDFS_HOME=/hadoop/hadoop-3.0.0,
> ZEPPELIN_RUNNER=/usr/lib/jvm/java-8-openjdk-amd64/bin/java,
> HADOOP_MAPRED_HOME=/hadoop/hadoop-3.0.0, PWD=/opt/zeppelin,
> HADOOP_COMMON_HOME=/hadoop/hadoop-3.0.0,
> HADOOP_INSTALL=/hadoop/hadoop-3.0.0, ZEPPELIN_HOME=/opt/zeppelin,
> LOG_TAG=[ZEPPELIN_0.9.0]:, ZEPPELIN_INTP_MEM=-Xms1024m -Xmx2048m,
> HADOOP_OPTS=-Djava.library.path=/hadoop/hadoop-3.0.0/lib/nativ,
> PYSPARK_DRIVER_PYTHON=python, ZEPPELIN_PID_DIR=/opt/zeppelin/run,
> ZEPPELIN_ANGULAR_WAR=/opt/zeppelin/zeppelin-web-angular-0.9.0.war,
> ZEPPELIN_MEM=-Xms1024m -Xmx1024m, HOSTNAME=d5ea32f1f431,
> LC_ALL=en_US.UTF-8, ZEPPELIN_IDENT_STRING=, PYSPARK_PIN_THREAD=true,
> HADOOP_HOME=/hadoop/hadoop-3.0.0, USE_HADOOP=true,
> HADOOP_COMMON_LIB_NATIVE_DIR=/hadoop/hadoop-3.0.0/lib/native,
> ZEPPELIN_ADDR=0.0.0.0,
> ZEPPELIN_INTERPRETER_REMOTE_RUNNER=bin/interpreter.sh, SHLVL=0,
> HOME=/opt/zeppelin}
> zeppelin_1  |  INFO [2021-04-01 23:18:36,445] ({SchedulerFactory4}
> ProcessLauncher.java[transition]:109) - Process state is transitioned to
> LAUNCHED
> zeppelin_1  |  INFO [2021-04-01 23:18:36,446] ({SchedulerFactory4}
> ProcessLauncher.java[launch]:96) - Process is launched:
> [/opt/zeppelin/bin/interpreter.sh, -d, /opt/zeppelin/interpreter/spark, -c,
> 172.2.0.2, -p, 46781, -r, :, -i, spark-shared_process, -l,
> /opt/zeppelin/local-repo/spark, -g, spark]
> zeppelin_1  |  WARN [2021-04-01 23:20:51,930] ({Exec Default Executor}
> RemoteInterpreterManagedProcess.java[onProcessComplete]:255) - Process is
> exited with exit value 0
> zeppelin_1  |  INFO [2021-04-01 23:20:51,933] ({Exec Default Executor}
> ProcessLauncher.java[transition]:109) - Process state is transitioned to
> COMPLETED
> zeppelin_1  |  INFO [2021-04-01 23:24:06,162] ({qtp418304857-11}
> VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark
> Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
> zeppelin_1  |  INFO [2021-04-01 23:24:15,933] ({qtp418304857-27}
> VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark
> Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
> zeppelin_1  |  WARN [2021-04-01 23:28:36,539] ({SchedulerFactory4}
> NotebookServer.java[onStatusChange]:1928) - Job 20180530-101750_1491737301
> is finished, status: ERROR, exception: null, result: %text
> org.apache.zeppelin.interpreter.InterpreterException: java.io.IOException:
> Fail to launch interpreter process:
> zeppelin_1  | Warning: Master yarn-cluster is deprecated since 2.0. Please
> use master "yarn" with specified deploy mode instead.
> zeppelin_1  | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> zeppelin_1  | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to
> ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new
> application from cluster with 4 NodeManagers
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Verifying our
> application has not requested more than the maximum memory capability of
> the cluster (16400 MB per container)
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM
> container, with 1408 MB memory including 384 MB overhead
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Setting up container
> launch context for our AM
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch
> environment for our AM container
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for
> our AM container
> zeppelin_1  | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars
> nor spark.yarn.archive is set, falling back to uploading libraries under
> SPARK_HOME.
> zeppelin_1  | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip
> zeppelin_1  | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar
> zeppelin_1  | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar
> zeppelin_1  | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar
> zeppelin_1  | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties
> zeppelin_1  | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip
> zeppelin_1  | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip
> zeppelin_1  | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip
> zeppelin_1  | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls to: zeppelin,sandbox
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls to: zeppelin,sandbox
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls groups to:
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls groups to:
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager:
> SecurityManager: authentication disabled; ui acls disabled; users  with
> view permissions: Set(zeppelin, sandbox); groups with view permissions:
> Set(); users  with modify permissions: Set(zeppelin, sandbox); groups with
> modify permissions: Set()
> zeppelin_1  | 21/04/01 23:20:51 INFO yarn.Client: Submitting application
> application_1617315347811_0170 to ResourceManager
> zeppelin_1  | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted
> application application_1617315347811_0170
> zeppelin_1  | 21/04/01 23:20:51 INFO yarn.Client: Application report for
> application_1617315347811_0170 (state: ACCEPTED)
> zeppelin_1  | 21/04/01 23:20:51 INFO yarn.Client:
> zeppelin_1  |    client token: N/A
> zeppelin_1  |    diagnostics: N/A
> zeppelin_1  |    ApplicationMaster host: N/A
> zeppelin_1  |    ApplicationMaster RPC port: -1
> zeppelin_1  |    queue: root.users.sandbox
> zeppelin_1  |    start time: 1617319251597
> zeppelin_1  |    final status: UNDEFINED
> zeppelin_1  |    tracking URL:
> http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/
> zeppelin_1  |    user: sandbox
> zeppelin_1  | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown
> hook called
> zeppelin_1  | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2
> zeppelin_1  | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440
> zeppelin_1  |
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:129)
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getFormType(RemoteInterpreter.java:271)
> zeppelin_1  |   at
> org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:444)
> zeppelin_1  |   at
> org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:72)
> zeppelin_1  |   at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
> zeppelin_1  |   at
> org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:132)
> zeppelin_1  |   at
> org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:182)
> zeppelin_1  |   at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> zeppelin_1  |   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> zeppelin_1  |   at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
> zeppelin_1  |   at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> zeppelin_1  |   at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> zeppelin_1  |   at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> zeppelin_1  |   at java.lang.Thread.run(Thread.java:748)
> zeppelin_1  | Caused by: java.io.IOException: Fail to launch interpreter
> process:
> zeppelin_1  | Warning: Master yarn-cluster is deprecated since 2.0. Please
> use master "yarn" with specified deploy mode instead.
> zeppelin_1  | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> zeppelin_1  | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to
> ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new
> application from cluster with 4 NodeManagers
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Verifying our
> application has not requested more than the maximum memory capability of
> the cluster (16400 MB per container)
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM
> container, with 1408 MB memory including 384 MB overhead
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Setting up container
> launch context for our AM
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch
> environment for our AM container
> zeppelin_1  | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for
> our AM container
> zeppelin_1  | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars
> nor spark.yarn.archive is set, falling back to uploading libraries under
> SPARK_HOME.
> zeppelin_1  | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip
> zeppelin_1  | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar
> zeppelin_1  | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar
> zeppelin_1  | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar
> zeppelin_1  | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties
> zeppelin_1  | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip
> zeppelin_1  | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip
> zeppelin_1  | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip
> zeppelin_1  | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls to: zeppelin,sandbox
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls to: zeppelin,sandbox
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls groups to:
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls groups to:
> zeppelin_1  | 21/04/01 23:20:46 INFO spark.SecurityManager:
> SecurityManager: authentication disabled; ui acls disabled; users  with
> view permissions: Set(zeppelin, sandbox); groups with view permissions:
> Set(); users  with modify permissions: Set(zeppelin, sandbox); groups with
> modify permissions: Set()
> zeppelin_1  | 21/04/01 23:20:51 INFO yarn.Client: Submitting application
> application_1617315347811_0170 to ResourceManager
> zeppelin_1  | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted
> application application_1617315347811_0170
> zeppelin_1  | 21/04/01 23:20:51 INFO yarn.Client: Application report for
> application_1617315347811_0170 (state: ACCEPTED)
> zeppelin_1  | 21/04/01 23:20:51 INFO yarn.Client:
> zeppelin_1  |    client token: N/A
> zeppelin_1  |    diagnostics: N/A
> zeppelin_1  |    ApplicationMaster host: N/A
> zeppelin_1  |    ApplicationMaster RPC port: -1
> zeppelin_1  |    queue: root.users.sandbox
> zeppelin_1  |    start time: 1617319251597
> zeppelin_1  |    final status: UNDEFINED
> zeppelin_1  |    tracking URL:
> http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/
> zeppelin_1  |    user: sandbox
> zeppelin_1  | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown
> hook called
> zeppelin_1  | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440
> zeppelin_1  | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2
> zeppelin_1  |
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterManagedProcess.start(RemoteInterpreterManagedProcess.java:126)
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.ManagedInterpreterGroup.getOrCreateInterpreterProcess(ManagedInterpreterGroup.java:68)
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getOrCreateInterpreterProcess(RemoteInterpreter.java:104)
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.internal_create(RemoteInterpreter.java:154)
> zeppelin_1  |   at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:126)
> zeppelin_1  |   ... 13 more
> zeppelin_1  |
> zeppelin_1  |  INFO [2021-04-01 23:28:36,542] ({SchedulerFactory4}
> VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark
> Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
>
>

-- 
Best Regards

Jeff Zhang