You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@zeppelin.apache.org by Theo Diefenthal <th...@scoop-software.de> on 2021/04/01 23:48:40 UTC
Connect Dockerized Zeppelin to dev Cloudera 6.3.1 cluster [Promise
timeout when Registering interpreter process]
Hi there,
I want to achieve the following usecase: Start Zeppelin 0.9.0 (in docker) on my local dev machine but let the Spark jobs in the notebook run on a remote cluster via YARN.
For a few hours already, I try to setup that environment with my companies Cloudera CDH 6.3.1 development cluster. That cluster is unsecured (despite that it can only be reached when connected to VPN). With a lot of trial and error I finally achieved a successful connection from my dockerized Zeppelin to the cluster. This means that when I start running a spark cell in Zeppelin, I can see a new application in YARN on the cluster-side [named spark-shared_process] . However, eventually the execution of the cell will fail with the following stack trace in the yarn application [1]. I have no idea where this timeout could potentially come from and I'd be happy if you could help me out here. In the said VPN to the dev cluster, there are no connection restrictions like firewalls or stuff like that engaged. The cell I run is the first one in "3. Spark SQL (Scala)" Zeppelin quick start notebooks with title "Create Dataset/DataFrame via SparkSession".
For reference, I also attach my docker-compose file [2] and my Dockerfile for building Zeppelin with Spark and Hadoop [3] (Note that I add hadoop conf files into the image because I'd like to distribute the image as ready-to-run for the other people in my project without needing them to copy over the hadoop conf files). After start of the container, I further change the interpreter settings by setting yarn-cluster in %spark interpreter settings and also set zeppelin.interpreter.connect.timeout to 600.000.
Best regards
Theo
PS: HDFS in general seems to work well. [4]
PPS: I also attach the docker container logs from an attempt [5]
[1]
INFO [2021-04-01 23:48:20,984] ({main} Logging.scala[logInfo]:54) - Registered signal handler for TERM
INFO [2021-04-01 23:48:21,005] ({main} Logging.scala[logInfo]:54) - Registered signal handler for HUP
INFO [2021-04-01 23:48:21,014] ({main} Logging.scala[logInfo]:54) - Registered signal handler for INT
INFO [2021-04-01 23:48:22,158] ({main} Logging.scala[logInfo]:54) - Changing view acls to: yarn,sandbox
INFO [2021-04-01 23:48:22,160] ({main} Logging.scala[logInfo]:54) - Changing modify acls to: yarn,sandbox
INFO [2021-04-01 23:48:22,161] ({main} Logging.scala[logInfo]:54) - Changing view acls groups to:
INFO [2021-04-01 23:48:22,162] ({main} Logging.scala[logInfo]:54) - Changing modify acls groups to:
INFO [2021-04-01 23:48:22,168] ({main} Logging.scala[logInfo]:54) - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(yarn, sandbox); groups with view permissions: Set(); users with modify permissions: Set(yarn, sandbox); groups with modify permissions: Set()
INFO [2021-04-01 23:48:25,388] ({main} Logging.scala[logInfo]:54) - Preparing Local resources
WARN [2021-04-01 23:48:28,111] ({main} NativeCodeLoader.java[<clinit>]:62) - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
INFO [2021-04-01 23:48:29,004] ({main} Logging.scala[logInfo]:54) - ApplicationAttemptId: appattempt_1617228950227_5781_000001
INFO [2021-04-01 23:48:29,041] ({main} Logging.scala[logInfo]:54) - Starting the user application in a separate Thread
INFO [2021-04-01 23:48:29,289] ({main} Logging.scala[logInfo]:54) - Waiting for spark context initialization...
INFO [2021-04-01 23:48:30,007] ({RegisterThread} RemoteInterpreterServer.java[run]:595) - Start registration
INFO [2021-04-01 23:48:30,009] ({RemoteInterpreterServer-Thread} RemoteInterpreterServer.java[run]:193) - Launching ThriftServer at 99.99.99.99:44802
INFO [2021-04-01 23:48:31,276] ({RegisterThread} RemoteInterpreterServer.java[run]:609) - Registering interpreter process
ERROR [2021-04-01 23:50:09,531] ({main} Logging.scala[logError]:91) - Uncaught exception:
java.util.concurrent.TimeoutException: Futures timed out after [100000 milliseconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:223)
at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:227)
at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:220)
at org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:469)
at org.apache.spark.deploy.yarn.ApplicationMaster.org$apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:305)
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply$mcV$sp(ApplicationMaster.scala:245)
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245)
at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245)
at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$3.run(ApplicationMaster.scala:780)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
at org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:779)
at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:244)
at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:804)
at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
INFO [2021-04-01 23:50:09,547] ({main} Logging.scala[logInfo]:54) - Final app status: FAILED, exitCode: 13, (reason: Uncaught exception: java.util.concurrent.TimeoutException: Futures timed out after [100000 milliseconds]
[2]
version: '3.7'
services:
zeppelin:
build: zeppelin-customized
ports:
- "9999:8080"
environment:
ZEPPELIN_PORT: 8080
ZEPPELIN_JAVA_OPTS: >-
-Dspark.driver.memory=1g
-Dspark.executor.memory=2g
HADOOP_USER_NAME: sandbox
volumes:
- zeppelindata:/zeppelin/data
- zeppelinnotebooks:/zeppelin/notebook
volumes:
zeppelindata:
zeppelinnotebooks:
[3]
FROM apache/zeppelin:0.9.0
# default user is 1000 in zeppelin base..
USER root
RUN mkdir /spark && chown 1000:1000 /spark && mkdir /hadoop && chown 1000:1000 /hadoop
USER 1000
# Add Spark
RUN cd /spark \
&& wget https://artfiles.org/apache.org/spark/spark-2.4.7/spark-2.4.7-bin-hadoop2.7.tgz \
&& tar xf spark-2.4.7-bin-hadoop2.7.tgz \
&& rm spark-2.4.7-bin-hadoop2.7.tgz \
&& cd ~
ENV SPARK_HOME /spark/spark-2.4.7-bin-hadoop2.7
ENV HADOOP_CONF_DIR /zeppelin/conf
# Add Hadoop
RUN cd /hadoop \
&& wget https://archive.apache.org/dist/hadoop/common/hadoop-3.0.0/hadoop-3.0.0.tar.gz \
&& tar xf hadoop-3.0.0.tar.gz \
&& rm hadoop-3.0.0.tar.gz \
&& cd ~
ENV HADOOP_HOME /hadoop/hadoop-3.0.0
ENV HADOOP_INSTALL=$HADOOP_HOME
ENV HADOOP_MAPRED_HOME=$HADOOP_HOME
ENV HADOOP_COMMON_HOME=$HADOOP_HOME
ENV HADOOP_HDFS_HOME=$HADOOP_HOME
ENV YARN_HOME=$HADOOP_HOME
ENV HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
ENV HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib/nativ"
ENV PATH="${HADOOP_HOME}/bin:${HADOOP_HOME}/sbin:${PATH}"
ENV USE_HADOOP=true
# Copy over /etc/hadoop/conf from one of the cluster nodes...
COPY cloudernode/conf/ /zeppelin/conf/
[4]
%sh
hdfs dfs -ls /user/sandbox
=> prints out properly.
[5]
zeppelin_1 | WARN [2021-04-01 23:18:36,440] ({SchedulerFactory4} SparkInterpreterLauncher.java[buildEnvFromProperties]:221) - spark-defaults.conf doesn't exist: /spark/spark-2.4.7-bin-hadoop2.7/conf/spark-defaults.conf
zeppelin_1 | INFO [2021-04-01 23:18:36,440] ({SchedulerFactory4} SparkInterpreterLauncher.java[buildEnvFromProperties]:224) - buildEnvFromProperties: {PATH=/hadoop/hadoop-3.0.0/bin:/hadoop/hadoop-3.0.0/sbin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin, ZEPPELIN_PORT=8080, HADOOP_CONF_DIR=/zeppelin/conf, ZEPPELIN_JAVA_OPTS=-Dspark.driver.memory=1g -Dspark.executor.memory=2g, ZEPPELIN_LOG_DIR=/opt/zeppelin/logs, MASTER=yarn, ZEPPELIN_WAR=/opt/zeppelin/zeppelin-web-0.9.0.war, ZEPPELIN_ENCODING=UTF-8, ZEPPELIN_SPARK_CONF= --conf spark.yarn.dist.archives=/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr --conf spark.yarn.isPython=true --conf spark.executor.instances=2 --conf spark.app.name=spark-shared_process --conf spark.webui.yarn.useProxy=false --conf spark.driver.cores=1 --conf spark.yarn.maxAppAttempts=1 --conf spark.executor.memory=2g --conf spark.master=yarn-cluster --conf spark.files=/opt/zeppelin/conf/log4j_yarn_cluster.properties --conf spark.driver.memory=1g --conf spark.jars=/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar,/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar --conf spark.executor.cores=1 --conf spark.yarn.submit.waitAppCompletion=false, JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64, JAVA_OPTS= -Dspark.driver.memory=1g -Dspark.executor.memory=2g -Dfile.encoding=UTF-8 -Xms1024m -Xmx1024m -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties -Dzeppelin.log.file=/opt/zeppelin/logs/zeppelin--d5ea32f1f431.log, INTERPRETER_GROUP_ID=spark-shared_process, Z_VERSION=0.9.0, LANG=en_US.UTF-8, JAVA_INTP_OPTS= -Dfile.encoding=UTF-8 -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties -Dlog4j.configurationFile=file:///opt/zeppelin/conf/log4j2.properties, PYSPARK_PYTHON=python, HADOOP_USER_NAME=sandbox, ZEPPELIN_SPARK_YARN_CLUSTER=true, Z_HOME=/opt/zeppelin, SPARK_HOME=/spark/spark-2.4.7-bin-hadoop2.7, ZEPPELIN_CONF_DIR=/opt/zeppelin/conf, YARN_HOME=/hadoop/hadoop-3.0.0, HADOOP_HDFS_HOME=/hadoop/hadoop-3.0.0, ZEPPELIN_RUNNER=/usr/lib/jvm/java-8-openjdk-amd64/bin/java, HADOOP_MAPRED_HOME=/hadoop/hadoop-3.0.0, PWD=/opt/zeppelin, HADOOP_COMMON_HOME=/hadoop/hadoop-3.0.0, HADOOP_INSTALL=/hadoop/hadoop-3.0.0, ZEPPELIN_HOME=/opt/zeppelin, LOG_TAG=[ZEPPELIN_0.9.0]:, ZEPPELIN_INTP_MEM=-Xms1024m -Xmx2048m, HADOOP_OPTS=-Djava.library.path=/hadoop/hadoop-3.0.0/lib/nativ, PYSPARK_DRIVER_PYTHON=python, ZEPPELIN_PID_DIR=/opt/zeppelin/run, ZEPPELIN_ANGULAR_WAR=/opt/zeppelin/zeppelin-web-angular-0.9.0.war, ZEPPELIN_MEM=-Xms1024m -Xmx1024m, HOSTNAME=d5ea32f1f431, LC_ALL=en_US.UTF-8, ZEPPELIN_IDENT_STRING=, PYSPARK_PIN_THREAD=true, HADOOP_HOME=/hadoop/hadoop-3.0.0, USE_HADOOP=true, HADOOP_COMMON_LIB_NATIVE_DIR=/hadoop/hadoop-3.0.0/lib/native, ZEPPELIN_ADDR=0.0.0.0, ZEPPELIN_INTERPRETER_REMOTE_RUNNER=bin/interpreter.sh, SHLVL=0, HOME=/opt/zeppelin}
zeppelin_1 | INFO [2021-04-01 23:18:36,445] ({SchedulerFactory4} ProcessLauncher.java[transition]:109) - Process state is transitioned to LAUNCHED
zeppelin_1 | INFO [2021-04-01 23:18:36,446] ({SchedulerFactory4} ProcessLauncher.java[launch]:96) - Process is launched: [/opt/zeppelin/bin/interpreter.sh, -d, /opt/zeppelin/interpreter/spark, -c, 172.2.0.2, -p, 46781, -r, :, -i, spark-shared_process, -l, /opt/zeppelin/local-repo/spark, -g, spark]
zeppelin_1 | WARN [2021-04-01 23:20:51,930] ({Exec Default Executor} RemoteInterpreterManagedProcess.java[onProcessComplete]:255) - Process is exited with exit value 0
zeppelin_1 | INFO [2021-04-01 23:20:51,933] ({Exec Default Executor} ProcessLauncher.java[transition]:109) - Process state is transitioned to COMPLETED
zeppelin_1 | INFO [2021-04-01 23:24:06,162] ({qtp418304857-11} VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
zeppelin_1 | INFO [2021-04-01 23:24:15,933] ({qtp418304857-27} VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
zeppelin_1 | WARN [2021-04-01 23:28:36,539] ({SchedulerFactory4} NotebookServer.java[onStatusChange]:1928) - Job 20180530-101750_1491737301 is finished, status: ERROR, exception: null, result: %text org.apache.zeppelin.interpreter.InterpreterException: java.io.IOException: Fail to launch interpreter process:
zeppelin_1 | Warning: Master yarn-cluster is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead.
zeppelin_1 | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
zeppelin_1 | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (16400 MB per container)
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM container, with 1408 MB memory including 384 MB overhead
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up container launch context for our AM
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch environment for our AM container
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for our AM container
zeppelin_1 | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME.
zeppelin_1 | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip
zeppelin_1 | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar
zeppelin_1 | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties
zeppelin_1 | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip
zeppelin_1 | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip
zeppelin_1 | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip
zeppelin_1 | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls to: zeppelin,sandbox
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls to: zeppelin,sandbox
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls groups to:
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls groups to:
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin, sandbox); groups with view permissions: Set(); users with modify permissions: Set(zeppelin, sandbox); groups with modify permissions: Set()
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Submitting application application_1617315347811_0170 to ResourceManager
zeppelin_1 | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted application application_1617315347811_0170
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Application report for application_1617315347811_0170 (state: ACCEPTED)
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client:
zeppelin_1 | client token: N/A
zeppelin_1 | diagnostics: N/A
zeppelin_1 | ApplicationMaster host: N/A
zeppelin_1 | ApplicationMaster RPC port: -1
zeppelin_1 | queue: root.users.sandbox
zeppelin_1 | start time: 1617319251597
zeppelin_1 | final status: UNDEFINED
zeppelin_1 | tracking URL: http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/
zeppelin_1 | user: sandbox
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown hook called
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440
zeppelin_1 |
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:129)
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getFormType(RemoteInterpreter.java:271)
zeppelin_1 | at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:444)
zeppelin_1 | at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:72)
zeppelin_1 | at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
zeppelin_1 | at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:132)
zeppelin_1 | at org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:182)
zeppelin_1 | at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
zeppelin_1 | at java.util.concurrent.FutureTask.run(FutureTask.java:266)
zeppelin_1 | at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
zeppelin_1 | at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
zeppelin_1 | at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
zeppelin_1 | at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
zeppelin_1 | at java.lang.Thread.run(Thread.java:748)
zeppelin_1 | Caused by: java.io.IOException: Fail to launch interpreter process:
zeppelin_1 | Warning: Master yarn-cluster is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead.
zeppelin_1 | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
zeppelin_1 | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new application from cluster with 4 NodeManagers
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (16400 MB per container)
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM container, with 1408 MB memory including 384 MB overhead
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up container launch context for our AM
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch environment for our AM container
zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for our AM container
zeppelin_1 | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME.
zeppelin_1 | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip
zeppelin_1 | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar
zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar
zeppelin_1 | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties
zeppelin_1 | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip
zeppelin_1 | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip
zeppelin_1 | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip
zeppelin_1 | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip -> hdfs://machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls to: zeppelin,sandbox
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls to: zeppelin,sandbox
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view acls groups to:
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing modify acls groups to:
zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin, sandbox); groups with view permissions: Set(); users with modify permissions: Set(zeppelin, sandbox); groups with modify permissions: Set()
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Submitting application application_1617315347811_0170 to ResourceManager
zeppelin_1 | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted application application_1617315347811_0170
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Application report for application_1617315347811_0170 (state: ACCEPTED)
zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client:
zeppelin_1 | client token: N/A
zeppelin_1 | diagnostics: N/A
zeppelin_1 | ApplicationMaster host: N/A
zeppelin_1 | ApplicationMaster RPC port: -1
zeppelin_1 | queue: root.users.sandbox
zeppelin_1 | start time: 1617319251597
zeppelin_1 | final status: UNDEFINED
zeppelin_1 | tracking URL: http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/
zeppelin_1 | user: sandbox
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown hook called
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440
zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2
zeppelin_1 |
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreterManagedProcess.start(RemoteInterpreterManagedProcess.java:126)
zeppelin_1 | at org.apache.zeppelin.interpreter.ManagedInterpreterGroup.getOrCreateInterpreterProcess(ManagedInterpreterGroup.java:68)
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getOrCreateInterpreterProcess(RemoteInterpreter.java:104)
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.internal_create(RemoteInterpreter.java:154)
zeppelin_1 | at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:126)
zeppelin_1 | ... 13 more
zeppelin_1 |
zeppelin_1 | INFO [2021-04-01 23:28:36,542] ({SchedulerFactory4} VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
Re: Connect Dockerized Zeppelin to dev Cloudera 6.3.1 cluster
[Promise timeout when Registering interpreter process]
Posted by Jeff Zhang <zj...@gmail.com>.
Most likely it is due to network issue, the connection between spark driver
(run in yarn container in yarn cluster mode) and zeppelin server is
bidirectional. It looks like it is due to your spark driver unable to
connect to the zeppelin server.
Theo Diefenthal <th...@scoop-software.de> 于2021年4月2日周五 上午7:48写道:
> Hi there,
>
> I want to achieve the following usecase: Start Zeppelin 0.9.0 (in docker)
> on my local dev machine but let the Spark jobs in the notebook run on a
> remote cluster via YARN.
>
> For a few hours already, I try to setup that environment with my companies
> Cloudera CDH 6.3.1 development cluster. That cluster is unsecured (despite
> that it can only be reached when connected to VPN). With a lot of trial and
> error I finally achieved a successful connection from my dockerized
> Zeppelin to the cluster. This means that when I start running a spark cell
> in Zeppelin, I can see a new application in YARN on the cluster-side [named
> spark-shared_process]. However, eventually the execution of the cell will
> fail with the following stack trace in the yarn application [1]. I have no
> idea where this timeout could potentially come from and I'd be happy if you
> could help me out here. In the said VPN to the dev cluster, there are no
> connection restrictions like firewalls or stuff like that engaged. The cell
> I run is the first one in "3. Spark SQL (Scala)" Zeppelin quick start
> notebooks with title "Create Dataset/DataFrame via SparkSession".
>
> For reference, I also attach my docker-compose file [2] and my Dockerfile
> for building Zeppelin with Spark and Hadoop [3] (Note that I add hadoop
> conf files into the image because I'd like to distribute the image as
> ready-to-run for the other people in my project without needing them to
> copy over the hadoop conf files). After start of the container, I further
> change the interpreter settings by setting yarn-cluster in %spark
> interpreter settings and also set zeppelin.interpreter.connect.timeout to
> 600.000.
>
> Best regards
> Theo
>
> PS: HDFS in general seems to work well. [4]
> PPS: I also attach the docker container logs from an attempt [5]
>
>
>
> [1]
> INFO [2021-04-01 23:48:20,984] ({main} Logging.scala[logInfo]:54) -
> Registered signal handler for TERM
> INFO [2021-04-01 23:48:21,005] ({main} Logging.scala[logInfo]:54) -
> Registered signal handler for HUP
> INFO [2021-04-01 23:48:21,014] ({main} Logging.scala[logInfo]:54) -
> Registered signal handler for INT
> INFO [2021-04-01 23:48:22,158] ({main} Logging.scala[logInfo]:54) -
> Changing view acls to: yarn,sandbox
> INFO [2021-04-01 23:48:22,160] ({main} Logging.scala[logInfo]:54) -
> Changing modify acls to: yarn,sandbox
> INFO [2021-04-01 23:48:22,161] ({main} Logging.scala[logInfo]:54) -
> Changing view acls groups to:
> INFO [2021-04-01 23:48:22,162] ({main} Logging.scala[logInfo]:54) -
> Changing modify acls groups to:
> INFO [2021-04-01 23:48:22,168] ({main} Logging.scala[logInfo]:54) -
> SecurityManager: authentication disabled; ui acls disabled; users with
> view permissions: Set(yarn, sandbox); groups with view permissions: Set();
> users with modify permissions: Set(yarn, sandbox); groups with modify
> permissions: Set()
> INFO [2021-04-01 23:48:25,388] ({main} Logging.scala[logInfo]:54) -
> Preparing Local resources
> WARN [2021-04-01 23:48:28,111] ({main}
> NativeCodeLoader.java[<clinit>]:62) - Unable to load native-hadoop library
> for your platform... using builtin-java classes where applicable
> INFO [2021-04-01 23:48:29,004] ({main} Logging.scala[logInfo]:54) -
> ApplicationAttemptId: appattempt_1617228950227_5781_000001
> INFO [2021-04-01 23:48:29,041] ({main} Logging.scala[logInfo]:54) -
> Starting the user application in a separate Thread
> INFO [2021-04-01 23:48:29,289] ({main} Logging.scala[logInfo]:54) -
> Waiting for spark context initialization...
> INFO [2021-04-01 23:48:30,007] ({RegisterThread}
> RemoteInterpreterServer.java[run]:595) - Start registration
> INFO [2021-04-01 23:48:30,009] ({RemoteInterpreterServer-Thread}
> RemoteInterpreterServer.java[run]:193) - Launching ThriftServer at
> 99.99.99.99:44802
> INFO [2021-04-01 23:48:31,276] ({RegisterThread}
> RemoteInterpreterServer.java[run]:609) - Registering interpreter process
> ERROR [2021-04-01 23:50:09,531] ({main} Logging.scala[logError]:91) -
> Uncaught exception:
> java.util.concurrent.TimeoutException: Futures timed out after [100000
> milliseconds]
> at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:223)
> at
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:227)
> at
> org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:220)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:469)
> at org.apache.spark.deploy.yarn.ApplicationMaster.org
> $apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:305)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply$mcV$sp(ApplicationMaster.scala:245)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$1.apply(ApplicationMaster.scala:245)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$3.run(ApplicationMaster.scala:780)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:779)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:244)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:804)
> at
> org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
> INFO [2021-04-01 23:50:09,547] ({main} Logging.scala[logInfo]:54) - Final
> app status: FAILED, exitCode: 13, (reason: Uncaught exception:
> java.util.concurrent.TimeoutException: Futures timed out after [100000
> milliseconds]
>
>
> [2]
> version: '3.7'
> services:
> zeppelin:
> build: zeppelin-customized
> ports:
> - "9999:8080"
> environment:
> ZEPPELIN_PORT: 8080
> ZEPPELIN_JAVA_OPTS: >-
> -Dspark.driver.memory=1g
> -Dspark.executor.memory=2g
> HADOOP_USER_NAME: sandbox
> volumes:
> - zeppelindata:/zeppelin/data
> - zeppelinnotebooks:/zeppelin/notebook
> volumes:
> zeppelindata:
> zeppelinnotebooks:
>
>
> [3]
> FROM apache/zeppelin:0.9.0
> # default user is 1000 in zeppelin base..
> USER root
> RUN mkdir /spark && chown 1000:1000 /spark && mkdir /hadoop && chown
> 1000:1000 /hadoop
> USER 1000
>
> # Add Spark
> RUN cd /spark \
> && wget
> https://artfiles.org/apache.org/spark/spark-2.4.7/spark-2.4.7-bin-hadoop2.7.tgz
> \
> && tar xf spark-2.4.7-bin-hadoop2.7.tgz \
> && rm spark-2.4.7-bin-hadoop2.7.tgz \
> && cd ~
> ENV SPARK_HOME /spark/spark-2.4.7-bin-hadoop2.7
> ENV HADOOP_CONF_DIR /zeppelin/conf
>
> # Add Hadoop
> RUN cd /hadoop \
> && wget
> https://archive.apache.org/dist/hadoop/common/hadoop-3.0.0/hadoop-3.0.0.tar.gz
> \
> && tar xf hadoop-3.0.0.tar.gz \
> && rm hadoop-3.0.0.tar.gz \
> && cd ~
> ENV HADOOP_HOME /hadoop/hadoop-3.0.0
> ENV HADOOP_INSTALL=$HADOOP_HOME
> ENV HADOOP_MAPRED_HOME=$HADOOP_HOME
> ENV HADOOP_COMMON_HOME=$HADOOP_HOME
> ENV HADOOP_HDFS_HOME=$HADOOP_HOME
> ENV YARN_HOME=$HADOOP_HOME
> ENV HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
> ENV HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib/nativ"
> ENV PATH="${HADOOP_HOME}/bin:${HADOOP_HOME}/sbin:${PATH}"
> ENV USE_HADOOP=true
>
> # Copy over /etc/hadoop/conf from one of the cluster nodes...
> COPY cloudernode/conf/ /zeppelin/conf/
>
>
> [4]
> %sh
> hdfs dfs -ls /user/sandbox
> => prints out properly.
>
>
> [5]
> zeppelin_1 | WARN [2021-04-01 23:18:36,440] ({SchedulerFactory4}
> SparkInterpreterLauncher.java[buildEnvFromProperties]:221) -
> spark-defaults.conf doesn't exist:
> /spark/spark-2.4.7-bin-hadoop2.7/conf/spark-defaults.conf
> zeppelin_1 | INFO [2021-04-01 23:18:36,440] ({SchedulerFactory4}
> SparkInterpreterLauncher.java[buildEnvFromProperties]:224) -
> buildEnvFromProperties:
> {PATH=/hadoop/hadoop-3.0.0/bin:/hadoop/hadoop-3.0.0/sbin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin,
> ZEPPELIN_PORT=8080, HADOOP_CONF_DIR=/zeppelin/conf,
> ZEPPELIN_JAVA_OPTS=-Dspark.driver.memory=1g -Dspark.executor.memory=2g,
> ZEPPELIN_LOG_DIR=/opt/zeppelin/logs, MASTER=yarn,
> ZEPPELIN_WAR=/opt/zeppelin/zeppelin-web-0.9.0.war, ZEPPELIN_ENCODING=UTF-8,
> ZEPPELIN_SPARK_CONF= --conf
> spark.yarn.dist.archives=/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr
> --conf spark.yarn.isPython=true --conf spark.executor.instances=2 --conf
> spark.app.name=spark-shared_process --conf
> spark.webui.yarn.useProxy=false --conf spark.driver.cores=1 --conf
> spark.yarn.maxAppAttempts=1 --conf spark.executor.memory=2g --conf
> spark.master=yarn-cluster --conf
> spark.files=/opt/zeppelin/conf/log4j_yarn_cluster.properties --conf
> spark.driver.memory=1g --conf
> spark.jars=/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar,/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar
> --conf spark.executor.cores=1 --conf
> spark.yarn.submit.waitAppCompletion=false,
> JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64, JAVA_OPTS=
> -Dspark.driver.memory=1g -Dspark.executor.memory=2g -Dfile.encoding=UTF-8
> -Xms1024m -Xmx1024m
> -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties
> -Dzeppelin.log.file=/opt/zeppelin/logs/zeppelin--d5ea32f1f431.log,
> INTERPRETER_GROUP_ID=spark-shared_process, Z_VERSION=0.9.0,
> LANG=en_US.UTF-8, JAVA_INTP_OPTS= -Dfile.encoding=UTF-8
> -Dlog4j.configuration=file:///opt/zeppelin/conf/log4j.properties
> -Dlog4j.configurationFile=file:///opt/zeppelin/conf/log4j2.properties,
> PYSPARK_PYTHON=python, HADOOP_USER_NAME=sandbox,
> ZEPPELIN_SPARK_YARN_CLUSTER=true, Z_HOME=/opt/zeppelin,
> SPARK_HOME=/spark/spark-2.4.7-bin-hadoop2.7,
> ZEPPELIN_CONF_DIR=/opt/zeppelin/conf, YARN_HOME=/hadoop/hadoop-3.0.0,
> HADOOP_HDFS_HOME=/hadoop/hadoop-3.0.0,
> ZEPPELIN_RUNNER=/usr/lib/jvm/java-8-openjdk-amd64/bin/java,
> HADOOP_MAPRED_HOME=/hadoop/hadoop-3.0.0, PWD=/opt/zeppelin,
> HADOOP_COMMON_HOME=/hadoop/hadoop-3.0.0,
> HADOOP_INSTALL=/hadoop/hadoop-3.0.0, ZEPPELIN_HOME=/opt/zeppelin,
> LOG_TAG=[ZEPPELIN_0.9.0]:, ZEPPELIN_INTP_MEM=-Xms1024m -Xmx2048m,
> HADOOP_OPTS=-Djava.library.path=/hadoop/hadoop-3.0.0/lib/nativ,
> PYSPARK_DRIVER_PYTHON=python, ZEPPELIN_PID_DIR=/opt/zeppelin/run,
> ZEPPELIN_ANGULAR_WAR=/opt/zeppelin/zeppelin-web-angular-0.9.0.war,
> ZEPPELIN_MEM=-Xms1024m -Xmx1024m, HOSTNAME=d5ea32f1f431,
> LC_ALL=en_US.UTF-8, ZEPPELIN_IDENT_STRING=, PYSPARK_PIN_THREAD=true,
> HADOOP_HOME=/hadoop/hadoop-3.0.0, USE_HADOOP=true,
> HADOOP_COMMON_LIB_NATIVE_DIR=/hadoop/hadoop-3.0.0/lib/native,
> ZEPPELIN_ADDR=0.0.0.0,
> ZEPPELIN_INTERPRETER_REMOTE_RUNNER=bin/interpreter.sh, SHLVL=0,
> HOME=/opt/zeppelin}
> zeppelin_1 | INFO [2021-04-01 23:18:36,445] ({SchedulerFactory4}
> ProcessLauncher.java[transition]:109) - Process state is transitioned to
> LAUNCHED
> zeppelin_1 | INFO [2021-04-01 23:18:36,446] ({SchedulerFactory4}
> ProcessLauncher.java[launch]:96) - Process is launched:
> [/opt/zeppelin/bin/interpreter.sh, -d, /opt/zeppelin/interpreter/spark, -c,
> 172.2.0.2, -p, 46781, -r, :, -i, spark-shared_process, -l,
> /opt/zeppelin/local-repo/spark, -g, spark]
> zeppelin_1 | WARN [2021-04-01 23:20:51,930] ({Exec Default Executor}
> RemoteInterpreterManagedProcess.java[onProcessComplete]:255) - Process is
> exited with exit value 0
> zeppelin_1 | INFO [2021-04-01 23:20:51,933] ({Exec Default Executor}
> ProcessLauncher.java[transition]:109) - Process state is transitioned to
> COMPLETED
> zeppelin_1 | INFO [2021-04-01 23:24:06,162] ({qtp418304857-11}
> VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark
> Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
> zeppelin_1 | INFO [2021-04-01 23:24:15,933] ({qtp418304857-27}
> VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark
> Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
> zeppelin_1 | WARN [2021-04-01 23:28:36,539] ({SchedulerFactory4}
> NotebookServer.java[onStatusChange]:1928) - Job 20180530-101750_1491737301
> is finished, status: ERROR, exception: null, result: %text
> org.apache.zeppelin.interpreter.InterpreterException: java.io.IOException:
> Fail to launch interpreter process:
> zeppelin_1 | Warning: Master yarn-cluster is deprecated since 2.0. Please
> use master "yarn" with specified deploy mode instead.
> zeppelin_1 | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> zeppelin_1 | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to
> ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new
> application from cluster with 4 NodeManagers
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Verifying our
> application has not requested more than the maximum memory capability of
> the cluster (16400 MB per container)
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM
> container, with 1408 MB memory including 384 MB overhead
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up container
> launch context for our AM
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch
> environment for our AM container
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for
> our AM container
> zeppelin_1 | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars
> nor spark.yarn.archive is set, falling back to uploading libraries under
> SPARK_HOME.
> zeppelin_1 | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip
> zeppelin_1 | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar
> zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar
> zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar
> zeppelin_1 | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties
> zeppelin_1 | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip
> zeppelin_1 | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip
> zeppelin_1 | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip
> zeppelin_1 | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls to: zeppelin,sandbox
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls to: zeppelin,sandbox
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls groups to:
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls groups to:
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager:
> SecurityManager: authentication disabled; ui acls disabled; users with
> view permissions: Set(zeppelin, sandbox); groups with view permissions:
> Set(); users with modify permissions: Set(zeppelin, sandbox); groups with
> modify permissions: Set()
> zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Submitting application
> application_1617315347811_0170 to ResourceManager
> zeppelin_1 | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted
> application application_1617315347811_0170
> zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Application report for
> application_1617315347811_0170 (state: ACCEPTED)
> zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client:
> zeppelin_1 | client token: N/A
> zeppelin_1 | diagnostics: N/A
> zeppelin_1 | ApplicationMaster host: N/A
> zeppelin_1 | ApplicationMaster RPC port: -1
> zeppelin_1 | queue: root.users.sandbox
> zeppelin_1 | start time: 1617319251597
> zeppelin_1 | final status: UNDEFINED
> zeppelin_1 | tracking URL:
> http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/
> zeppelin_1 | user: sandbox
> zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown
> hook called
> zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2
> zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440
> zeppelin_1 |
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:129)
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getFormType(RemoteInterpreter.java:271)
> zeppelin_1 | at
> org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:444)
> zeppelin_1 | at
> org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:72)
> zeppelin_1 | at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
> zeppelin_1 | at
> org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:132)
> zeppelin_1 | at
> org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:182)
> zeppelin_1 | at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> zeppelin_1 | at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> zeppelin_1 | at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
> zeppelin_1 | at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> zeppelin_1 | at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> zeppelin_1 | at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> zeppelin_1 | at java.lang.Thread.run(Thread.java:748)
> zeppelin_1 | Caused by: java.io.IOException: Fail to launch interpreter
> process:
> zeppelin_1 | Warning: Master yarn-cluster is deprecated since 2.0. Please
> use master "yarn" with specified deploy mode instead.
> zeppelin_1 | 21/04/01 23:18:44 WARN util.NativeCodeLoader: Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> zeppelin_1 | 21/04/01 23:18:44 INFO client.RMProxy: Connecting to
> ResourceManager at machine1.REMOVEDDOMAIN.de/99.99.99.99:8032
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Requesting a new
> application from cluster with 4 NodeManagers
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Verifying our
> application has not requested more than the maximum memory capability of
> the cluster (16400 MB per container)
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Will allocate AM
> container, with 1408 MB memory including 384 MB overhead
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up container
> launch context for our AM
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Setting up the launch
> environment for our AM container
> zeppelin_1 | 21/04/01 23:18:45 INFO yarn.Client: Preparing resources for
> our AM container
> zeppelin_1 | 21/04/01 23:18:45 WARN yarn.Client: Neither spark.yarn.jars
> nor spark.yarn.archive is set, falling back to uploading libraries under
> SPARK_HOME.
> zeppelin_1 | 21/04/01 23:18:53 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_libs__5266504625643101044.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_libs__5266504625643101044.zip
> zeppelin_1 | 21/04/01 23:20:09 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/spark-interpreter-0.9.0.jar -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-interpreter-0.9.0.jar
> zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/spark/scala-2.11/spark-scala-2.11-0.9.0.jar
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/spark-scala-2.11-0.9.0.jar
> zeppelin_1 | 21/04/01 23:20:35 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/interpreter/zeppelin-interpreter-shaded-0.9.0.jar ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/zeppelin-interpreter-shaded-0.9.0.jar
> zeppelin_1 | 21/04/01 23:20:41 INFO yarn.Client: Uploading resource
> file:/opt/zeppelin/conf/log4j_yarn_cluster.properties -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/log4j_yarn_cluster.properties
> zeppelin_1 | 21/04/01 23:20:42 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/R/lib/sparkr.zip#sparkr -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/sparkr.zip
> zeppelin_1 | 21/04/01 23:20:43 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/pyspark.zip -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/pyspark.zip
> zeppelin_1 | 21/04/01 23:20:44 INFO yarn.Client: Uploading resource
> file:/spark/spark-2.4.7-bin-hadoop2.7/python/lib/py4j-0.10.7-src.zip ->
> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/py4j-0.10.7-src.zip
> zeppelin_1 | 21/04/01 23:20:45 INFO yarn.Client: Uploading resource
> file:/tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440/__spark_conf__8289533000141907930.zip
> -> hdfs://
> machine1.REMOVEDDOMAIN.de:8020/user/sandbox/.sparkStaging/application_1617315347811_0170/__spark_conf__.zip
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls to: zeppelin,sandbox
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls to: zeppelin,sandbox
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing view
> acls groups to:
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager: Changing
> modify acls groups to:
> zeppelin_1 | 21/04/01 23:20:46 INFO spark.SecurityManager:
> SecurityManager: authentication disabled; ui acls disabled; users with
> view permissions: Set(zeppelin, sandbox); groups with view permissions:
> Set(); users with modify permissions: Set(zeppelin, sandbox); groups with
> modify permissions: Set()
> zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Submitting application
> application_1617315347811_0170 to ResourceManager
> zeppelin_1 | 21/04/01 23:20:51 INFO impl.YarnClientImpl: Submitted
> application application_1617315347811_0170
> zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client: Application report for
> application_1617315347811_0170 (state: ACCEPTED)
> zeppelin_1 | 21/04/01 23:20:51 INFO yarn.Client:
> zeppelin_1 | client token: N/A
> zeppelin_1 | diagnostics: N/A
> zeppelin_1 | ApplicationMaster host: N/A
> zeppelin_1 | ApplicationMaster RPC port: -1
> zeppelin_1 | queue: root.users.sandbox
> zeppelin_1 | start time: 1617319251597
> zeppelin_1 | final status: UNDEFINED
> zeppelin_1 | tracking URL:
> http://machine1.REMOVEDDOMAIN.de:8088/proxy/application_1617315347811_0170/
> zeppelin_1 | user: sandbox
> zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Shutdown
> hook called
> zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-4c2bf1a1-2e67-42a9-8524-7810e1448440
> zeppelin_1 | 21/04/01 23:20:51 INFO util.ShutdownHookManager: Deleting
> directory /tmp/spark-1d86bc2c-eade-48f5-9650-423eef0fbda2
> zeppelin_1 |
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterManagedProcess.start(RemoteInterpreterManagedProcess.java:126)
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.ManagedInterpreterGroup.getOrCreateInterpreterProcess(ManagedInterpreterGroup.java:68)
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.getOrCreateInterpreterProcess(RemoteInterpreter.java:104)
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.internal_create(RemoteInterpreter.java:154)
> zeppelin_1 | at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.open(RemoteInterpreter.java:126)
> zeppelin_1 | ... 13 more
> zeppelin_1 |
> zeppelin_1 | INFO [2021-04-01 23:28:36,542] ({SchedulerFactory4}
> VFSNotebookRepo.java[save]:144) - Saving note 2EYUV26VR to Spark
> Tutorial/3. Spark SQL (Scala)_2EYUV26VR.zpln
>
>
--
Best Regards
Jeff Zhang