You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Christopher Piggott <cp...@gmail.com> on 2017/12/26 16:00:56 UTC
Spark 2.2.1 worker invocation
I need to set java.library.path to get access to some native code.
Following directions, I made a spark-env.sh:
#!/usr/bin/env bash
export
LD_LIBRARY_PATH="/usr/local/lib/libcdfNativeLibrary.so:/usr/local/lib/libcdf.so:${LD_LIBRARY_PATH}"
export SPARK_WORKER_OPTS=-Djava.library.path=/usr/local/lib
export SPARK_WORKER_MEMORY=2g
to no avail. (I tried both with and without exporting the environment).
Looking at how the worker actually starts up:
/usr/lib/jvm/default/bin/java -cp /home/spark/conf/:/home/spark/jars/*
-Xmx1024M -Dspark.driver.port=37219
org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url spark://
CoarseGrainedScheduler@10.1.1.1:37219 --executor-id 113 --hostname 10.2.2.1
--cores 8 --app-id app-20171225145607-0003 --worker-url spark://
Worker@10.2.2.1:35449
It doesn't seem to take any options. I put an 'echo' in just to confirm
that spark-env.sh is getting invoked (and it is).
So, just out of curiosity, I tried to troubleshoot this:
spark@node2-1:~$ grep -R SPARK_WORKER_OPTS *
conf/spark-env.sh:export
SPARK_WORKER_OPTS=-Djava.library.path=/usr/local/lib
conf/spark-env.sh.template:# - SPARK_WORKER_OPTS, to set config
properties only for the worker (e.g. "-Dx=y")
The variable doesn't seem to get referenced anywhere in the spark
distribution. I checked a number of other options in spark-env.sh.template
and they didn't seem to be referenced either. I expected to find them in
various startup scripts.
I can probably "fix" my problem by hacking the lower-level startup scripts,
but first I'd like to inquire about what's going on here. How and where
are these variables actually used?
Re: Spark 2.2.1 worker invocation
Posted by Felix Cheung <fe...@hotmail.com>.
I think you are looking for spark.executor.extraJavaOptions
https://spark.apache.org/docs/latest/configuration.html#runtime-environment
________________________________
From: Christopher Piggott <cp...@gmail.com>
Sent: Tuesday, December 26, 2017 8:00:56 AM
To: user@spark.apache.org
Subject: Spark 2.2.1 worker invocation
I need to set java.library.path to get access to some native code. Following directions, I made a spark-env.sh:
#!/usr/bin/env bash
export LD_LIBRARY_PATH="/usr/local/lib/libcdfNativeLibrary.so:/usr/local/lib/libcdf.so:${LD_LIBRARY_PATH}"
export SPARK_WORKER_OPTS=-Djava.library.path=/usr/local/lib
export SPARK_WORKER_MEMORY=2g
to no avail. (I tried both with and without exporting the environment). Looking at how the worker actually starts up:
/usr/lib/jvm/default/bin/java -cp /home/spark/conf/:/home/spark/jars/* -Xmx1024M -Dspark.driver.port=37219 org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url spark://CoarseGrainedScheduler@10.1.1.1:37219<http://CoarseGrainedScheduler@10.1.1.1:37219> --executor-id 113 --hostname 10.2.2.1 --cores 8 --app-id app-20171225145607-0003 --worker-url spark://Worker@10.2.2.1:35449<http://Worker@10.2.2.1:35449>
It doesn't seem to take any options. I put an 'echo' in just to confirm that spark-env.sh is getting invoked (and it is).
So, just out of curiosity, I tried to troubleshoot this:
spark@node2-1:~$ grep -R SPARK_WORKER_OPTS *
conf/spark-env.sh:export SPARK_WORKER_OPTS=-Djava.library.path=/usr/local/lib
conf/spark-env.sh.template:# - SPARK_WORKER_OPTS, to set config properties only for the worker (e.g. "-Dx=y")
The variable doesn't seem to get referenced anywhere in the spark distribution. I checked a number of other options in spark-env.sh.template and they didn't seem to be referenced either. I expected to find them in various startup scripts.
I can probably "fix" my problem by hacking the lower-level startup scripts, but first I'd like to inquire about what's going on here. How and where are these variables actually used?