You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Manoj Samel <ma...@gmail.com> on 2014/01/21 00:14:11 UTC

spark-shell on standalone cluster gives error " no mesos in java.library.path"

Hi

I deployed spark 0.8.1 on standalone cluster per
https://spark.incubator.apache.org/docs/0.8.1/spark-standalone.html

When i start a spark-shell , I get following error

I thought mesos should not be required for standalone cluster. Do I have to
change any parameters in make-distribution.sh that I used to build the
spark distribution for this cluster ? I left all to default (and noticed
that the default HADOOP version is 1.0.4 which is not my hadoop version -
but I am not using Hadoop here).

Creating SparkContext...
Failed to load native Mesos library from
java.lang.UnsatisfiedLinkError: no mesos in java.library.path
at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1738)
at java.lang.Runtime.loadLibrary0(Runtime.java:823)
at java.lang.System.loadLibrary(System.java:1028)
at org.apache.mesos.MesosNativeLibrary.load(MesosNativeLibrary.java:52)
at org.apache.mesos.MesosNativeLibrary.load(MesosNativeLibrary.java:64)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:260)
at org.apache.spark.repl.SparkILoop.createSparkContext(SparkILoop.scala:862)

Re: spark-shell on standalone cluster gives error " no mesos in java.library.path"

Posted by Manoj Samel <ma...@gmail.com>.
Please ignore this error - I found the issue.

Thanks !


On Mon, Jan 20, 2014 at 3:14 PM, Manoj Samel <ma...@gmail.com>wrote:

> Hi
>
> I deployed spark 0.8.1 on standalone cluster per
> https://spark.incubator.apache.org/docs/0.8.1/spark-standalone.html
>
> When i start a spark-shell , I get following error
>
> I thought mesos should not be required for standalone cluster. Do I have
> to change any parameters in make-distribution.sh that I used to build the
> spark distribution for this cluster ? I left all to default (and noticed
> that the default HADOOP version is 1.0.4 which is not my hadoop version -
> but I am not using Hadoop here).
>
> Creating SparkContext...
> Failed to load native Mesos library from
> java.lang.UnsatisfiedLinkError: no mesos in java.library.path
> at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1738)
>  at java.lang.Runtime.loadLibrary0(Runtime.java:823)
> at java.lang.System.loadLibrary(System.java:1028)
>  at org.apache.mesos.MesosNativeLibrary.load(MesosNativeLibrary.java:52)
> at org.apache.mesos.MesosNativeLibrary.load(MesosNativeLibrary.java:64)
>  at org.apache.spark.SparkContext.<init>(SparkContext.scala:260)
> at
> org.apache.spark.repl.SparkILoop.createSparkContext(SparkILoop.scala:862)
>