You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by Fairiz Azizi <co...@gmail.com> on 2014/10/06 08:19:04 UTC

Spark on Mesos 0.20

The Spark online docs indicate that Spark is compatible with Mesos 0.18.1

I've gotten it to work just fine on 0.18.1 and 0.18.2

Has anyone tried Spark on a newer version of Mesos, i.e. Mesos v0.20.0?

-Fi

Re: Spark on Mesos 0.20

Posted by Fairiz Azizi <co...@gmail.com>.
Hello,

Sorry for the delay (again), we were busy upgrading our cluster from MAPR
3.0.x to Mapr 3.1.1.26113.GA

I updated my builds to include referencing the native hadoop libraries by
this distribution as well as installing SNAPPY (I no longer see the 'unable
to load native hadoop libraries' and also see that it loads the SNAPPY
library).

I ran the example against a directory of ApacheLog files containing about
4.4GB, and things seem to work fine.

time MASTER="mesos://xxxxxxxx*:5050*" /opt/spark/current/bin/run-example
LogQuery "maprfs:///user/hive/warehouse/apachelog/dt=20141017/16"

14/10/18 05:23:21 INFO scheduler.DAGScheduler: Stage 0 (collect at
LogQuery.scala:80) finished in 1.704 s
14/10/18 05:23:21 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool default
14/10/18 05:23:21 INFO spark.SparkContext: Job finished: collect at
LogQuery.scala:80, took 40.533904277 s
(null,null,null) bytes=0 n=16682940

real 0m51.393s
user 0m19.130s
sys 0m4.120s

So this combination of software seems to work fine for me!

Spark 1.1.0
Mesos 0.20.1
MAPR 3.1.1.26113.GA
spark-1.1.0-bin-mapr3.tgz

Note: one thing you might try is increasing your spark.executor.memory
setting
Mine was set to 8GB in the spark-defaults.conf file.

Hope this helps,
Fi


Fairiz "Fi" Azizi

On Thu, Oct 9, 2014 at 11:35 PM, Gurvinder Singh <gurvinder.singh@uninett.no
> wrote:

> On 10/10/2014 06:11 AM, Fairiz Azizi wrote:
> > Hello,
> >
> > Sorry for the late reply.
> >
> > When I tried the LogQuery example this time, things now seem to be fine!
> >
> > ...
> >
> > 14/10/10 04:01:21 INFO scheduler.DAGScheduler: Stage 0 (collect at
> > LogQuery.scala:80) finished in 0.429 s
> >
> > 14/10/10 04:01:21 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
> > whose tasks have all completed, from pool defa
> >
> > 14/10/10 04:01:21 INFO spark.SparkContext: Job finished: collect at
> > LogQuery.scala:80, took 12.802743914 s
> >
> > (10.10.10.10,"FRED",GET http://images.com/2013/Generic.jpg HTTP/1.1)
> > bytes=621       n=2
> >
> >
> > Not sure if this is the correct response for that example.
> >
> > Our mesos/spark builds have since been updated since I last wrote.
> >
> > Possibly, the JDK version was updated to 1.7.0_67
> >
> > If you are using an older JDK, maybe try updating that?
> I have tested on current JDK 7 and now I am running JDK 8, the problem
> still exist. Can you run logquery on data of size say 100+ GB, so that
> you have more map tasks. As we start to see the issue on larger tasks.
>
> - Gurvinder
> >
> >
> > - Fi
> >
> >
> >
> > Fairiz "Fi" Azizi
> >
> > On Wed, Oct 8, 2014 at 7:54 AM, RJ Nowling <rnowling@gmail.com
> > <ma...@gmail.com>> wrote:
> >
> >     Yep!  That's the example I was talking about.
> >
> >     Is an error message printed when it hangs? I get :
> >
> >     14/09/30 13:23:14 ERROR BlockManagerMasterActor: Got two different
> block manager registrations on 20140930-131734-1723727882-5050-1895-1
> >
> >
> >
> >     On Tue, Oct 7, 2014 at 8:36 PM, Fairiz Azizi <coderfi@gmail.com
> >     <ma...@gmail.com>> wrote:
> >
> >         Sure, could you point me to the example?
> >
> >         The only thing I could find was
> >
> https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/LogQuery.scala
> >
> >         So do you mean running it like:
> >            MASTER="mesos://xxxxxxx_:5050_" ./run-example LogQuery
> >
> >         I tried that and I can see the job run and the tasks complete on
> >         the slave nodes, but the client process seems to hang forever,
> >         it's probably a different problem. BTW, only a dozen or so tasks
> >         kick off.
> >
> >         I actually haven't done much with Scala and Spark (it's been all
> >         python).
> >
> >         Fi
> >
> >
> >
> >         Fairiz "Fi" Azizi
> >
> >         On Tue, Oct 7, 2014 at 6:29 AM, RJ Nowling <rnowling@gmail.com
> >         <ma...@gmail.com>> wrote:
> >
> >             I was able to reproduce it on a small 4 node cluster (1
> >             mesos master and 3 mesos slaves) with relatively low-end
> >             specs.  As I said, I just ran the log query examples with
> >             the fine-grained mesos mode.
> >
> >             Spark 1.1.0 and mesos 0.20.1.
> >
> >             Fairiz, could you try running the logquery example included
> >             with Spark and see what you get?
> >
> >             Thanks!
> >
> >             On Mon, Oct 6, 2014 at 8:07 PM, Fairiz Azizi
> >             <coderfi@gmail.com <ma...@gmail.com>> wrote:
> >
> >                 That's what great about Spark, the community is so
> >                 active! :)
> >
> >                 I compiled Mesos 0.20.1 from the source tarball.
> >
> >                 Using the Mapr3 Spark 1.1.0 distribution from the Spark
> >                 downloads page  (spark-1.1.0-bin-mapr3.tgz).
> >
> >                 I see no problems for the workloads we are trying.
> >
> >                 However, the cluster is small (less than 100 cores
> >                 across 3 nodes).
> >
> >                 The workloads reads in just a few gigabytes from HDFS,
> >                 via an ipython notebook spark shell.
> >
> >                 thanks,
> >                 Fi
> >
> >
> >
> >                 Fairiz "Fi" Azizi
> >
> >                 On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen
> >                 <tnachen@gmail.com <ma...@gmail.com>> wrote:
> >
> >                     Ok I created SPARK-3817 to track this, will try to
> >                     repro it as well.
> >
> >                     Tim
> >
> >                     On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling
> >                     <rnowling@gmail.com <ma...@gmail.com>>
> wrote:
> >                     > I've recently run into this issue as well. I get
> >                     it from running Spark
> >                     > examples such as log query.  Maybe that'll help
> >                     reproduce the issue.
> >                     >
> >                     >
> >                     > On Monday, October 6, 2014, Gurvinder Singh
> >                     <gurvinder.singh@uninett.no
> >                     <ma...@uninett.no>>
> >                     > wrote:
> >                     >>
> >                     >> The issue does not occur if the task at hand has
> >                     small number of map
> >                     >> tasks. I have a task which has 978 map tasks and
> >                     I see this error as
> >                     >>
> >                     >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor:
> >                     Got two different block
> >                     >> manager registrations on
> >                     20140711-081617-711206558-5050-2543-5
> >                     >>
> >                     >> Here is the log from the mesos-slave where this
> >                     container was running.
> >                     >>
> >                     >> http://pastebin.com/Q1Cuzm6Q
> >                     >>
> >                     >> If you look for the code from where error
> >                     produced by spark, you will
> >                     >> see that it simply exit and saying in comments
> >                     "this should never
> >                     >> happen, lets just quit" :-)
> >                     >>
> >                     >> - Gurvinder
> >                     >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
> >                     >> > (Hit enter too soon...)
> >                     >> >
> >                     >> > What is your setup and steps to repro this?
> >                     >> >
> >                     >> > Tim
> >                     >> >
> >                     >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen
> >                     <tnachen@gmail.com <ma...@gmail.com>>
> wrote:
> >                     >> >> Hi Gurvinder,
> >                     >> >>
> >                     >> >> I tried fine grain mode before and didn't get
> >                     into that problem.
> >                     >> >>
> >                     >> >>
> >                     >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder
> Singh
> >                     >> >> <gurvinder.singh@uninett.no
> >                     <ma...@uninett.no>> wrote:
> >                     >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
> >                     >> >>>> The Spark online docs indicate that Spark is
> >                     compatible with Mesos
> >                     >> >>>> 0.18.1
> >                     >> >>>>
> >                     >> >>>> I've gotten it to work just fine on 0.18.1
> >                     and 0.18.2
> >                     >> >>>>
> >                     >> >>>> Has anyone tried Spark on a newer version of
> >                     Mesos, i.e. Mesos
> >                     >> >>>> v0.20.0?
> >                     >> >>>>
> >                     >> >>>> -Fi
> >                     >> >>>>
> >                     >> >>> Yeah we are using Spark 1.1.0 with Mesos
> >                     0.20.1. It runs fine in
> >                     >> >>> coarse
> >                     >> >>> mode, in fine grain mode there is an issue
> >                     with blockmanager names
> >                     >> >>> conflict. I have been waiting for it to be
> >                     fixed but it is still
> >                     >> >>> there.
> >                     >> >>>
> >                     >> >>> -Gurvinder
> >                     >> >>>
> >                     >> >>>
> >
>  ---------------------------------------------------------------------
> >                     >> >>> To unsubscribe, e-mail:
> >                     dev-unsubscribe@spark.apache.org
> >                     <ma...@spark.apache.org>
> >                     >> >>> For additional commands, e-mail:
> >                     dev-help@spark.apache.org
> >                     <ma...@spark.apache.org>
> >                     >> >>>
> >                     >>
> >                     >>
> >                     >>
> >
>  ---------------------------------------------------------------------
> >                     >> To unsubscribe, e-mail:
> >                     dev-unsubscribe@spark.apache.org
> >                     <ma...@spark.apache.org>
> >                     >> For additional commands, e-mail:
> >                     dev-help@spark.apache.org
> >                     <ma...@spark.apache.org>
> >                     >>
> >                     >
> >                     >
> >                     > --
> >                     > em rnowling@gmail.com <ma...@gmail.com>
> >                     > c 954.496.2314 <tel:954.496.2314>
> >
> >
> >
> >
> >
> >             --
> >             em rnowling@gmail.com <ma...@gmail.com>
> >             c 954.496.2314 <tel:954.496.2314>
> >
> >
> >
> >
> >
> >     --
> >     em rnowling@gmail.com <ma...@gmail.com>
> >     c 954.496.2314 <tel:954.496.2314>
> >
> >
>
>

Re: Spark on Mesos 0.20

Posted by Gurvinder Singh <gu...@uninett.no>.
On 10/10/2014 06:11 AM, Fairiz Azizi wrote:
> Hello,
> 
> Sorry for the late reply.
> 
> When I tried the LogQuery example this time, things now seem to be fine!
> 
> ...
> 
> 14/10/10 04:01:21 INFO scheduler.DAGScheduler: Stage 0 (collect at
> LogQuery.scala:80) finished in 0.429 s
> 
> 14/10/10 04:01:21 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
> whose tasks have all completed, from pool defa
> 
> 14/10/10 04:01:21 INFO spark.SparkContext: Job finished: collect at
> LogQuery.scala:80, took 12.802743914 s
> 
> (10.10.10.10,"FRED",GET http://images.com/2013/Generic.jpg HTTP/1.1)   
> bytes=621       n=2
> 
> 
> Not sure if this is the correct response for that example.
> 
> Our mesos/spark builds have since been updated since I last wrote.
> 
> Possibly, the JDK version was updated to 1.7.0_67
> 
> If you are using an older JDK, maybe try updating that?
I have tested on current JDK 7 and now I am running JDK 8, the problem
still exist. Can you run logquery on data of size say 100+ GB, so that
you have more map tasks. As we start to see the issue on larger tasks.

- Gurvinder
> 
> 
> - Fi
> 
> 
> 
> Fairiz "Fi" Azizi
> 
> On Wed, Oct 8, 2014 at 7:54 AM, RJ Nowling <rnowling@gmail.com
> <ma...@gmail.com>> wrote:
> 
>     Yep!  That's the example I was talking about.
> 
>     Is an error message printed when it hangs? I get :
> 
>     14/09/30 13:23:14 ERROR BlockManagerMasterActor: Got two different block manager registrations on 20140930-131734-1723727882-5050-1895-1
> 
> 
> 
>     On Tue, Oct 7, 2014 at 8:36 PM, Fairiz Azizi <coderfi@gmail.com
>     <ma...@gmail.com>> wrote:
> 
>         Sure, could you point me to the example?
> 
>         The only thing I could find was
>         https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/LogQuery.scala
> 
>         So do you mean running it like:
>            MASTER="mesos://xxxxxxx_:5050_" ./run-example LogQuery
> 
>         I tried that and I can see the job run and the tasks complete on
>         the slave nodes, but the client process seems to hang forever,
>         it's probably a different problem. BTW, only a dozen or so tasks
>         kick off.
> 
>         I actually haven't done much with Scala and Spark (it's been all
>         python).
> 
>         Fi
> 
> 
> 
>         Fairiz "Fi" Azizi
> 
>         On Tue, Oct 7, 2014 at 6:29 AM, RJ Nowling <rnowling@gmail.com
>         <ma...@gmail.com>> wrote:
> 
>             I was able to reproduce it on a small 4 node cluster (1
>             mesos master and 3 mesos slaves) with relatively low-end
>             specs.  As I said, I just ran the log query examples with
>             the fine-grained mesos mode.
> 
>             Spark 1.1.0 and mesos 0.20.1.
> 
>             Fairiz, could you try running the logquery example included
>             with Spark and see what you get?
> 
>             Thanks!
> 
>             On Mon, Oct 6, 2014 at 8:07 PM, Fairiz Azizi
>             <coderfi@gmail.com <ma...@gmail.com>> wrote:
> 
>                 That's what great about Spark, the community is so
>                 active! :)
> 
>                 I compiled Mesos 0.20.1 from the source tarball.
> 
>                 Using the Mapr3 Spark 1.1.0 distribution from the Spark
>                 downloads page  (spark-1.1.0-bin-mapr3.tgz).
> 
>                 I see no problems for the workloads we are trying. 
> 
>                 However, the cluster is small (less than 100 cores
>                 across 3 nodes).
> 
>                 The workloads reads in just a few gigabytes from HDFS,
>                 via an ipython notebook spark shell.
> 
>                 thanks,
>                 Fi
> 
> 
> 
>                 Fairiz "Fi" Azizi
> 
>                 On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen
>                 <tnachen@gmail.com <ma...@gmail.com>> wrote:
> 
>                     Ok I created SPARK-3817 to track this, will try to
>                     repro it as well.
> 
>                     Tim
> 
>                     On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling
>                     <rnowling@gmail.com <ma...@gmail.com>> wrote:
>                     > I've recently run into this issue as well. I get
>                     it from running Spark
>                     > examples such as log query.  Maybe that'll help
>                     reproduce the issue.
>                     >
>                     >
>                     > On Monday, October 6, 2014, Gurvinder Singh
>                     <gurvinder.singh@uninett.no
>                     <ma...@uninett.no>>
>                     > wrote:
>                     >>
>                     >> The issue does not occur if the task at hand has
>                     small number of map
>                     >> tasks. I have a task which has 978 map tasks and
>                     I see this error as
>                     >>
>                     >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor:
>                     Got two different block
>                     >> manager registrations on
>                     20140711-081617-711206558-5050-2543-5
>                     >>
>                     >> Here is the log from the mesos-slave where this
>                     container was running.
>                     >>
>                     >> http://pastebin.com/Q1Cuzm6Q
>                     >>
>                     >> If you look for the code from where error
>                     produced by spark, you will
>                     >> see that it simply exit and saying in comments
>                     "this should never
>                     >> happen, lets just quit" :-)
>                     >>
>                     >> - Gurvinder
>                     >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
>                     >> > (Hit enter too soon...)
>                     >> >
>                     >> > What is your setup and steps to repro this?
>                     >> >
>                     >> > Tim
>                     >> >
>                     >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen
>                     <tnachen@gmail.com <ma...@gmail.com>> wrote:
>                     >> >> Hi Gurvinder,
>                     >> >>
>                     >> >> I tried fine grain mode before and didn't get
>                     into that problem.
>                     >> >>
>                     >> >>
>                     >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>                     >> >> <gurvinder.singh@uninett.no
>                     <ma...@uninett.no>> wrote:
>                     >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>                     >> >>>> The Spark online docs indicate that Spark is
>                     compatible with Mesos
>                     >> >>>> 0.18.1
>                     >> >>>>
>                     >> >>>> I've gotten it to work just fine on 0.18.1
>                     and 0.18.2
>                     >> >>>>
>                     >> >>>> Has anyone tried Spark on a newer version of
>                     Mesos, i.e. Mesos
>                     >> >>>> v0.20.0?
>                     >> >>>>
>                     >> >>>> -Fi
>                     >> >>>>
>                     >> >>> Yeah we are using Spark 1.1.0 with Mesos
>                     0.20.1. It runs fine in
>                     >> >>> coarse
>                     >> >>> mode, in fine grain mode there is an issue
>                     with blockmanager names
>                     >> >>> conflict. I have been waiting for it to be
>                     fixed but it is still
>                     >> >>> there.
>                     >> >>>
>                     >> >>> -Gurvinder
>                     >> >>>
>                     >> >>>
>                     ---------------------------------------------------------------------
>                     >> >>> To unsubscribe, e-mail:
>                     dev-unsubscribe@spark.apache.org
>                     <ma...@spark.apache.org>
>                     >> >>> For additional commands, e-mail:
>                     dev-help@spark.apache.org
>                     <ma...@spark.apache.org>
>                     >> >>>
>                     >>
>                     >>
>                     >>
>                     ---------------------------------------------------------------------
>                     >> To unsubscribe, e-mail:
>                     dev-unsubscribe@spark.apache.org
>                     <ma...@spark.apache.org>
>                     >> For additional commands, e-mail:
>                     dev-help@spark.apache.org
>                     <ma...@spark.apache.org>
>                     >>
>                     >
>                     >
>                     > --
>                     > em rnowling@gmail.com <ma...@gmail.com>
>                     > c 954.496.2314 <tel:954.496.2314>
> 
> 
> 
> 
> 
>             -- 
>             em rnowling@gmail.com <ma...@gmail.com>
>             c 954.496.2314 <tel:954.496.2314>
> 
> 
> 
> 
> 
>     -- 
>     em rnowling@gmail.com <ma...@gmail.com>
>     c 954.496.2314 <tel:954.496.2314>
> 
> 


---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org


Re: Spark on Mesos 0.20

Posted by Fairiz Azizi <co...@gmail.com>.
Hello,

Sorry for the late reply.

When I tried the LogQuery example this time, things now seem to be fine!

...

14/10/10 04:01:21 INFO scheduler.DAGScheduler: Stage 0 (collect at
LogQuery.scala:80) finished in 0.429 s

14/10/10 04:01:21 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 0.0,
whose tasks have all completed, from pool defa

14/10/10 04:01:21 INFO spark.SparkContext: Job finished: collect at
LogQuery.scala:80, took 12.802743914 s

(10.10.10.10,"FRED",GET http://images.com/2013/Generic.jpg HTTP/1.1)
bytes=621       n=2


Not sure if this is the correct response for that example.

Our mesos/spark builds have since been updated since I last wrote.

Possibly, the JDK version was updated to 1.7.0_67

If you are using an older JDK, maybe try updating that?


- Fi



Fairiz "Fi" Azizi

On Wed, Oct 8, 2014 at 7:54 AM, RJ Nowling <rn...@gmail.com> wrote:

> Yep!  That's the example I was talking about.
>
> Is an error message printed when it hangs? I get :
>
> 14/09/30 13:23:14 ERROR BlockManagerMasterActor: Got two different block manager registrations on 20140930-131734-1723727882-5050-1895-1
>
>
>
> On Tue, Oct 7, 2014 at 8:36 PM, Fairiz Azizi <co...@gmail.com> wrote:
>
>> Sure, could you point me to the example?
>>
>> The only thing I could find was
>>
>> https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/LogQuery.scala
>>
>> So do you mean running it like:
>>    MASTER="mesos://xxxxxxx*:5050*" ./run-example LogQuery
>>
>> I tried that and I can see the job run and the tasks complete on the
>> slave nodes, but the client process seems to hang forever, it's probably a
>> different problem. BTW, only a dozen or so tasks kick off.
>>
>> I actually haven't done much with Scala and Spark (it's been all python).
>>
>> Fi
>>
>>
>>
>> Fairiz "Fi" Azizi
>>
>> On Tue, Oct 7, 2014 at 6:29 AM, RJ Nowling <rn...@gmail.com> wrote:
>>
>>> I was able to reproduce it on a small 4 node cluster (1 mesos master and
>>> 3 mesos slaves) with relatively low-end specs.  As I said, I just ran the
>>> log query examples with the fine-grained mesos mode.
>>>
>>> Spark 1.1.0 and mesos 0.20.1.
>>>
>>> Fairiz, could you try running the logquery example included with Spark
>>> and see what you get?
>>>
>>> Thanks!
>>>
>>> On Mon, Oct 6, 2014 at 8:07 PM, Fairiz Azizi <co...@gmail.com> wrote:
>>>
>>>> That's what great about Spark, the community is so active! :)
>>>>
>>>> I compiled Mesos 0.20.1 from the source tarball.
>>>>
>>>> Using the Mapr3 Spark 1.1.0 distribution from the Spark downloads page
>>>>  (spark-1.1.0-bin-mapr3.tgz).
>>>>
>>>> I see no problems for the workloads we are trying.
>>>>
>>>> However, the cluster is small (less than 100 cores across 3 nodes).
>>>>
>>>> The workloads reads in just a few gigabytes from HDFS, via an ipython
>>>> notebook spark shell.
>>>>
>>>> thanks,
>>>> Fi
>>>>
>>>>
>>>>
>>>> Fairiz "Fi" Azizi
>>>>
>>>> On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen <tn...@gmail.com> wrote:
>>>>
>>>>> Ok I created SPARK-3817 to track this, will try to repro it as well.
>>>>>
>>>>> Tim
>>>>>
>>>>> On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling <rn...@gmail.com> wrote:
>>>>> > I've recently run into this issue as well. I get it from running
>>>>> Spark
>>>>> > examples such as log query.  Maybe that'll help reproduce the issue.
>>>>> >
>>>>> >
>>>>> > On Monday, October 6, 2014, Gurvinder Singh <
>>>>> gurvinder.singh@uninett.no>
>>>>> > wrote:
>>>>> >>
>>>>> >> The issue does not occur if the task at hand has small number of map
>>>>> >> tasks. I have a task which has 978 map tasks and I see this error as
>>>>> >>
>>>>> >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different
>>>>> block
>>>>> >> manager registrations on 20140711-081617-711206558-5050-2543-5
>>>>> >>
>>>>> >> Here is the log from the mesos-slave where this container was
>>>>> running.
>>>>> >>
>>>>> >> http://pastebin.com/Q1Cuzm6Q
>>>>> >>
>>>>> >> If you look for the code from where error produced by spark, you
>>>>> will
>>>>> >> see that it simply exit and saying in comments "this should never
>>>>> >> happen, lets just quit" :-)
>>>>> >>
>>>>> >> - Gurvinder
>>>>> >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
>>>>> >> > (Hit enter too soon...)
>>>>> >> >
>>>>> >> > What is your setup and steps to repro this?
>>>>> >> >
>>>>> >> > Tim
>>>>> >> >
>>>>> >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com>
>>>>> wrote:
>>>>> >> >> Hi Gurvinder,
>>>>> >> >>
>>>>> >> >> I tried fine grain mode before and didn't get into that problem.
>>>>> >> >>
>>>>> >> >>
>>>>> >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>>>>> >> >> <gu...@uninett.no> wrote:
>>>>> >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>>>>> >> >>>> The Spark online docs indicate that Spark is compatible with
>>>>> Mesos
>>>>> >> >>>> 0.18.1
>>>>> >> >>>>
>>>>> >> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>>>>> >> >>>>
>>>>> >> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
>>>>> >> >>>> v0.20.0?
>>>>> >> >>>>
>>>>> >> >>>> -Fi
>>>>> >> >>>>
>>>>> >> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in
>>>>> >> >>> coarse
>>>>> >> >>> mode, in fine grain mode there is an issue with blockmanager
>>>>> names
>>>>> >> >>> conflict. I have been waiting for it to be fixed but it is still
>>>>> >> >>> there.
>>>>> >> >>>
>>>>> >> >>> -Gurvinder
>>>>> >> >>>
>>>>> >> >>>
>>>>> ---------------------------------------------------------------------
>>>>> >> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>>>> >> >>> For additional commands, e-mail: dev-help@spark.apache.org
>>>>> >> >>>
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> ---------------------------------------------------------------------
>>>>> >> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>>>> >> For additional commands, e-mail: dev-help@spark.apache.org
>>>>> >>
>>>>> >
>>>>> >
>>>>> > --
>>>>> > em rnowling@gmail.com
>>>>> > c 954.496.2314
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> em rnowling@gmail.com
>>> c 954.496.2314
>>>
>>
>>
>
>
> --
> em rnowling@gmail.com
> c 954.496.2314
>

Re: Spark on Mesos 0.20

Posted by RJ Nowling <rn...@gmail.com>.
Yep!  That's the example I was talking about.

Is an error message printed when it hangs? I get :

14/09/30 13:23:14 ERROR BlockManagerMasterActor: Got two different
block manager registrations on 20140930-131734-1723727882-5050-1895-1



On Tue, Oct 7, 2014 at 8:36 PM, Fairiz Azizi <co...@gmail.com> wrote:

> Sure, could you point me to the example?
>
> The only thing I could find was
>
> https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/LogQuery.scala
>
> So do you mean running it like:
>    MASTER="mesos://xxxxxxx*:5050*" ./run-example LogQuery
>
> I tried that and I can see the job run and the tasks complete on the slave
> nodes, but the client process seems to hang forever, it's probably a
> different problem. BTW, only a dozen or so tasks kick off.
>
> I actually haven't done much with Scala and Spark (it's been all python).
>
> Fi
>
>
>
> Fairiz "Fi" Azizi
>
> On Tue, Oct 7, 2014 at 6:29 AM, RJ Nowling <rn...@gmail.com> wrote:
>
>> I was able to reproduce it on a small 4 node cluster (1 mesos master and
>> 3 mesos slaves) with relatively low-end specs.  As I said, I just ran the
>> log query examples with the fine-grained mesos mode.
>>
>> Spark 1.1.0 and mesos 0.20.1.
>>
>> Fairiz, could you try running the logquery example included with Spark
>> and see what you get?
>>
>> Thanks!
>>
>> On Mon, Oct 6, 2014 at 8:07 PM, Fairiz Azizi <co...@gmail.com> wrote:
>>
>>> That's what great about Spark, the community is so active! :)
>>>
>>> I compiled Mesos 0.20.1 from the source tarball.
>>>
>>> Using the Mapr3 Spark 1.1.0 distribution from the Spark downloads page
>>>  (spark-1.1.0-bin-mapr3.tgz).
>>>
>>> I see no problems for the workloads we are trying.
>>>
>>> However, the cluster is small (less than 100 cores across 3 nodes).
>>>
>>> The workloads reads in just a few gigabytes from HDFS, via an ipython
>>> notebook spark shell.
>>>
>>> thanks,
>>> Fi
>>>
>>>
>>>
>>> Fairiz "Fi" Azizi
>>>
>>> On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen <tn...@gmail.com> wrote:
>>>
>>>> Ok I created SPARK-3817 to track this, will try to repro it as well.
>>>>
>>>> Tim
>>>>
>>>> On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling <rn...@gmail.com> wrote:
>>>> > I've recently run into this issue as well. I get it from running Spark
>>>> > examples such as log query.  Maybe that'll help reproduce the issue.
>>>> >
>>>> >
>>>> > On Monday, October 6, 2014, Gurvinder Singh <
>>>> gurvinder.singh@uninett.no>
>>>> > wrote:
>>>> >>
>>>> >> The issue does not occur if the task at hand has small number of map
>>>> >> tasks. I have a task which has 978 map tasks and I see this error as
>>>> >>
>>>> >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different
>>>> block
>>>> >> manager registrations on 20140711-081617-711206558-5050-2543-5
>>>> >>
>>>> >> Here is the log from the mesos-slave where this container was
>>>> running.
>>>> >>
>>>> >> http://pastebin.com/Q1Cuzm6Q
>>>> >>
>>>> >> If you look for the code from where error produced by spark, you will
>>>> >> see that it simply exit and saying in comments "this should never
>>>> >> happen, lets just quit" :-)
>>>> >>
>>>> >> - Gurvinder
>>>> >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
>>>> >> > (Hit enter too soon...)
>>>> >> >
>>>> >> > What is your setup and steps to repro this?
>>>> >> >
>>>> >> > Tim
>>>> >> >
>>>> >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com>
>>>> wrote:
>>>> >> >> Hi Gurvinder,
>>>> >> >>
>>>> >> >> I tried fine grain mode before and didn't get into that problem.
>>>> >> >>
>>>> >> >>
>>>> >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>>>> >> >> <gu...@uninett.no> wrote:
>>>> >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>>>> >> >>>> The Spark online docs indicate that Spark is compatible with
>>>> Mesos
>>>> >> >>>> 0.18.1
>>>> >> >>>>
>>>> >> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>>>> >> >>>>
>>>> >> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
>>>> >> >>>> v0.20.0?
>>>> >> >>>>
>>>> >> >>>> -Fi
>>>> >> >>>>
>>>> >> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in
>>>> >> >>> coarse
>>>> >> >>> mode, in fine grain mode there is an issue with blockmanager
>>>> names
>>>> >> >>> conflict. I have been waiting for it to be fixed but it is still
>>>> >> >>> there.
>>>> >> >>>
>>>> >> >>> -Gurvinder
>>>> >> >>>
>>>> >> >>>
>>>> ---------------------------------------------------------------------
>>>> >> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>>> >> >>> For additional commands, e-mail: dev-help@spark.apache.org
>>>> >> >>>
>>>> >>
>>>> >>
>>>> >> ---------------------------------------------------------------------
>>>> >> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>>> >> For additional commands, e-mail: dev-help@spark.apache.org
>>>> >>
>>>> >
>>>> >
>>>> > --
>>>> > em rnowling@gmail.com
>>>> > c 954.496.2314
>>>>
>>>
>>>
>>
>>
>> --
>> em rnowling@gmail.com
>> c 954.496.2314
>>
>
>


-- 
em rnowling@gmail.com
c 954.496.2314

Re: Spark on Mesos 0.20

Posted by Fairiz Azizi <co...@gmail.com>.
Sure, could you point me to the example?

The only thing I could find was
https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/LogQuery.scala

So do you mean running it like:
   MASTER="mesos://xxxxxxx*:5050*" ./run-example LogQuery

I tried that and I can see the job run and the tasks complete on the slave
nodes, but the client process seems to hang forever, it's probably a
different problem. BTW, only a dozen or so tasks kick off.

I actually haven't done much with Scala and Spark (it's been all python).

Fi



Fairiz "Fi" Azizi

On Tue, Oct 7, 2014 at 6:29 AM, RJ Nowling <rn...@gmail.com> wrote:

> I was able to reproduce it on a small 4 node cluster (1 mesos master and 3
> mesos slaves) with relatively low-end specs.  As I said, I just ran the log
> query examples with the fine-grained mesos mode.
>
> Spark 1.1.0 and mesos 0.20.1.
>
> Fairiz, could you try running the logquery example included with Spark and
> see what you get?
>
> Thanks!
>
> On Mon, Oct 6, 2014 at 8:07 PM, Fairiz Azizi <co...@gmail.com> wrote:
>
>> That's what great about Spark, the community is so active! :)
>>
>> I compiled Mesos 0.20.1 from the source tarball.
>>
>> Using the Mapr3 Spark 1.1.0 distribution from the Spark downloads page
>>  (spark-1.1.0-bin-mapr3.tgz).
>>
>> I see no problems for the workloads we are trying.
>>
>> However, the cluster is small (less than 100 cores across 3 nodes).
>>
>> The workloads reads in just a few gigabytes from HDFS, via an ipython
>> notebook spark shell.
>>
>> thanks,
>> Fi
>>
>>
>>
>> Fairiz "Fi" Azizi
>>
>> On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen <tn...@gmail.com> wrote:
>>
>>> Ok I created SPARK-3817 to track this, will try to repro it as well.
>>>
>>> Tim
>>>
>>> On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling <rn...@gmail.com> wrote:
>>> > I've recently run into this issue as well. I get it from running Spark
>>> > examples such as log query.  Maybe that'll help reproduce the issue.
>>> >
>>> >
>>> > On Monday, October 6, 2014, Gurvinder Singh <
>>> gurvinder.singh@uninett.no>
>>> > wrote:
>>> >>
>>> >> The issue does not occur if the task at hand has small number of map
>>> >> tasks. I have a task which has 978 map tasks and I see this error as
>>> >>
>>> >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different
>>> block
>>> >> manager registrations on 20140711-081617-711206558-5050-2543-5
>>> >>
>>> >> Here is the log from the mesos-slave where this container was running.
>>> >>
>>> >> http://pastebin.com/Q1Cuzm6Q
>>> >>
>>> >> If you look for the code from where error produced by spark, you will
>>> >> see that it simply exit and saying in comments "this should never
>>> >> happen, lets just quit" :-)
>>> >>
>>> >> - Gurvinder
>>> >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
>>> >> > (Hit enter too soon...)
>>> >> >
>>> >> > What is your setup and steps to repro this?
>>> >> >
>>> >> > Tim
>>> >> >
>>> >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com>
>>> wrote:
>>> >> >> Hi Gurvinder,
>>> >> >>
>>> >> >> I tried fine grain mode before and didn't get into that problem.
>>> >> >>
>>> >> >>
>>> >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>>> >> >> <gu...@uninett.no> wrote:
>>> >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>>> >> >>>> The Spark online docs indicate that Spark is compatible with
>>> Mesos
>>> >> >>>> 0.18.1
>>> >> >>>>
>>> >> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>>> >> >>>>
>>> >> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
>>> >> >>>> v0.20.0?
>>> >> >>>>
>>> >> >>>> -Fi
>>> >> >>>>
>>> >> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in
>>> >> >>> coarse
>>> >> >>> mode, in fine grain mode there is an issue with blockmanager names
>>> >> >>> conflict. I have been waiting for it to be fixed but it is still
>>> >> >>> there.
>>> >> >>>
>>> >> >>> -Gurvinder
>>> >> >>>
>>> >> >>>
>>> ---------------------------------------------------------------------
>>> >> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>> >> >>> For additional commands, e-mail: dev-help@spark.apache.org
>>> >> >>>
>>> >>
>>> >>
>>> >> ---------------------------------------------------------------------
>>> >> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>> >> For additional commands, e-mail: dev-help@spark.apache.org
>>> >>
>>> >
>>> >
>>> > --
>>> > em rnowling@gmail.com
>>> > c 954.496.2314
>>>
>>
>>
>
>
> --
> em rnowling@gmail.com
> c 954.496.2314
>

Re: Spark on Mesos 0.20

Posted by RJ Nowling <rn...@gmail.com>.
I was able to reproduce it on a small 4 node cluster (1 mesos master and 3
mesos slaves) with relatively low-end specs.  As I said, I just ran the log
query examples with the fine-grained mesos mode.

Spark 1.1.0 and mesos 0.20.1.

Fairiz, could you try running the logquery example included with Spark and
see what you get?

Thanks!

On Mon, Oct 6, 2014 at 8:07 PM, Fairiz Azizi <co...@gmail.com> wrote:

> That's what great about Spark, the community is so active! :)
>
> I compiled Mesos 0.20.1 from the source tarball.
>
> Using the Mapr3 Spark 1.1.0 distribution from the Spark downloads page
>  (spark-1.1.0-bin-mapr3.tgz).
>
> I see no problems for the workloads we are trying.
>
> However, the cluster is small (less than 100 cores across 3 nodes).
>
> The workloads reads in just a few gigabytes from HDFS, via an ipython
> notebook spark shell.
>
> thanks,
> Fi
>
>
>
> Fairiz "Fi" Azizi
>
> On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen <tn...@gmail.com> wrote:
>
>> Ok I created SPARK-3817 to track this, will try to repro it as well.
>>
>> Tim
>>
>> On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling <rn...@gmail.com> wrote:
>> > I've recently run into this issue as well. I get it from running Spark
>> > examples such as log query.  Maybe that'll help reproduce the issue.
>> >
>> >
>> > On Monday, October 6, 2014, Gurvinder Singh <gurvinder.singh@uninett.no
>> >
>> > wrote:
>> >>
>> >> The issue does not occur if the task at hand has small number of map
>> >> tasks. I have a task which has 978 map tasks and I see this error as
>> >>
>> >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different
>> block
>> >> manager registrations on 20140711-081617-711206558-5050-2543-5
>> >>
>> >> Here is the log from the mesos-slave where this container was running.
>> >>
>> >> http://pastebin.com/Q1Cuzm6Q
>> >>
>> >> If you look for the code from where error produced by spark, you will
>> >> see that it simply exit and saying in comments "this should never
>> >> happen, lets just quit" :-)
>> >>
>> >> - Gurvinder
>> >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
>> >> > (Hit enter too soon...)
>> >> >
>> >> > What is your setup and steps to repro this?
>> >> >
>> >> > Tim
>> >> >
>> >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com>
>> wrote:
>> >> >> Hi Gurvinder,
>> >> >>
>> >> >> I tried fine grain mode before and didn't get into that problem.
>> >> >>
>> >> >>
>> >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>> >> >> <gu...@uninett.no> wrote:
>> >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>> >> >>>> The Spark online docs indicate that Spark is compatible with Mesos
>> >> >>>> 0.18.1
>> >> >>>>
>> >> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>> >> >>>>
>> >> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
>> >> >>>> v0.20.0?
>> >> >>>>
>> >> >>>> -Fi
>> >> >>>>
>> >> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in
>> >> >>> coarse
>> >> >>> mode, in fine grain mode there is an issue with blockmanager names
>> >> >>> conflict. I have been waiting for it to be fixed but it is still
>> >> >>> there.
>> >> >>>
>> >> >>> -Gurvinder
>> >> >>>
>> >> >>>
>> ---------------------------------------------------------------------
>> >> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>> >> >>> For additional commands, e-mail: dev-help@spark.apache.org
>> >> >>>
>> >>
>> >>
>> >> ---------------------------------------------------------------------
>> >> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>> >> For additional commands, e-mail: dev-help@spark.apache.org
>> >>
>> >
>> >
>> > --
>> > em rnowling@gmail.com
>> > c 954.496.2314
>>
>
>


-- 
em rnowling@gmail.com
c 954.496.2314

Re: Spark on Mesos 0.20

Posted by Fairiz Azizi <co...@gmail.com>.
That's what great about Spark, the community is so active! :)

I compiled Mesos 0.20.1 from the source tarball.

Using the Mapr3 Spark 1.1.0 distribution from the Spark downloads page
 (spark-1.1.0-bin-mapr3.tgz).

I see no problems for the workloads we are trying.

However, the cluster is small (less than 100 cores across 3 nodes).

The workloads reads in just a few gigabytes from HDFS, via an ipython
notebook spark shell.

thanks,
Fi



Fairiz "Fi" Azizi

On Mon, Oct 6, 2014 at 9:20 AM, Timothy Chen <tn...@gmail.com> wrote:

> Ok I created SPARK-3817 to track this, will try to repro it as well.
>
> Tim
>
> On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling <rn...@gmail.com> wrote:
> > I've recently run into this issue as well. I get it from running Spark
> > examples such as log query.  Maybe that'll help reproduce the issue.
> >
> >
> > On Monday, October 6, 2014, Gurvinder Singh <gu...@uninett.no>
> > wrote:
> >>
> >> The issue does not occur if the task at hand has small number of map
> >> tasks. I have a task which has 978 map tasks and I see this error as
> >>
> >> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different block
> >> manager registrations on 20140711-081617-711206558-5050-2543-5
> >>
> >> Here is the log from the mesos-slave where this container was running.
> >>
> >> http://pastebin.com/Q1Cuzm6Q
> >>
> >> If you look for the code from where error produced by spark, you will
> >> see that it simply exit and saying in comments "this should never
> >> happen, lets just quit" :-)
> >>
> >> - Gurvinder
> >> On 10/06/2014 09:30 AM, Timothy Chen wrote:
> >> > (Hit enter too soon...)
> >> >
> >> > What is your setup and steps to repro this?
> >> >
> >> > Tim
> >> >
> >> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com>
> wrote:
> >> >> Hi Gurvinder,
> >> >>
> >> >> I tried fine grain mode before and didn't get into that problem.
> >> >>
> >> >>
> >> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
> >> >> <gu...@uninett.no> wrote:
> >> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
> >> >>>> The Spark online docs indicate that Spark is compatible with Mesos
> >> >>>> 0.18.1
> >> >>>>
> >> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
> >> >>>>
> >> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
> >> >>>> v0.20.0?
> >> >>>>
> >> >>>> -Fi
> >> >>>>
> >> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in
> >> >>> coarse
> >> >>> mode, in fine grain mode there is an issue with blockmanager names
> >> >>> conflict. I have been waiting for it to be fixed but it is still
> >> >>> there.
> >> >>>
> >> >>> -Gurvinder
> >> >>>
> >> >>>
> ---------------------------------------------------------------------
> >> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
> >> >>> For additional commands, e-mail: dev-help@spark.apache.org
> >> >>>
> >>
> >>
> >> ---------------------------------------------------------------------
> >> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
> >> For additional commands, e-mail: dev-help@spark.apache.org
> >>
> >
> >
> > --
> > em rnowling@gmail.com
> > c 954.496.2314
>

Re: Spark on Mesos 0.20

Posted by Timothy Chen <tn...@gmail.com>.
Ok I created SPARK-3817 to track this, will try to repro it as well.

Tim

On Mon, Oct 6, 2014 at 6:08 AM, RJ Nowling <rn...@gmail.com> wrote:
> I've recently run into this issue as well. I get it from running Spark
> examples such as log query.  Maybe that'll help reproduce the issue.
>
>
> On Monday, October 6, 2014, Gurvinder Singh <gu...@uninett.no>
> wrote:
>>
>> The issue does not occur if the task at hand has small number of map
>> tasks. I have a task which has 978 map tasks and I see this error as
>>
>> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different block
>> manager registrations on 20140711-081617-711206558-5050-2543-5
>>
>> Here is the log from the mesos-slave where this container was running.
>>
>> http://pastebin.com/Q1Cuzm6Q
>>
>> If you look for the code from where error produced by spark, you will
>> see that it simply exit and saying in comments "this should never
>> happen, lets just quit" :-)
>>
>> - Gurvinder
>> On 10/06/2014 09:30 AM, Timothy Chen wrote:
>> > (Hit enter too soon...)
>> >
>> > What is your setup and steps to repro this?
>> >
>> > Tim
>> >
>> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com> wrote:
>> >> Hi Gurvinder,
>> >>
>> >> I tried fine grain mode before and didn't get into that problem.
>> >>
>> >>
>> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>> >> <gu...@uninett.no> wrote:
>> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>> >>>> The Spark online docs indicate that Spark is compatible with Mesos
>> >>>> 0.18.1
>> >>>>
>> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>> >>>>
>> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
>> >>>> v0.20.0?
>> >>>>
>> >>>> -Fi
>> >>>>
>> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in
>> >>> coarse
>> >>> mode, in fine grain mode there is an issue with blockmanager names
>> >>> conflict. I have been waiting for it to be fixed but it is still
>> >>> there.
>> >>>
>> >>> -Gurvinder
>> >>>
>> >>> ---------------------------------------------------------------------
>> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>> >>> For additional commands, e-mail: dev-help@spark.apache.org
>> >>>
>>
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>> For additional commands, e-mail: dev-help@spark.apache.org
>>
>
>
> --
> em rnowling@gmail.com
> c 954.496.2314

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org


Re: Spark on Mesos 0.20

Posted by RJ Nowling <rn...@gmail.com>.
I've recently run into this issue as well. I get it from running Spark
examples such as log query.  Maybe that'll help reproduce the issue.

On Monday, October 6, 2014, Gurvinder Singh <gu...@uninett.no>
wrote:

> The issue does not occur if the task at hand has small number of map
> tasks. I have a task which has 978 map tasks and I see this error as
>
> 14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different block
> manager registrations on 20140711-081617-711206558-5050-2543-5
>
> Here is the log from the mesos-slave where this container was running.
>
> http://pastebin.com/Q1Cuzm6Q
>
> If you look for the code from where error produced by spark, you will
> see that it simply exit and saying in comments "this should never
> happen, lets just quit" :-)
>
> - Gurvinder
> On 10/06/2014 09:30 AM, Timothy Chen wrote:
> > (Hit enter too soon...)
> >
> > What is your setup and steps to repro this?
> >
> > Tim
> >
> > On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tnachen@gmail.com
> <javascript:;>> wrote:
> >> Hi Gurvinder,
> >>
> >> I tried fine grain mode before and didn't get into that problem.
> >>
> >>
> >> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
> >> <gurvinder.singh@uninett.no <javascript:;>> wrote:
> >>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
> >>>> The Spark online docs indicate that Spark is compatible with Mesos
> 0.18.1
> >>>>
> >>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
> >>>>
> >>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos
> v0.20.0?
> >>>>
> >>>> -Fi
> >>>>
> >>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in coarse
> >>> mode, in fine grain mode there is an issue with blockmanager names
> >>> conflict. I have been waiting for it to be fixed but it is still there.
> >>>
> >>> -Gurvinder
> >>>
> >>> ---------------------------------------------------------------------
> >>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
> <javascript:;>
> >>> For additional commands, e-mail: dev-help@spark.apache.org
> <javascript:;>
> >>>
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org <javascript:;>
> For additional commands, e-mail: dev-help@spark.apache.org <javascript:;>
>
>

-- 
em rnowling@gmail.com
c 954.496.2314

Re: Spark on Mesos 0.20

Posted by Gurvinder Singh <gu...@uninett.no>.
The issue does not occur if the task at hand has small number of map
tasks. I have a task which has 978 map tasks and I see this error as

14/10/06 09:34:40 ERROR BlockManagerMasterActor: Got two different block
manager registrations on 20140711-081617-711206558-5050-2543-5

Here is the log from the mesos-slave where this container was running.

http://pastebin.com/Q1Cuzm6Q

If you look for the code from where error produced by spark, you will
see that it simply exit and saying in comments "this should never
happen, lets just quit" :-)

- Gurvinder
On 10/06/2014 09:30 AM, Timothy Chen wrote:
> (Hit enter too soon...)
> 
> What is your setup and steps to repro this?
> 
> Tim
> 
> On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com> wrote:
>> Hi Gurvinder,
>>
>> I tried fine grain mode before and didn't get into that problem.
>>
>>
>> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
>> <gu...@uninett.no> wrote:
>>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>>>> The Spark online docs indicate that Spark is compatible with Mesos 0.18.1
>>>>
>>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>>>>
>>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos v0.20.0?
>>>>
>>>> -Fi
>>>>
>>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in coarse
>>> mode, in fine grain mode there is an issue with blockmanager names
>>> conflict. I have been waiting for it to be fixed but it is still there.
>>>
>>> -Gurvinder
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>>> For additional commands, e-mail: dev-help@spark.apache.org
>>>


---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org


Re: Spark on Mesos 0.20

Posted by Timothy Chen <tn...@gmail.com>.
(Hit enter too soon...)

What is your setup and steps to repro this?

Tim

On Mon, Oct 6, 2014 at 12:30 AM, Timothy Chen <tn...@gmail.com> wrote:
> Hi Gurvinder,
>
> I tried fine grain mode before and didn't get into that problem.
>
>
> On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
> <gu...@uninett.no> wrote:
>> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>>> The Spark online docs indicate that Spark is compatible with Mesos 0.18.1
>>>
>>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>>>
>>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos v0.20.0?
>>>
>>> -Fi
>>>
>> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in coarse
>> mode, in fine grain mode there is an issue with blockmanager names
>> conflict. I have been waiting for it to be fixed but it is still there.
>>
>> -Gurvinder
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
>> For additional commands, e-mail: dev-help@spark.apache.org
>>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org


Re: Spark on Mesos 0.20

Posted by Timothy Chen <tn...@gmail.com>.
Hi Gurvinder,

I tried fine grain mode before and didn't get into that problem.


On Sun, Oct 5, 2014 at 11:44 PM, Gurvinder Singh
<gu...@uninett.no> wrote:
> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
>> The Spark online docs indicate that Spark is compatible with Mesos 0.18.1
>>
>> I've gotten it to work just fine on 0.18.1 and 0.18.2
>>
>> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos v0.20.0?
>>
>> -Fi
>>
> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in coarse
> mode, in fine grain mode there is an issue with blockmanager names
> conflict. I have been waiting for it to be fixed but it is still there.
>
> -Gurvinder
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
> For additional commands, e-mail: dev-help@spark.apache.org
>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org


Re: Spark on Mesos 0.20

Posted by Andrew Ash <an...@andrewash.com>.
Hi Gurvinder,

Is there a SPARK ticket tracking the issue you describe?

On Mon, Oct 6, 2014 at 2:44 AM, Gurvinder Singh <gu...@uninett.no>
wrote:

> On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
> > The Spark online docs indicate that Spark is compatible with Mesos 0.18.1
> >
> > I've gotten it to work just fine on 0.18.1 and 0.18.2
> >
> > Has anyone tried Spark on a newer version of Mesos, i.e. Mesos v0.20.0?
> >
> > -Fi
> >
> Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in coarse
> mode, in fine grain mode there is an issue with blockmanager names
> conflict. I have been waiting for it to be fixed but it is still there.
>
> -Gurvinder
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
> For additional commands, e-mail: dev-help@spark.apache.org
>
>

Re: Spark on Mesos 0.20

Posted by Gurvinder Singh <gu...@uninett.no>.
On 10/06/2014 08:19 AM, Fairiz Azizi wrote:
> The Spark online docs indicate that Spark is compatible with Mesos 0.18.1
> 
> I've gotten it to work just fine on 0.18.1 and 0.18.2
> 
> Has anyone tried Spark on a newer version of Mesos, i.e. Mesos v0.20.0?
> 
> -Fi
> 
Yeah we are using Spark 1.1.0 with Mesos 0.20.1. It runs fine in coarse
mode, in fine grain mode there is an issue with blockmanager names
conflict. I have been waiting for it to be fixed but it is still there.

-Gurvinder

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org