You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by Alessandro Baretta <al...@gmail.com> on 2015/01/11 01:40:43 UTC

Job priority

Is it possible to specify a priority level for a job, such that the active
jobs might be scheduled in order of priority?

Alex

Re: Job priority

Posted by Alessandro Baretta <al...@gmail.com>.
Ok, will do.

Thanks for providing some context on this topic.

Alex

On Sun, Jan 11, 2015 at 8:34 PM, Patrick Wendell <pw...@gmail.com> wrote:

> Priority scheduling isn't something we've supported in Spark and we've
> opted to support FIFO and Fair scheduling and asked users to try and
> fit these to the needs of their applications.
>
> In practice from what I've seen of priority schedulers, such as the
> linux CPU scheduler, is that strict priority scheduling is never used
> in practice because of priority starvation and other issues. So you
> have this second tier of heuristics that exist to deal with issues
> like starvation, priority inversion, etc, and these become very
> complex over time.
>
> That said, I looked a this a bit with @kayousterhout and I don't think
> it would be very hard to implement a simple priority scheduler in the
> current architecture. My main concern would be additional complexity
> that would develop over time, based on looking at previous
> implementations in the wild.
>
> Alessandro, would you be able to open a JIRA and list some of your
> requirements there? That way we could hear whether other people have
> similar needs.
>
> - Patrick
>
> On Sun, Jan 11, 2015 at 10:07 AM, Mark Hamstra <ma...@clearstorydata.com>
> wrote:
> > Yes, if you are asking about developing a new priority queue job
> scheduling
> > feature and not just about how job scheduling currently works in Spark,
> the
> > that's a dev list issue.  The current job scheduling priority is at the
> > granularity of pools containing jobs, not the jobs themselves; so if you
> > require strictly job-level priority queuing, that would require a new
> > development effort -- and one that I expect will involve a lot of tricky
> > corner cases.
> >
> > Sorry for misreading the nature of your initial inquiry.
> >
> > On Sun, Jan 11, 2015 at 7:36 AM, Alessandro Baretta <
> alexbaretta@gmail.com>
> > wrote:
> >
> >> Cody,
> >>
> >> While I might be able to improve the scheduling of my jobs by using a
> few
> >> different pools with weights equal to, say, 1, 1e3 and 1e6, effectively
> >> getting a small handful of priority classes. Still, this is really not
> >> quite what I am describing. This is why my original post was on the dev
> >> list. Let me then ask if there is any interest in having priority queue
> job
> >> scheduling in Spark. This is something I might be able to pull off.
> >>
> >> Alex
> >>
> >> On Sun, Jan 11, 2015 at 6:21 AM, Cody Koeninger <co...@koeninger.org>
> >> wrote:
> >>
> >>> If you set up a number of pools equal to the number of different
> priority
> >>> levels you want, make the relative weights of those pools very
> different,
> >>> and submit a job to the pool representing its priority, I think youll
> get
> >>> behavior equivalent to a priority queue. Try it and see.
> >>>
> >>> If I'm misunderstandng what youre trying to do, then I don't know.
> >>>
> >>>
> >>> On Sunday, January 11, 2015, Alessandro Baretta <alexbaretta@gmail.com
> >
> >>> wrote:
> >>>
> >>>> Cody,
> >>>>
> >>>> Maybe I'm not getting this, but it doesn't look like this page is
> >>>> describing a priority queue scheduling policy. What this section
> discusses
> >>>> is how resources are shared between queues. A weight-1000 pool will
> get
> >>>> 1000 times more resources allocated to it than a priority 1 queue.
> Great,
> >>>> but not what I want. I want to be able to define an Ordering on make
> my
> >>>> tasks representing their priority, and have Spark allocate all
> resources to
> >>>> the job that has the highest priority.
> >>>>
> >>>> Alex
> >>>>
> >>>> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <co...@koeninger.org>
> >>>> wrote:
> >>>>
> >>>>>
> >>>>>
> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
> >>>>>
> >>>>> "Setting a high weight such as 1000 also makes it possible to
> >>>>> implement *priority* between pools--in essence, the weight-1000 pool
> >>>>> will always get to launch tasks first whenever it has jobs active."
> >>>>>
> >>>>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
> >>>>> alexbaretta@gmail.com> wrote:
> >>>>>
> >>>>>> Mark,
> >>>>>>
> >>>>>> Thanks, but I don't see how this documentation solves my problem.
> You
> >>>>>> are referring me to documentation of fair scheduling; whereas, I am
> asking
> >>>>>> about as unfair a scheduling policy as can be: a priority queue.
> >>>>>>
> >>>>>> Alex
> >>>>>>
> >>>>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <
> mark@clearstorydata.com
> >>>>>> > wrote:
> >>>>>>
> >>>>>>> -dev, +user
> >>>>>>>
> >>>>>>> http://spark.apache.org/docs/latest/job-scheduling.html
> >>>>>>>
> >>>>>>>
> >>>>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
> >>>>>>> alexbaretta@gmail.com> wrote:
> >>>>>>>
> >>>>>>>> Is it possible to specify a priority level for a job, such that
> the
> >>>>>>>> active
> >>>>>>>> jobs might be scheduled in order of priority?
> >>>>>>>>
> >>>>>>>> Alex
> >>>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>
> >>
>

Re: Job priority

Posted by Patrick Wendell <pw...@gmail.com>.
Priority scheduling isn't something we've supported in Spark and we've
opted to support FIFO and Fair scheduling and asked users to try and
fit these to the needs of their applications.

In practice from what I've seen of priority schedulers, such as the
linux CPU scheduler, is that strict priority scheduling is never used
in practice because of priority starvation and other issues. So you
have this second tier of heuristics that exist to deal with issues
like starvation, priority inversion, etc, and these become very
complex over time.

That said, I looked a this a bit with @kayousterhout and I don't think
it would be very hard to implement a simple priority scheduler in the
current architecture. My main concern would be additional complexity
that would develop over time, based on looking at previous
implementations in the wild.

Alessandro, would you be able to open a JIRA and list some of your
requirements there? That way we could hear whether other people have
similar needs.

- Patrick

On Sun, Jan 11, 2015 at 10:07 AM, Mark Hamstra <ma...@clearstorydata.com> wrote:
> Yes, if you are asking about developing a new priority queue job scheduling
> feature and not just about how job scheduling currently works in Spark, the
> that's a dev list issue.  The current job scheduling priority is at the
> granularity of pools containing jobs, not the jobs themselves; so if you
> require strictly job-level priority queuing, that would require a new
> development effort -- and one that I expect will involve a lot of tricky
> corner cases.
>
> Sorry for misreading the nature of your initial inquiry.
>
> On Sun, Jan 11, 2015 at 7:36 AM, Alessandro Baretta <al...@gmail.com>
> wrote:
>
>> Cody,
>>
>> While I might be able to improve the scheduling of my jobs by using a few
>> different pools with weights equal to, say, 1, 1e3 and 1e6, effectively
>> getting a small handful of priority classes. Still, this is really not
>> quite what I am describing. This is why my original post was on the dev
>> list. Let me then ask if there is any interest in having priority queue job
>> scheduling in Spark. This is something I might be able to pull off.
>>
>> Alex
>>
>> On Sun, Jan 11, 2015 at 6:21 AM, Cody Koeninger <co...@koeninger.org>
>> wrote:
>>
>>> If you set up a number of pools equal to the number of different priority
>>> levels you want, make the relative weights of those pools very different,
>>> and submit a job to the pool representing its priority, I think youll get
>>> behavior equivalent to a priority queue. Try it and see.
>>>
>>> If I'm misunderstandng what youre trying to do, then I don't know.
>>>
>>>
>>> On Sunday, January 11, 2015, Alessandro Baretta <al...@gmail.com>
>>> wrote:
>>>
>>>> Cody,
>>>>
>>>> Maybe I'm not getting this, but it doesn't look like this page is
>>>> describing a priority queue scheduling policy. What this section discusses
>>>> is how resources are shared between queues. A weight-1000 pool will get
>>>> 1000 times more resources allocated to it than a priority 1 queue. Great,
>>>> but not what I want. I want to be able to define an Ordering on make my
>>>> tasks representing their priority, and have Spark allocate all resources to
>>>> the job that has the highest priority.
>>>>
>>>> Alex
>>>>
>>>> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <co...@koeninger.org>
>>>> wrote:
>>>>
>>>>>
>>>>> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
>>>>>
>>>>> "Setting a high weight such as 1000 also makes it possible to
>>>>> implement *priority* between pools--in essence, the weight-1000 pool
>>>>> will always get to launch tasks first whenever it has jobs active."
>>>>>
>>>>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
>>>>> alexbaretta@gmail.com> wrote:
>>>>>
>>>>>> Mark,
>>>>>>
>>>>>> Thanks, but I don't see how this documentation solves my problem. You
>>>>>> are referring me to documentation of fair scheduling; whereas, I am asking
>>>>>> about as unfair a scheduling policy as can be: a priority queue.
>>>>>>
>>>>>> Alex
>>>>>>
>>>>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <mark@clearstorydata.com
>>>>>> > wrote:
>>>>>>
>>>>>>> -dev, +user
>>>>>>>
>>>>>>> http://spark.apache.org/docs/latest/job-scheduling.html
>>>>>>>
>>>>>>>
>>>>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>>>>>>> alexbaretta@gmail.com> wrote:
>>>>>>>
>>>>>>>> Is it possible to specify a priority level for a job, such that the
>>>>>>>> active
>>>>>>>> jobs might be scheduled in order of priority?
>>>>>>>>
>>>>>>>> Alex
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@spark.apache.org
For additional commands, e-mail: dev-help@spark.apache.org


Re: Job priority

Posted by Mark Hamstra <ma...@clearstorydata.com>.
Yes, if you are asking about developing a new priority queue job scheduling
feature and not just about how job scheduling currently works in Spark, the
that's a dev list issue.  The current job scheduling priority is at the
granularity of pools containing jobs, not the jobs themselves; so if you
require strictly job-level priority queuing, that would require a new
development effort -- and one that I expect will involve a lot of tricky
corner cases.

Sorry for misreading the nature of your initial inquiry.

On Sun, Jan 11, 2015 at 7:36 AM, Alessandro Baretta <al...@gmail.com>
wrote:

> Cody,
>
> While I might be able to improve the scheduling of my jobs by using a few
> different pools with weights equal to, say, 1, 1e3 and 1e6, effectively
> getting a small handful of priority classes. Still, this is really not
> quite what I am describing. This is why my original post was on the dev
> list. Let me then ask if there is any interest in having priority queue job
> scheduling in Spark. This is something I might be able to pull off.
>
> Alex
>
> On Sun, Jan 11, 2015 at 6:21 AM, Cody Koeninger <co...@koeninger.org>
> wrote:
>
>> If you set up a number of pools equal to the number of different priority
>> levels you want, make the relative weights of those pools very different,
>> and submit a job to the pool representing its priority, I think youll get
>> behavior equivalent to a priority queue. Try it and see.
>>
>> If I'm misunderstandng what youre trying to do, then I don't know.
>>
>>
>> On Sunday, January 11, 2015, Alessandro Baretta <al...@gmail.com>
>> wrote:
>>
>>> Cody,
>>>
>>> Maybe I'm not getting this, but it doesn't look like this page is
>>> describing a priority queue scheduling policy. What this section discusses
>>> is how resources are shared between queues. A weight-1000 pool will get
>>> 1000 times more resources allocated to it than a priority 1 queue. Great,
>>> but not what I want. I want to be able to define an Ordering on make my
>>> tasks representing their priority, and have Spark allocate all resources to
>>> the job that has the highest priority.
>>>
>>> Alex
>>>
>>> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <co...@koeninger.org>
>>> wrote:
>>>
>>>>
>>>> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
>>>>
>>>> "Setting a high weight such as 1000 also makes it possible to
>>>> implement *priority* between pools—in essence, the weight-1000 pool
>>>> will always get to launch tasks first whenever it has jobs active."
>>>>
>>>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
>>>> alexbaretta@gmail.com> wrote:
>>>>
>>>>> Mark,
>>>>>
>>>>> Thanks, but I don't see how this documentation solves my problem. You
>>>>> are referring me to documentation of fair scheduling; whereas, I am asking
>>>>> about as unfair a scheduling policy as can be: a priority queue.
>>>>>
>>>>> Alex
>>>>>
>>>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <mark@clearstorydata.com
>>>>> > wrote:
>>>>>
>>>>>> -dev, +user
>>>>>>
>>>>>> http://spark.apache.org/docs/latest/job-scheduling.html
>>>>>>
>>>>>>
>>>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>>>>>> alexbaretta@gmail.com> wrote:
>>>>>>
>>>>>>> Is it possible to specify a priority level for a job, such that the
>>>>>>> active
>>>>>>> jobs might be scheduled in order of priority?
>>>>>>>
>>>>>>> Alex
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>

Re: Job priority

Posted by Alessandro Baretta <al...@gmail.com>.
Cody,

While I might be able to improve the scheduling of my jobs by using a few
different pools with weights equal to, say, 1, 1e3 and 1e6, effectively
getting a small handful of priority classes. Still, this is really not
quite what I am describing. This is why my original post was on the dev
list. Let me then ask if there is any interest in having priority queue job
scheduling in Spark. This is something I might be able to pull off.

Alex

On Sun, Jan 11, 2015 at 6:21 AM, Cody Koeninger <co...@koeninger.org> wrote:

> If you set up a number of pools equal to the number of different priority
> levels you want, make the relative weights of those pools very different,
> and submit a job to the pool representing its priority, I think youll get
> behavior equivalent to a priority queue. Try it and see.
>
> If I'm misunderstandng what youre trying to do, then I don't know.
>
>
> On Sunday, January 11, 2015, Alessandro Baretta <al...@gmail.com>
> wrote:
>
>> Cody,
>>
>> Maybe I'm not getting this, but it doesn't look like this page is
>> describing a priority queue scheduling policy. What this section discusses
>> is how resources are shared between queues. A weight-1000 pool will get
>> 1000 times more resources allocated to it than a priority 1 queue. Great,
>> but not what I want. I want to be able to define an Ordering on make my
>> tasks representing their priority, and have Spark allocate all resources to
>> the job that has the highest priority.
>>
>> Alex
>>
>> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <co...@koeninger.org>
>> wrote:
>>
>>>
>>> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
>>>
>>> "Setting a high weight such as 1000 also makes it possible to implement
>>> *priority* between pools—in essence, the weight-1000 pool will always
>>> get to launch tasks first whenever it has jobs active."
>>>
>>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
>>> alexbaretta@gmail.com> wrote:
>>>
>>>> Mark,
>>>>
>>>> Thanks, but I don't see how this documentation solves my problem. You
>>>> are referring me to documentation of fair scheduling; whereas, I am asking
>>>> about as unfair a scheduling policy as can be: a priority queue.
>>>>
>>>> Alex
>>>>
>>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <ma...@clearstorydata.com>
>>>> wrote:
>>>>
>>>>> -dev, +user
>>>>>
>>>>> http://spark.apache.org/docs/latest/job-scheduling.html
>>>>>
>>>>>
>>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>>>>> alexbaretta@gmail.com> wrote:
>>>>>
>>>>>> Is it possible to specify a priority level for a job, such that the
>>>>>> active
>>>>>> jobs might be scheduled in order of priority?
>>>>>>
>>>>>> Alex
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>

Re: Job priority

Posted by Alessandro Baretta <al...@gmail.com>.
Cody,

While I might be able to improve the scheduling of my jobs by using a few
different pools with weights equal to, say, 1, 1e3 and 1e6, effectively
getting a small handful of priority classes. Still, this is really not
quite what I am describing. This is why my original post was on the dev
list. Let me then ask if there is any interest in having priority queue job
scheduling in Spark. This is something I might be able to pull off.

Alex

On Sun, Jan 11, 2015 at 6:21 AM, Cody Koeninger <co...@koeninger.org> wrote:

> If you set up a number of pools equal to the number of different priority
> levels you want, make the relative weights of those pools very different,
> and submit a job to the pool representing its priority, I think youll get
> behavior equivalent to a priority queue. Try it and see.
>
> If I'm misunderstandng what youre trying to do, then I don't know.
>
>
> On Sunday, January 11, 2015, Alessandro Baretta <al...@gmail.com>
> wrote:
>
>> Cody,
>>
>> Maybe I'm not getting this, but it doesn't look like this page is
>> describing a priority queue scheduling policy. What this section discusses
>> is how resources are shared between queues. A weight-1000 pool will get
>> 1000 times more resources allocated to it than a priority 1 queue. Great,
>> but not what I want. I want to be able to define an Ordering on make my
>> tasks representing their priority, and have Spark allocate all resources to
>> the job that has the highest priority.
>>
>> Alex
>>
>> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <co...@koeninger.org>
>> wrote:
>>
>>>
>>> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
>>>
>>> "Setting a high weight such as 1000 also makes it possible to implement
>>> *priority* between pools—in essence, the weight-1000 pool will always
>>> get to launch tasks first whenever it has jobs active."
>>>
>>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
>>> alexbaretta@gmail.com> wrote:
>>>
>>>> Mark,
>>>>
>>>> Thanks, but I don't see how this documentation solves my problem. You
>>>> are referring me to documentation of fair scheduling; whereas, I am asking
>>>> about as unfair a scheduling policy as can be: a priority queue.
>>>>
>>>> Alex
>>>>
>>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <ma...@clearstorydata.com>
>>>> wrote:
>>>>
>>>>> -dev, +user
>>>>>
>>>>> http://spark.apache.org/docs/latest/job-scheduling.html
>>>>>
>>>>>
>>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>>>>> alexbaretta@gmail.com> wrote:
>>>>>
>>>>>> Is it possible to specify a priority level for a job, such that the
>>>>>> active
>>>>>> jobs might be scheduled in order of priority?
>>>>>>
>>>>>> Alex
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>

Re: Job priority

Posted by Cody Koeninger <co...@koeninger.org>.
If you set up a number of pools equal to the number of different priority
levels you want, make the relative weights of those pools very different,
and submit a job to the pool representing its priority, I think youll get
behavior equivalent to a priority queue. Try it and see.

If I'm misunderstandng what youre trying to do, then I don't know.

On Sunday, January 11, 2015, Alessandro Baretta <al...@gmail.com>
wrote:

> Cody,
>
> Maybe I'm not getting this, but it doesn't look like this page is
> describing a priority queue scheduling policy. What this section discusses
> is how resources are shared between queues. A weight-1000 pool will get
> 1000 times more resources allocated to it than a priority 1 queue. Great,
> but not what I want. I want to be able to define an Ordering on make my
> tasks representing their priority, and have Spark allocate all resources to
> the job that has the highest priority.
>
> Alex
>
> On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <cody@koeninger.org
> <javascript:_e(%7B%7D,'cvml','cody@koeninger.org');>> wrote:
>
>>
>> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
>>
>> "Setting a high weight such as 1000 also makes it possible to implement
>> *priority* between pools—in essence, the weight-1000 pool will always
>> get to launch tasks first whenever it has jobs active."
>>
>> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
>> alexbaretta@gmail.com
>> <javascript:_e(%7B%7D,'cvml','alexbaretta@gmail.com');>> wrote:
>>
>>> Mark,
>>>
>>> Thanks, but I don't see how this documentation solves my problem. You
>>> are referring me to documentation of fair scheduling; whereas, I am asking
>>> about as unfair a scheduling policy as can be: a priority queue.
>>>
>>> Alex
>>>
>>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <mark@clearstorydata.com
>>> <javascript:_e(%7B%7D,'cvml','mark@clearstorydata.com');>> wrote:
>>>
>>>> -dev, +user
>>>>
>>>> http://spark.apache.org/docs/latest/job-scheduling.html
>>>>
>>>>
>>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>>>> alexbaretta@gmail.com
>>>> <javascript:_e(%7B%7D,'cvml','alexbaretta@gmail.com');>> wrote:
>>>>
>>>>> Is it possible to specify a priority level for a job, such that the
>>>>> active
>>>>> jobs might be scheduled in order of priority?
>>>>>
>>>>> Alex
>>>>>
>>>>
>>>>
>>>
>>
>

Re: Job priority

Posted by Alessandro Baretta <al...@gmail.com>.
Cody,

Maybe I'm not getting this, but it doesn't look like this page is
describing a priority queue scheduling policy. What this section discusses
is how resources are shared between queues. A weight-1000 pool will get
1000 times more resources allocated to it than a priority 1 queue. Great,
but not what I want. I want to be able to define an Ordering on make my
tasks representing their priority, and have Spark allocate all resources to
the job that has the highest priority.

Alex

On Sat, Jan 10, 2015 at 10:11 PM, Cody Koeninger <co...@koeninger.org> wrote:

>
> http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties
>
> "Setting a high weight such as 1000 also makes it possible to implement
> *priority* between pools—in essence, the weight-1000 pool will always get
> to launch tasks first whenever it has jobs active."
>
> On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <
> alexbaretta@gmail.com> wrote:
>
>> Mark,
>>
>> Thanks, but I don't see how this documentation solves my problem. You are
>> referring me to documentation of fair scheduling; whereas, I am asking
>> about as unfair a scheduling policy as can be: a priority queue.
>>
>> Alex
>>
>> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <ma...@clearstorydata.com>
>> wrote:
>>
>>> -dev, +user
>>>
>>> http://spark.apache.org/docs/latest/job-scheduling.html
>>>
>>>
>>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>>> alexbaretta@gmail.com> wrote:
>>>
>>>> Is it possible to specify a priority level for a job, such that the
>>>> active
>>>> jobs might be scheduled in order of priority?
>>>>
>>>> Alex
>>>>
>>>
>>>
>>
>

Re: Job priority

Posted by Cody Koeninger <co...@koeninger.org>.
http://spark.apache.org/docs/latest/job-scheduling.html#configuring-pool-properties

"Setting a high weight such as 1000 also makes it possible to implement
*priority* between pools—in essence, the weight-1000 pool will always get
to launch tasks first whenever it has jobs active."

On Sat, Jan 10, 2015 at 11:57 PM, Alessandro Baretta <al...@gmail.com>
wrote:

> Mark,
>
> Thanks, but I don't see how this documentation solves my problem. You are
> referring me to documentation of fair scheduling; whereas, I am asking
> about as unfair a scheduling policy as can be: a priority queue.
>
> Alex
>
> On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <ma...@clearstorydata.com>
> wrote:
>
>> -dev, +user
>>
>> http://spark.apache.org/docs/latest/job-scheduling.html
>>
>>
>> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <
>> alexbaretta@gmail.com> wrote:
>>
>>> Is it possible to specify a priority level for a job, such that the
>>> active
>>> jobs might be scheduled in order of priority?
>>>
>>> Alex
>>>
>>
>>
>

Re: Job priority

Posted by Alessandro Baretta <al...@gmail.com>.
Mark,

Thanks, but I don't see how this documentation solves my problem. You are
referring me to documentation of fair scheduling; whereas, I am asking
about as unfair a scheduling policy as can be: a priority queue.

Alex

On Sat, Jan 10, 2015 at 5:00 PM, Mark Hamstra <ma...@clearstorydata.com>
wrote:

> -dev, +user
>
> http://spark.apache.org/docs/latest/job-scheduling.html
>
>
> On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <alexbaretta@gmail.com
> > wrote:
>
>> Is it possible to specify a priority level for a job, such that the active
>> jobs might be scheduled in order of priority?
>>
>> Alex
>>
>
>

Re: Job priority

Posted by Mark Hamstra <ma...@clearstorydata.com>.
-dev, +user

http://spark.apache.org/docs/latest/job-scheduling.html


On Sat, Jan 10, 2015 at 4:40 PM, Alessandro Baretta <al...@gmail.com>
wrote:

> Is it possible to specify a priority level for a job, such that the active
> jobs might be scheduled in order of priority?
>
> Alex
>