You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@mesos.apache.org by Benjamin Mahler <be...@gmail.com> on 2015/05/14 02:03:12 UTC

Re: [jira] [Updated] (MESOS-2020) mesos should send docker failure messages to scheduler

+tim, jake

What does pending closed mean? I noticed that this had become the default
resolution yesterday. Did something change Jake?

Tim, should this be resolved?

On Tue, May 12, 2015 at 12:28 AM, Timothy Chen (JIRA) <ji...@apache.org>
wrote:

>
>      [
> https://issues.apache.org/jira/browse/MESOS-2020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
> ]
>
> Timothy Chen updated MESOS-2020:
> --------------------------------
>     Fix Version/s: 0.23.0
>
> > mesos should send docker failure messages to scheduler
> > ------------------------------------------------------
> >
> >                 Key: MESOS-2020
> >                 URL: https://issues.apache.org/jira/browse/MESOS-2020
> >             Project: Mesos
> >          Issue Type: Bug
> >          Components: docker, slave
> >    Affects Versions: 0.20.1
> >         Environment: mesos-slaves running in docker
> >            Reporter: Ian Babrou
> >            Assignee: Jay Buffington
> >             Fix For: 0.23.0
> >
> >
> > I tried to start container that cannot actually start for some reason,
> like this (from slave logs):
> > {"log":"I1031 08:58:02.692958    60 slave.cpp:1112] Launching task
> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 for framework
> 20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.692986684Z"}
> > {"log":"I1031 08:58:02.693696    60 slave.cpp:1222] Queuing task
> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' for executor
> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 of framework
> '20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.693734418Z"}
> > {"log":"I1031 08:58:02.695612    64 docker.cpp:743] Starting container
> '02a786c9-556c-4ed9-80d2-1850a49030fe' for task
> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' (and executor
> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799') of framework
> '20141003-172543-3892422848-5050-1-0000'\n","stream":"stderr","time":"2014-10-31T08:58:02.6956675Z"}
> > {"log":"E1031 08:58:05.272902    65 slave.cpp:2485] Container
> '02a786c9-556c-4ed9-80d2-1850a49030fe' for executor
> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' of framework
> '20141003-172543-3892422848-5050-1-0000' failed to start: Failed to 'docker
> run -d -c 204 -m 67108864 -e PORT=31459 -e PORT0=31459 -e PORTS=31459 -e
> HOST=web544 -e MESOS_SANDBOX=/mnt/mesos/sandbox -v
> /var/lib/mesos/slave/slaves/20141028-073834-3909200064-5050-1-75/frameworks/20141003-172543-3892422848-5050-1-0000/executors/topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799/runs/02a786c9-556c-4ed9-80d2-1850a49030fe:/mnt/mesos/sandbox
> --net host --name mesos-02a786c9-556c-4ed9-80d2-1850a49030fe
> docker.core.tf/demo:1': exit status = exited with status 1 stderr =
> 2014/10/31 08:58:04 Error response from daemon: Cannot start container
> 05666763bff98ad70f35968add3923018a9d457a1f4e9dab0981936841093d2f: exec:
> \"/server.js\": permission
> denied\n","stream":"stderr","time":"2014-10-31T08:58:05.273016253Z"}
> > But when I go to task sandbox from mesos ui, stdout and stderr are empty.
> > Marathon keeps scheduling tasks, but they all silently fail, this is
> very misleading.
>
>
>
> --
> This message was sent by Atlassian JIRA
> (v6.3.4#6332)
>

Re: [jira] [Updated] (MESOS-2020) mesos should send docker failure messages to scheduler

Posted by Benjamin Mahler <bm...@twitter.com.INVALID>.
Jake, seems like default resolution is "pending closed" now. Previously the
default resolution was "fixed". In this case, I re-opened to get it into
the "fixed" state.

Did the default resolution switch to "pending closed"?

https://confluence.atlassian.com/display/JIRA/Defining+Resolution+Field+Values

On Thu, May 14, 2015 at 6:16 AM, Jake Farrell <jf...@apache.org> wrote:

> Nothing in the workflow changed as far as i'm aware, looks like it was
> reopened and then set as resolved
>
> -Jake
>
>
> On Wed, May 13, 2015 at 9:13 PM, Timothy Chen <tn...@gmail.com> wrote:
>
> > I have no idea too, I wanted to resolve the ticket but it seems to be
> > stuck on this state.
> >
> > Tim
> >
> > > On May 13, 2015, at 5:03 PM, Benjamin Mahler <
> benjamin.mahler@gmail.com>
> > wrote:
> > >
> > > +tim, jake
> > >
> > > What does pending closed mean? I noticed that this had become the
> default
> > > resolution yesterday. Did something change Jake?
> > >
> > > Tim, should this be resolved?
> > >
> > > On Tue, May 12, 2015 at 12:28 AM, Timothy Chen (JIRA) <jira@apache.org
> >
> > > wrote:
> > >
> > >>
> > >>     [
> > >>
> >
> https://issues.apache.org/jira/browse/MESOS-2020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
> > >> ]
> > >>
> > >> Timothy Chen updated MESOS-2020:
> > >> --------------------------------
> > >>    Fix Version/s: 0.23.0
> > >>
> > >>> mesos should send docker failure messages to scheduler
> > >>> ------------------------------------------------------
> > >>>
> > >>>                Key: MESOS-2020
> > >>>                URL: https://issues.apache.org/jira/browse/MESOS-2020
> > >>>            Project: Mesos
> > >>>         Issue Type: Bug
> > >>>         Components: docker, slave
> > >>>   Affects Versions: 0.20.1
> > >>>        Environment: mesos-slaves running in docker
> > >>>           Reporter: Ian Babrou
> > >>>           Assignee: Jay Buffington
> > >>>            Fix For: 0.23.0
> > >>>
> > >>>
> > >>> I tried to start container that cannot actually start for some
> reason,
> > >> like this (from slave logs):
> > >>> {"log":"I1031 08:58:02.692958    60 slave.cpp:1112] Launching task
> > >> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 for framework
> > >>
> >
> 20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.692986684Z"}
> > >>> {"log":"I1031 08:58:02.693696    60 slave.cpp:1222] Queuing task
> > >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' for executor
> > >> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 of framework
> > >>
> >
> '20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.693734418Z"}
> > >>> {"log":"I1031 08:58:02.695612    64 docker.cpp:743] Starting
> container
> > >> '02a786c9-556c-4ed9-80d2-1850a49030fe' for task
> > >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' (and executor
> > >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799') of framework
> > >>
> >
> '20141003-172543-3892422848-5050-1-0000'\n","stream":"stderr","time":"2014-10-31T08:58:02.6956675Z"}
> > >>> {"log":"E1031 08:58:05.272902    65 slave.cpp:2485] Container
> > >> '02a786c9-556c-4ed9-80d2-1850a49030fe' for executor
> > >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' of framework
> > >> '20141003-172543-3892422848-5050-1-0000' failed to start: Failed to
> > 'docker
> > >> run -d -c 204 -m 67108864 -e PORT=31459 -e PORT0=31459 -e PORTS=31459
> -e
> > >> HOST=web544 -e MESOS_SANDBOX=/mnt/mesos/sandbox -v
> > >>
> >
> /var/lib/mesos/slave/slaves/20141028-073834-3909200064-5050-1-75/frameworks/20141003-172543-3892422848-5050-1-0000/executors/topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799/runs/02a786c9-556c-4ed9-80d2-1850a49030fe:/mnt/mesos/sandbox
> > >> --net host --name mesos-02a786c9-556c-4ed9-80d2-1850a49030fe
> > >> docker.core.tf/demo:1': exit status = exited with status 1 stderr =
> > >> 2014/10/31 08:58:04 Error response from daemon: Cannot start container
> > >> 05666763bff98ad70f35968add3923018a9d457a1f4e9dab0981936841093d2f:
> exec:
> > >> \"/server.js\": permission
> > >> denied\n","stream":"stderr","time":"2014-10-31T08:58:05.273016253Z"}
> > >>> But when I go to task sandbox from mesos ui, stdout and stderr are
> > empty.
> > >>> Marathon keeps scheduling tasks, but they all silently fail, this is
> > >> very misleading.
> > >>
> > >>
> > >>
> > >> --
> > >> This message was sent by Atlassian JIRA
> > >> (v6.3.4#6332)
> > >>
> >
>

Re: [jira] [Updated] (MESOS-2020) mesos should send docker failure messages to scheduler

Posted by Jake Farrell <jf...@apache.org>.
Nothing in the workflow changed as far as i'm aware, looks like it was
reopened and then set as resolved

-Jake


On Wed, May 13, 2015 at 9:13 PM, Timothy Chen <tn...@gmail.com> wrote:

> I have no idea too, I wanted to resolve the ticket but it seems to be
> stuck on this state.
>
> Tim
>
> > On May 13, 2015, at 5:03 PM, Benjamin Mahler <be...@gmail.com>
> wrote:
> >
> > +tim, jake
> >
> > What does pending closed mean? I noticed that this had become the default
> > resolution yesterday. Did something change Jake?
> >
> > Tim, should this be resolved?
> >
> > On Tue, May 12, 2015 at 12:28 AM, Timothy Chen (JIRA) <ji...@apache.org>
> > wrote:
> >
> >>
> >>     [
> >>
> https://issues.apache.org/jira/browse/MESOS-2020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
> >> ]
> >>
> >> Timothy Chen updated MESOS-2020:
> >> --------------------------------
> >>    Fix Version/s: 0.23.0
> >>
> >>> mesos should send docker failure messages to scheduler
> >>> ------------------------------------------------------
> >>>
> >>>                Key: MESOS-2020
> >>>                URL: https://issues.apache.org/jira/browse/MESOS-2020
> >>>            Project: Mesos
> >>>         Issue Type: Bug
> >>>         Components: docker, slave
> >>>   Affects Versions: 0.20.1
> >>>        Environment: mesos-slaves running in docker
> >>>           Reporter: Ian Babrou
> >>>           Assignee: Jay Buffington
> >>>            Fix For: 0.23.0
> >>>
> >>>
> >>> I tried to start container that cannot actually start for some reason,
> >> like this (from slave logs):
> >>> {"log":"I1031 08:58:02.692958    60 slave.cpp:1112] Launching task
> >> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 for framework
> >>
> 20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.692986684Z"}
> >>> {"log":"I1031 08:58:02.693696    60 slave.cpp:1222] Queuing task
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' for executor
> >> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 of framework
> >>
> '20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.693734418Z"}
> >>> {"log":"I1031 08:58:02.695612    64 docker.cpp:743] Starting container
> >> '02a786c9-556c-4ed9-80d2-1850a49030fe' for task
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' (and executor
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799') of framework
> >>
> '20141003-172543-3892422848-5050-1-0000'\n","stream":"stderr","time":"2014-10-31T08:58:02.6956675Z"}
> >>> {"log":"E1031 08:58:05.272902    65 slave.cpp:2485] Container
> >> '02a786c9-556c-4ed9-80d2-1850a49030fe' for executor
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' of framework
> >> '20141003-172543-3892422848-5050-1-0000' failed to start: Failed to
> 'docker
> >> run -d -c 204 -m 67108864 -e PORT=31459 -e PORT0=31459 -e PORTS=31459 -e
> >> HOST=web544 -e MESOS_SANDBOX=/mnt/mesos/sandbox -v
> >>
> /var/lib/mesos/slave/slaves/20141028-073834-3909200064-5050-1-75/frameworks/20141003-172543-3892422848-5050-1-0000/executors/topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799/runs/02a786c9-556c-4ed9-80d2-1850a49030fe:/mnt/mesos/sandbox
> >> --net host --name mesos-02a786c9-556c-4ed9-80d2-1850a49030fe
> >> docker.core.tf/demo:1': exit status = exited with status 1 stderr =
> >> 2014/10/31 08:58:04 Error response from daemon: Cannot start container
> >> 05666763bff98ad70f35968add3923018a9d457a1f4e9dab0981936841093d2f: exec:
> >> \"/server.js\": permission
> >> denied\n","stream":"stderr","time":"2014-10-31T08:58:05.273016253Z"}
> >>> But when I go to task sandbox from mesos ui, stdout and stderr are
> empty.
> >>> Marathon keeps scheduling tasks, but they all silently fail, this is
> >> very misleading.
> >>
> >>
> >>
> >> --
> >> This message was sent by Atlassian JIRA
> >> (v6.3.4#6332)
> >>
>

Re: [jira] [Updated] (MESOS-2020) mesos should send docker failure messages to scheduler

Posted by Jake Farrell <jf...@apache.org>.
Nothing in the workflow changed as far as i'm aware, looks like it was
reopened and then set as resolved

-Jake


On Wed, May 13, 2015 at 9:13 PM, Timothy Chen <tn...@gmail.com> wrote:

> I have no idea too, I wanted to resolve the ticket but it seems to be
> stuck on this state.
>
> Tim
>
> > On May 13, 2015, at 5:03 PM, Benjamin Mahler <be...@gmail.com>
> wrote:
> >
> > +tim, jake
> >
> > What does pending closed mean? I noticed that this had become the default
> > resolution yesterday. Did something change Jake?
> >
> > Tim, should this be resolved?
> >
> > On Tue, May 12, 2015 at 12:28 AM, Timothy Chen (JIRA) <ji...@apache.org>
> > wrote:
> >
> >>
> >>     [
> >>
> https://issues.apache.org/jira/browse/MESOS-2020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
> >> ]
> >>
> >> Timothy Chen updated MESOS-2020:
> >> --------------------------------
> >>    Fix Version/s: 0.23.0
> >>
> >>> mesos should send docker failure messages to scheduler
> >>> ------------------------------------------------------
> >>>
> >>>                Key: MESOS-2020
> >>>                URL: https://issues.apache.org/jira/browse/MESOS-2020
> >>>            Project: Mesos
> >>>         Issue Type: Bug
> >>>         Components: docker, slave
> >>>   Affects Versions: 0.20.1
> >>>        Environment: mesos-slaves running in docker
> >>>           Reporter: Ian Babrou
> >>>           Assignee: Jay Buffington
> >>>            Fix For: 0.23.0
> >>>
> >>>
> >>> I tried to start container that cannot actually start for some reason,
> >> like this (from slave logs):
> >>> {"log":"I1031 08:58:02.692958    60 slave.cpp:1112] Launching task
> >> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 for framework
> >>
> 20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.692986684Z"}
> >>> {"log":"I1031 08:58:02.693696    60 slave.cpp:1222] Queuing task
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' for executor
> >> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 of framework
> >>
> '20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.693734418Z"}
> >>> {"log":"I1031 08:58:02.695612    64 docker.cpp:743] Starting container
> >> '02a786c9-556c-4ed9-80d2-1850a49030fe' for task
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' (and executor
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799') of framework
> >>
> '20141003-172543-3892422848-5050-1-0000'\n","stream":"stderr","time":"2014-10-31T08:58:02.6956675Z"}
> >>> {"log":"E1031 08:58:05.272902    65 slave.cpp:2485] Container
> >> '02a786c9-556c-4ed9-80d2-1850a49030fe' for executor
> >> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' of framework
> >> '20141003-172543-3892422848-5050-1-0000' failed to start: Failed to
> 'docker
> >> run -d -c 204 -m 67108864 -e PORT=31459 -e PORT0=31459 -e PORTS=31459 -e
> >> HOST=web544 -e MESOS_SANDBOX=/mnt/mesos/sandbox -v
> >>
> /var/lib/mesos/slave/slaves/20141028-073834-3909200064-5050-1-75/frameworks/20141003-172543-3892422848-5050-1-0000/executors/topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799/runs/02a786c9-556c-4ed9-80d2-1850a49030fe:/mnt/mesos/sandbox
> >> --net host --name mesos-02a786c9-556c-4ed9-80d2-1850a49030fe
> >> docker.core.tf/demo:1': exit status = exited with status 1 stderr =
> >> 2014/10/31 08:58:04 Error response from daemon: Cannot start container
> >> 05666763bff98ad70f35968add3923018a9d457a1f4e9dab0981936841093d2f: exec:
> >> \"/server.js\": permission
> >> denied\n","stream":"stderr","time":"2014-10-31T08:58:05.273016253Z"}
> >>> But when I go to task sandbox from mesos ui, stdout and stderr are
> empty.
> >>> Marathon keeps scheduling tasks, but they all silently fail, this is
> >> very misleading.
> >>
> >>
> >>
> >> --
> >> This message was sent by Atlassian JIRA
> >> (v6.3.4#6332)
> >>
>

Re: [jira] [Updated] (MESOS-2020) mesos should send docker failure messages to scheduler

Posted by Timothy Chen <tn...@gmail.com>.
I have no idea too, I wanted to resolve the ticket but it seems to be stuck on this state.

Tim

> On May 13, 2015, at 5:03 PM, Benjamin Mahler <be...@gmail.com> wrote:
> 
> +tim, jake
> 
> What does pending closed mean? I noticed that this had become the default
> resolution yesterday. Did something change Jake?
> 
> Tim, should this be resolved?
> 
> On Tue, May 12, 2015 at 12:28 AM, Timothy Chen (JIRA) <ji...@apache.org>
> wrote:
> 
>> 
>>     [
>> https://issues.apache.org/jira/browse/MESOS-2020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
>> ]
>> 
>> Timothy Chen updated MESOS-2020:
>> --------------------------------
>>    Fix Version/s: 0.23.0
>> 
>>> mesos should send docker failure messages to scheduler
>>> ------------------------------------------------------
>>> 
>>>                Key: MESOS-2020
>>>                URL: https://issues.apache.org/jira/browse/MESOS-2020
>>>            Project: Mesos
>>>         Issue Type: Bug
>>>         Components: docker, slave
>>>   Affects Versions: 0.20.1
>>>        Environment: mesos-slaves running in docker
>>>           Reporter: Ian Babrou
>>>           Assignee: Jay Buffington
>>>            Fix For: 0.23.0
>>> 
>>> 
>>> I tried to start container that cannot actually start for some reason,
>> like this (from slave logs):
>>> {"log":"I1031 08:58:02.692958    60 slave.cpp:1112] Launching task
>> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 for framework
>> 20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.692986684Z"}
>>> {"log":"I1031 08:58:02.693696    60 slave.cpp:1222] Queuing task
>> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' for executor
>> topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799 of framework
>> '20141003-172543-3892422848-5050-1-0000\n","stream":"stderr","time":"2014-10-31T08:58:02.693734418Z"}
>>> {"log":"I1031 08:58:02.695612    64 docker.cpp:743] Starting container
>> '02a786c9-556c-4ed9-80d2-1850a49030fe' for task
>> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' (and executor
>> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799') of framework
>> '20141003-172543-3892422848-5050-1-0000'\n","stream":"stderr","time":"2014-10-31T08:58:02.6956675Z"}
>>> {"log":"E1031 08:58:05.272902    65 slave.cpp:2485] Container
>> '02a786c9-556c-4ed9-80d2-1850a49030fe' for executor
>> 'topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799' of framework
>> '20141003-172543-3892422848-5050-1-0000' failed to start: Failed to 'docker
>> run -d -c 204 -m 67108864 -e PORT=31459 -e PORT0=31459 -e PORTS=31459 -e
>> HOST=web544 -e MESOS_SANDBOX=/mnt/mesos/sandbox -v
>> /var/lib/mesos/slave/slaves/20141028-073834-3909200064-5050-1-75/frameworks/20141003-172543-3892422848-5050-1-0000/executors/topface_demo.055a8f8f-60dc-11e4-b4ad-56847afe9799/runs/02a786c9-556c-4ed9-80d2-1850a49030fe:/mnt/mesos/sandbox
>> --net host --name mesos-02a786c9-556c-4ed9-80d2-1850a49030fe
>> docker.core.tf/demo:1': exit status = exited with status 1 stderr =
>> 2014/10/31 08:58:04 Error response from daemon: Cannot start container
>> 05666763bff98ad70f35968add3923018a9d457a1f4e9dab0981936841093d2f: exec:
>> \"/server.js\": permission
>> denied\n","stream":"stderr","time":"2014-10-31T08:58:05.273016253Z"}
>>> But when I go to task sandbox from mesos ui, stdout and stderr are empty.
>>> Marathon keeps scheduling tasks, but they all silently fail, this is
>> very misleading.
>> 
>> 
>> 
>> --
>> This message was sent by Atlassian JIRA
>> (v6.3.4#6332)
>>