You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Chris Hokamp <ch...@gmail.com> on 2013/04/13 17:15:52 UTC

Mapper always hangs at the same spot

Hello,

We have a job where all mappers finish except for one, which always hangs
at the same spot (i.e. reaches 49%, then never progresses).

This is likely due to a bug in the wiki parser in our Pig UDF. We can
afford to lose the data this mapper is working on if it would allow the job
to finish. Question: is there a hadoop configuration parameter similar to
mapred.skip.map.max.skip.records that would let us skip a map that doesn't
progress after X amount of time? Any other possible workarounds for this
case would also be useful.

We are currently using hadoop 1.1.0 and Pig 0.10.1.

Thanks,
Chris

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
The UDF and our Pig scripts work fine for most languages' wikidumps, and
this hanging mapper issue only pops up with English wikidumps. It is
certainly an issue with the wikiparser getting stuck in a recursive loop,
and it must be a markup-related bug since this only happens with English.
We're working on tracking it down now. Thanks for the help and advice!

CH


On Sun, Apr 14, 2013 at 1:20 AM, Azuryy Yu <az...@gmail.com> wrote:

> agree. just check your app. or paste map code here.
>
> --Send from my Sony mobile.
> On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:
>
>> Your application logic is likely stuck in a loop.
>>
>>
>> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>>
>>> >When you say "never progresses", do you see the MR framework kill it
>>> >automatically after 10 minutes of inactivity or does it never ever
>>> >exit?
>>>
>>> The latter -- it never exits. Killing it manually seems like a good
>>> option for now. We already have mapred.max.map.failures.percent set to
>>> a non-zero value, but because the task never fails, this never comes into
>>> effect.
>>>
>>> Thanks for the help,
>>> Chris
>>>
>>>
>>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>>
>>>> When you say "never progresses", do you see the MR framework kill it
>>>> automatically after 10 minutes of inactivity or does it never ever
>>>> exit?
>>>>
>>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>>> in msec. You could also set mapred.max.map.failures.percent to a
>>>> non-zero value to allow that much percentage of tasks to fail without
>>>> also marking the whole job as a failure.
>>>>
>>>> If the task itself does not get killed by the framework due to
>>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>>> manually.
>>>>
>>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>>> wrote:
>>>> > Hello,
>>>> >
>>>> > We have a job where all mappers finish except for one, which always
>>>> hangs at
>>>> > the same spot (i.e. reaches 49%, then never progresses).
>>>> >
>>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>>> afford
>>>> > to lose the data this mapper is working on if it would allow the job
>>>> to
>>>> > finish. Question: is there a hadoop configuration parameter similar to
>>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>>> doesn't
>>>> > progress after X amount of time? Any other possible workarounds for
>>>> this
>>>> > case would also be useful.
>>>> >
>>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>>> >
>>>> > Thanks,
>>>> > Chris
>>>>
>>>>
>>>>
>>>> --
>>>> Harsh J
>>>>
>>>
>>>
>>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
The UDF and our Pig scripts work fine for most languages' wikidumps, and
this hanging mapper issue only pops up with English wikidumps. It is
certainly an issue with the wikiparser getting stuck in a recursive loop,
and it must be a markup-related bug since this only happens with English.
We're working on tracking it down now. Thanks for the help and advice!

CH


On Sun, Apr 14, 2013 at 1:20 AM, Azuryy Yu <az...@gmail.com> wrote:

> agree. just check your app. or paste map code here.
>
> --Send from my Sony mobile.
> On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:
>
>> Your application logic is likely stuck in a loop.
>>
>>
>> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>>
>>> >When you say "never progresses", do you see the MR framework kill it
>>> >automatically after 10 minutes of inactivity or does it never ever
>>> >exit?
>>>
>>> The latter -- it never exits. Killing it manually seems like a good
>>> option for now. We already have mapred.max.map.failures.percent set to
>>> a non-zero value, but because the task never fails, this never comes into
>>> effect.
>>>
>>> Thanks for the help,
>>> Chris
>>>
>>>
>>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>>
>>>> When you say "never progresses", do you see the MR framework kill it
>>>> automatically after 10 minutes of inactivity or does it never ever
>>>> exit?
>>>>
>>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>>> in msec. You could also set mapred.max.map.failures.percent to a
>>>> non-zero value to allow that much percentage of tasks to fail without
>>>> also marking the whole job as a failure.
>>>>
>>>> If the task itself does not get killed by the framework due to
>>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>>> manually.
>>>>
>>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>>> wrote:
>>>> > Hello,
>>>> >
>>>> > We have a job where all mappers finish except for one, which always
>>>> hangs at
>>>> > the same spot (i.e. reaches 49%, then never progresses).
>>>> >
>>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>>> afford
>>>> > to lose the data this mapper is working on if it would allow the job
>>>> to
>>>> > finish. Question: is there a hadoop configuration parameter similar to
>>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>>> doesn't
>>>> > progress after X amount of time? Any other possible workarounds for
>>>> this
>>>> > case would also be useful.
>>>> >
>>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>>> >
>>>> > Thanks,
>>>> > Chris
>>>>
>>>>
>>>>
>>>> --
>>>> Harsh J
>>>>
>>>
>>>
>>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
The UDF and our Pig scripts work fine for most languages' wikidumps, and
this hanging mapper issue only pops up with English wikidumps. It is
certainly an issue with the wikiparser getting stuck in a recursive loop,
and it must be a markup-related bug since this only happens with English.
We're working on tracking it down now. Thanks for the help and advice!

CH


On Sun, Apr 14, 2013 at 1:20 AM, Azuryy Yu <az...@gmail.com> wrote:

> agree. just check your app. or paste map code here.
>
> --Send from my Sony mobile.
> On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:
>
>> Your application logic is likely stuck in a loop.
>>
>>
>> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>>
>>> >When you say "never progresses", do you see the MR framework kill it
>>> >automatically after 10 minutes of inactivity or does it never ever
>>> >exit?
>>>
>>> The latter -- it never exits. Killing it manually seems like a good
>>> option for now. We already have mapred.max.map.failures.percent set to
>>> a non-zero value, but because the task never fails, this never comes into
>>> effect.
>>>
>>> Thanks for the help,
>>> Chris
>>>
>>>
>>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>>
>>>> When you say "never progresses", do you see the MR framework kill it
>>>> automatically after 10 minutes of inactivity or does it never ever
>>>> exit?
>>>>
>>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>>> in msec. You could also set mapred.max.map.failures.percent to a
>>>> non-zero value to allow that much percentage of tasks to fail without
>>>> also marking the whole job as a failure.
>>>>
>>>> If the task itself does not get killed by the framework due to
>>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>>> manually.
>>>>
>>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>>> wrote:
>>>> > Hello,
>>>> >
>>>> > We have a job where all mappers finish except for one, which always
>>>> hangs at
>>>> > the same spot (i.e. reaches 49%, then never progresses).
>>>> >
>>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>>> afford
>>>> > to lose the data this mapper is working on if it would allow the job
>>>> to
>>>> > finish. Question: is there a hadoop configuration parameter similar to
>>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>>> doesn't
>>>> > progress after X amount of time? Any other possible workarounds for
>>>> this
>>>> > case would also be useful.
>>>> >
>>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>>> >
>>>> > Thanks,
>>>> > Chris
>>>>
>>>>
>>>>
>>>> --
>>>> Harsh J
>>>>
>>>
>>>
>>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
The UDF and our Pig scripts work fine for most languages' wikidumps, and
this hanging mapper issue only pops up with English wikidumps. It is
certainly an issue with the wikiparser getting stuck in a recursive loop,
and it must be a markup-related bug since this only happens with English.
We're working on tracking it down now. Thanks for the help and advice!

CH


On Sun, Apr 14, 2013 at 1:20 AM, Azuryy Yu <az...@gmail.com> wrote:

> agree. just check your app. or paste map code here.
>
> --Send from my Sony mobile.
> On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:
>
>> Your application logic is likely stuck in a loop.
>>
>>
>> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>>
>>> >When you say "never progresses", do you see the MR framework kill it
>>> >automatically after 10 minutes of inactivity or does it never ever
>>> >exit?
>>>
>>> The latter -- it never exits. Killing it manually seems like a good
>>> option for now. We already have mapred.max.map.failures.percent set to
>>> a non-zero value, but because the task never fails, this never comes into
>>> effect.
>>>
>>> Thanks for the help,
>>> Chris
>>>
>>>
>>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>>
>>>> When you say "never progresses", do you see the MR framework kill it
>>>> automatically after 10 minutes of inactivity or does it never ever
>>>> exit?
>>>>
>>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>>> in msec. You could also set mapred.max.map.failures.percent to a
>>>> non-zero value to allow that much percentage of tasks to fail without
>>>> also marking the whole job as a failure.
>>>>
>>>> If the task itself does not get killed by the framework due to
>>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>>> manually.
>>>>
>>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>>> wrote:
>>>> > Hello,
>>>> >
>>>> > We have a job where all mappers finish except for one, which always
>>>> hangs at
>>>> > the same spot (i.e. reaches 49%, then never progresses).
>>>> >
>>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>>> afford
>>>> > to lose the data this mapper is working on if it would allow the job
>>>> to
>>>> > finish. Question: is there a hadoop configuration parameter similar to
>>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>>> doesn't
>>>> > progress after X amount of time? Any other possible workarounds for
>>>> this
>>>> > case would also be useful.
>>>> >
>>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>>> >
>>>> > Thanks,
>>>> > Chris
>>>>
>>>>
>>>>
>>>> --
>>>> Harsh J
>>>>
>>>
>>>
>>

Re: Mapper always hangs at the same spot

Posted by Azuryy Yu <az...@gmail.com>.
agree. just check your app. or paste map code here.

--Send from my Sony mobile.
On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:

> Your application logic is likely stuck in a loop.
>
>
> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>
>> >When you say "never progresses", do you see the MR framework kill it
>> >automatically after 10 minutes of inactivity or does it never ever
>> >exit?
>>
>> The latter -- it never exits. Killing it manually seems like a good
>> option for now. We already have mapred.max.map.failures.percent set to a
>> non-zero value, but because the task never fails, this never comes into
>> effect.
>>
>> Thanks for the help,
>> Chris
>>
>>
>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>
>>> When you say "never progresses", do you see the MR framework kill it
>>> automatically after 10 minutes of inactivity or does it never ever
>>> exit?
>>>
>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>> in msec. You could also set mapred.max.map.failures.percent to a
>>> non-zero value to allow that much percentage of tasks to fail without
>>> also marking the whole job as a failure.
>>>
>>> If the task itself does not get killed by the framework due to
>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>> manually.
>>>
>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>> wrote:
>>> > Hello,
>>> >
>>> > We have a job where all mappers finish except for one, which always
>>> hangs at
>>> > the same spot (i.e. reaches 49%, then never progresses).
>>> >
>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>> afford
>>> > to lose the data this mapper is working on if it would allow the job to
>>> > finish. Question: is there a hadoop configuration parameter similar to
>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>> doesn't
>>> > progress after X amount of time? Any other possible workarounds for
>>> this
>>> > case would also be useful.
>>> >
>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>> >
>>> > Thanks,
>>> > Chris
>>>
>>>
>>>
>>> --
>>> Harsh J
>>>
>>
>>
>

Re: Mapper always hangs at the same spot

Posted by Azuryy Yu <az...@gmail.com>.
agree. just check your app. or paste map code here.

--Send from my Sony mobile.
On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:

> Your application logic is likely stuck in a loop.
>
>
> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>
>> >When you say "never progresses", do you see the MR framework kill it
>> >automatically after 10 minutes of inactivity or does it never ever
>> >exit?
>>
>> The latter -- it never exits. Killing it manually seems like a good
>> option for now. We already have mapred.max.map.failures.percent set to a
>> non-zero value, but because the task never fails, this never comes into
>> effect.
>>
>> Thanks for the help,
>> Chris
>>
>>
>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>
>>> When you say "never progresses", do you see the MR framework kill it
>>> automatically after 10 minutes of inactivity or does it never ever
>>> exit?
>>>
>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>> in msec. You could also set mapred.max.map.failures.percent to a
>>> non-zero value to allow that much percentage of tasks to fail without
>>> also marking the whole job as a failure.
>>>
>>> If the task itself does not get killed by the framework due to
>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>> manually.
>>>
>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>> wrote:
>>> > Hello,
>>> >
>>> > We have a job where all mappers finish except for one, which always
>>> hangs at
>>> > the same spot (i.e. reaches 49%, then never progresses).
>>> >
>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>> afford
>>> > to lose the data this mapper is working on if it would allow the job to
>>> > finish. Question: is there a hadoop configuration parameter similar to
>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>> doesn't
>>> > progress after X amount of time? Any other possible workarounds for
>>> this
>>> > case would also be useful.
>>> >
>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>> >
>>> > Thanks,
>>> > Chris
>>>
>>>
>>>
>>> --
>>> Harsh J
>>>
>>
>>
>

Re: Mapper always hangs at the same spot

Posted by Azuryy Yu <az...@gmail.com>.
agree. just check your app. or paste map code here.

--Send from my Sony mobile.
On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:

> Your application logic is likely stuck in a loop.
>
>
> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>
>> >When you say "never progresses", do you see the MR framework kill it
>> >automatically after 10 minutes of inactivity or does it never ever
>> >exit?
>>
>> The latter -- it never exits. Killing it manually seems like a good
>> option for now. We already have mapred.max.map.failures.percent set to a
>> non-zero value, but because the task never fails, this never comes into
>> effect.
>>
>> Thanks for the help,
>> Chris
>>
>>
>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>
>>> When you say "never progresses", do you see the MR framework kill it
>>> automatically after 10 minutes of inactivity or does it never ever
>>> exit?
>>>
>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>> in msec. You could also set mapred.max.map.failures.percent to a
>>> non-zero value to allow that much percentage of tasks to fail without
>>> also marking the whole job as a failure.
>>>
>>> If the task itself does not get killed by the framework due to
>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>> manually.
>>>
>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>> wrote:
>>> > Hello,
>>> >
>>> > We have a job where all mappers finish except for one, which always
>>> hangs at
>>> > the same spot (i.e. reaches 49%, then never progresses).
>>> >
>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>> afford
>>> > to lose the data this mapper is working on if it would allow the job to
>>> > finish. Question: is there a hadoop configuration parameter similar to
>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>> doesn't
>>> > progress after X amount of time? Any other possible workarounds for
>>> this
>>> > case would also be useful.
>>> >
>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>> >
>>> > Thanks,
>>> > Chris
>>>
>>>
>>>
>>> --
>>> Harsh J
>>>
>>
>>
>

Re: Mapper always hangs at the same spot

Posted by Azuryy Yu <az...@gmail.com>.
agree. just check your app. or paste map code here.

--Send from my Sony mobile.
On Apr 14, 2013 4:08 AM, "Edward Capriolo" <ed...@gmail.com> wrote:

> Your application logic is likely stuck in a loop.
>
>
> On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:
>
>> >When you say "never progresses", do you see the MR framework kill it
>> >automatically after 10 minutes of inactivity or does it never ever
>> >exit?
>>
>> The latter -- it never exits. Killing it manually seems like a good
>> option for now. We already have mapred.max.map.failures.percent set to a
>> non-zero value, but because the task never fails, this never comes into
>> effect.
>>
>> Thanks for the help,
>> Chris
>>
>>
>> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>>
>>> When you say "never progresses", do you see the MR framework kill it
>>> automatically after 10 minutes of inactivity or does it never ever
>>> exit?
>>>
>>> You can lower the timeout period on tasks via mapred.task.timeout set
>>> in msec. You could also set mapred.max.map.failures.percent to a
>>> non-zero value to allow that much percentage of tasks to fail without
>>> also marking the whole job as a failure.
>>>
>>> If the task itself does not get killed by the framework due to
>>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>>> manually.
>>>
>>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>>> wrote:
>>> > Hello,
>>> >
>>> > We have a job where all mappers finish except for one, which always
>>> hangs at
>>> > the same spot (i.e. reaches 49%, then never progresses).
>>> >
>>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>>> afford
>>> > to lose the data this mapper is working on if it would allow the job to
>>> > finish. Question: is there a hadoop configuration parameter similar to
>>> > mapred.skip.map.max.skip.records that would let us skip a map that
>>> doesn't
>>> > progress after X amount of time? Any other possible workarounds for
>>> this
>>> > case would also be useful.
>>> >
>>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>>> >
>>> > Thanks,
>>> > Chris
>>>
>>>
>>>
>>> --
>>> Harsh J
>>>
>>
>>
>

Re: Mapper always hangs at the same spot

Posted by Edward Capriolo <ed...@gmail.com>.
Your application logic is likely stuck in a loop.


On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:

> >When you say "never progresses", do you see the MR framework kill it
> >automatically after 10 minutes of inactivity or does it never ever
> >exit?
>
> The latter -- it never exits. Killing it manually seems like a good option
> for now. We already have mapred.max.map.failures.percent set to a
> non-zero value, but because the task never fails, this never comes into
> effect.
>
> Thanks for the help,
> Chris
>
>
> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>
>> When you say "never progresses", do you see the MR framework kill it
>> automatically after 10 minutes of inactivity or does it never ever
>> exit?
>>
>> You can lower the timeout period on tasks via mapred.task.timeout set
>> in msec. You could also set mapred.max.map.failures.percent to a
>> non-zero value to allow that much percentage of tasks to fail without
>> also marking the whole job as a failure.
>>
>> If the task itself does not get killed by the framework due to
>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>> manually.
>>
>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>> wrote:
>> > Hello,
>> >
>> > We have a job where all mappers finish except for one, which always
>> hangs at
>> > the same spot (i.e. reaches 49%, then never progresses).
>> >
>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>> afford
>> > to lose the data this mapper is working on if it would allow the job to
>> > finish. Question: is there a hadoop configuration parameter similar to
>> > mapred.skip.map.max.skip.records that would let us skip a map that
>> doesn't
>> > progress after X amount of time? Any other possible workarounds for this
>> > case would also be useful.
>> >
>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>> >
>> > Thanks,
>> > Chris
>>
>>
>>
>> --
>> Harsh J
>>
>
>

Re: Mapper always hangs at the same spot

Posted by Edward Capriolo <ed...@gmail.com>.
Your application logic is likely stuck in a loop.


On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:

> >When you say "never progresses", do you see the MR framework kill it
> >automatically after 10 minutes of inactivity or does it never ever
> >exit?
>
> The latter -- it never exits. Killing it manually seems like a good option
> for now. We already have mapred.max.map.failures.percent set to a
> non-zero value, but because the task never fails, this never comes into
> effect.
>
> Thanks for the help,
> Chris
>
>
> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>
>> When you say "never progresses", do you see the MR framework kill it
>> automatically after 10 minutes of inactivity or does it never ever
>> exit?
>>
>> You can lower the timeout period on tasks via mapred.task.timeout set
>> in msec. You could also set mapred.max.map.failures.percent to a
>> non-zero value to allow that much percentage of tasks to fail without
>> also marking the whole job as a failure.
>>
>> If the task itself does not get killed by the framework due to
>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>> manually.
>>
>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>> wrote:
>> > Hello,
>> >
>> > We have a job where all mappers finish except for one, which always
>> hangs at
>> > the same spot (i.e. reaches 49%, then never progresses).
>> >
>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>> afford
>> > to lose the data this mapper is working on if it would allow the job to
>> > finish. Question: is there a hadoop configuration parameter similar to
>> > mapred.skip.map.max.skip.records that would let us skip a map that
>> doesn't
>> > progress after X amount of time? Any other possible workarounds for this
>> > case would also be useful.
>> >
>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>> >
>> > Thanks,
>> > Chris
>>
>>
>>
>> --
>> Harsh J
>>
>
>

Re: Mapper always hangs at the same spot

Posted by Edward Capriolo <ed...@gmail.com>.
Your application logic is likely stuck in a loop.


On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:

> >When you say "never progresses", do you see the MR framework kill it
> >automatically after 10 minutes of inactivity or does it never ever
> >exit?
>
> The latter -- it never exits. Killing it manually seems like a good option
> for now. We already have mapred.max.map.failures.percent set to a
> non-zero value, but because the task never fails, this never comes into
> effect.
>
> Thanks for the help,
> Chris
>
>
> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>
>> When you say "never progresses", do you see the MR framework kill it
>> automatically after 10 minutes of inactivity or does it never ever
>> exit?
>>
>> You can lower the timeout period on tasks via mapred.task.timeout set
>> in msec. You could also set mapred.max.map.failures.percent to a
>> non-zero value to allow that much percentage of tasks to fail without
>> also marking the whole job as a failure.
>>
>> If the task itself does not get killed by the framework due to
>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>> manually.
>>
>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>> wrote:
>> > Hello,
>> >
>> > We have a job where all mappers finish except for one, which always
>> hangs at
>> > the same spot (i.e. reaches 49%, then never progresses).
>> >
>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>> afford
>> > to lose the data this mapper is working on if it would allow the job to
>> > finish. Question: is there a hadoop configuration parameter similar to
>> > mapred.skip.map.max.skip.records that would let us skip a map that
>> doesn't
>> > progress after X amount of time? Any other possible workarounds for this
>> > case would also be useful.
>> >
>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>> >
>> > Thanks,
>> > Chris
>>
>>
>>
>> --
>> Harsh J
>>
>
>

Re: Mapper always hangs at the same spot

Posted by Edward Capriolo <ed...@gmail.com>.
Your application logic is likely stuck in a loop.


On Sat, Apr 13, 2013 at 12:47 PM, Chris Hokamp <ch...@gmail.com>wrote:

> >When you say "never progresses", do you see the MR framework kill it
> >automatically after 10 minutes of inactivity or does it never ever
> >exit?
>
> The latter -- it never exits. Killing it manually seems like a good option
> for now. We already have mapred.max.map.failures.percent set to a
> non-zero value, but because the task never fails, this never comes into
> effect.
>
> Thanks for the help,
> Chris
>
>
> On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:
>
>> When you say "never progresses", do you see the MR framework kill it
>> automatically after 10 minutes of inactivity or does it never ever
>> exit?
>>
>> You can lower the timeout period on tasks via mapred.task.timeout set
>> in msec. You could also set mapred.max.map.failures.percent to a
>> non-zero value to allow that much percentage of tasks to fail without
>> also marking the whole job as a failure.
>>
>> If the task itself does not get killed by the framework due to
>> inactiveness, try doing a hadoop job -fail-task on its attempt ID
>> manually.
>>
>> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
>> wrote:
>> > Hello,
>> >
>> > We have a job where all mappers finish except for one, which always
>> hangs at
>> > the same spot (i.e. reaches 49%, then never progresses).
>> >
>> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
>> afford
>> > to lose the data this mapper is working on if it would allow the job to
>> > finish. Question: is there a hadoop configuration parameter similar to
>> > mapred.skip.map.max.skip.records that would let us skip a map that
>> doesn't
>> > progress after X amount of time? Any other possible workarounds for this
>> > case would also be useful.
>> >
>> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
>> >
>> > Thanks,
>> > Chris
>>
>>
>>
>> --
>> Harsh J
>>
>
>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
>When you say "never progresses", do you see the MR framework kill it
>automatically after 10 minutes of inactivity or does it never ever
>exit?

The latter -- it never exits. Killing it manually seems like a good option
for now. We already have mapred.max.map.failures.percent set to a non-zero
value, but because the task never fails, this never comes into effect.

Thanks for the help,
Chris


On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:

> When you say "never progresses", do you see the MR framework kill it
> automatically after 10 minutes of inactivity or does it never ever
> exit?
>
> You can lower the timeout period on tasks via mapred.task.timeout set
> in msec. You could also set mapred.max.map.failures.percent to a
> non-zero value to allow that much percentage of tasks to fail without
> also marking the whole job as a failure.
>
> If the task itself does not get killed by the framework due to
> inactiveness, try doing a hadoop job -fail-task on its attempt ID
> manually.
>
> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
> wrote:
> > Hello,
> >
> > We have a job where all mappers finish except for one, which always
> hangs at
> > the same spot (i.e. reaches 49%, then never progresses).
> >
> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
> afford
> > to lose the data this mapper is working on if it would allow the job to
> > finish. Question: is there a hadoop configuration parameter similar to
> > mapred.skip.map.max.skip.records that would let us skip a map that
> doesn't
> > progress after X amount of time? Any other possible workarounds for this
> > case would also be useful.
> >
> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
> >
> > Thanks,
> > Chris
>
>
>
> --
> Harsh J
>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
>When you say "never progresses", do you see the MR framework kill it
>automatically after 10 minutes of inactivity or does it never ever
>exit?

The latter -- it never exits. Killing it manually seems like a good option
for now. We already have mapred.max.map.failures.percent set to a non-zero
value, but because the task never fails, this never comes into effect.

Thanks for the help,
Chris


On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:

> When you say "never progresses", do you see the MR framework kill it
> automatically after 10 minutes of inactivity or does it never ever
> exit?
>
> You can lower the timeout period on tasks via mapred.task.timeout set
> in msec. You could also set mapred.max.map.failures.percent to a
> non-zero value to allow that much percentage of tasks to fail without
> also marking the whole job as a failure.
>
> If the task itself does not get killed by the framework due to
> inactiveness, try doing a hadoop job -fail-task on its attempt ID
> manually.
>
> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
> wrote:
> > Hello,
> >
> > We have a job where all mappers finish except for one, which always
> hangs at
> > the same spot (i.e. reaches 49%, then never progresses).
> >
> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
> afford
> > to lose the data this mapper is working on if it would allow the job to
> > finish. Question: is there a hadoop configuration parameter similar to
> > mapred.skip.map.max.skip.records that would let us skip a map that
> doesn't
> > progress after X amount of time? Any other possible workarounds for this
> > case would also be useful.
> >
> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
> >
> > Thanks,
> > Chris
>
>
>
> --
> Harsh J
>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
>When you say "never progresses", do you see the MR framework kill it
>automatically after 10 minutes of inactivity or does it never ever
>exit?

The latter -- it never exits. Killing it manually seems like a good option
for now. We already have mapred.max.map.failures.percent set to a non-zero
value, but because the task never fails, this never comes into effect.

Thanks for the help,
Chris


On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:

> When you say "never progresses", do you see the MR framework kill it
> automatically after 10 minutes of inactivity or does it never ever
> exit?
>
> You can lower the timeout period on tasks via mapred.task.timeout set
> in msec. You could also set mapred.max.map.failures.percent to a
> non-zero value to allow that much percentage of tasks to fail without
> also marking the whole job as a failure.
>
> If the task itself does not get killed by the framework due to
> inactiveness, try doing a hadoop job -fail-task on its attempt ID
> manually.
>
> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
> wrote:
> > Hello,
> >
> > We have a job where all mappers finish except for one, which always
> hangs at
> > the same spot (i.e. reaches 49%, then never progresses).
> >
> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
> afford
> > to lose the data this mapper is working on if it would allow the job to
> > finish. Question: is there a hadoop configuration parameter similar to
> > mapred.skip.map.max.skip.records that would let us skip a map that
> doesn't
> > progress after X amount of time? Any other possible workarounds for this
> > case would also be useful.
> >
> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
> >
> > Thanks,
> > Chris
>
>
>
> --
> Harsh J
>

Re: Mapper always hangs at the same spot

Posted by Chris Hokamp <ch...@gmail.com>.
>When you say "never progresses", do you see the MR framework kill it
>automatically after 10 minutes of inactivity or does it never ever
>exit?

The latter -- it never exits. Killing it manually seems like a good option
for now. We already have mapred.max.map.failures.percent set to a non-zero
value, but because the task never fails, this never comes into effect.

Thanks for the help,
Chris


On Sat, Apr 13, 2013 at 5:00 PM, Harsh J <ha...@cloudera.com> wrote:

> When you say "never progresses", do you see the MR framework kill it
> automatically after 10 minutes of inactivity or does it never ever
> exit?
>
> You can lower the timeout period on tasks via mapred.task.timeout set
> in msec. You could also set mapred.max.map.failures.percent to a
> non-zero value to allow that much percentage of tasks to fail without
> also marking the whole job as a failure.
>
> If the task itself does not get killed by the framework due to
> inactiveness, try doing a hadoop job -fail-task on its attempt ID
> manually.
>
> On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com>
> wrote:
> > Hello,
> >
> > We have a job where all mappers finish except for one, which always
> hangs at
> > the same spot (i.e. reaches 49%, then never progresses).
> >
> > This is likely due to a bug in the wiki parser in our Pig UDF. We can
> afford
> > to lose the data this mapper is working on if it would allow the job to
> > finish. Question: is there a hadoop configuration parameter similar to
> > mapred.skip.map.max.skip.records that would let us skip a map that
> doesn't
> > progress after X amount of time? Any other possible workarounds for this
> > case would also be useful.
> >
> > We are currently using hadoop 1.1.0 and Pig 0.10.1.
> >
> > Thanks,
> > Chris
>
>
>
> --
> Harsh J
>

Re: Mapper always hangs at the same spot

Posted by Harsh J <ha...@cloudera.com>.
When you say "never progresses", do you see the MR framework kill it
automatically after 10 minutes of inactivity or does it never ever
exit?

You can lower the timeout period on tasks via mapred.task.timeout set
in msec. You could also set mapred.max.map.failures.percent to a
non-zero value to allow that much percentage of tasks to fail without
also marking the whole job as a failure.

If the task itself does not get killed by the framework due to
inactiveness, try doing a hadoop job -fail-task on its attempt ID
manually.

On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com> wrote:
> Hello,
>
> We have a job where all mappers finish except for one, which always hangs at
> the same spot (i.e. reaches 49%, then never progresses).
>
> This is likely due to a bug in the wiki parser in our Pig UDF. We can afford
> to lose the data this mapper is working on if it would allow the job to
> finish. Question: is there a hadoop configuration parameter similar to
> mapred.skip.map.max.skip.records that would let us skip a map that doesn't
> progress after X amount of time? Any other possible workarounds for this
> case would also be useful.
>
> We are currently using hadoop 1.1.0 and Pig 0.10.1.
>
> Thanks,
> Chris



--
Harsh J

Re: Mapper always hangs at the same spot

Posted by Harsh J <ha...@cloudera.com>.
When you say "never progresses", do you see the MR framework kill it
automatically after 10 minutes of inactivity or does it never ever
exit?

You can lower the timeout period on tasks via mapred.task.timeout set
in msec. You could also set mapred.max.map.failures.percent to a
non-zero value to allow that much percentage of tasks to fail without
also marking the whole job as a failure.

If the task itself does not get killed by the framework due to
inactiveness, try doing a hadoop job -fail-task on its attempt ID
manually.

On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com> wrote:
> Hello,
>
> We have a job where all mappers finish except for one, which always hangs at
> the same spot (i.e. reaches 49%, then never progresses).
>
> This is likely due to a bug in the wiki parser in our Pig UDF. We can afford
> to lose the data this mapper is working on if it would allow the job to
> finish. Question: is there a hadoop configuration parameter similar to
> mapred.skip.map.max.skip.records that would let us skip a map that doesn't
> progress after X amount of time? Any other possible workarounds for this
> case would also be useful.
>
> We are currently using hadoop 1.1.0 and Pig 0.10.1.
>
> Thanks,
> Chris



--
Harsh J

Re: Mapper always hangs at the same spot

Posted by Harsh J <ha...@cloudera.com>.
When you say "never progresses", do you see the MR framework kill it
automatically after 10 minutes of inactivity or does it never ever
exit?

You can lower the timeout period on tasks via mapred.task.timeout set
in msec. You could also set mapred.max.map.failures.percent to a
non-zero value to allow that much percentage of tasks to fail without
also marking the whole job as a failure.

If the task itself does not get killed by the framework due to
inactiveness, try doing a hadoop job -fail-task on its attempt ID
manually.

On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com> wrote:
> Hello,
>
> We have a job where all mappers finish except for one, which always hangs at
> the same spot (i.e. reaches 49%, then never progresses).
>
> This is likely due to a bug in the wiki parser in our Pig UDF. We can afford
> to lose the data this mapper is working on if it would allow the job to
> finish. Question: is there a hadoop configuration parameter similar to
> mapred.skip.map.max.skip.records that would let us skip a map that doesn't
> progress after X amount of time? Any other possible workarounds for this
> case would also be useful.
>
> We are currently using hadoop 1.1.0 and Pig 0.10.1.
>
> Thanks,
> Chris



--
Harsh J

Re: Mapper always hangs at the same spot

Posted by Harsh J <ha...@cloudera.com>.
When you say "never progresses", do you see the MR framework kill it
automatically after 10 minutes of inactivity or does it never ever
exit?

You can lower the timeout period on tasks via mapred.task.timeout set
in msec. You could also set mapred.max.map.failures.percent to a
non-zero value to allow that much percentage of tasks to fail without
also marking the whole job as a failure.

If the task itself does not get killed by the framework due to
inactiveness, try doing a hadoop job -fail-task on its attempt ID
manually.

On Sat, Apr 13, 2013 at 8:45 PM, Chris Hokamp <ch...@gmail.com> wrote:
> Hello,
>
> We have a job where all mappers finish except for one, which always hangs at
> the same spot (i.e. reaches 49%, then never progresses).
>
> This is likely due to a bug in the wiki parser in our Pig UDF. We can afford
> to lose the data this mapper is working on if it would allow the job to
> finish. Question: is there a hadoop configuration parameter similar to
> mapred.skip.map.max.skip.records that would let us skip a map that doesn't
> progress after X amount of time? Any other possible workarounds for this
> case would also be useful.
>
> We are currently using hadoop 1.1.0 and Pig 0.10.1.
>
> Thanks,
> Chris



--
Harsh J