You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Jun Young Kim <ju...@gmail.com> on 2011/02/18 03:29:42 UTC

I got errors from hdfs about DataStreamer Exceptions.

hi, all.

I got errors from hdfs.

2011-02-18 11:21:29[WARN ][DFSOutputStream.java]run()(519) : DataStreamer Exception: java.io.IOException: Unable to create new block.
	at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:832)
	at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)

2011-02-18 11:21:29[WARN ][DFSOutputStream.java]setupPipelineForAppendOrRecovery()(730) : Could not get block locations. Source file "/user/test/51/output/ehshop00newsvc-r-00000" - Aborting...
2011-02-18 11:21:29[WARN ][Child.java]main()(234) : Exception running child : java.io.EOFException
	at java.io.DataInputStream.readShort(DataInputStream.java:298)
	at org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Status.read(DataTransferProtocol.java:113)
	at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:881)
	at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
	at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)

2011-02-18 11:21:29[INFO ][Task.java]taskCleanup()(996) : Runnning cleanup for the task



I think this one is also not different error.

org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: blk_-2325764274016776017_8292 file=/user/test/51/input/kids.txt

         at org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:559)

         at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:367)

         at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)

         at java.io.DataInputStream.read(DataInputStream.java:83)

         at org.apache.hadoop.util.LineReader.readLine(LineReader.java:138)

         at org.apache.hadoop.mapreduce.lib.input.LineRecordReader.nextKeyValue(LineRecordReader.java:149)

         at org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:465)

         at org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)

         at org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:90)

         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)


--> I've checked the file '/user/test/51/input/kids.txt ', but, there is 
not strange ones. this file is healthy.

Does anybody know about this error?
How could I fix this one?

thanks.

-- 
Junyoung Kim (juneng603@gmail.com)


Re: I got errors from hdfs about DataStreamer Exceptions.

Posted by 茅旭峰 <m9...@gmail.com>.
I double checked the cluster, all of the disk have plenty of free space.
I could not put any data into the cluster. The cluster summary shows

==
Cluster Summary***1987209 files and directories, 1993327 blocks = 3980536
total. Heap Size is 4.2 GB / 5.56 GB (75%)
*
Configured Capacity : 120.88 TB DFS Used : 49.55 TB Non DFS Used :
6.12 TBDFS Remaining:65.21 TBDFS Used%:40.99 %DFS Remaining%:53.95 %
Live
Nodes <http://cdn-hadoop-01:50070/dfsnodelist.jsp?whatNodes=LIVE> : 14 Dead
Nodes <http://cdn-hadoop-01:50070/dfsnodelist.jsp?whatNodes=DEAD> : 0
==

and hadoop fsck / shows good
==
Status: HEALTHY
 Total size:    17900052452875 B
 Total dirs:    35613
 Total files:   1951594 (Files currently being written: 1)
 Total blocks (validated):      1992835 (avg. block size 8982204 B)
 Minimally replicated blocks:   1992835 (100.0 %)
 Over-replicated blocks:        0 (0.0 %)
 Under-replicated blocks:       5 (2.5089885E-4 %)
 Mis-replicated blocks:         0 (0.0 %)
 Default replication factor:    3
 Average block replication:     3.0000422
 Corrupt blocks:                0
 Missing replicas:              21 (3.5125346E-4 %)
 Number of data-nodes:          14
 Number of racks:               1


The filesystem under path '/' is HEALTHY

==

Any suggestions? Thanks!

On Sat, Apr 16, 2011 at 9:48 PM, Sonal Goyal <so...@gmail.com> wrote:

> Hi,
>
> What is the cluster free disk space and current usage?
>
> Thanks and Regards,
> Sonal
> <https://github.com/sonalgoyal/hiho>Hadoop ETL and Data Integration<https://github.com/sonalgoyal/hiho>
> Nube Technologies <http://www.nubetech.co>
>
> <http://in.linkedin.com/in/sonalgoyal>
>
>
>
>
>
>
> On Sat, Apr 16, 2011 at 9:18 AM, 茅旭峰 <m9...@gmail.com> wrote:
>
>> Hi Jun Young,
>>
>> It looks like I'm also running into the same issue you've got.
>> Have you found more info? I googled
>> 'java.io.IOException: Unable to create new block.', some guys
>> suggested that it might be running out of fd, but it's not my case.
>> I verified the fds in use by lsof and ls /proc/pid/fd.
>> Do you have any hints?
>>
>> Thanks!
>>
>> Mao Xu-Feng
>>
>> On Fri, Feb 18, 2011 at 10:51 AM, Jun Young Kim <ju...@gmail.com>
>> wrote:
>>
>> > hi, harsh.
>> > you're always giving a response very quickly. ;)
>> >
>> > I am using a version 0.21.0 now.
>> > before asking about this problem, I've checked already file system
>> healthy.
>> >
>> > $> hadoop fsck /
>> > .....
>> > .....
>> > Status: HEALTHY
>> >  Total size:    24231595038 B
>> >  Total dirs:    43818
>> >  Total files:   41193 (Files currently being written: 2178)
>> >  Total blocks (validated):      40941 (avg. block size 591866 B) (Total
>> > open file blocks (not validated): 224)
>> >  Minimally replicated blocks:   40941 (100.0 %)
>> >  Over-replicated blocks:        1 (0.0024425392 %)
>> >  Under-replicated blocks:       2 (0.0048850784 %)
>> >  Mis-replicated blocks:         0 (0.0 %)
>> >  Default replication factor:    2
>> >  Average block replication:     2.1106226
>> >  Corrupt blocks:                0
>> >  Missing replicas:              4 (0.00462904 %)
>> >  Number of data-nodes:          8
>> >  Number of racks:               1
>> >
>> > The filesystem under path '/' is HEALTHY
>> >
>> > additionally, I found a little different error. here it is.
>> >
>> > java.io.IOException: Bad connect ack with firstBadLink as
>> > 10.25.241.107:50010 at
>> >
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:889)
>> > at
>> >
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
>> > at
>> >
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>> >
>> >
>> > here is my execution environment.
>> >
>> > average job count : 20
>> > max map capacity : 128
>> > max reduce capacity : 128
>> > avg/slot per node : 32
>> >
>> > avg input file size per job : 200M ~ 1G
>> >
>> >
>> > thanks.
>> >
>> > Junyoung Kim (juneng603@gmail.com)
>> >
>> >
>> > On 02/18/2011 11:43 AM, Harsh J wrote:
>> >
>> >> You may want to check your HDFS health stat via 'fsck'
>> >> (http://namenode/fsck or `hadoop fsck`). There may be a few corrupt
>> >> files or bad DNs.
>> >>
>> >> Would also be good to know what exact version of Hadoop you're running.
>> >>
>> >> On Fri, Feb 18, 2011 at 7:59 AM, Jun Young Kim<ju...@gmail.com>
>> >>  wrote:
>> >>
>> >>> hi, all.
>> >>>
>> >>> I got errors from hdfs.
>> >>>
>> >>> 2011-02-18 11:21:29[WARN ][DFSOutputStream.java]run()(519) :
>> DataStreamer
>> >>> Exception: java.io.IOException: Unable to create new block.
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:832)
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>> >>>
>> >>> 2011-02-18 11:21:29[WARN
>> >>> ][DFSOutputStream.java]setupPipelineForAppendOrRecovery()(730) : Could
>> >>> not
>> >>> get block locations. Source file
>> >>> "/user/test/51/output/ehshop00newsvc-r-00000" - Aborting...
>> >>> 2011-02-18 11:21:29[WARN ][Child.java]main()(234) : Exception running
>> >>> child
>> >>> : java.io.EOFException
>> >>>        at java.io.DataInputStream.readShort(DataInputStream.java:298)
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Status.read(DataTransferProtocol.java:113)
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:881)
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>> >>>
>> >>> 2011-02-18 11:21:29[INFO ][Task.java]taskCleanup()(996) : Runnning
>> >>> cleanup
>> >>> for the task
>> >>>
>> >>>
>> >>>
>> >>> I think this one is also not different error.
>> >>>
>> >>> org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
>> >>> blk_-2325764274016776017_8292 file=/user/test/51/input/kids.txt
>> >>>
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:559)
>> >>>
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:367)
>> >>>
>> >>>        at
>> >>> org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)
>> >>>
>> >>>        at java.io.DataInputStream.read(DataInputStream.java:83)
>> >>>
>> >>>        at
>> org.apache.hadoop.util.LineReader.readLine(LineReader.java:138)
>> >>>
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.mapreduce.lib.input.LineRecordReader.nextKeyValue(LineRecordReader.java:149)
>> >>>
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:465)
>> >>>
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
>> >>>
>> >>>        at
>> >>>
>> >>>
>> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:90)
>> >>>
>> >>>        at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
>> >>>
>> >>>
>> >>> -->  I've checked the file '/user/test/51/input/kids.txt ', but, there
>> is
>> >>> not
>> >>> strange ones. this file is healthy.
>> >>>
>> >>> Does anybody know about this error?
>> >>> How could I fix this one?
>> >>>
>> >>> thanks.
>> >>>
>> >>> --
>> >>> Junyoung Kim (juneng603@gmail.com)
>> >>>
>> >>>
>> >>>
>> >>
>> >>
>>
>
>

Re: I got errors from hdfs about DataStreamer Exceptions.

Posted by 茅旭峰 <m9...@gmail.com>.
Hi Jun Young,

It looks like I'm also running into the same issue you've got.
Have you found more info? I googled
'java.io.IOException: Unable to create new block.', some guys
suggested that it might be running out of fd, but it's not my case.
I verified the fds in use by lsof and ls /proc/pid/fd.
Do you have any hints?

Thanks!

Mao Xu-Feng

On Fri, Feb 18, 2011 at 10:51 AM, Jun Young Kim <ju...@gmail.com> wrote:

> hi, harsh.
> you're always giving a response very quickly. ;)
>
> I am using a version 0.21.0 now.
> before asking about this problem, I've checked already file system healthy.
>
> $> hadoop fsck /
> .....
> .....
> Status: HEALTHY
>  Total size:    24231595038 B
>  Total dirs:    43818
>  Total files:   41193 (Files currently being written: 2178)
>  Total blocks (validated):      40941 (avg. block size 591866 B) (Total
> open file blocks (not validated): 224)
>  Minimally replicated blocks:   40941 (100.0 %)
>  Over-replicated blocks:        1 (0.0024425392 %)
>  Under-replicated blocks:       2 (0.0048850784 %)
>  Mis-replicated blocks:         0 (0.0 %)
>  Default replication factor:    2
>  Average block replication:     2.1106226
>  Corrupt blocks:                0
>  Missing replicas:              4 (0.00462904 %)
>  Number of data-nodes:          8
>  Number of racks:               1
>
> The filesystem under path '/' is HEALTHY
>
> additionally, I found a little different error. here it is.
>
> java.io.IOException: Bad connect ack with firstBadLink as
> 10.25.241.107:50010 at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:889)
> at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
> at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>
>
> here is my execution environment.
>
> average job count : 20
> max map capacity : 128
> max reduce capacity : 128
> avg/slot per node : 32
>
> avg input file size per job : 200M ~ 1G
>
>
> thanks.
>
> Junyoung Kim (juneng603@gmail.com)
>
>
> On 02/18/2011 11:43 AM, Harsh J wrote:
>
>> You may want to check your HDFS health stat via 'fsck'
>> (http://namenode/fsck or `hadoop fsck`). There may be a few corrupt
>> files or bad DNs.
>>
>> Would also be good to know what exact version of Hadoop you're running.
>>
>> On Fri, Feb 18, 2011 at 7:59 AM, Jun Young Kim<ju...@gmail.com>
>>  wrote:
>>
>>> hi, all.
>>>
>>> I got errors from hdfs.
>>>
>>> 2011-02-18 11:21:29[WARN ][DFSOutputStream.java]run()(519) : DataStreamer
>>> Exception: java.io.IOException: Unable to create new block.
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:832)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>>>
>>> 2011-02-18 11:21:29[WARN
>>> ][DFSOutputStream.java]setupPipelineForAppendOrRecovery()(730) : Could
>>> not
>>> get block locations. Source file
>>> "/user/test/51/output/ehshop00newsvc-r-00000" - Aborting...
>>> 2011-02-18 11:21:29[WARN ][Child.java]main()(234) : Exception running
>>> child
>>> : java.io.EOFException
>>>        at java.io.DataInputStream.readShort(DataInputStream.java:298)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Status.read(DataTransferProtocol.java:113)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:881)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>>>
>>> 2011-02-18 11:21:29[INFO ][Task.java]taskCleanup()(996) : Runnning
>>> cleanup
>>> for the task
>>>
>>>
>>>
>>> I think this one is also not different error.
>>>
>>> org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
>>> blk_-2325764274016776017_8292 file=/user/test/51/input/kids.txt
>>>
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:559)
>>>
>>>        at
>>>
>>> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:367)
>>>
>>>        at
>>> org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)
>>>
>>>        at java.io.DataInputStream.read(DataInputStream.java:83)
>>>
>>>        at org.apache.hadoop.util.LineReader.readLine(LineReader.java:138)
>>>
>>>        at
>>>
>>> org.apache.hadoop.mapreduce.lib.input.LineRecordReader.nextKeyValue(LineRecordReader.java:149)
>>>
>>>        at
>>>
>>> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:465)
>>>
>>>        at
>>>
>>> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
>>>
>>>        at
>>>
>>> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:90)
>>>
>>>        at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
>>>
>>>
>>> -->  I've checked the file '/user/test/51/input/kids.txt ', but, there is
>>> not
>>> strange ones. this file is healthy.
>>>
>>> Does anybody know about this error?
>>> How could I fix this one?
>>>
>>> thanks.
>>>
>>> --
>>> Junyoung Kim (juneng603@gmail.com)
>>>
>>>
>>>
>>
>>

Re: I got errors from hdfs about DataStreamer Exceptions.

Posted by Jun Young Kim <ju...@gmail.com>.
hi, harsh.
you're always giving a response very quickly. ;)

I am using a version 0.21.0 now.
before asking about this problem, I've checked already file system healthy.

$> hadoop fsck /
.....
.....
Status: HEALTHY
  Total size:    24231595038 B
  Total dirs:    43818
  Total files:   41193 (Files currently being written: 2178)
  Total blocks (validated):      40941 (avg. block size 591866 B) (Total 
open file blocks (not validated): 224)
  Minimally replicated blocks:   40941 (100.0 %)
  Over-replicated blocks:        1 (0.0024425392 %)
  Under-replicated blocks:       2 (0.0048850784 %)
  Mis-replicated blocks:         0 (0.0 %)
  Default replication factor:    2
  Average block replication:     2.1106226
  Corrupt blocks:                0
  Missing replicas:              4 (0.00462904 %)
  Number of data-nodes:          8
  Number of racks:               1

The filesystem under path '/' is HEALTHY

additionally, I found a little different error. here it is.

java.io.IOException: Bad connect ack with firstBadLink as 
10.25.241.107:50010 at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:889) 
at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820) 
at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)


here is my execution environment.

average job count : 20
max map capacity : 128
max reduce capacity : 128
avg/slot per node : 32

avg input file size per job : 200M ~ 1G

thanks.

Junyoung Kim (juneng603@gmail.com)


On 02/18/2011 11:43 AM, Harsh J wrote:
> You may want to check your HDFS health stat via 'fsck'
> (http://namenode/fsck or `hadoop fsck`). There may be a few corrupt
> files or bad DNs.
>
> Would also be good to know what exact version of Hadoop you're running.
>
> On Fri, Feb 18, 2011 at 7:59 AM, Jun Young Kim<ju...@gmail.com>  wrote:
>> hi, all.
>>
>> I got errors from hdfs.
>>
>> 2011-02-18 11:21:29[WARN ][DFSOutputStream.java]run()(519) : DataStreamer
>> Exception: java.io.IOException: Unable to create new block.
>>         at
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:832)
>>         at
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>>
>> 2011-02-18 11:21:29[WARN
>> ][DFSOutputStream.java]setupPipelineForAppendOrRecovery()(730) : Could not
>> get block locations. Source file
>> "/user/test/51/output/ehshop00newsvc-r-00000" - Aborting...
>> 2011-02-18 11:21:29[WARN ][Child.java]main()(234) : Exception running child
>> : java.io.EOFException
>>         at java.io.DataInputStream.readShort(DataInputStream.java:298)
>>         at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Status.read(DataTransferProtocol.java:113)
>>         at
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:881)
>>         at
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
>>         at
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>>
>> 2011-02-18 11:21:29[INFO ][Task.java]taskCleanup()(996) : Runnning cleanup
>> for the task
>>
>>
>>
>> I think this one is also not different error.
>>
>> org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
>> blk_-2325764274016776017_8292 file=/user/test/51/input/kids.txt
>>
>>         at
>> org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:559)
>>
>>         at
>> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:367)
>>
>>         at
>> org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)
>>
>>         at java.io.DataInputStream.read(DataInputStream.java:83)
>>
>>         at org.apache.hadoop.util.LineReader.readLine(LineReader.java:138)
>>
>>         at
>> org.apache.hadoop.mapreduce.lib.input.LineRecordReader.nextKeyValue(LineRecordReader.java:149)
>>
>>         at
>> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:465)
>>
>>         at
>> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
>>
>>         at
>> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:90)
>>
>>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
>>
>>
>> -->  I've checked the file '/user/test/51/input/kids.txt ', but, there is not
>> strange ones. this file is healthy.
>>
>> Does anybody know about this error?
>> How could I fix this one?
>>
>> thanks.
>>
>> --
>> Junyoung Kim (juneng603@gmail.com)
>>
>>
>
>

Re: I got errors from hdfs about DataStreamer Exceptions.

Posted by Harsh J <qw...@gmail.com>.
You may want to check your HDFS health stat via 'fsck'
(http://namenode/fsck or `hadoop fsck`). There may be a few corrupt
files or bad DNs.

Would also be good to know what exact version of Hadoop you're running.

On Fri, Feb 18, 2011 at 7:59 AM, Jun Young Kim <ju...@gmail.com> wrote:
> hi, all.
>
> I got errors from hdfs.
>
> 2011-02-18 11:21:29[WARN ][DFSOutputStream.java]run()(519) : DataStreamer
> Exception: java.io.IOException: Unable to create new block.
>        at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:832)
>        at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>
> 2011-02-18 11:21:29[WARN
> ][DFSOutputStream.java]setupPipelineForAppendOrRecovery()(730) : Could not
> get block locations. Source file
> "/user/test/51/output/ehshop00newsvc-r-00000" - Aborting...
> 2011-02-18 11:21:29[WARN ][Child.java]main()(234) : Exception running child
> : java.io.EOFException
>        at java.io.DataInputStream.readShort(DataInputStream.java:298)
>        at
> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$Status.read(DataTransferProtocol.java:113)
>        at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:881)
>        at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:820)
>        at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:427)
>
> 2011-02-18 11:21:29[INFO ][Task.java]taskCleanup()(996) : Runnning cleanup
> for the task
>
>
>
> I think this one is also not different error.
>
> org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block:
> blk_-2325764274016776017_8292 file=/user/test/51/input/kids.txt
>
>        at
> org.apache.hadoop.hdfs.DFSInputStream.chooseDataNode(DFSInputStream.java:559)
>
>        at
> org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo(DFSInputStream.java:367)
>
>        at
> org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:514)
>
>        at java.io.DataInputStream.read(DataInputStream.java:83)
>
>        at org.apache.hadoop.util.LineReader.readLine(LineReader.java:138)
>
>        at
> org.apache.hadoop.mapreduce.lib.input.LineRecordReader.nextKeyValue(LineRecordReader.java:149)
>
>        at
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:465)
>
>        at
> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
>
>        at
> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:90)
>
>        at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
>
>
> --> I've checked the file '/user/test/51/input/kids.txt ', but, there is not
> strange ones. this file is healthy.
>
> Does anybody know about this error?
> How could I fix this one?
>
> thanks.
>
> --
> Junyoung Kim (juneng603@gmail.com)
>
>



-- 
Harsh J
www.harshj.com