You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Michael Basnight <mb...@gmail.com> on 2009/05/05 21:34:30 UTC

DFS # of blocks

howdy all, im doing some hadoop testing (so im still new to it), and  
im running into an error. ( DataStreamer Exception:  
java.io.IOException: Unable to create new block.) My DFS is not large  
(971 files and directories, 3906 blocks = 4877 total. Heap Size is  
13.9 MB / 966.69 MB (1%) ), and definitely not full ;), and hadoop is  
writing to a separate file for each reducer output key (roughly 4K  
keys), but after like 900 it just dies. I am using a child of  
MultipleTextOutputFormat<Text, Text> for my output format that creates  
a file for the key and puts the list of values in it in the same dfs  
folder. Below is the last 20 lines of logs for the node that it failed  
on. Any thoughts?

eTask: Read 122146089 bytes from map-output for  
attempt_200905051459_0001_m_000000_0
2009-05-05 15:09:59,672 INFO org.apache.hadoop.mapred.ReduceTask: Rec  
#1 from attempt_200905051459_0001_m_000000_0 -> (19, 2771) from hadoop2
2009-05-05 15:10:00,652 INFO org.apache.hadoop.mapred.ReduceTask:  
GetMapEventsThread exiting
2009-05-05 15:10:00,652 INFO org.apache.hadoop.mapred.ReduceTask:  
getMapsEventsThread joined.
2009-05-05 15:10:00,652 INFO org.apache.hadoop.mapred.ReduceTask:  
Closed ram manager
2009-05-05 15:10:00,652 INFO org.apache.hadoop.mapred.ReduceTask:  
Interleaved on-disk merge complete: 0 files left.
2009-05-05 15:10:00,652 INFO org.apache.hadoop.mapred.ReduceTask: In- 
memory merge complete: 2 files left.
2009-05-05 15:10:00,780 INFO org.apache.hadoop.mapred.Merger: Merging  
2 sorted segments
2009-05-05 15:10:00,780 INFO org.apache.hadoop.mapred.Merger: Down to  
the last merge-pass, with 2 segments left of total size: 245292507 bytes
2009-05-05 15:10:06,276 INFO org.apache.hadoop.mapred.ReduceTask:  
Merged 2 segments, 245292507 bytes to disk to satisfy reduce memory  
limit
2009-05-05 15:10:06,280 INFO org.apache.hadoop.mapred.ReduceTask:  
Merging 1 files, 245292509 bytes from disk
2009-05-05 15:10:06,284 INFO org.apache.hadoop.mapred.ReduceTask:  
Merging 0 segments, 0 bytes from memory into reduce
2009-05-05 15:10:06,284 INFO org.apache.hadoop.mapred.Merger: Merging  
1 sorted segments
2009-05-05 15:10:06,312 INFO org.apache.hadoop.mapred.Merger: Down to  
the last merge-pass, with 1 segments left of total size: 245292505 bytes
2009-05-05 15:10:35,346 INFO org.apache.hadoop.hdfs.DFSClient:  
Exception in createBlockOutputStream java.io.EOFException
2009-05-05 15:10:35,374 INFO org.apache.hadoop.hdfs.DFSClient:  
Abandoning block blk_-2874778794594289753_8565
2009-05-05 15:10:41,402 INFO org.apache.hadoop.hdfs.DFSClient:  
Exception in createBlockOutputStream java.io.EOFException
2009-05-05 15:10:41,402 INFO org.apache.hadoop.hdfs.DFSClient:  
Abandoning block blk_243746846946054460_8565
2009-05-05 15:10:47,411 INFO org.apache.hadoop.hdfs.DFSClient:  
Exception in createBlockOutputStream java.io.EOFException
2009-05-05 15:10:47,411 INFO org.apache.hadoop.hdfs.DFSClient:  
Abandoning block blk_-4508716893999737242_8565
2009-05-05 15:10:53,419 INFO org.apache.hadoop.hdfs.DFSClient:  
Exception in createBlockOutputStream java.io.EOFException
2009-05-05 15:10:53,419 INFO org.apache.hadoop.hdfs.DFSClient:  
Abandoning block blk_-2679897353336358687_8565
2009-05-05 15:10:59,423 WARN org.apache.hadoop.hdfs.DFSClient:  
DataStreamer Exception: java.io.IOException: Unable to create new block.
	at org.apache.hadoop.hdfs.DFSClient 
$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2781)
	at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access 
$2000(DFSClient.java:2046)
	at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream 
$DataStreamer.run(DFSClient.java:2232)

2009-05-05 15:10:59,423 WARN org.apache.hadoop.hdfs.DFSClient: Error  
Recovery for block blk_-2679897353336358687_8565 bad datanode[0] nodes  
== null
2009-05-05 15:10:59,423 WARN org.apache.hadoop.hdfs.DFSClient: Could  
not get block locations. Source file "/testing/output/medium_output/ 
_temporary/_attempt_200905051459_0001_r_000000_3/sometestingkey" -  
Aborting...
2009-05-05 15:12:24,960 WARN org.apache.hadoop.mapred.TaskTracker:  
Error running child
java.io.EOFException
	at java.io.DataInputStream.readByte(Unknown Source)
	at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:298)
	at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:319)
	at org.apache.hadoop.io.Text.readString(Text.java:400)
	at org.apache.hadoop.hdfs.DFSClient 
$DFSOutputStream.createBlockOutputStream(DFSClient.java:2837)
	at org.apache.hadoop.hdfs.DFSClient 
$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2762)
	at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access 
$2000(DFSClient.java:2046)
	at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream 
$DataStreamer.run(DFSClient.java:2232)
2009-05-05 15:12:24,972 INFO org.apache.hadoop.mapred.TaskRunner:  
Runnning cleanup for the task