You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Mayuran Yogarajah <ma...@casalemedia.com> on 2009/06/07 02:55:35 UTC

Every time the mapping phase finishes I see this

There are always a few 'Failed/Killed Task Attempts' and when I view the 
logs for
these I see:

- some that are empty, ie stdout/stderr/syslog logs are all blank
- several that say:

2009-06-06 20:47:15,309 WARN org.apache.hadoop.mapred.TaskTracker: Error running child
java.io.IOException: Filesystem closed
	at org.apache.hadoop.dfs.DFSClient.checkOpen(DFSClient.java:195)
	at org.apache.hadoop.dfs.DFSClient.access$600(DFSClient.java:59)
	at org.apache.hadoop.dfs.DFSClient$DFSInputStream.close(DFSClient.java:1359)
	at java.io.FilterInputStream.close(FilterInputStream.java:159)
	at org.apache.hadoop.mapred.LineRecordReader$LineReader.close(LineRecordReader.java:103)
	at org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:301)
	at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:173)
	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:231)
	at org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2198)



Any idea why this happens? I don't understand why I'd be seeing these 
only as the mappers get to
100%.

thanks

Re: Every time the mapping phase finishes I see this

Posted by Mayuran Yogarajah <ma...@casalemedia.com>.
I should mention..these are Hadoop streaming jobs, Hadoop version 
hadoop-0.18.3.

Any idea about the empty stdout/stderr/syslog logs? I have no way to 
really track down
whats causing them. 

thanks


Steve Loughran wrote:
> Mayuran Yogarajah wrote:
>   
>> There are always a few 'Failed/Killed Task Attempts' and when I view the
>> logs for
>> these I see:
>>
>> - some that are empty, ie stdout/stderr/syslog logs are all blank
>> - several that say:
>>
>> 2009-06-06 20:47:15,309 WARN org.apache.hadoop.mapred.TaskTracker: Error
>> running child
>> java.io.IOException: Filesystem closed
>>     at org.apache.hadoop.dfs.DFSClient.checkOpen(DFSClient.java:195)
>>     at org.apache.hadoop.dfs.DFSClient.access$600(DFSClient.java:59)
>>     at
>> org.apache.hadoop.dfs.DFSClient$DFSInputStream.close(DFSClient.java:1359)
>>     at java.io.FilterInputStream.close(FilterInputStream.java:159)
>>     at
>> org.apache.hadoop.mapred.LineRecordReader$LineReader.close(LineRecordReader.java:103)
>>
>>     at
>> org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:301)
>>     at
>> org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:173)
>>
>>     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:231)
>>     at
>> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2198)
>>
>>
>>
>> Any idea why this happens? I don't understand why I'd be seeing these
>> only as the mappers get to
>> 100%.
>>     
>
> Seen this when something in the same process got a FileSystem reference
> by FileSystem.get() and then called close() on it -it closes the client
> for every thread/class that has a reference to the same object.
>
>
> We're planning on adding more diagnostics, by tracking who closed the
> filesystem
> https://issues.apache.org/jira/browse/HADOOP-5933
>   


Re: Every time the mapping phase finishes I see this

Posted by Steve Loughran <st...@apache.org>.
Mayuran Yogarajah wrote:
> There are always a few 'Failed/Killed Task Attempts' and when I view the 
> logs for
> these I see:
> 
> - some that are empty, ie stdout/stderr/syslog logs are all blank
> - several that say:
> 
> 2009-06-06 20:47:15,309 WARN org.apache.hadoop.mapred.TaskTracker: Error 
> running child
> java.io.IOException: Filesystem closed
>     at org.apache.hadoop.dfs.DFSClient.checkOpen(DFSClient.java:195)
>     at org.apache.hadoop.dfs.DFSClient.access$600(DFSClient.java:59)
>     at 
> org.apache.hadoop.dfs.DFSClient$DFSInputStream.close(DFSClient.java:1359)
>     at java.io.FilterInputStream.close(FilterInputStream.java:159)
>     at 
> org.apache.hadoop.mapred.LineRecordReader$LineReader.close(LineRecordReader.java:103) 
> 
>     at 
> org.apache.hadoop.mapred.LineRecordReader.close(LineRecordReader.java:301)
>     at 
> org.apache.hadoop.mapred.MapTask$TrackedRecordReader.close(MapTask.java:173) 
> 
>     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:231)
>     at 
> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2198)
> 
> 
> 
> Any idea why this happens? I don't understand why I'd be seeing these 
> only as the mappers get to
> 100%.

Seen this when something in the same process got a FileSystem reference 
by FileSystem.get() and then called close() on it -it closes the client 
for every thread/class that has a reference to the same object.


We're planning on adding more diagnostics, by tracking who closed the 
filesystem
https://issues.apache.org/jira/browse/HADOOP-5933