You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-issues@hadoop.apache.org by "Devaraj K (JIRA)" <ji...@apache.org> on 2011/05/10 15:50:47 UTC

[jira] [Commented] (MAPREDUCE-2481) SocketTimeoutException is coming in the reduce task when the data size is very high

    [ https://issues.apache.org/jira/browse/MAPREDUCE-2481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13031188#comment-13031188 ] 

Devaraj K commented on MAPREDUCE-2481:
--------------------------------------

It has already read many other fields and so there is no chance of a network disconnect, because always problem is coming while reading same field (isMap property in TaskCompletionEvent object).

{code:title=TaskCompletionEvent.java|borderStyle=solid}
  public void readFields(DataInput in) throws IOException {
    taskId.readFields(in); 
    idWithinJob = WritableUtils.readVInt(in);
    isMap = in.readBoolean();
{code}

> SocketTimeoutException is coming in the reduce task when the data size is very high
> -----------------------------------------------------------------------------------
>
>                 Key: MAPREDUCE-2481
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2481
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>          Components: task
>    Affects Versions: 0.20.2
>            Reporter: Devaraj K
>
> SocketTimeoutException is coming when reduce task tries to read MapTaskCompletionEventsUpdate object from task tracker, it is able to read reset, TaskCompletionEvent.taskId, TaskCompletionEvent.idWithinJob properties and it is failing for reading the property isMap in TaskCompletionEvent which is of type boolean. This exception is coming multiple times.
> {code}
> 2011-04-20 15:58:03,037 FATAL mapred.TaskTracker (TaskTracker.java:fatalError(2812)) - Task: attempt_201104201115_0010_r_000002_0 - Killed : java.io.IOException:  Tried for the max ping retries On TimeOut :1
> 	at org.apache.hadoop.ipc.Client.checkPingRetries(Client.java:1342)
> 	at org.apache.hadoop.ipc.Client$Connection$PingInputStream.read(Client.java:402)
> 	at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
> 	at java.io.BufferedInputStream.read(BufferedInputStream.java:237)
> 	at java.io.DataInputStream.readBoolean(DataInputStream.java:225)
> 	at org.apache.hadoop.mapred.TaskCompletionEvent.readFields(TaskCompletionEvent.java:230)
> 	at org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate.readFields(MapTaskCompletionEventsUpdate.java:64)
> 	at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:245)
> 	at org.apache.hadoop.io.ObjectWritable.readFields(ObjectWritable.java:69)
> 	at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:698)
> 	at org.apache.hadoop.ipc.Client$Connection.run(Client.java:593)
> Caused by: java.net.SocketTimeoutException: 60000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/127.0.0.1:45798 remote=/127.0.0.1:35419]
> 	at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:165)
> 	at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
> 	at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
> 	at java.io.FilterInputStream.read(FilterInputStream.java:116)
> 	at org.apache.hadoop.ipc.Client$Connection$PingInputStream.read(Client.java:397)
> 	... 9 more
> {code}
>  org.mortbay.jetty.EofException is also coming many times in the logs as described in MAPREDUCE-5.
> {code}
> 2011-04-20 15:57:20,748 WARN  mapred.TaskTracker (TaskTracker.java:doGet(3164)) - getMapOutput(attempt_201104201115_0010_m_000038_0,4) failed :
> org.mortbay.jetty.EofException
> 	at org.mortbay.jetty.HttpGenerator.flush(HttpGenerator.java:787)
> {code}

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira