You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Iyappan Srinivasan (JIRA)" <ji...@apache.org> on 2009/04/03 10:29:12 UTC

[jira] Commented: (HADOOP-5599) Unable to run jobs when all the nodes in rack are down

    [ https://issues.apache.org/jira/browse/HADOOP-5599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12695292#action_12695292 ] 

Iyappan Srinivasan commented on HADOOP-5599:
--------------------------------------------

I was able to reproduce the above said issue.

1) In the cluster, generate data using randomwriter
2) Get one rack and kill all the datanodes in that rack only.
3) Run sort job. It fails.
4) Run Fsck from root. It says data corrupt.

I have attached the logs of these .

> Unable to run jobs when all the nodes in rack are down
> ------------------------------------------------------
>
>                 Key: HADOOP-5599
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5599
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.20.0
>            Reporter: Ramya R
>             Fix For: 0.20.0
>
>         Attachments: 5599log.txt
>
>
> Jobs such as randomwriter, sort, validator fail when all the datanodes in a rack are down.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.