You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-issues@hadoop.apache.org by "Andras Bokor (JIRA)" <ji...@apache.org> on 2018/04/10 11:39:00 UTC

[jira] [Resolved] (HADOOP-5342) DataNodes do not start up because InconsistentFSStateException on just part of the disks in use

     [ https://issues.apache.org/jira/browse/HADOOP-5342?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Andras Bokor resolved HADOOP-5342.
----------------------------------
    Resolution: Cannot Reproduce

Last reported occurrence was in 2010 so closing as Cannot Reproduce. Please reopen if you still experience this.

> DataNodes do not start up because InconsistentFSStateException on just part of the disks in use
> -----------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-5342
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5342
>             Project: Hadoop Common
>          Issue Type: Bug
>    Affects Versions: 0.18.2
>            Reporter: Christian Kunz
>            Assignee: Hairong Kuang
>            Priority: Critical
>
> After restarting a cluster (including rebooting) the dfs got corrupted because many DataNodes did not start up, running into the following exception:
> 2009-02-26 22:33:53,774 ERROR org.apache.hadoop.dfs.DataNode: org.apache.hadoop.dfs.InconsistentFSStateException: Directory xxx  is in an inconsistent state: version file in current directory is missing.
> 	at org.apache.hadoop.dfs.Storage$StorageDirectory.analyzeStorage(Storage.java:326)
> 	at org.apache.hadoop.dfs.DataStorage.recoverTransitionRead(DataStorage.java:105)
> 	at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:306)
> 	at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:223)
> 	at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:3030)
> 	at org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2985)
> 	at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2993)
> 	at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3115)
> This happens when using multiple disks with at least one previously marked as read-only, such that the storage version became out-dated, but after reboot it was mounted read-write, resulting in the DataNode not starting because of out-dated version.
> This is a big headache. If a DataNode has multiple disks of which at least one has the correct storage version then out-dated versions should not bring down the DataNode.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-issues-help@hadoop.apache.org