You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-user@hadoop.apache.org by Manish Bhoge <ma...@rocketmail.com> on 2013/08/07 09:00:13 UTC

Namenode is failing with expception to join

I have all configuration fine. But whenever i start namenode it fails with a below exception. No clue where to fix this?


"2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete. 2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.

Re: Namenode is failing with expception to join

Posted by Jitendra Yadav <je...@gmail.com>.
Hi,

Did you configured your Name Node to store multiple copies of its metadata?.

You can recover your name node in that situation.

#hadoop namenode -recover

it will ask you whether you want to continue or not, Please follow the
instructions.

Thanks
On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

>  I am not using HA here. All I am trying here is to make a 2 node
> cluster. But before that i wanted to make sure that i am setting up
> everything right and make the HDFS up on Pseudo distributed mode. However,
> I am suspecting a mistake in my /etc/hosts file. As, I have rename the
> local host to myhost-1
>
> Please suggest.
>
>   ------------------------------
> *From:* Azuryy Yu <az...@gmail.com>
> *To:* user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com>
> *Sent:* Wednesday, 7 August 2013 1:08 PM
> *Subject:* Re: Namenode is failing with expception to join
>
>  Manish,
>
> you stop HDFS then start HDFS on the standby name node right?
>
> please looked at https://issues.apache.org/jira/browse/HDFS-5058
>
> there are two solutions:
> 1) start HDFS on the active name node, nor SBN
> 2) copy {namenode.name.dir}/* to the SBN
>
> I advice #1.
>
>
>
>
> On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:
>
>  I have all configuration fine. But whenever i start namenode it fails
> with a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>
>
>
>

Re: Namenode is failing with expception to join

Posted by Jitendra Yadav <je...@gmail.com>.
Hi,

Did you configured your Name Node to store multiple copies of its metadata?.

You can recover your name node in that situation.

#hadoop namenode -recover

it will ask you whether you want to continue or not, Please follow the
instructions.

Thanks
On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

>  I am not using HA here. All I am trying here is to make a 2 node
> cluster. But before that i wanted to make sure that i am setting up
> everything right and make the HDFS up on Pseudo distributed mode. However,
> I am suspecting a mistake in my /etc/hosts file. As, I have rename the
> local host to myhost-1
>
> Please suggest.
>
>   ------------------------------
> *From:* Azuryy Yu <az...@gmail.com>
> *To:* user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com>
> *Sent:* Wednesday, 7 August 2013 1:08 PM
> *Subject:* Re: Namenode is failing with expception to join
>
>  Manish,
>
> you stop HDFS then start HDFS on the standby name node right?
>
> please looked at https://issues.apache.org/jira/browse/HDFS-5058
>
> there are two solutions:
> 1) start HDFS on the active name node, nor SBN
> 2) copy {namenode.name.dir}/* to the SBN
>
> I advice #1.
>
>
>
>
> On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:
>
>  I have all configuration fine. But whenever i start namenode it fails
> with a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>
>
>
>

Re: Namenode is failing with expception to join

Posted by Jitendra Yadav <je...@gmail.com>.
Hi,

Did you configured your Name Node to store multiple copies of its metadata?.

You can recover your name node in that situation.

#hadoop namenode -recover

it will ask you whether you want to continue or not, Please follow the
instructions.

Thanks
On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

>  I am not using HA here. All I am trying here is to make a 2 node
> cluster. But before that i wanted to make sure that i am setting up
> everything right and make the HDFS up on Pseudo distributed mode. However,
> I am suspecting a mistake in my /etc/hosts file. As, I have rename the
> local host to myhost-1
>
> Please suggest.
>
>   ------------------------------
> *From:* Azuryy Yu <az...@gmail.com>
> *To:* user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com>
> *Sent:* Wednesday, 7 August 2013 1:08 PM
> *Subject:* Re: Namenode is failing with expception to join
>
>  Manish,
>
> you stop HDFS then start HDFS on the standby name node right?
>
> please looked at https://issues.apache.org/jira/browse/HDFS-5058
>
> there are two solutions:
> 1) start HDFS on the active name node, nor SBN
> 2) copy {namenode.name.dir}/* to the SBN
>
> I advice #1.
>
>
>
>
> On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:
>
>  I have all configuration fine. But whenever i start namenode it fails
> with a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>
>
>
>

Re: Namenode is failing with expception to join

Posted by Jitendra Yadav <je...@gmail.com>.
Hi,

Did you configured your Name Node to store multiple copies of its metadata?.

You can recover your name node in that situation.

#hadoop namenode -recover

it will ask you whether you want to continue or not, Please follow the
instructions.

Thanks
On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

>  I am not using HA here. All I am trying here is to make a 2 node
> cluster. But before that i wanted to make sure that i am setting up
> everything right and make the HDFS up on Pseudo distributed mode. However,
> I am suspecting a mistake in my /etc/hosts file. As, I have rename the
> local host to myhost-1
>
> Please suggest.
>
>   ------------------------------
> *From:* Azuryy Yu <az...@gmail.com>
> *To:* user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com>
> *Sent:* Wednesday, 7 August 2013 1:08 PM
> *Subject:* Re: Namenode is failing with expception to join
>
>  Manish,
>
> you stop HDFS then start HDFS on the standby name node right?
>
> please looked at https://issues.apache.org/jira/browse/HDFS-5058
>
> there are two solutions:
> 1) start HDFS on the active name node, nor SBN
> 2) copy {namenode.name.dir}/* to the SBN
>
> I advice #1.
>
>
>
>
> On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:
>
>  I have all configuration fine. But whenever i start namenode it fails
> with a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>
>
>
>

Re: Namenode is failing with expception to join

Posted by Manish Bhoge <ma...@rocketmail.com>.
I am not using HA here. All I am trying here is to make a 2 node cluster. But before that i wanted to make sure that i am setting up everything right and make the HDFS up on Pseudo distributed mode. However, I am suspecting a mistake in my /etc/hosts file. As, I have rename the local host to myhost-1
 

Please suggest.



________________________________
 From: Azuryy Yu <az...@gmail.com>
To: user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com> 
Sent: Wednesday, 7 August 2013 1:08 PM
Subject: Re: Namenode is failing with expception to join
 


Manish,
 
you stop HDFS then start HDFS on the standby name node right?  
 
please looked at https://issues.apache.org/jira/browse/HDFS-5058
 
there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN 
 
I advice #1.
 
 



On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com> wrote:

I have all configuration fine. But whenever i start namenode it fails with a below exception. No clue where to fix this?
>
>
>
>"2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete. 2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.

Re: Namenode is failing with expception to join

Posted by Manish Bhoge <ma...@rocketmail.com>.
I am not using HA here. All I am trying here is to make a 2 node cluster. But before that i wanted to make sure that i am setting up everything right and make the HDFS up on Pseudo distributed mode. However, I am suspecting a mistake in my /etc/hosts file. As, I have rename the local host to myhost-1
 

Please suggest.



________________________________
 From: Azuryy Yu <az...@gmail.com>
To: user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com> 
Sent: Wednesday, 7 August 2013 1:08 PM
Subject: Re: Namenode is failing with expception to join
 


Manish,
 
you stop HDFS then start HDFS on the standby name node right?  
 
please looked at https://issues.apache.org/jira/browse/HDFS-5058
 
there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN 
 
I advice #1.
 
 



On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com> wrote:

I have all configuration fine. But whenever i start namenode it fails with a below exception. No clue where to fix this?
>
>
>
>"2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete. 2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.

Re: Namenode is failing with expception to join

Posted by Manish Bhoge <ma...@rocketmail.com>.
I am not using HA here. All I am trying here is to make a 2 node cluster. But before that i wanted to make sure that i am setting up everything right and make the HDFS up on Pseudo distributed mode. However, I am suspecting a mistake in my /etc/hosts file. As, I have rename the local host to myhost-1
 

Please suggest.



________________________________
 From: Azuryy Yu <az...@gmail.com>
To: user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com> 
Sent: Wednesday, 7 August 2013 1:08 PM
Subject: Re: Namenode is failing with expception to join
 


Manish,
 
you stop HDFS then start HDFS on the standby name node right?  
 
please looked at https://issues.apache.org/jira/browse/HDFS-5058
 
there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN 
 
I advice #1.
 
 



On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com> wrote:

I have all configuration fine. But whenever i start namenode it fails with a below exception. No clue where to fix this?
>
>
>
>"2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete. 2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.

Re: Namenode is failing with expception to join

Posted by Manish Bhoge <ma...@rocketmail.com>.
I am not using HA here. All I am trying here is to make a 2 node cluster. But before that i wanted to make sure that i am setting up everything right and make the HDFS up on Pseudo distributed mode. However, I am suspecting a mistake in my /etc/hosts file. As, I have rename the local host to myhost-1
 

Please suggest.



________________________________
 From: Azuryy Yu <az...@gmail.com>
To: user@hadoop.apache.org; Manish Bhoge <ma...@rocketmail.com> 
Sent: Wednesday, 7 August 2013 1:08 PM
Subject: Re: Namenode is failing with expception to join
 


Manish,
 
you stop HDFS then start HDFS on the standby name node right?  
 
please looked at https://issues.apache.org/jira/browse/HDFS-5058
 
there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN 
 
I advice #1.
 
 



On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com> wrote:

I have all configuration fine. But whenever i start namenode it fails with a below exception. No clue where to fix this?
>
>
>
>"2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete. 2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.

Re: Namenode is failing with expception to join

Posted by Azuryy Yu <az...@gmail.com>.
Manish,

you stop HDFS then start HDFS on the standby name node right?

please looked at https://issues.apache.org/jira/browse/HDFS-5058

there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN

I advice #1.




On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

> I have all configuration fine. But whenever i start namenode it fails with
> a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>

Re: Namenode is failing with expception to join

Posted by Azuryy Yu <az...@gmail.com>.
Manish,

you stop HDFS then start HDFS on the standby name node right?

please looked at https://issues.apache.org/jira/browse/HDFS-5058

there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN

I advice #1.




On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

> I have all configuration fine. But whenever i start namenode it fails with
> a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>

Re: Namenode is failing with expception to join

Posted by Azuryy Yu <az...@gmail.com>.
Manish,

you stop HDFS then start HDFS on the standby name node right?

please looked at https://issues.apache.org/jira/browse/HDFS-5058

there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN

I advice #1.




On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

> I have all configuration fine. But whenever i start namenode it fails with
> a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>

Re: Namenode is failing with expception to join

Posted by Azuryy Yu <az...@gmail.com>.
Manish,

you stop HDFS then start HDFS on the standby name node right?

please looked at https://issues.apache.org/jira/browse/HDFS-5058

there are two solutions:
1) start HDFS on the active name node, nor SBN
2) copy {namenode.name.dir}/* to the SBN

I advice #1.




On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <ma...@rocketmail.com>wrote:

> I have all configuration fine. But whenever i start namenode it fails with
> a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream: Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255' to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1, but got txid 515247.
>
>