You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Lars George <la...@worldlingo.com> on 2008/01/04 20:19:54 UTC
WrongRegionException - How do I fix it?
Hi,
I have inserted about 3.5m documents in a single two column table in
HBase running on 32 nodes. So far I was able to insert most data, but
with the last million or so I am stuck with this error:
org.apache.hadoop.hbase.WrongRegionException: Requested row out of range
for HRegion docs,DC20020099792_20020725,9149203683830573099,
startKey='DC20020099792_20020725', endKey='DC20020099792_20020725',
row='DC20020099841_20020725'
Querying for the document returns nothing, means it looks like the
document does not exist - although I am sure I tried inserting it a few
times. Deleting or trying to re-insert returns the above error, both
through the API (using HTable) as well as through the HBase shell.
I tried a restart of Hadoop/HBase to no avail. How do fix this problem?
Any help is appreciated.
Best regards,
Lars
----
Lars George, CTO
WorldLingo
Re: WrongRegionException - How do I fix it?
Posted by Lars George <la...@worldlingo.com>.
Hi Stack,
That also does not seem right?
192.168.105.55:60020 -6321685367535019868 requests: 0 regions:
64
192.168.105.51:60020 -4174238310567886366 requests: 0 regions:
51
192.168.105.17:60020 5581663697624462721 requests: 0 regions:
156
192.168.105.57:60020 -5211607822000744542 requests: 0 regions:
62
192.168.105.31:60020 7888193037635443812 requests: 0 regions:
287
192.168.105.59:60020 -46386689243367454 requests: 0 regions:
95
192.168.105.45:60020 8533079767141988197 requests: 0 regions:
42
192.168.105.77:60020 7803898403387937725 requests: 0 regions:
67
192.168.105.49:60020 -163550390447606491 requests: 0 regions:
217
192.168.105.75:60020 -6446348712848682453 requests: 0 regions:
44
192.168.105.27:60020 4779581083944377871 requests: 0 regions:
109
192.168.105.65:60020 4901865573283507122 requests: 0 regions:
54
192.168.105.41:60020 4787302286434914341 requests: 0 regions:
114
192.168.105.33:60020 -2508294300014104969 requests: 0 regions:
66
192.168.105.73:60020 -2334633498376763895 requests: 0 regions:
88
192.168.105.63:60020 8674021859672682545 requests: 0 regions:
48
192.168.105.35:60020 1407450539566532332 requests: 4296 regions:
1
192.168.105.47:60020 -3402160042637529057 requests: 0 regions:
35
192.168.105.37:60020 8133142711395683731 requests: 0 regions:
67
192.168.105.71:60020 -9207688945963624521 requests: 0 regions:
57
192.168.105.61:60020 -7699687997693377026 requests: 0 regions:
33
192.168.105.23:60020 4216774296721487713 requests: 0 regions:
89
192.168.105.13:60020 -6032091814002308094 requests: 0 regions:
44
192.168.105.15:60020 -4136342318896158593 requests: 0 regions:
320
192.168.105.67:60020 5042396379505131550 requests: 0 regions:
34
192.168.105.21:60020 -7677313271932692345 requests: 5 regions:
172
192.168.105.53:60020 1635187659254945320 requests: 0 regions:
0
192.168.105.69:60020 5310455167394956263 requests: 0 regions:
157
192.168.105.25:60020 1421797545576665774 requests: 0 regions:
97
192.168.105.43:60020 -6942740213227675121 requests: 0 regions:
79
192.168.105.29:60020 3244292531316078025 requests: 0 regions:
98
192.168.105.39:60020 -2915937415126913637 requests: 0 regions:
134
See that .35 server. Way different than the others. This is server 26 in
the email I sent before, the one with all sorts of errors not being able
to find regions, delete blocks etc.
Can I somehow wipe this clean?
Lars
stack wrote:
> Lars George wrote:
>> Hi,
>>
>> I have inserted about 3.5m documents in a single two column table in
>> HBase running on 32 nodes. So far I was able to insert most data, but
>> with the last million or so I am stuck with this error:
>>
>> org.apache.hadoop.hbase.WrongRegionException: Requested row out of
>> range for HRegion docs,DC20020099792_20020725,9149203683830573099,
>> startKey='DC20020099792_20020725', endKey='DC20020099792_20020725',
>> row='DC20020099841_20020725'
>
> This happens every time you try to do an insert?
>
>> Querying for the document returns nothing, means it looks like the
>> document does not exist - although I am sure I tried inserting it a
>> few times. Deleting or trying to re-insert returns the above error,
>> both through the API (using HTable) as well as through the HBase shell.
>>
>> I tried a restart of Hadoop/HBase to no avail. How do fix this
>> problem? Any help is appreciated.
>
> You have DEBUG enabled for hbase (See
> http://wiki.apache.org/lucene-hadoop/Hbase/FAQ#4). Do the logs tell
> you anything more: e.g. any interesting exceptions?
>
> Which hbase version?
>
> St.Ack
>
Re: WrongRegionException - How do I fix it?
Posted by Lars George <la...@worldlingo.com>.
Hi Stack,
Yes, I read that on Nabble today too, that is why I was asking actually.
What is the timeline for this? I will reply to Jim's message too.
Thanks for pointing that out.
Lars
stack wrote:
> stack wrote:
>> Regards your having to do this again in the near future, hopefully
>> not...
>
> Related, see
> http://www.nabble.com/Question-for-HBase-users-tc14607732.html.
> St.Ack
>
Re: WrongRegionException - How do I fix it?
Posted by stack <st...@duboce.net>.
stack wrote:
> Regards your having to do this again in the near future, hopefully not...
Related, see http://www.nabble.com/Question-for-HBase-users-tc14607732.html.
St.Ack
Re: WrongRegionException - How do I fix it?
Posted by stack <st...@duboce.net>.
Lars George wrote:
> Eeek, means what? Replace jar files with the trunk version and the
> reformat the whole Hadoop dfs (since I only have HBase on top of it)
> and then reimporting all 4.5m documents? What are my chances that
> there are more profound changes coming that require me to do that
> again in the near future? Just asking.
Yes. Start over especially if you are running hbase only. That'd be
cleanest and quickest path to hbase TRUNK nirvana.
Regards your having to do this again in the near future, hopefully not,
but please allow for the fact that hbase is beta software. There is at
least one well-known means of losing data (HADOOP-1700) and probably
more that we have yet to discover. Because of this, we want to be clear
that folks should be prepared for data loss in their hbase installations.
St.Ack
P.S. Lars, there is an #hbase channel up on irc.freenode.net if you need
more immediate help or need suggestions for how you might faster
populate hbase.
Re: WrongRegionException - How do I fix it?
Posted by Lars George <la...@worldlingo.com>.
Stack,
Eeek, means what? Replace jar files with the trunk version and the
reformat the whole Hadoop dfs (since I only have HBase on top of it) and
then reimporting all 4.5m documents? What are my chances that there are
more profound changes coming that require me to do that again in the
near future? Just asking.
Lars
stack wrote:
> Lars George wrote:
>> Hi Stack,
>>
>> Can and will do, but does that make the error go away, i.e.
>> automagically fix it? Or is it broken and nothing can be done about
>> it now?
> Your current install is broke. We could try spending time getting it
> back into a healthy state but TRUNK is more robust than the first
> hbase release, 0.15.x. It also lacks many of the bugs you reproduced
> in your earlier mail. I'd suggest that our time would be better spent
> working with the latest hbase particularly as patches to hbase in the
> 0.15 branch are unlikely.
>
> Thanks Lars,
> St.Ack
>
Re: WrongRegionException - How do I fix it?
Posted by stack <st...@duboce.net>.
Lars George wrote:
> Hi Stack,
>
> Can and will do, but does that make the error go away, i.e.
> automagically fix it? Or is it broken and nothing can be done about it
> now?
Your current install is broke. We could try spending time getting it
back into a healthy state but TRUNK is more robust than the first hbase
release, 0.15.x. It also lacks many of the bugs you reproduced in your
earlier mail. I'd suggest that our time would be better spent working
with the latest hbase particularly as patches to hbase in the 0.15
branch are unlikely.
Thanks Lars,
St.Ack
Re: WrongRegionException - How do I fix it?
Posted by Lars George <la...@worldlingo.com>.
Hi Stack,
Can and will do, but does that make the error go away, i.e.
automagically fix it? Or is it broken and nothing can be done about it now?
Lars
stack wrote:
> If possible, please move to TRUNK instead. Most of the below have
> been addressed there (I can send you a patch if you want to run hbase
> TRUNK on hadoop 0.15.x).
>
> Further comments inline below:
>
>
> Lars George wrote:
>> Hi Stack,
>>
>> Yes, it happens every time I insert particular rows. Before it would
>> fail every now and so often, but since now all "good" rows are
>> inserted I am stuck with the ones that do not insert. And I am sure
>> they did once, with no error. So they are in there in limbo, but I
>> cannot retrieve nor delete or insert them.
>>
>> It mentions in the FAQ that I can switch on debugging through the UI,
>> but I cannot see where. I am using version 0.15.1, is that supposed
>> to have that option or do I need to go the log4j.properties plus
>> restart route?
>
> This is a post-0.15 release feature (It says post-0.15.x in the FAQ).
>
>
>> I have errors all the time - which quite frankly worry me. Here a
>> list of what I see so far:
>>
>> 1. At startup
>>
>> ==>
>> /usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-40.worldlingo.com.log
>> <==
>> 2008-01-03 14:11:22,512 WARN org.apache.hadoop.util.NativeCodeLoader:
>> Unable to load native-hadoop library for your platform... using
>> builtin-java classes where applicable
>> 2008-01-03 14:11:29,808 WARN org.apache.hadoop.hbase.HRegionServer:
>> Processing message (Retry: 0)
>> java.io.IOException: java.io.IOException:
>> java.util.ConcurrentModificationException
> ....
>
>
> Fixed in TRUNK
>
>
>>
>>
>> 2. Sporadically
>>
>> 2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode:
>> Unexpected error trying to delete block blk_-8931657506153335343.
>> Block not found in blockMap.
>> 2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode:
>> Unexpected error trying to delete block blk_3775459202881005176.
>> Block not found in blockMap.
>> 2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode:
>> Unexpected error trying to delete block blk_-283089329129695997.
>> Block not found in blockMap.
>> 2008-01-03 21:32:00,644 WARN org.apache.hadoop.dfs.DataNode:
>> java.io.IOException: Error in deleting blocks.
>> at org.apache.hadoop.dfs.FSDataset.invalidate(FSDataset.java:719)
>> at
>> org.apache.hadoop.dfs.DataNode.processCommand(DataNode.java:625)
>> at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:528)
>> at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1494)
>> at java.lang.Thread.run(Thread.java:595)
>>
>> These come with all sorts of blocks, they do not appear too often,
>> but on a regular basis.
>>
> 0.15.x was doing updates against files that had been removed by
> another thread. Made for strange errors in hdfs. That said, I don't
> recall having seen the above. Do you have namenode at DEBUG level?
> If so, try tracing the above problematic blocks therein; see if you
> can figure out a story as to what happened with these blocks.
>
> If HDFS is in an inconsistent state, hbase will be inconsistent too.
>
>> 3. Misc
>>
>> I see these too from this morning
>>
>> 2008-01-04 08:23:30,616 ERROR org.apache.hadoop.hbase.HRegionServer:
>> unable to process message: MSG_REGION_OPEN : regionname: do
>> cs,DC20020096869_20020725,43610073395851568, startKey:
>> <DC20020096869_20020725>, tableDesc: {name: docs, families:
>> {contents:={n
>> ame: contents, max versions: 3, compression: NONE, in memory: false,
>> max length: 2147483647, bloom filter: none}, language:={name: l
>> anguage, max versions: 3, compression: NONE, in memory: false, max
>> length: 2147483647, bloom filter: none}, mimetype:={name: mimetyp
>> e, max versions: 3, compression: NONE, in memory: false, max length:
>> 2147483647, bloom filter: none}}}
>> java.io.IOException: java.io.IOException: Cannot open filename
>> /hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/co
>> ntents/info/1501965039462307633
>> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
>> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>> at java.lang.reflect.Method.invoke(Method.java:585)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>>
>> at
>> sun.reflect.GeneratedConstructorAccessor7.newInstance(Unknown Source)
>> at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>>
>> at
>> java.lang.reflect.Constructor.newInstance(Constructor.java:494)
>> at
>> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
>>
>> at
>> org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
>>
>> at
>> org.apache.hadoop.hbase.HRegionServer$Worker.run(HRegionServer.java:903)
>> at java.lang.Thread.run(Thread.java:595)
>>
>>
>
> I'd guess this file is made of the above cited problematic blocks.
> Can you find it on hdfs? Can you download it? (Try doing a
> './bin/hadoop fs fsck /HBASE_DIR' -- see what it says).
>
>
>
>> Another one is this:
>>
>> ==>
>> /usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-62.worldlingo.com.log
>> <==
>> 2008-01-04 08:16:32,001 WARN org.apache.hadoop.hbase.HRegion: Region
>> docs,DC20020099792_20020725,9149203683830573099 is NOT spli
>> table though its aggregate size is 111.4m and desired size is 64.0m
>>
>> These come up with different region numbers.
>>
> This is ok. We want to split the region because its > 64MB but this
> region has outstanding references to another, parent region, so is not
> yet splittable (It should have turned splittable a little later in
> your log).
>
>
>> And another one:
>>
>> 2008-01-03 11:27:55,437 WARN org.apache.hadoop.hbase.HStore: Failed
>> getting store size
>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File does
>> not exist
>> at
>> org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:489)
>> at
>> org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1360)
>> at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:428)
>> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>> at java.lang.reflect.Method.invoke(Method.java:585)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>>
> Fixed in TRUNK
>
>
>>
>> And another one:
>>
>> 2008-01-03 15:43:39,590 WARN org.apache.hadoop.dfs.DataNode: Got
>> exception while serving blk_3676251342939485484 to /192.168.105.21:
>> java.io.IOException: Block blk_3676251342939485484 is not valid.
>> at
>> org.apache.hadoop.dfs.FSDataset.getBlockFile(FSDataset.java:528)
>> at
>> org.apache.hadoop.dfs.DataNode$BlockSender.<init>(DataNode.java:1051)
>> at
>> org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:843)
>> at
>> org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:801)
>> at java.lang.Thread.run(Thread.java:595)
>>
>> Again, they come with different block numbers.
>>
> This hdfs exception may be recoverable IIRC; hdfs gets the block
> elsewhere.
>
>
>> 4. Inserting document with errors
>>
>> If I try to add one of those documents, where I get an error back,
>> this is what I see in the logs so far:
>>
>> Uploading -> DC20020099841_20020725
>>
>> Server 20:
>>
>>
> ....
>> 2007-12-30 17:26:42,392 INFO org.mortbay.util.Container: Started
>> HttpContext[/static,/static]
>> 2007-12-30 17:26:42,395 INFO org.mortbay.http.SocketListener: Started
>> SocketListener on 0.0.0.0:60030
>> 2007-12-30 17:26:42,395 INFO org.mortbay.util.Container: Started
>> org.mortbay.jetty.Server@1fd0fafc
>> 2007-12-30 17:26:42,396 INFO org.apache.hadoop.ipc.Server: IPC Server
>> listener on 60020: starting
>> 2007-12-30 17:26:42,397 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 0 on 60020: starting
>> 2007-12-30 17:26:42,402 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 1 on 60020: starting
>> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 3 on 60020: starting
>> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 4 on 60020: starting
>> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 5 on 60020: starting
>> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 6 on 60020: starting
>>
>
>
> Please set these to run at DEBUG level, at least while we are trying
> to figure out whats going on.
>
>>
>> Server 26:
>>
>> 2008-01-04 12:18:28,125 WARN org.apache.hadoop.hbase.HRegionServer:
>> java.io.IOException: java.io.IOException: Cannot open filename /
>> hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
>>
>> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
>> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>> at java.lang.reflect.Method.invoke(Method.java:585)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>>
>> 2008-01-04 12:18:28,131 WARN org.apache.hadoop.hbase.HRegionServer:
>> java.io.IOException: java.io.IOException: Cannot open filename /
>> hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
>>
>> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
>> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>> at java.lang.reflect.Method.invoke(Method.java:585)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>>
>> 2008-01-04 12:18:28,147 WARN org.apache.hadoop.hbase.HRegionServer:
>> java.io.IOException: java.io.IOException: Cannot open filename /
>> hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
>>
>> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
>> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>
>> at java.lang.reflect.Method.invoke(Method.java:585)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>>
> The above are bad errors. hdfs lost our files or, more likely, we
> mangled their writing.
>
> ...
>>
>> Does this help?
>>
>
> Yes.
>
> St.Ack
>
>
>> Thanks,
>> Lars
>>
>>
>>
>> stack wrote:
>>> Lars George wrote:
>>>> Hi,
>>>>
>>>> I have inserted about 3.5m documents in a single two column table
>>>> in HBase running on 32 nodes. So far I was able to insert most
>>>> data, but with the last million or so I am stuck with this error:
>>>>
>>>> org.apache.hadoop.hbase.WrongRegionException: Requested row out of
>>>> range for HRegion docs,DC20020099792_20020725,9149203683830573099,
>>>> startKey='DC20020099792_20020725', endKey='DC20020099792_20020725',
>>>> row='DC20020099841_20020725'
>>>
>>> This happens every time you try to do an insert?
>>>
>>>> Querying for the document returns nothing, means it looks like the
>>>> document does not exist - although I am sure I tried inserting it a
>>>> few times. Deleting or trying to re-insert returns the above error,
>>>> both through the API (using HTable) as well as through the HBase
>>>> shell.
>>>>
>>>> I tried a restart of Hadoop/HBase to no avail. How do fix this
>>>> problem? Any help is appreciated.
>>>
>>> You have DEBUG enabled for hbase (See
>>> http://wiki.apache.org/lucene-hadoop/Hbase/FAQ#4). Do the logs tell
>>> you anything more: e.g. any interesting exceptions?
>>>
>>> Which hbase version?
>>>
>>> St.Ack
>>>
>
Re: WrongRegionException - How do I fix it?
Posted by stack <st...@duboce.net>.
If possible, please move to TRUNK instead. Most of the below have been
addressed there (I can send you a patch if you want to run hbase TRUNK
on hadoop 0.15.x).
Further comments inline below:
Lars George wrote:
> Hi Stack,
>
> Yes, it happens every time I insert particular rows. Before it would
> fail every now and so often, but since now all "good" rows are
> inserted I am stuck with the ones that do not insert. And I am sure
> they did once, with no error. So they are in there in limbo, but I
> cannot retrieve nor delete or insert them.
>
> It mentions in the FAQ that I can switch on debugging through the UI,
> but I cannot see where. I am using version 0.15.1, is that supposed to
> have that option or do I need to go the log4j.properties plus restart
> route?
This is a post-0.15 release feature (It says post-0.15.x in the FAQ).
> I have errors all the time - which quite frankly worry me. Here a
> list of what I see so far:
>
> 1. At startup
>
> ==>
> /usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-40.worldlingo.com.log
> <==
> 2008-01-03 14:11:22,512 WARN org.apache.hadoop.util.NativeCodeLoader:
> Unable to load native-hadoop library for your platform... using
> builtin-java classes where applicable
> 2008-01-03 14:11:29,808 WARN org.apache.hadoop.hbase.HRegionServer:
> Processing message (Retry: 0)
> java.io.IOException: java.io.IOException:
> java.util.ConcurrentModificationException
....
Fixed in TRUNK
>
>
> 2. Sporadically
>
> 2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode:
> Unexpected error trying to delete block blk_-8931657506153335343.
> Block not found in blockMap.
> 2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode:
> Unexpected error trying to delete block blk_3775459202881005176. Block
> not found in blockMap.
> 2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode:
> Unexpected error trying to delete block blk_-283089329129695997. Block
> not found in blockMap.
> 2008-01-03 21:32:00,644 WARN org.apache.hadoop.dfs.DataNode:
> java.io.IOException: Error in deleting blocks.
> at org.apache.hadoop.dfs.FSDataset.invalidate(FSDataset.java:719)
> at
> org.apache.hadoop.dfs.DataNode.processCommand(DataNode.java:625)
> at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:528)
> at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1494)
> at java.lang.Thread.run(Thread.java:595)
>
> These come with all sorts of blocks, they do not appear too often, but
> on a regular basis.
>
0.15.x was doing updates against files that had been removed by another
thread. Made for strange errors in hdfs. That said, I don't recall
having seen the above. Do you have namenode at DEBUG level? If so, try
tracing the above problematic blocks therein; see if you can figure out
a story as to what happened with these blocks.
If HDFS is in an inconsistent state, hbase will be inconsistent too.
> 3. Misc
>
> I see these too from this morning
>
> 2008-01-04 08:23:30,616 ERROR org.apache.hadoop.hbase.HRegionServer:
> unable to process message: MSG_REGION_OPEN : regionname: do
> cs,DC20020096869_20020725,43610073395851568, startKey:
> <DC20020096869_20020725>, tableDesc: {name: docs, families: {contents:={n
> ame: contents, max versions: 3, compression: NONE, in memory: false,
> max length: 2147483647, bloom filter: none}, language:={name: l
> anguage, max versions: 3, compression: NONE, in memory: false, max
> length: 2147483647, bloom filter: none}, mimetype:={name: mimetyp
> e, max versions: 3, compression: NONE, in memory: false, max length:
> 2147483647, bloom filter: none}}}
> java.io.IOException: java.io.IOException: Cannot open filename
> /hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/co
> ntents/info/1501965039462307633
> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>
> at java.lang.reflect.Method.invoke(Method.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>
> at
> sun.reflect.GeneratedConstructorAccessor7.newInstance(Unknown Source)
> at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>
> at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
> at
> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
>
> at
> org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
>
> at
> org.apache.hadoop.hbase.HRegionServer$Worker.run(HRegionServer.java:903)
> at java.lang.Thread.run(Thread.java:595)
>
>
I'd guess this file is made of the above cited problematic blocks. Can
you find it on hdfs? Can you download it? (Try doing a './bin/hadoop
fs fsck /HBASE_DIR' -- see what it says).
> Another one is this:
>
> ==>
> /usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-62.worldlingo.com.log
> <==
> 2008-01-04 08:16:32,001 WARN org.apache.hadoop.hbase.HRegion: Region
> docs,DC20020099792_20020725,9149203683830573099 is NOT spli
> table though its aggregate size is 111.4m and desired size is 64.0m
>
> These come up with different region numbers.
>
This is ok. We want to split the region because its > 64MB but this
region has outstanding references to another, parent region, so is not
yet splittable (It should have turned splittable a little later in your
log).
> And another one:
>
> 2008-01-03 11:27:55,437 WARN org.apache.hadoop.hbase.HStore: Failed
> getting store size
> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File does
> not exist
> at
> org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:489)
> at
> org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1360)
> at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:428)
> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>
> at java.lang.reflect.Method.invoke(Method.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>
Fixed in TRUNK
>
> And another one:
>
> 2008-01-03 15:43:39,590 WARN org.apache.hadoop.dfs.DataNode: Got
> exception while serving blk_3676251342939485484 to /192.168.105.21:
> java.io.IOException: Block blk_3676251342939485484 is not valid.
> at
> org.apache.hadoop.dfs.FSDataset.getBlockFile(FSDataset.java:528)
> at
> org.apache.hadoop.dfs.DataNode$BlockSender.<init>(DataNode.java:1051)
> at
> org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:843)
> at
> org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:801)
> at java.lang.Thread.run(Thread.java:595)
>
> Again, they come with different block numbers.
>
This hdfs exception may be recoverable IIRC; hdfs gets the block elsewhere.
> 4. Inserting document with errors
>
> If I try to add one of those documents, where I get an error back,
> this is what I see in the logs so far:
>
> Uploading -> DC20020099841_20020725
>
> Server 20:
>
>
....
> 2007-12-30 17:26:42,392 INFO org.mortbay.util.Container: Started
> HttpContext[/static,/static]
> 2007-12-30 17:26:42,395 INFO org.mortbay.http.SocketListener: Started
> SocketListener on 0.0.0.0:60030
> 2007-12-30 17:26:42,395 INFO org.mortbay.util.Container: Started
> org.mortbay.jetty.Server@1fd0fafc
> 2007-12-30 17:26:42,396 INFO org.apache.hadoop.ipc.Server: IPC Server
> listener on 60020: starting
> 2007-12-30 17:26:42,397 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 0 on 60020: starting
> 2007-12-30 17:26:42,402 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 1 on 60020: starting
> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 3 on 60020: starting
> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 4 on 60020: starting
> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 5 on 60020: starting
> 2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 6 on 60020: starting
>
Please set these to run at DEBUG level, at least while we are trying to
figure out whats going on.
>
> Server 26:
>
> 2008-01-04 12:18:28,125 WARN org.apache.hadoop.hbase.HRegionServer:
> java.io.IOException: java.io.IOException: Cannot open filename /
> hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
>
> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>
> at java.lang.reflect.Method.invoke(Method.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>
> 2008-01-04 12:18:28,131 WARN org.apache.hadoop.hbase.HRegionServer:
> java.io.IOException: java.io.IOException: Cannot open filename /
> hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
>
> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>
> at java.lang.reflect.Method.invoke(Method.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>
> 2008-01-04 12:18:28,147 WARN org.apache.hadoop.hbase.HRegionServer:
> java.io.IOException: java.io.IOException: Cannot open filename /
> hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
>
> at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
> at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>
> at java.lang.reflect.Method.invoke(Method.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
>
The above are bad errors. hdfs lost our files or, more likely, we
mangled their writing.
...
>
> Does this help?
>
Yes.
St.Ack
> Thanks,
> Lars
>
>
>
> stack wrote:
>> Lars George wrote:
>>> Hi,
>>>
>>> I have inserted about 3.5m documents in a single two column table in
>>> HBase running on 32 nodes. So far I was able to insert most data,
>>> but with the last million or so I am stuck with this error:
>>>
>>> org.apache.hadoop.hbase.WrongRegionException: Requested row out of
>>> range for HRegion docs,DC20020099792_20020725,9149203683830573099,
>>> startKey='DC20020099792_20020725', endKey='DC20020099792_20020725',
>>> row='DC20020099841_20020725'
>>
>> This happens every time you try to do an insert?
>>
>>> Querying for the document returns nothing, means it looks like the
>>> document does not exist - although I am sure I tried inserting it a
>>> few times. Deleting or trying to re-insert returns the above error,
>>> both through the API (using HTable) as well as through the HBase shell.
>>>
>>> I tried a restart of Hadoop/HBase to no avail. How do fix this
>>> problem? Any help is appreciated.
>>
>> You have DEBUG enabled for hbase (See
>> http://wiki.apache.org/lucene-hadoop/Hbase/FAQ#4). Do the logs tell
>> you anything more: e.g. any interesting exceptions?
>>
>> Which hbase version?
>>
>> St.Ack
>>
Re: WrongRegionException - How do I fix it?
Posted by Lars George <la...@worldlingo.com>.
Hi Stack,
Yes, it happens every time I insert particular rows. Before it would
fail every now and so often, but since now all "good" rows are inserted
I am stuck with the ones that do not insert. And I am sure they did
once, with no error. So they are in there in limbo, but I cannot
retrieve nor delete or insert them.
It mentions in the FAQ that I can switch on debugging through the UI,
but I cannot see where. I am using version 0.15.1, is that supposed to
have that option or do I need to go the log4j.properties plus restart route?
I have errors all the time - which quite frankly worry me. Here a list
of what I see so far:
1. At startup
==>
/usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-40.worldlingo.com.log
<==
2008-01-03 14:11:22,512 WARN org.apache.hadoop.util.NativeCodeLoader:
Unable to load native-hadoop library for your platform... using
builtin-java classes where applicable
2008-01-03 14:11:29,808 WARN org.apache.hadoop.hbase.HRegionServer:
Processing message (Retry: 0)
java.io.IOException: java.io.IOException:
java.util.ConcurrentModificationException
at java.util.HashMap$HashIterator.nextEntry(HashMap.java:841)
at java.util.HashMap$EntryIterator.next(HashMap.java:883)
at java.util.HashMap$EntryIterator.next(HashMap.java:881)
at
org.apache.hadoop.hbase.HMaster.getRegionsToAssign(HMaster.java:1739)
at org.apache.hadoop.hbase.HMaster.assignRegions(HMaster.java:1597)
at org.apache.hadoop.hbase.HMaster.processMsgs(HMaster.java:1582)
at
org.apache.hadoop.hbase.HMaster.regionServerReport(HMaster.java:1400)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
at org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:532)
at java.lang.Thread.run(Thread.java:595)
2. Sporadically
2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode: Unexpected
error trying to delete block blk_-8931657506153335343. Block not found
in blockMap.
2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode: Unexpected
error trying to delete block blk_3775459202881005176. Block not found in
blockMap.
2008-01-03 21:32:00,639 WARN org.apache.hadoop.dfs.DataNode: Unexpected
error trying to delete block blk_-283089329129695997. Block not found in
blockMap.
2008-01-03 21:32:00,644 WARN org.apache.hadoop.dfs.DataNode:
java.io.IOException: Error in deleting blocks.
at org.apache.hadoop.dfs.FSDataset.invalidate(FSDataset.java:719)
at org.apache.hadoop.dfs.DataNode.processCommand(DataNode.java:625)
at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:528)
at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1494)
at java.lang.Thread.run(Thread.java:595)
These come with all sorts of blocks, they do not appear too often, but
on a regular basis.
3. Misc
I see these too from this morning
2008-01-04 08:23:30,616 ERROR org.apache.hadoop.hbase.HRegionServer:
unable to process message: MSG_REGION_OPEN : regionname: do
cs,DC20020096869_20020725,43610073395851568, startKey:
<DC20020096869_20020725>, tableDesc: {name: docs, families: {contents:={n
ame: contents, max versions: 3, compression: NONE, in memory: false, max
length: 2147483647, bloom filter: none}, language:={name: l
anguage, max versions: 3, compression: NONE, in memory: false, max
length: 2147483647, bloom filter: none}, mimetype:={name: mimetyp
e, max versions: 3, compression: NONE, in memory: false, max length:
2147483647, bloom filter: none}}}
java.io.IOException: java.io.IOException: Cannot open filename
/hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/co
ntents/info/1501965039462307633
at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
at sun.reflect.GeneratedConstructorAccessor7.newInstance(Unknown
Source)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
at
org.apache.hadoop.hbase.HRegionServer$Worker.run(HRegionServer.java:903)
at java.lang.Thread.run(Thread.java:595)
Another one is this:
==>
/usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-62.worldlingo.com.log
<==
2008-01-04 08:16:32,001 WARN org.apache.hadoop.hbase.HRegion: Region
docs,DC20020099792_20020725,9149203683830573099 is NOT spli
table though its aggregate size is 111.4m and desired size is 64.0m
These come up with different region numbers.
And another one:
2008-01-03 11:27:55,437 WARN org.apache.hadoop.hbase.HStore: Failed
getting store size
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File does
not exist
at
org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:489)
at
org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1360)
at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:428)
at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
at org.apache.hadoop.ipc.Client.call(Client.java:482)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:184)
at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source)
at org.apache.hadoop.dfs.DFSClient.getFileInfo(DFSClient.java:438)
at
org.apache.hadoop.dfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:334)
at org.apache.hadoop.hbase.HStoreFile.length(HStoreFile.java:971)
at org.apache.hadoop.hbase.HStore.size(HStore.java:1322)
at org.apache.hadoop.hbase.HRegion.largestHStore(HRegion.java:630)
at org.apache.hadoop.hbase.HRegion.needsSplit(HRegion.java:590)
at
org.apache.hadoop.hbase.HRegionServer$SplitOrCompactChecker.checkForSplitsOrCompactions(HRegionServer.java:201)
at
org.apache.hadoop.hbase.HRegionServer$SplitOrCompactChecker.chore(HRegionServer.java:186)
at org.apache.hadoop.hbase.Chore.run(Chore.java:58)
And another one:
2008-01-03 15:43:39,590 WARN org.apache.hadoop.dfs.DataNode: Got
exception while serving blk_3676251342939485484 to /192.168.105.21:
java.io.IOException: Block blk_3676251342939485484 is not valid.
at org.apache.hadoop.dfs.FSDataset.getBlockFile(FSDataset.java:528)
at
org.apache.hadoop.dfs.DataNode$BlockSender.<init>(DataNode.java:1051)
at
org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:843)
at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:801)
at java.lang.Thread.run(Thread.java:595)
Again, they come with different block numbers.
4. Inserting document with errors
If I try to add one of those documents, where I get an error back, this
is what I see in the logs so far:
Uploading -> DC20020099841_20020725
Server 20:
2008-01-04 12:18:01,987 WARN org.apache.hadoop.hbase.HRegion: Region
docs,DC20020099792_20020725,9149203683830573099 is NOT spli
table though its aggregate size is 111.4m and desired size is 64.0m
2008-01-04 12:18:16,983 WARN org.apache.hadoop.hbase.HRegion: Region
docs,DC20020099792_20020725,3506350021119702948 is NOT spli
table though its aggregate size is 111.4m and desired size is 64.0m
2008-01-04 12:18:16,988 WARN org.apache.hadoop.hbase.HRegion: Region
docs,DC20020099792_20020725,9149203683830573099 is NOT spli
table though its aggregate size is 111.4m and desired size is 64.0m
2008-01-04 12:18:31,985 WARN org.apache.hadoop.hbase.HRegion: Region
docs,DC20020099792_20020725,3506350021119702948 is NOT spli
table though its aggregate size is 111.4m and desired size is 64.0m
2008-01-04 12:18:31,990 WARN org.apache.hadoop.hbase.HRegion: Region
docs,DC20020099792_20020725,9149203683830573099 is NOT spli
table though its aggregate size is 111.4m and desired size is 64.0m
==>
/usr/local/hadoop/logs/hbase-pdc-regionserver-lv1-xen-pdc-20.worldlingo.com.log.2007-12-30
<==
2007-12-30 17:26:40,932 INFO org.apache.hadoop.hbase.HRegionServer: Root
dir: /hbase
2007-12-30 17:26:41,091 WARN org.apache.hadoop.util.NativeCodeLoader:
Unable to load native-hadoop library for your platform... usin
g builtin-java classes where applicable
2007-12-30 17:26:41,093 INFO org.apache.hadoop.hbase.HLog: new log
writer created at /hbase/log_192.168.105.29_2760247535651987904_6
0020/hlog.dat.000
2007-12-30 17:26:41,166 INFO org.mortbay.util.Credential: Checking
Resource aliases
2007-12-30 17:26:41,328 INFO org.mortbay.http.HttpServer: Version
Jetty/5.1.4
2007-12-30 17:26:42,276 INFO org.mortbay.util.Container: Started
org.mortbay.jetty.servlet.WebApplicationHandler@69d6065
2007-12-30 17:26:42,392 INFO org.mortbay.util.Container: Started
WebApplicationContext[/,/]
2007-12-30 17:26:42,392 INFO org.mortbay.util.Container: Started
HttpContext[/logs,/logs]
2007-12-30 17:26:42,392 INFO org.mortbay.util.Container: Started
HttpContext[/static,/static]
2007-12-30 17:26:42,395 INFO org.mortbay.http.SocketListener: Started
SocketListener on 0.0.0.0:60030
2007-12-30 17:26:42,395 INFO org.mortbay.util.Container: Started
org.mortbay.jetty.Server@1fd0fafc
2007-12-30 17:26:42,396 INFO org.apache.hadoop.ipc.Server: IPC Server
listener on 60020: starting
2007-12-30 17:26:42,397 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 0 on 60020: starting
2007-12-30 17:26:42,402 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 1 on 60020: starting
2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 3 on 60020: starting
2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 4 on 60020: starting
2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 5 on 60020: starting
2007-12-30 17:26:42,403 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 6 on 60020: starting
Server 26:
2008-01-04 12:18:28,125 WARN org.apache.hadoop.hbase.HRegionServer:
java.io.IOException: java.io.IOException: Cannot open filename /
hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
2008-01-04 12:18:28,131 WARN org.apache.hadoop.hbase.HRegionServer:
java.io.IOException: java.io.IOException: Cannot open filename /
hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
2008-01-04 12:18:28,147 WARN org.apache.hadoop.hbase.HRegionServer:
java.io.IOException: java.io.IOException: Cannot open filename /
hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/contents/info/1501965039462307633
at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
...
2008-01-04 12:18:28,181 ERROR org.apache.hadoop.hbase.HRegionServer:
unable to process message: MSG_REGION_OPEN : regionname: do
cs,DC20020095856_20020725,2673659877920938250, startKey:
<DC20020095856_20020725>, tableDesc: {name: docs, families: {contents:=
{name: contents, max versions: 3, compression: NONE, in memory: false,
max length: 2147483647, bloom filter: none}, language:={name:
language, max versions: 3, compression: NONE, in memory: false, max
length: 2147483647, bloom filter: none}, mimetype:={name: mimet
ype, max versions: 3, compression: NONE, in memory: false, max length:
2147483647, bloom filter: none}}}
java.io.IOException: java.io.IOException: Cannot open filename
/hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/co
ntents/info/1501965039462307633
at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
at
sun.reflect.GeneratedConstructorAccessor12.newInstance(Unknown Source)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
at
org.apache.hadoop.hbase.HRegionServer$Worker.run(HRegionServer.java:903)
at java.lang.Thread.run(Thread.java:595)
2008-01-04 12:18:28,187 ERROR org.apache.hadoop.hbase.HRegionServer:
unable to process message: MSG_REGION_OPEN : regionname: do
cs,DC20020096869_20020725,43610073395851568, startKey:
<DC20020096869_20020725>, tableDesc: {name: docs, families: {contents:={n
ame: contents, max versions: 3, compression: NONE, in memory: false, max
length: 2147483647, bloom filter: none}, language:={name: l
anguage, max versions: 3, compression: NONE, in memory: false, max
length: 2147483647, bloom filter: none}, mimetype:={name: mimetyp
e, max versions: 3, compression: NONE, in memory: false, max length:
2147483647, bloom filter: none}}}
java.io.IOException: java.io.IOException: Cannot open filename
/hbase/hregion_docs,DC20020095856_20020725,7894263634108415584/co
ntents/info/1501965039462307633
at org.apache.hadoop.dfs.NameNode.open(NameNode.java:238)
at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown Source)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)
at
sun.reflect.GeneratedConstructorAccessor12.newInstance(Unknown Source)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
at
org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
at
org.apache.hadoop.hbase.HRegionServer$Worker.run(HRegionServer.java:903)
at java.lang.Thread.run(Thread.java:595)
Does this help?
Thanks,
Lars
stack wrote:
> Lars George wrote:
>> Hi,
>>
>> I have inserted about 3.5m documents in a single two column table in
>> HBase running on 32 nodes. So far I was able to insert most data, but
>> with the last million or so I am stuck with this error:
>>
>> org.apache.hadoop.hbase.WrongRegionException: Requested row out of
>> range for HRegion docs,DC20020099792_20020725,9149203683830573099,
>> startKey='DC20020099792_20020725', endKey='DC20020099792_20020725',
>> row='DC20020099841_20020725'
>
> This happens every time you try to do an insert?
>
>> Querying for the document returns nothing, means it looks like the
>> document does not exist - although I am sure I tried inserting it a
>> few times. Deleting or trying to re-insert returns the above error,
>> both through the API (using HTable) as well as through the HBase shell.
>>
>> I tried a restart of Hadoop/HBase to no avail. How do fix this
>> problem? Any help is appreciated.
>
> You have DEBUG enabled for hbase (See
> http://wiki.apache.org/lucene-hadoop/Hbase/FAQ#4). Do the logs tell
> you anything more: e.g. any interesting exceptions?
>
> Which hbase version?
>
> St.Ack
>
Re: WrongRegionException - How do I fix it?
Posted by stack <st...@duboce.net>.
Lars George wrote:
> Hi,
>
> I have inserted about 3.5m documents in a single two column table in
> HBase running on 32 nodes. So far I was able to insert most data, but
> with the last million or so I am stuck with this error:
>
> org.apache.hadoop.hbase.WrongRegionException: Requested row out of
> range for HRegion docs,DC20020099792_20020725,9149203683830573099,
> startKey='DC20020099792_20020725', endKey='DC20020099792_20020725',
> row='DC20020099841_20020725'
This happens every time you try to do an insert?
> Querying for the document returns nothing, means it looks like the
> document does not exist - although I am sure I tried inserting it a
> few times. Deleting or trying to re-insert returns the above error,
> both through the API (using HTable) as well as through the HBase shell.
>
> I tried a restart of Hadoop/HBase to no avail. How do fix this
> problem? Any help is appreciated.
You have DEBUG enabled for hbase (See
http://wiki.apache.org/lucene-hadoop/Hbase/FAQ#4). Do the logs tell you
anything more: e.g. any interesting exceptions?
Which hbase version?
St.Ack