You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Arthur van Hoff <av...@ellerdale.com> on 2008/11/12 17:16:03 UTC

Multiple start regions in a table

Hi,

I have a table which appears to have two regions without a start key.
Is that expected? My job is getting stuck scanning these two regions.

Any suggestions?

Thanks.
-- 
Arthur van Hoff - Grand Master of Alphabetical Order
The Ellerdale Project, Menlo Park, CA
avh@ellerdale.com, 650-283-0842

Re: xceiverCount 257 exceeds the limit of concurrent xcievers 256

Posted by Michael Stack <st...@duboce.net>.
Try upping the limit on your datanodes.  Set dfs.datanode.max.xcievers 
up to 1024 or more.
St.Ack

Dru Jensen wrote:
> hbase-users,
>
> I have been running MR processes for several days against HBase with 
> success until recently the region servers shut themselves down.
> Hadoop 0.18.1
> Hbase 0.18.1
> 3 node cluster
>
> Checking the region server logs, I see the following Exception before 
> shutdown:
>
> 2008-11-11 19:55:52,451 INFO org.apache.hadoop.dfs.DFSClient: 
> Exception in createBlockOutputStream java.io.IOException: Could not 
> read from stream
> 2008-11-11 19:55:52,451 INFO org.apache.hadoop.dfs.DFSClient: 
> Abandoning block blk_-5467014108758633036_595771
> 2008-11-11 19:55:58,455 WARN org.apache.hadoop.dfs.DFSClient: 
> DataStreamer Exception: java.io.IOException: Unable to create new block.
> 2008-11-11 19:55:58,455 WARN org.apache.hadoop.dfs.DFSClient: Error 
> Recovery for block blk_-5467014108758633036_595771 bad datanode[0]
> 2008-11-11 19:55:58,482 FATAL 
> org.apache.hadoop.hbase.regionserver.Flusher: Replay of hlog required. 
> Forcing server shutdown
>
> Hadoop datanode:
>
> ERROR org.apache.hadoop.dfs.DataNode: 
> DatanodeRegistration(10.10.10.53:50010, 
> storageID=DS-1570581820-10.10.10.53-50010-1224117842339, 
> infoPort=50075, ipcPort=50020):DataXceiver: java.io.IOException: 
> xceiverCount 258 exceeds the limit of concurrent xcievers 256
>
> Also, I can't seem to recover from this.
>
> At restart HBase regionserver log:
>
> ERROR org.apache.hadoop.hbase.regionserver.HRegionServer: error 
> opening region table,key,1226427697843
> java.io.IOException: Could not obtain block: 
> blk_-7717647125888341834_589063 
> file=/hbase/table/1691377931/title/mapfiles/7939912624709782499/index
>
> Hadoop datanode:
>
> ERROR org.apache.hadoop.dfs.DataNode: 
> DatanodeRegistration(10.10.10.51:50010, 
> storageID=DS-2079833884-10.10.10.51-50010-1224117832456, 
> infoPort=50075, ipcPort=50020):DataXceiver: java.io.IOException: 
> xceiverCount 257 exceeds the limit of concurrent xcievers 256
>     at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:1030)
>     at java.lang.Thread.run(Thread.java:595)
>
> Any help is appreciated.
>
> Thanks,
> Dru
>


xceiverCount 257 exceeds the limit of concurrent xcievers 256

Posted by Dru Jensen <dr...@gmail.com>.
hbase-users,

I have been running MR processes for several days against HBase with  
success until recently the region servers shut themselves down.
Hadoop 0.18.1
Hbase 0.18.1
3 node cluster

Checking the region server logs, I see the following Exception before  
shutdown:

2008-11-11 19:55:52,451 INFO org.apache.hadoop.dfs.DFSClient:  
Exception in createBlockOutputStream java.io.IOException: Could not  
read from stream
2008-11-11 19:55:52,451 INFO org.apache.hadoop.dfs.DFSClient:  
Abandoning block blk_-5467014108758633036_595771
2008-11-11 19:55:58,455 WARN org.apache.hadoop.dfs.DFSClient:  
DataStreamer Exception: java.io.IOException: Unable to create new block.
2008-11-11 19:55:58,455 WARN org.apache.hadoop.dfs.DFSClient: Error  
Recovery for block blk_-5467014108758633036_595771 bad datanode[0]
2008-11-11 19:55:58,482 FATAL  
org.apache.hadoop.hbase.regionserver.Flusher: Replay of hlog required.  
Forcing server shutdown

Hadoop datanode:

ERROR org.apache.hadoop.dfs.DataNode:  
DatanodeRegistration(10.10.10.53:50010,  
storageID=DS-1570581820-10.10.10.53-50010-1224117842339,  
infoPort=50075, ipcPort=50020):DataXceiver: java.io.IOException:  
xceiverCount 258 exceeds the limit of concurrent xcievers 256

Also, I can't seem to recover from this.

At restart HBase regionserver log:

ERROR org.apache.hadoop.hbase.regionserver.HRegionServer: error  
opening region table,key,1226427697843
java.io.IOException: Could not obtain block:  
blk_-7717647125888341834_589063 file=/hbase/table/1691377931/title/ 
mapfiles/7939912624709782499/index

Hadoop datanode:

ERROR org.apache.hadoop.dfs.DataNode:  
DatanodeRegistration(10.10.10.51:50010,  
storageID=DS-2079833884-10.10.10.51-50010-1224117832456,  
infoPort=50075, ipcPort=50020):DataXceiver: java.io.IOException:  
xceiverCount 257 exceeds the limit of concurrent xcievers 256
	at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:1030)
	at java.lang.Thread.run(Thread.java:595)

Any help is appreciated.

Thanks,
Dru


Re: Multiple start regions in a table

Posted by stack <st...@duboce.net>.
Arthur van Hoff wrote:
> We are using hbase 0.18.1. The count seems happy to count the first
> 100,000 rows...
> I was able to merge the two regions.
>
> However, I noticed that there are many overlapping regions in this
> table. Is that expected?
>   
No.  Only time you should see overlap is if the first region in list of 
regions out of .META. is offline and parent of the next two regions that 
follow.

> Here are some of the start/end keys:
>
>                     -0000010225 (newly merged)
> 0000003046-0000005771 (overlaps with the previous section)
> 0000005771-0000008992
> 0000008992-0000012909
> 0000010225-0000020097 (overlaps with the previous section)
> 0000012909-0000015835 (overlaps with the previous section)
>   

Do your MR jobs work now?

> ...
>
> How can this happen?
>   

Should never happen.

Could happen if edits to .META. were dropped.  Have seen rare case of 
this split events but the gaping holes should be closed at least in the 
version you are running.

If you can't have your MR run to completion, and if overlapping region 
is not offlined, may have to do more repairs.  Then, run with DEBUG 
enabled and if it happens again, should be able to see by examination of 
logs how it came about.

St.Ack


> On Wed, Nov 12, 2008 at 11:05 AM, stack <st...@duboce.net> wrote:
>   
>> Arthur van Hoff wrote:
>>     
>>> Below are the two regions in the .META. scan. I don't think either of
>>> them is disabled.
>>>
>>>       
>> Right.  One should be offlined.  Which version of hbase is this?
>>
>>     
>>> The table is too large to count, and I am not sure how to enable DEBUG.
>>>
>>>
>>>       
>> Well, looks like problem is in first region so you could start the count and
>> looks like issue should pop up quick.
>>
>> To enable DEBUG, http://wiki.apache.org/hadoop/Hbase/FAQ#5.
>>
>> To 'repair', your table, shut it down and then run:
>>
>> $ ./bin/hbase org.apache.hadoop.hbase.util.Merge
>> Usage: bin/hbase merge <table-name> <region-1> <region-2>
>>
>> Supply table name, and two regions to merge.
>>
>> Let us know how it goes.
>>
>> St.Ack
>>
>>     
>>>  wikipedia,,1226039865543    column=historian:assignment,
>>> timestamp=1226471612375, value=Region assigned to server
>>> 192.168.33.31:60020
>>>  wikipedia,,1226039865543    column=historian:open,
>>> timestamp=1226471627434, value=Region opened on server :
>>> gore.int.ellerdale.com
>>>  wikipedia,,1226039865543    column=historian:split,
>>> timestamp=1226039870630, value=Region split from  :
>>> wikipedia,,1226039822665
>>>  wikipedia,,1226039865543    column=info:regioninfo,
>>> timestamp=1226039870633, value=REGION => {NAME =>
>>> 'wikipedia,,1226039865543', STARTKEY => '', ENDKEY => '0000020097',
>>> ENCODED => 1423314116, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
>>> 'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
>>> => 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
>>> '2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
>>> 'false'}]}}
>>>  wikipedia,,1226039865543    column=info:server,
>>> timestamp=1226471627433, value=192.168.33.31:60020
>>>  wikipedia,,1226039865543    column=info:serverstartcode,
>>> timestamp=1226471627433, value=1226471577772
>>>  wikipedia,,1226469112210    column=historian:assignment,
>>> timestamp=1226471612379, value=Region assigned to server
>>> 192.168.33.31:60020
>>>  wikipedia,,1226469112210    column=historian:open,
>>> timestamp=1226471630431, value=Region opened on server :
>>> gore.int.ellerdale.com
>>>  wikipedia,,1226469112210    column=historian:split,
>>> timestamp=1226469113121, value=Region split from  :
>>> wikipedia,,1226356665025
>>>  wikipedia,,1226469112210    column=info:regioninfo,
>>> timestamp=1226469113125, value=REGION => {NAME =>
>>> 'wikipedia,,1226469112210', STARTKEY => '', ENDKEY => '0000003046',
>>> ENCODED => 566585206, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
>>> 'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
>>> => 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
>>> '2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
>>> 'false'}]}}
>>>  wikipedia,,1226469112210    column=info:server,
>>> timestamp=1226471630430, value=192.168.33.31:60020
>>>  wikipedia,,1226469112210    column=info:serverstartcode,
>>> timestamp=1226471630430, value=1226471577772
>>>
>>>
>>> On Wed, Nov 12, 2008 at 10:09 AM, stack <st...@duboce.net> wrote:
>>>
>>>       
>>>> Are they both enabled?
>>>>
>>>> You can see by doing "scan '.META.'" in the shell and then look at the
>>>> particular regions in the output.  Disabled regions will have a disabled
>>>> attribute.
>>>>
>>>> What error are you seeing when in your job?
>>>>
>>>> You could enable DEBUG and then in shell try running "count 'TABLENAME'".
>>>>  With DEBUG enabled, you'll see the regions loading and you'll be able to
>>>> pinpoint the problematic one.  Or, if your table is large, run a "scan
>>>> 'TABLENAME', ['NONEXISTANT_COLUMNNAME']".  That'll cause the scanner to spin
>>>> through all regions quickly -- there'll be no values found as it scans --
>>>> and again with DEBUG you can see the regions loading and figure the bad one.
>>>>
>>>> When you've found the bad region, look at its listing in the '.META.'
>>>> output.  Or paste it in here and we'll take a look for you.
>>>>
>>>> Thanks Arthur,
>>>> St.Ack
>>>>
>>>>
>>>> Arthur van Hoff wrote:
>>>>
>>>>         
>>>>> Hi,
>>>>>
>>>>> I have a table which appears to have two regions without a start key.
>>>>> Is that expected? My job is getting stuck scanning these two regions.
>>>>>
>>>>> Any suggestions?
>>>>>
>>>>> Thanks.
>>>>>
>>>>>
>>>>>           
>>>
>>> --
>>> Arthur van Hoff - Grand Master of Alphabetical Order
>>> The Ellerdale Project, Menlo Park, CA
>>> avh@ellerdale.com, 650-283-0842
>>>
>>>       
>>     
>
>
>
>   


Re: Multiple start regions in a table

Posted by Arthur van Hoff <av...@ellerdale.com>.
We are using hbase 0.18.1. The count seems happy to count the first
100,000 rows...
I was able to merge the two regions.

However, I noticed that there are many overlapping regions in this
table. Is that expected?
Here are some of the start/end keys:

                    -0000010225 (newly merged)
0000003046-0000005771 (overlaps with the previous section)
0000005771-0000008992
0000008992-0000012909
0000010225-0000020097 (overlaps with the previous section)
0000012909-0000015835 (overlaps with the previous section)
...

How can this happen?

On Wed, Nov 12, 2008 at 11:05 AM, stack <st...@duboce.net> wrote:
> Arthur van Hoff wrote:
>>
>> Below are the two regions in the .META. scan. I don't think either of
>> them is disabled.
>>
>
> Right.  One should be offlined.  Which version of hbase is this?
>
>> The table is too large to count, and I am not sure how to enable DEBUG.
>>
>>
>
> Well, looks like problem is in first region so you could start the count and
> looks like issue should pop up quick.
>
> To enable DEBUG, http://wiki.apache.org/hadoop/Hbase/FAQ#5.
>
> To 'repair', your table, shut it down and then run:
>
> $ ./bin/hbase org.apache.hadoop.hbase.util.Merge
> Usage: bin/hbase merge <table-name> <region-1> <region-2>
>
> Supply table name, and two regions to merge.
>
> Let us know how it goes.
>
> St.Ack
>
>>  wikipedia,,1226039865543    column=historian:assignment,
>> timestamp=1226471612375, value=Region assigned to server
>> 192.168.33.31:60020
>>  wikipedia,,1226039865543    column=historian:open,
>> timestamp=1226471627434, value=Region opened on server :
>> gore.int.ellerdale.com
>>  wikipedia,,1226039865543    column=historian:split,
>> timestamp=1226039870630, value=Region split from  :
>> wikipedia,,1226039822665
>>  wikipedia,,1226039865543    column=info:regioninfo,
>> timestamp=1226039870633, value=REGION => {NAME =>
>> 'wikipedia,,1226039865543', STARTKEY => '', ENDKEY => '0000020097',
>> ENCODED => 1423314116, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
>> 'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
>> => 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
>> '2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
>> 'false'}]}}
>>  wikipedia,,1226039865543    column=info:server,
>> timestamp=1226471627433, value=192.168.33.31:60020
>>  wikipedia,,1226039865543    column=info:serverstartcode,
>> timestamp=1226471627433, value=1226471577772
>>  wikipedia,,1226469112210    column=historian:assignment,
>> timestamp=1226471612379, value=Region assigned to server
>> 192.168.33.31:60020
>>  wikipedia,,1226469112210    column=historian:open,
>> timestamp=1226471630431, value=Region opened on server :
>> gore.int.ellerdale.com
>>  wikipedia,,1226469112210    column=historian:split,
>> timestamp=1226469113121, value=Region split from  :
>> wikipedia,,1226356665025
>>  wikipedia,,1226469112210    column=info:regioninfo,
>> timestamp=1226469113125, value=REGION => {NAME =>
>> 'wikipedia,,1226469112210', STARTKEY => '', ENDKEY => '0000003046',
>> ENCODED => 566585206, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
>> 'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
>> => 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
>> '2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
>> 'false'}]}}
>>  wikipedia,,1226469112210    column=info:server,
>> timestamp=1226471630430, value=192.168.33.31:60020
>>  wikipedia,,1226469112210    column=info:serverstartcode,
>> timestamp=1226471630430, value=1226471577772
>>
>>
>> On Wed, Nov 12, 2008 at 10:09 AM, stack <st...@duboce.net> wrote:
>>
>>>
>>> Are they both enabled?
>>>
>>> You can see by doing "scan '.META.'" in the shell and then look at the
>>> particular regions in the output.  Disabled regions will have a disabled
>>> attribute.
>>>
>>> What error are you seeing when in your job?
>>>
>>> You could enable DEBUG and then in shell try running "count 'TABLENAME'".
>>>  With DEBUG enabled, you'll see the regions loading and you'll be able to
>>> pinpoint the problematic one.  Or, if your table is large, run a "scan
>>> 'TABLENAME', ['NONEXISTANT_COLUMNNAME']".  That'll cause the scanner to spin
>>> through all regions quickly -- there'll be no values found as it scans --
>>> and again with DEBUG you can see the regions loading and figure the bad one.
>>>
>>> When you've found the bad region, look at its listing in the '.META.'
>>> output.  Or paste it in here and we'll take a look for you.
>>>
>>> Thanks Arthur,
>>> St.Ack
>>>
>>>
>>> Arthur van Hoff wrote:
>>>
>>>>
>>>> Hi,
>>>>
>>>> I have a table which appears to have two regions without a start key.
>>>> Is that expected? My job is getting stuck scanning these two regions.
>>>>
>>>> Any suggestions?
>>>>
>>>> Thanks.
>>>>
>>>>
>>
>>
>>
>> --
>> Arthur van Hoff - Grand Master of Alphabetical Order
>> The Ellerdale Project, Menlo Park, CA
>> avh@ellerdale.com, 650-283-0842
>>
>
>



-- 
Arthur van Hoff - Grand Master of Alphabetical Order
The Ellerdale Project, Menlo Park, CA
avh@ellerdale.com, 650-283-0842

Re: Multiple start regions in a table

Posted by stack <st...@duboce.net>.
Arthur van Hoff wrote:
> Below are the two regions in the .META. scan. I don't think either of
> them is disabled.
>   
Right.  One should be offlined.  Which version of hbase is this?

> The table is too large to count, and I am not sure how to enable DEBUG.
>
>   
Well, looks like problem is in first region so you could start the count 
and looks like issue should pop up quick.

To enable DEBUG, http://wiki.apache.org/hadoop/Hbase/FAQ#5.

To 'repair', your table, shut it down and then run:

$ ./bin/hbase org.apache.hadoop.hbase.util.Merge
Usage: bin/hbase merge <table-name> <region-1> <region-2>

Supply table name, and two regions to merge.

Let us know how it goes.

St.Ack

>  wikipedia,,1226039865543    column=historian:assignment,
> timestamp=1226471612375, value=Region assigned to server
> 192.168.33.31:60020
>  wikipedia,,1226039865543    column=historian:open,
> timestamp=1226471627434, value=Region opened on server :
> gore.int.ellerdale.com
>  wikipedia,,1226039865543    column=historian:split,
> timestamp=1226039870630, value=Region split from  :
> wikipedia,,1226039822665
>  wikipedia,,1226039865543    column=info:regioninfo,
> timestamp=1226039870633, value=REGION => {NAME =>
> 'wikipedia,,1226039865543', STARTKEY => '', ENDKEY => '0000020097',
> ENCODED => 1423314116, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
> 'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
> => 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
> '2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
> 'false'}]}}
>  wikipedia,,1226039865543    column=info:server,
> timestamp=1226471627433, value=192.168.33.31:60020
>  wikipedia,,1226039865543    column=info:serverstartcode,
> timestamp=1226471627433, value=1226471577772
>  wikipedia,,1226469112210    column=historian:assignment,
> timestamp=1226471612379, value=Region assigned to server
> 192.168.33.31:60020
>  wikipedia,,1226469112210    column=historian:open,
> timestamp=1226471630431, value=Region opened on server :
> gore.int.ellerdale.com
>  wikipedia,,1226469112210    column=historian:split,
> timestamp=1226469113121, value=Region split from  :
> wikipedia,,1226356665025
>  wikipedia,,1226469112210    column=info:regioninfo,
> timestamp=1226469113125, value=REGION => {NAME =>
> 'wikipedia,,1226469112210', STARTKEY => '', ENDKEY => '0000003046',
> ENCODED => 566585206, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
> 'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
> => 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
> '2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
> 'false'}]}}
>  wikipedia,,1226469112210    column=info:server,
> timestamp=1226471630430, value=192.168.33.31:60020
>  wikipedia,,1226469112210    column=info:serverstartcode,
> timestamp=1226471630430, value=1226471577772
>
>
> On Wed, Nov 12, 2008 at 10:09 AM, stack <st...@duboce.net> wrote:
>   
>> Are they both enabled?
>>
>> You can see by doing "scan '.META.'" in the shell and then look at the particular regions in the output.  Disabled regions will have a disabled attribute.
>>
>> What error are you seeing when in your job?
>>
>> You could enable DEBUG and then in shell try running "count 'TABLENAME'".  With DEBUG enabled, you'll see the regions loading and you'll be able to pinpoint the problematic one.  Or, if your table is large, run a "scan 'TABLENAME', ['NONEXISTANT_COLUMNNAME']".  That'll cause the scanner to spin through all regions quickly -- there'll be no values found as it scans -- and again with DEBUG you can see the regions loading and figure the bad one.
>>
>> When you've found the bad region, look at its listing in the '.META.' output.  Or paste it in here and we'll take a look for you.
>>
>> Thanks Arthur,
>> St.Ack
>>
>>
>> Arthur van Hoff wrote:
>>     
>>> Hi,
>>>
>>> I have a table which appears to have two regions without a start key.
>>> Is that expected? My job is getting stuck scanning these two regions.
>>>
>>> Any suggestions?
>>>
>>> Thanks.
>>>
>>>       
>
>
>
> --
> Arthur van Hoff - Grand Master of Alphabetical Order
> The Ellerdale Project, Menlo Park, CA
> avh@ellerdale.com, 650-283-0842
>   


Re: Multiple start regions in a table

Posted by Arthur van Hoff <av...@ellerdale.com>.
Below are the two regions in the .META. scan. I don't think either of
them is disabled.
The table is too large to count, and I am not sure how to enable DEBUG.

 wikipedia,,1226039865543    column=historian:assignment,
timestamp=1226471612375, value=Region assigned to server
192.168.33.31:60020
 wikipedia,,1226039865543    column=historian:open,
timestamp=1226471627434, value=Region opened on server :
gore.int.ellerdale.com
 wikipedia,,1226039865543    column=historian:split,
timestamp=1226039870630, value=Region split from  :
wikipedia,,1226039822665
 wikipedia,,1226039865543    column=info:regioninfo,
timestamp=1226039870633, value=REGION => {NAME =>
'wikipedia,,1226039865543', STARTKEY => '', ENDKEY => '0000020097',
ENCODED => 1423314116, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
=> 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
'2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
'false'}]}}
 wikipedia,,1226039865543    column=info:server,
timestamp=1226471627433, value=192.168.33.31:60020
 wikipedia,,1226039865543    column=info:serverstartcode,
timestamp=1226471627433, value=1226471577772
 wikipedia,,1226469112210    column=historian:assignment,
timestamp=1226471612379, value=Region assigned to server
192.168.33.31:60020
 wikipedia,,1226469112210    column=historian:open,
timestamp=1226471630431, value=Region opened on server :
gore.int.ellerdale.com
 wikipedia,,1226469112210    column=historian:split,
timestamp=1226469113121, value=Region split from  :
wikipedia,,1226356665025
 wikipedia,,1226469112210    column=info:regioninfo,
timestamp=1226469113125, value=REGION => {NAME =>
'wikipedia,,1226469112210', STARTKEY => '', ENDKEY => '0000003046',
ENCODED => 566585206, TABLE => {{NAME => 'wikipedia', IS_ROOT =>
'false', IS_META => 'false', FAMILIES => [{NAME => 'data', BLOOMFILTER
=> 'false', VERSIONS => '3', COMPRESSION => 'NONE', LENGTH =>
'2147483647', TTL => '-1', IN_MEMORY => 'false', BLOCKCACHE =>
'false'}]}}
 wikipedia,,1226469112210    column=info:server,
timestamp=1226471630430, value=192.168.33.31:60020
 wikipedia,,1226469112210    column=info:serverstartcode,
timestamp=1226471630430, value=1226471577772


On Wed, Nov 12, 2008 at 10:09 AM, stack <st...@duboce.net> wrote:
>
> Are they both enabled?
>
> You can see by doing "scan '.META.'" in the shell and then look at the particular regions in the output.  Disabled regions will have a disabled attribute.
>
> What error are you seeing when in your job?
>
> You could enable DEBUG and then in shell try running "count 'TABLENAME'".  With DEBUG enabled, you'll see the regions loading and you'll be able to pinpoint the problematic one.  Or, if your table is large, run a "scan 'TABLENAME', ['NONEXISTANT_COLUMNNAME']".  That'll cause the scanner to spin through all regions quickly -- there'll be no values found as it scans -- and again with DEBUG you can see the regions loading and figure the bad one.
>
> When you've found the bad region, look at its listing in the '.META.' output.  Or paste it in here and we'll take a look for you.
>
> Thanks Arthur,
> St.Ack
>
>
> Arthur van Hoff wrote:
>>
>> Hi,
>>
>> I have a table which appears to have two regions without a start key.
>> Is that expected? My job is getting stuck scanning these two regions.
>>
>> Any suggestions?
>>
>> Thanks.
>>
>



--
Arthur van Hoff - Grand Master of Alphabetical Order
The Ellerdale Project, Menlo Park, CA
avh@ellerdale.com, 650-283-0842

Re: Multiple start regions in a table

Posted by stack <st...@duboce.net>.
Are they both enabled?

You can see by doing "scan '.META.'" in the shell and then look at the 
particular regions in the output.  Disabled regions will have a disabled 
attribute.

What error are you seeing when in your job?

You could enable DEBUG and then in shell try running "count 
'TABLENAME'".  With DEBUG enabled, you'll see the regions loading and 
you'll be able to pinpoint the problematic one.  Or, if your table is 
large, run a "scan 'TABLENAME', ['NONEXISTANT_COLUMNNAME']".  That'll 
cause the scanner to spin through all regions quickly -- there'll be no 
values found as it scans -- and again with DEBUG you can see the regions 
loading and figure the bad one.

When you've found the bad region, look at its listing in the '.META.' 
output.  Or paste it in here and we'll take a look for you.

Thanks Arthur,
St.Ack


Arthur van Hoff wrote:
> Hi,
>
> I have a table which appears to have two regions without a start key.
> Is that expected? My job is getting stuck scanning these two regions.
>
> Any suggestions?
>
> Thanks.
>