You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Imran M Yousuf <im...@gmail.com> on 2010/08/24 05:35:26 UTC

Best way to get multiple non-sequential rows

Hi,

I am using the HBase client API to interact with HBase. I have noticed
that HTableInterface has operations such as put(List<Put>),
delete(List<Delete>), but there is no similar method for Get. Using
scan it is possible to load a range of rows, i.e. sequential rows. My
question is -
how would it be most efficient to load N non-sequential rows?

Currently I am using get(Get) method N times.

-- 
Imran M Yousuf
Blog: http://imyousuf-tech.blogs.smartitengineering.com/
Mobile: +880-1711402557

Re: Best way to get multiple non-sequential rows

Posted by Imran M Yousuf <im...@gmail.com>.
Thanks St.Ack.

On Tue, Aug 24, 2010 at 9:54 AM, Stack <st...@duboce.net> wrote:
> There is no multiget currently.  You'll have to effect it yourself at
> a higher level spinning threads each doing a single Get.
> St.Ack
>
> On Mon, Aug 23, 2010 at 8:35 PM, Imran M Yousuf <im...@gmail.com> wrote:
>> Hi,
>>
>> I am using the HBase client API to interact with HBase. I have noticed
>> that HTableInterface has operations such as put(List<Put>),
>> delete(List<Delete>), but there is no similar method for Get. Using
>> scan it is possible to load a range of rows, i.e. sequential rows. My
>> question is -
>> how would it be most efficient to load N non-sequential rows?
>>
>> Currently I am using get(Get) method N times.
>>
>> --
>> Imran M Yousuf
>> Blog: http://imyousuf-tech.blogs.smartitengineering.com/
>> Mobile: +880-1711402557
>>
>



-- 
Imran M Yousuf
Entrepreneur & CEO
Smart IT Engineering Ltd.
Dhaka, Bangladesh
Email: imran@smartitengineering.com
Blog: http://imyousuf-tech.blogs.smartitengineering.com/
Mobile: +880-1711402557

Re: Best way to get multiple non-sequential rows

Posted by Stack <st...@duboce.net>.
There is no multiget currently.  You'll have to effect it yourself at
a higher level spinning threads each doing a single Get.
St.Ack

On Mon, Aug 23, 2010 at 8:35 PM, Imran M Yousuf <im...@gmail.com> wrote:
> Hi,
>
> I am using the HBase client API to interact with HBase. I have noticed
> that HTableInterface has operations such as put(List<Put>),
> delete(List<Delete>), but there is no similar method for Get. Using
> scan it is possible to load a range of rows, i.e. sequential rows. My
> question is -
> how would it be most efficient to load N non-sequential rows?
>
> Currently I am using get(Get) method N times.
>
> --
> Imran M Yousuf
> Blog: http://imyousuf-tech.blogs.smartitengineering.com/
> Mobile: +880-1711402557
>

Re: Best way to get multiple non-sequential rows

Posted by Imran M Yousuf <im...@gmail.com>.
Thanks for the suggestions Michael.

On Tue, Aug 24, 2010 at 5:37 PM, Michael Segel
<mi...@hotmail.com> wrote:
>
> Hi,
>
> Non sequential rows?
>
> Short answer... it depends.  :-)
>
> Longer answer... how 'non-sequential' ?
>
> If you're using a key that is hashed (SHA-1)  then your rows will be fairly random and 'non-sequential.
> Here you're best bet is to fetch each row via a get().  In order to do the get you have to know the specific key so the fetch should be fairly quick and consistent regardless of the size of the database. (near linear scalability). This works great if you know your key.
>
> If you're using some key that isn't hashed but the rows aren't sequential, you may want to do a range scan and then drop
> the rows that are not needed. This may be faster in some specific situations where all of your data is within one or two regions of a large, large table.
> (But its so specific, I don't know of the value in terms of a generic query.)
>
> An extreme and bad example... suppose you want to find all of the shops along a specific street and in part of the key you include the street side but is also based on the address.
> If you did a scan, you'd end up with a list where you may want every other entry.  So here it would be faster to do a sequential scan with a partial key to put a boundary on which regions to scan.  (Again this is a bad example.)
> If you also write your own custom filter, you can get it to return only the rows you want.
>
> Again, I apologize for the bad example... it was the first thing I could think of before I finished my first cup of coffee in the morning.
>
> HTH
>
> -Mike
>
>
>> Date: Tue, 24 Aug 2010 09:35:26 +0600
>> Subject: Best way to get multiple non-sequential rows
>> From: imyousuf@gmail.com
>> To: user@hbase.apache.org
>>
>> Hi,
>>
>> I am using the HBase client API to interact with HBase. I have noticed
>> that HTableInterface has operations such as put(List<Put>),
>> delete(List<Delete>), but there is no similar method for Get. Using
>> scan it is possible to load a range of rows, i.e. sequential rows. My
>> question is -
>> how would it be most efficient to load N non-sequential rows?
>>
>> Currently I am using get(Get) method N times.
>>
>> --
>> Imran M Yousuf
>> Blog: http://imyousuf-tech.blogs.smartitengineering.com/
>> Mobile: +880-1711402557
>



-- 
Imran M Yousuf
Entrepreneur & CEO
Smart IT Engineering Ltd.
Dhaka, Bangladesh
Email: imran@smartitengineering.com
Blog: http://imyousuf-tech.blogs.smartitengineering.com/
Mobile: +880-1711402557

Re: org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Ryan Rawson <ry...@gmail.com>.
my theory here is that those messages are just a side effect of the
real problem... they get kicked out when we are doing block reads
which involves a decompress of the data. If you are seeing slow reads
or slow decompressions you will see more 'brand-new...', because we
attempt to reuse them as much as possible.

Any other 'error' messages in the datanode logs?  Or namenode?

-ryan

On Tue, Nov 9, 2010 at 11:51 PM, Seraph Imalia <se...@eisp.co.za> wrote:
> These...
>
> cat logs/hbase-root-regionserver-dynobuntu17.log.2010-11-09 | grep xciever
> cat logs/hbase-root-master-dynobuntu17.log.2010-11-09 | grep xciever
> cat logs/hbase-root-master-dynobuntu17.log | grep xciever
> cat logs/hbase-root-regionserver-dynobuntu17.log | grep xciever
>
>
> And these (cause on the link you sent it is spelt both ways)...
>
> cat logs/hbase-root-regionserver-dynobuntu17.log.2010-11-09 | grep xceiver
> cat logs/hbase-root-master-dynobuntu17.log.2010-11-09 | grep xceiver
> cat logs/hbase-root-master-dynobuntu17.log | grep xceiver
> cat logs/hbase-root-regionserver-dynobuntu17.log | grep xceiver
>
> Both came back with nothing at all :(
>
>
> I also scanned every log for the past 7 days and the "Got brand-new
> decompressor" has only ever happened last night.  Whilst that does not
> seem to be an error message, it may lead us to what really caused it.
>  Under what conditions would it "Get a new decompressor"?
>
> Scanning the logs also revealed that
> "649681515:java.net.SocketTimeoutException: 60000 millis timeout
> while waiting for channel to be ready for connect. ch
> : java.nio.channels.SocketChannel[connection-pending remote=/192.168.2.97:5
> 0010]" started happening 2 hours before the first "Got brand-new
> decompressor" (about 10 SocketTimeoutExceptions every 5 minutes).  The
> message also shows three times on the 4th Nov, once on 5th Nov and about
> 10 times on 8th - but were not as frequent or as dense as last night's
> problem.
>
> It is also interesting to note that this happened during a time when we
> are only at about 40% load to what it normally is during the day.
>
> Seraph
>
>
>
> On 2010/11/10 12:25 AM, "Ryan Rawson" <ry...@gmail.com> wrote:
>
>>This sounds like it could be the dreaded 'xciever count' issue.
>>Threads are your resources here. See:
>>
>>http://wiki.apache.org/hadoop/Hbase/Troubleshooting#A5
>>
>>Let me know if you see anything like that.
>>
>>
>>
>>On Tue, Nov 9, 2010 at 2:22 PM, Seraph Imalia <se...@eisp.co.za> wrote:
>>> Hi Ryan,
>>>
>>> Thanks for replying so soon.
>>>
>>> Whatever it was, it has stopped happening, so I am breathing normally
>>> again and it is not so urgent anymore.  I need to try figure out what
>>> caused this though.  I get the feeling it is server resource related -
>>> almost like something using the HDD or CPU heavily.  atop did not show
>>> anything unusual, but the 1 regionserver/datanode was sluggish while I
>>>was
>>> debugging the problem.  It has stopped being sluggish and it seems too
>>> much of a coincidence that it is sluggish at the same time hbase gave
>>> those errors.  Also, the mention of codec and compression in the logs
>>> makes me thing it is related to CPU rather than HDD.  Syslog and Kernel
>>> logs also reveal nothing unusual.  Any ideas on how to figure out what
>>> happened?
>>>
>>> Logs in hadoop seem normal.  Both datanodes are showing the following:
>>>
>>> 2010-11-10 00:06:48,510 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36783, bytes: 15480, op:
>>> HDFS_READ, cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:48,621 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36784, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:48,688 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36785, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:48,791 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36786, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:48,940 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36787, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,039 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36788, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,110 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36789, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,204 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36790, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,291 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36791, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,375 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36792, bytes: 1548, op:
>>> HDFS_READ, cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,449 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36793, bytes: 516, op:
>>>HDFS_READ,
>>> cliID: DFSClient_1620748290, srvID:
>>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>>> blk_3714134476848125077_129818
>>> 2010-11-10 00:06:49,555 INFO
>>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>>> /192.168.2.97:50010, dest: /192.168.2.97:36794, bytes: 516, op:
>>>
>>>
>>> Namenode looks like this:
>>>
>>> 2010-11-10 00:03:17,947 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
>>> 192.168.2.90
>>> 2010-11-10 00:05:47,774 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null
>>>perm=null
>>> 2010-11-10 00:05:47,775 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT-
>>>dst=null pe
>>> rm=null
>>> 2010-11-10 00:05:47,775 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META.
>>>dst=null pe
>>> rm=null
>>> 2010-11-10 00:05:47,776 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/ChannelUIDTable ds
>>> t=null perm=null
>>> 2010-11-10 00:05:47,777 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex
>>>dst=null
>>> perm=null
>>> 2010-11-10 00:05:47,820 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/UrlIndex-hostCount
>>>  dst=null perm=null
>>> 2010-11-10 00:05:47,820 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/UrlIndex-indexHost
>>>  dst=null perm=null
>>> 2010-11-10 00:05:47,864 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/UrlIndex-indexUrlU
>>> ID dst=null perm=null
>>> 2010-11-10 00:08:17,953 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
>>> 192.168.2.90
>>> 2010-11-10 00:10:43,052 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null
>>>perm=null
>>> 2010-11-10 00:10:43,053 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT-
>>>dst=null pe
>>> rm=null
>>> 2010-11-10 00:10:43,054 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META.
>>>dst=null pe
>>> rm=null
>>> 2010-11-10 00:10:43,054 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/ChannelUIDTable ds
>>> t=null perm=null
>>> 2010-11-10 00:10:43,056 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex
>>>dst=null
>>> perm=null
>>> 2010-11-10 00:10:43,100 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/UrlIndex-hostCount
>>>  dst=null perm=null
>>> 2010-11-10 00:10:43,101 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/UrlIndex-indexHost
>>>  dst=null perm=null
>>> 2010-11-10 00:10:43,143 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>>src=/hbase/UrlIndex-indexUrlU
>>> ID dst=null perm=null
>>> 2010-11-10 00:13:17,960 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
>>> 192.168.2.90
>>>
>>>
>>> Regards,
>>> Seraph
>>>
>>>
>>>
>>> On 2010/11/10 12:08 AM, "Ryan Rawson" <ry...@gmail.com> wrote:
>>>
>>>>Looks like you are running into HDFS issues, can you check the
>>>>datanode logs for errors?
>>>>
>>>>-ryan
>>>>
>>>>On Tue, Nov 9, 2010 at 2:06 PM, Seraph Imalia <se...@eisp.co.za> wrote:
>>>>> Hi,
>>>>>
>>>>> Some more info: That same Region server just showed the following in
>>>>>the
>>>>> logs too - hope this explains it?
>>>>>
>>>>> Regards,
>>>>> Seraph
>>>>>
>>>>> 649681515:java.net.SocketTimeoutException: 60000 millis timeout while
>>>>> waiting for channel to be ready for connect. ch :
>>>>> java.nio.channels.SocketChannel[connection-pending
>>>>> remote=/192.168.2.97:50010]
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.j
>>>>>av
>>>>>a:
>>>>> 213)
>>>>>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFS
>>>>>Cl
>>>>>ie
>>>>> nt.java:1848)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:192
>>>>>2)
>>>>>  at
>>>>>org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bound
>>>>>ed
>>>>>Ra
>>>>> ngeFileInputStream.java:101)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bound
>>>>>ed
>>>>>Ra
>>>>> ngeFileInputStream.java:88)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bound
>>>>>ed
>>>>>Ra
>>>>> ngeFileInputStream.java:81)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.io.compress.BlockDecompressorStream.rawReadInt(BlockD
>>>>>ec
>>>>>om
>>>>> pressorStream.java:121)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockD
>>>>>ec
>>>>>om
>>>>> pressorStream.java:66)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStrea
>>>>>m.
>>>>>ja
>>>>> va:74)
>>>>>  at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>>>>>  at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>>>>  at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:100)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:101
>>>>>8)
>>>>>  at
>>>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:966)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.next(HFile.java:1
>>>>>15
>>>>>9)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileSca
>>>>>nn
>>>>>er
>>>>> .java:58)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.jav
>>>>>a:
>>>>>79
>>>>> )
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.jav
>>>>>a:
>>>>>23
>>>>> 6)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.jav
>>>>>a:
>>>>>10
>>>>> 6)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.nextInternal
>>>>>(H
>>>>>Re
>>>>> gion.java:1915)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.next(HRegion
>>>>>.j
>>>>>av
>>>>> a:1879)
>>>>>  at
>>>>>org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2500)
>>>>>  at
>>>>>org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2486)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.ja
>>>>>va
>>>>>:1
>>>>> 733)
>>>>>  at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>>>>>  at
>>>>>
>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccesso
>>>>>rI
>>>>>mp
>>>>> l.java:25)
>>>>>  at java.lang.reflect.Method.invoke(Method.java:597)
>>>>>  at
>>>>>org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)
>>>>>  at
>>>>>
>>>>>org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:91
>>>>>5)
>>>>>
>>>>> 2010-11-10 00:03:57,903 DEBUG
>>>>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>> Total=66.45012MB (69678000), Free=341.48737MB (358075472),
>>>>>Max=407.9375MB
>>>>> (427753472), Counts: Blocks=2147, Access=42032, Hit=39143, Miss=2889,
>>>>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.12666654586792%, Miss
>>>>> Ratio=6.8733349442481995%, Evicted/Run=NaN
>>>>> 2010-11-10 00:04:57,903 DEBUG
>>>>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>> Total=69.27812MB (72643376), Free=338.65936MB (355110096),
>>>>>Max=407.9375MB
>>>>> (427753472), Counts: Blocks=2192, Access=43926, Hit=40999, Miss=2927,
>>>>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.33652257919312%, Miss
>>>>> Ratio=6.663479655981064%, Evicted/Run=NaN
>>>>>
>>>>>
>>>>>
>>>>> On 2010/11/09 11:59 PM, "Seraph Imalia" <se...@eisp.co.za> wrote:
>>>>>
>>>>>>Hi,
>>>>>>
>>>>>>One of our region servers keeps doing the following - it has only just
>>>>>>started doing this since 40 minutes ago.  Our clients are able to get
>>>>>>data
>>>>>>from hBase, but after a short while, threads lock up and they start
>>>>>>waiting indefinitely for data to be returned.  What is wrong? - What
>>>>>>do
>>>>>>we
>>>>>>do? - I am desperate, please help as quick as you can.
>>>>>>
>>>>>>Regards,
>>>>>>Seraph
>>>>>>
>>>>>>2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:50:57,903 DEBUG
>>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>>Total=25.640144MB (26885640), Free=382.29736MB (400867832),
>>>>>>Max=407.9375MB
>>>>>>(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
>>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
>>>>>>Ratio=7.142169773578644%, Evicted/Run=NaN
>>>>>>2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:51:57,903 DEBUG
>>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>>Total=28.028427MB (29389936), Free=379.90906MB (398363536),
>>>>>>Max=407.9375MB
>>>>>>(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
>>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
>>>>>>Ratio=7.254531979560852%, Evicted/Run=NaN
>>>>>>2010-11-09 23:52:57,903 DEBUG
>>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>>Total=31.233871MB (32751088), Free=376.7036MB (395002384),
>>>>>>Max=407.9375MB
>>>>>>(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
>>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
>>>>>>Ratio=7.353174686431885%, Evicted/Run=NaN
>>>>>>2010-11-09 23:53:57,903 DEBUG
>>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>>Total=34.532898MB (36210368), Free=373.4046MB (391543104),
>>>>>>Max=407.9375MB
>>>>>>(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
>>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
>>>>>>Ratio=7.484499365091324%, Evicted/Run=NaN
>>>>>>2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:54:57,903 DEBUG
>>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>>Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
>>>>>>(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
>>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
>>>>>>Ratio=7.609961926937103%, Evicted/Run=NaN
>>>>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>>Got
>>>>>>brand-new decompressor
>>>>>>2010-11-09 23:55:57,903 DEBUG
>>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>>Total=40.266388MB (42222368), Free=367.6711MB (385531104),
>>>>>>Max=407.9375MB
>>>>>>(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
>>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
>>>>>>Ratio=7.300348579883575%, Evicted/Run=NaN
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>
>>>
>>>
>
>
>
>

Re: org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Seraph Imalia <se...@eisp.co.za>.
These...

cat logs/hbase-root-regionserver-dynobuntu17.log.2010-11-09 | grep xciever
cat logs/hbase-root-master-dynobuntu17.log.2010-11-09 | grep xciever
cat logs/hbase-root-master-dynobuntu17.log | grep xciever
cat logs/hbase-root-regionserver-dynobuntu17.log | grep xciever


And these (cause on the link you sent it is spelt both ways)...

cat logs/hbase-root-regionserver-dynobuntu17.log.2010-11-09 | grep xceiver
cat logs/hbase-root-master-dynobuntu17.log.2010-11-09 | grep xceiver
cat logs/hbase-root-master-dynobuntu17.log | grep xceiver
cat logs/hbase-root-regionserver-dynobuntu17.log | grep xceiver

Both came back with nothing at all :(


I also scanned every log for the past 7 days and the "Got brand-new
decompressor" has only ever happened last night.  Whilst that does not
seem to be an error message, it may lead us to what really caused it.
 Under what conditions would it "Get a new decompressor"?

Scanning the logs also revealed that
"649681515:java.net.SocketTimeoutException: 60000 millis timeout
while waiting for channel to be ready for connect. ch
: java.nio.channels.SocketChannel[connection-pending remote=/192.168.2.97:5
0010]" started happening 2 hours before the first "Got brand-new
decompressor" (about 10 SocketTimeoutExceptions every 5 minutes).  The
message also shows three times on the 4th Nov, once on 5th Nov and about
10 times on 8th - but were not as frequent or as dense as last night's
problem.

It is also interesting to note that this happened during a time when we
are only at about 40% load to what it normally is during the day.

Seraph



On 2010/11/10 12:25 AM, "Ryan Rawson" <ry...@gmail.com> wrote:

>This sounds like it could be the dreaded 'xciever count' issue.
>Threads are your resources here. See:
>
>http://wiki.apache.org/hadoop/Hbase/Troubleshooting#A5
>
>Let me know if you see anything like that.
>
>
>
>On Tue, Nov 9, 2010 at 2:22 PM, Seraph Imalia <se...@eisp.co.za> wrote:
>> Hi Ryan,
>>
>> Thanks for replying so soon.
>>
>> Whatever it was, it has stopped happening, so I am breathing normally
>> again and it is not so urgent anymore.  I need to try figure out what
>> caused this though.  I get the feeling it is server resource related -
>> almost like something using the HDD or CPU heavily.  atop did not show
>> anything unusual, but the 1 regionserver/datanode was sluggish while I
>>was
>> debugging the problem.  It has stopped being sluggish and it seems too
>> much of a coincidence that it is sluggish at the same time hbase gave
>> those errors.  Also, the mention of codec and compression in the logs
>> makes me thing it is related to CPU rather than HDD.  Syslog and Kernel
>> logs also reveal nothing unusual.  Any ideas on how to figure out what
>> happened?
>>
>> Logs in hadoop seem normal.  Both datanodes are showing the following:
>>
>> 2010-11-10 00:06:48,510 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36783, bytes: 15480, op:
>> HDFS_READ, cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:48,621 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36784, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:48,688 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36785, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:48,791 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36786, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:48,940 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36787, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,039 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36788, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,110 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36789, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,204 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36790, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,291 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36791, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,375 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36792, bytes: 1548, op:
>> HDFS_READ, cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,449 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36793, bytes: 516, op:
>>HDFS_READ,
>> cliID: DFSClient_1620748290, srvID:
>> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
>> blk_3714134476848125077_129818
>> 2010-11-10 00:06:49,555 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
>> /192.168.2.97:50010, dest: /192.168.2.97:36794, bytes: 516, op:
>>
>>
>> Namenode looks like this:
>>
>> 2010-11-10 00:03:17,947 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
>> 192.168.2.90
>> 2010-11-10 00:05:47,774 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null
>>perm=null
>> 2010-11-10 00:05:47,775 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT-
>>dst=null pe
>> rm=null
>> 2010-11-10 00:05:47,775 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META.
>>dst=null pe
>> rm=null
>> 2010-11-10 00:05:47,776 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/ChannelUIDTable ds
>> t=null perm=null
>> 2010-11-10 00:05:47,777 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex
>>dst=null
>> perm=null
>> 2010-11-10 00:05:47,820 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/UrlIndex-hostCount
>>  dst=null perm=null
>> 2010-11-10 00:05:47,820 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/UrlIndex-indexHost
>>  dst=null perm=null
>> 2010-11-10 00:05:47,864 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/UrlIndex-indexUrlU
>> ID dst=null perm=null
>> 2010-11-10 00:08:17,953 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
>> 192.168.2.90
>> 2010-11-10 00:10:43,052 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null
>>perm=null
>> 2010-11-10 00:10:43,053 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT-
>>dst=null pe
>> rm=null
>> 2010-11-10 00:10:43,054 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META.
>>dst=null pe
>> rm=null
>> 2010-11-10 00:10:43,054 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/ChannelUIDTable ds
>> t=null perm=null
>> 2010-11-10 00:10:43,056 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex
>>dst=null
>> perm=null
>> 2010-11-10 00:10:43,100 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/UrlIndex-hostCount
>>  dst=null perm=null
>> 2010-11-10 00:10:43,101 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/UrlIndex-indexHost
>>  dst=null perm=null
>> 2010-11-10 00:10:43,143 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=root,root ip=/192.168.2.97 cmd=listStatus
>>src=/hbase/UrlIndex-indexUrlU
>> ID dst=null perm=null
>> 2010-11-10 00:13:17,960 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
>> 192.168.2.90
>>
>>
>> Regards,
>> Seraph
>>
>>
>>
>> On 2010/11/10 12:08 AM, "Ryan Rawson" <ry...@gmail.com> wrote:
>>
>>>Looks like you are running into HDFS issues, can you check the
>>>datanode logs for errors?
>>>
>>>-ryan
>>>
>>>On Tue, Nov 9, 2010 at 2:06 PM, Seraph Imalia <se...@eisp.co.za> wrote:
>>>> Hi,
>>>>
>>>> Some more info: That same Region server just showed the following in
>>>>the
>>>> logs too - hope this explains it?
>>>>
>>>> Regards,
>>>> Seraph
>>>>
>>>> 649681515:java.net.SocketTimeoutException: 60000 millis timeout while
>>>> waiting for channel to be ready for connect. ch :
>>>> java.nio.channels.SocketChannel[connection-pending
>>>> remote=/192.168.2.97:50010]
>>>>  at
>>>>
>>>>org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.j
>>>>av
>>>>a:
>>>> 213)
>>>>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
>>>>  at
>>>>
>>>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFS
>>>>Cl
>>>>ie
>>>> nt.java:1848)
>>>>  at
>>>>
>>>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:192
>>>>2)
>>>>  at
>>>>org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bound
>>>>ed
>>>>Ra
>>>> ngeFileInputStream.java:101)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bound
>>>>ed
>>>>Ra
>>>> ngeFileInputStream.java:88)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bound
>>>>ed
>>>>Ra
>>>> ngeFileInputStream.java:81)
>>>>  at
>>>>
>>>>org.apache.hadoop.io.compress.BlockDecompressorStream.rawReadInt(BlockD
>>>>ec
>>>>om
>>>> pressorStream.java:121)
>>>>  at
>>>>
>>>>org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockD
>>>>ec
>>>>om
>>>> pressorStream.java:66)
>>>>  at
>>>>
>>>>org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStrea
>>>>m.
>>>>ja
>>>> va:74)
>>>>  at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>>>>  at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>>>  at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:100)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:101
>>>>8)
>>>>  at
>>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:966)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.next(HFile.java:1
>>>>15
>>>>9)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileSca
>>>>nn
>>>>er
>>>> .java:58)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.jav
>>>>a:
>>>>79
>>>> )
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.jav
>>>>a:
>>>>23
>>>> 6)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.jav
>>>>a:
>>>>10
>>>> 6)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.nextInternal
>>>>(H
>>>>Re
>>>> gion.java:1915)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.next(HRegion
>>>>.j
>>>>av
>>>> a:1879)
>>>>  at 
>>>>org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2500)
>>>>  at 
>>>>org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2486)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.ja
>>>>va
>>>>:1
>>>> 733)
>>>>  at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>>>>  at
>>>>
>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccesso
>>>>rI
>>>>mp
>>>> l.java:25)
>>>>  at java.lang.reflect.Method.invoke(Method.java:597)
>>>>  at 
>>>>org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)
>>>>  at
>>>>
>>>>org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:91
>>>>5)
>>>>
>>>> 2010-11-10 00:03:57,903 DEBUG
>>>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>> Total=66.45012MB (69678000), Free=341.48737MB (358075472),
>>>>Max=407.9375MB
>>>> (427753472), Counts: Blocks=2147, Access=42032, Hit=39143, Miss=2889,
>>>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.12666654586792%, Miss
>>>> Ratio=6.8733349442481995%, Evicted/Run=NaN
>>>> 2010-11-10 00:04:57,903 DEBUG
>>>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>> Total=69.27812MB (72643376), Free=338.65936MB (355110096),
>>>>Max=407.9375MB
>>>> (427753472), Counts: Blocks=2192, Access=43926, Hit=40999, Miss=2927,
>>>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.33652257919312%, Miss
>>>> Ratio=6.663479655981064%, Evicted/Run=NaN
>>>>
>>>>
>>>>
>>>> On 2010/11/09 11:59 PM, "Seraph Imalia" <se...@eisp.co.za> wrote:
>>>>
>>>>>Hi,
>>>>>
>>>>>One of our region servers keeps doing the following - it has only just
>>>>>started doing this since 40 minutes ago.  Our clients are able to get
>>>>>data
>>>>>from hBase, but after a short while, threads lock up and they start
>>>>>waiting indefinitely for data to be returned.  What is wrong? - What
>>>>>do
>>>>>we
>>>>>do? - I am desperate, please help as quick as you can.
>>>>>
>>>>>Regards,
>>>>>Seraph
>>>>>
>>>>>2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:50:57,903 DEBUG
>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>Total=25.640144MB (26885640), Free=382.29736MB (400867832),
>>>>>Max=407.9375MB
>>>>>(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
>>>>>Ratio=7.142169773578644%, Evicted/Run=NaN
>>>>>2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:51:57,903 DEBUG
>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>Total=28.028427MB (29389936), Free=379.90906MB (398363536),
>>>>>Max=407.9375MB
>>>>>(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
>>>>>Ratio=7.254531979560852%, Evicted/Run=NaN
>>>>>2010-11-09 23:52:57,903 DEBUG
>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>Total=31.233871MB (32751088), Free=376.7036MB (395002384),
>>>>>Max=407.9375MB
>>>>>(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
>>>>>Ratio=7.353174686431885%, Evicted/Run=NaN
>>>>>2010-11-09 23:53:57,903 DEBUG
>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>Total=34.532898MB (36210368), Free=373.4046MB (391543104),
>>>>>Max=407.9375MB
>>>>>(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
>>>>>Ratio=7.484499365091324%, Evicted/Run=NaN
>>>>>2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:54:57,903 DEBUG
>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
>>>>>(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
>>>>>Ratio=7.609961926937103%, Evicted/Run=NaN
>>>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>>Got
>>>>>brand-new decompressor
>>>>>2010-11-09 23:55:57,903 DEBUG
>>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>>Total=40.266388MB (42222368), Free=367.6711MB (385531104),
>>>>>Max=407.9375MB
>>>>>(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
>>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
>>>>>Ratio=7.300348579883575%, Evicted/Run=NaN
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>>
>>>>
>>
>>
>>




Re: org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Ryan Rawson <ry...@gmail.com>.
This sounds like it could be the dreaded 'xciever count' issue.
Threads are your resources here. See:

http://wiki.apache.org/hadoop/Hbase/Troubleshooting#A5

Let me know if you see anything like that.



On Tue, Nov 9, 2010 at 2:22 PM, Seraph Imalia <se...@eisp.co.za> wrote:
> Hi Ryan,
>
> Thanks for replying so soon.
>
> Whatever it was, it has stopped happening, so I am breathing normally
> again and it is not so urgent anymore.  I need to try figure out what
> caused this though.  I get the feeling it is server resource related -
> almost like something using the HDD or CPU heavily.  atop did not show
> anything unusual, but the 1 regionserver/datanode was sluggish while I was
> debugging the problem.  It has stopped being sluggish and it seems too
> much of a coincidence that it is sluggish at the same time hbase gave
> those errors.  Also, the mention of codec and compression in the logs
> makes me thing it is related to CPU rather than HDD.  Syslog and Kernel
> logs also reveal nothing unusual.  Any ideas on how to figure out what
> happened?
>
> Logs in hadoop seem normal.  Both datanodes are showing the following:
>
> 2010-11-10 00:06:48,510 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36783, bytes: 15480, op:
> HDFS_READ, cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:48,621 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36784, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:48,688 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36785, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:48,791 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36786, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:48,940 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36787, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,039 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36788, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,110 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36789, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,204 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36790, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,291 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36791, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,375 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36792, bytes: 1548, op:
> HDFS_READ, cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,449 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36793, bytes: 516, op: HDFS_READ,
> cliID: DFSClient_1620748290, srvID:
> DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
> blk_3714134476848125077_129818
> 2010-11-10 00:06:49,555 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
> /192.168.2.97:50010, dest: /192.168.2.97:36794, bytes: 516, op:
>
>
> Namenode looks like this:
>
> 2010-11-10 00:03:17,947 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
> 192.168.2.90
> 2010-11-10 00:05:47,774 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null perm=null
> 2010-11-10 00:05:47,775 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT- dst=null pe
> rm=null
> 2010-11-10 00:05:47,775 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META. dst=null pe
> rm=null
> 2010-11-10 00:05:47,776 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/ChannelUIDTable ds
> t=null perm=null
> 2010-11-10 00:05:47,777 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex dst=null
> perm=null
> 2010-11-10 00:05:47,820 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-hostCount
>  dst=null perm=null
> 2010-11-10 00:05:47,820 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexHost
>  dst=null perm=null
> 2010-11-10 00:05:47,864 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexUrlU
> ID dst=null perm=null
> 2010-11-10 00:08:17,953 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
> 192.168.2.90
> 2010-11-10 00:10:43,052 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null perm=null
> 2010-11-10 00:10:43,053 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT- dst=null pe
> rm=null
> 2010-11-10 00:10:43,054 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META. dst=null pe
> rm=null
> 2010-11-10 00:10:43,054 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/ChannelUIDTable ds
> t=null perm=null
> 2010-11-10 00:10:43,056 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex dst=null
> perm=null
> 2010-11-10 00:10:43,100 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-hostCount
>  dst=null perm=null
> 2010-11-10 00:10:43,101 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexHost
>  dst=null perm=null
> 2010-11-10 00:10:43,143 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
> ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexUrlU
> ID dst=null perm=null
> 2010-11-10 00:13:17,960 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
> 192.168.2.90
>
>
> Regards,
> Seraph
>
>
>
> On 2010/11/10 12:08 AM, "Ryan Rawson" <ry...@gmail.com> wrote:
>
>>Looks like you are running into HDFS issues, can you check the
>>datanode logs for errors?
>>
>>-ryan
>>
>>On Tue, Nov 9, 2010 at 2:06 PM, Seraph Imalia <se...@eisp.co.za> wrote:
>>> Hi,
>>>
>>> Some more info: That same Region server just showed the following in the
>>> logs too - hope this explains it?
>>>
>>> Regards,
>>> Seraph
>>>
>>> 649681515:java.net.SocketTimeoutException: 60000 millis timeout while
>>> waiting for channel to be ready for connect. ch :
>>> java.nio.channels.SocketChannel[connection-pending
>>> remote=/192.168.2.97:50010]
>>>  at
>>>
>>>org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.jav
>>>a:
>>> 213)
>>>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
>>>  at
>>>
>>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFSCl
>>>ie
>>> nt.java:1848)
>>>  at
>>>
>>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1922)
>>>  at
>>>org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
>>>  at
>>>
>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bounded
>>>Ra
>>> ngeFileInputStream.java:101)
>>>  at
>>>
>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bounded
>>>Ra
>>> ngeFileInputStream.java:88)
>>>  at
>>>
>>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bounded
>>>Ra
>>> ngeFileInputStream.java:81)
>>>  at
>>>
>>>org.apache.hadoop.io.compress.BlockDecompressorStream.rawReadInt(BlockDec
>>>om
>>> pressorStream.java:121)
>>>  at
>>>
>>>org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockDec
>>>om
>>> pressorStream.java:66)
>>>  at
>>>
>>>org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.
>>>ja
>>> va:74)
>>>  at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>>>  at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>>  at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:100)
>>>  at
>>>
>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:1018)
>>>  at
>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:966)
>>>  at
>>>
>>>org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.next(HFile.java:115
>>>9)
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScann
>>>er
>>> .java:58)
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:
>>>79
>>> )
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:
>>>23
>>> 6)
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:
>>>10
>>> 6)
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.nextInternal(H
>>>Re
>>> gion.java:1915)
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.next(HRegion.j
>>>av
>>> a:1879)
>>>  at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2500)
>>>  at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2486)
>>>  at
>>>
>>>org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.java
>>>:1
>>> 733)
>>>  at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>>>  at
>>>
>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
>>>mp
>>> l.java:25)
>>>  at java.lang.reflect.Method.invoke(Method.java:597)
>>>  at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)
>>>  at
>>>
>>>org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915)
>>>
>>> 2010-11-10 00:03:57,903 DEBUG
>>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>> Total=66.45012MB (69678000), Free=341.48737MB (358075472),
>>>Max=407.9375MB
>>> (427753472), Counts: Blocks=2147, Access=42032, Hit=39143, Miss=2889,
>>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.12666654586792%, Miss
>>> Ratio=6.8733349442481995%, Evicted/Run=NaN
>>> 2010-11-10 00:04:57,903 DEBUG
>>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>> Total=69.27812MB (72643376), Free=338.65936MB (355110096),
>>>Max=407.9375MB
>>> (427753472), Counts: Blocks=2192, Access=43926, Hit=40999, Miss=2927,
>>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.33652257919312%, Miss
>>> Ratio=6.663479655981064%, Evicted/Run=NaN
>>>
>>>
>>>
>>> On 2010/11/09 11:59 PM, "Seraph Imalia" <se...@eisp.co.za> wrote:
>>>
>>>>Hi,
>>>>
>>>>One of our region servers keeps doing the following - it has only just
>>>>started doing this since 40 minutes ago.  Our clients are able to get
>>>>data
>>>>from hBase, but after a short while, threads lock up and they start
>>>>waiting indefinitely for data to be returned.  What is wrong? - What do
>>>>we
>>>>do? - I am desperate, please help as quick as you can.
>>>>
>>>>Regards,
>>>>Seraph
>>>>
>>>>2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:50:57,903 DEBUG
>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>Total=25.640144MB (26885640), Free=382.29736MB (400867832),
>>>>Max=407.9375MB
>>>>(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
>>>>Ratio=7.142169773578644%, Evicted/Run=NaN
>>>>2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:51:57,903 DEBUG
>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>Total=28.028427MB (29389936), Free=379.90906MB (398363536),
>>>>Max=407.9375MB
>>>>(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
>>>>Ratio=7.254531979560852%, Evicted/Run=NaN
>>>>2010-11-09 23:52:57,903 DEBUG
>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>Total=31.233871MB (32751088), Free=376.7036MB (395002384),
>>>>Max=407.9375MB
>>>>(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
>>>>Ratio=7.353174686431885%, Evicted/Run=NaN
>>>>2010-11-09 23:53:57,903 DEBUG
>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>Total=34.532898MB (36210368), Free=373.4046MB (391543104),
>>>>Max=407.9375MB
>>>>(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
>>>>Ratio=7.484499365091324%, Evicted/Run=NaN
>>>>2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:54:57,903 DEBUG
>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
>>>>(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
>>>>Ratio=7.609961926937103%, Evicted/Run=NaN
>>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool:
>>>>Got
>>>>brand-new decompressor
>>>>2010-11-09 23:55:57,903 DEBUG
>>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>>Total=40.266388MB (42222368), Free=367.6711MB (385531104),
>>>>Max=407.9375MB
>>>>(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
>>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
>>>>Ratio=7.300348579883575%, Evicted/Run=NaN
>>>>
>>>>
>>>>
>>>
>>>
>>>
>>>
>
>
>
>

Re: org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Seraph Imalia <se...@eisp.co.za>.
Hi Ryan,

Thanks for replying so soon.

Whatever it was, it has stopped happening, so I am breathing normally
again and it is not so urgent anymore.  I need to try figure out what
caused this though.  I get the feeling it is server resource related -
almost like something using the HDD or CPU heavily.  atop did not show
anything unusual, but the 1 regionserver/datanode was sluggish while I was
debugging the problem.  It has stopped being sluggish and it seems too
much of a coincidence that it is sluggish at the same time hbase gave
those errors.  Also, the mention of codec and compression in the logs
makes me thing it is related to CPU rather than HDD.  Syslog and Kernel
logs also reveal nothing unusual.  Any ideas on how to figure out what
happened?

Logs in hadoop seem normal.  Both datanodes are showing the following:

2010-11-10 00:06:48,510 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36783, bytes: 15480, op:
HDFS_READ, cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:48,621 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36784, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:48,688 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36785, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:48,791 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36786, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:48,940 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36787, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,039 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36788, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,110 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36789, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,204 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36790, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,291 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36791, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,375 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36792, bytes: 1548, op:
HDFS_READ, cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,449 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36793, bytes: 516, op: HDFS_READ,
cliID: DFSClient_1620748290, srvID:
DS-1090448426-192.168.2.97-50010-1282311128239, blockid:
blk_3714134476848125077_129818
2010-11-10 00:06:49,555 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src:
/192.168.2.97:50010, dest: /192.168.2.97:36794, bytes: 516, op: 


Namenode looks like this:

2010-11-10 00:03:17,947 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
192.168.2.90
2010-11-10 00:05:47,774 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null perm=null
2010-11-10 00:05:47,775 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT- dst=null pe
rm=null
2010-11-10 00:05:47,775 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META. dst=null pe
rm=null
2010-11-10 00:05:47,776 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/ChannelUIDTable ds
t=null perm=null
2010-11-10 00:05:47,777 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex dst=null 
perm=null
2010-11-10 00:05:47,820 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-hostCount
 dst=null perm=null
2010-11-10 00:05:47,820 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexHost
 dst=null perm=null
2010-11-10 00:05:47,864 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexUrlU
ID dst=null perm=null
2010-11-10 00:08:17,953 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
192.168.2.90
2010-11-10 00:10:43,052 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase dst=null perm=null
2010-11-10 00:10:43,053 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/-ROOT- dst=null pe
rm=null
2010-11-10 00:10:43,054 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/.META. dst=null pe
rm=null
2010-11-10 00:10:43,054 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/ChannelUIDTable ds
t=null perm=null
2010-11-10 00:10:43,056 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex dst=null 
perm=null
2010-11-10 00:10:43,100 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-hostCount
 dst=null perm=null
2010-11-10 00:10:43,101 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexHost
 dst=null perm=null
2010-11-10 00:10:43,143 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
ugi=root,root ip=/192.168.2.97 cmd=listStatus src=/hbase/UrlIndex-indexUrlU
ID dst=null perm=null
2010-11-10 00:13:17,960 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from
192.168.2.90


Regards,
Seraph



On 2010/11/10 12:08 AM, "Ryan Rawson" <ry...@gmail.com> wrote:

>Looks like you are running into HDFS issues, can you check the
>datanode logs for errors?
>
>-ryan
>
>On Tue, Nov 9, 2010 at 2:06 PM, Seraph Imalia <se...@eisp.co.za> wrote:
>> Hi,
>>
>> Some more info: That same Region server just showed the following in the
>> logs too - hope this explains it?
>>
>> Regards,
>> Seraph
>>
>> 649681515:java.net.SocketTimeoutException: 60000 millis timeout while
>> waiting for channel to be ready for connect. ch :
>> java.nio.channels.SocketChannel[connection-pending
>> remote=/192.168.2.97:50010]
>>  at
>> 
>>org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.jav
>>a:
>> 213)
>>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
>>  at
>> 
>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFSCl
>>ie
>> nt.java:1848)
>>  at
>> 
>>org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1922)
>>  at 
>>org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
>>  at
>> 
>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bounded
>>Ra
>> ngeFileInputStream.java:101)
>>  at
>> 
>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bounded
>>Ra
>> ngeFileInputStream.java:88)
>>  at
>> 
>>org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(Bounded
>>Ra
>> ngeFileInputStream.java:81)
>>  at
>> 
>>org.apache.hadoop.io.compress.BlockDecompressorStream.rawReadInt(BlockDec
>>om
>> pressorStream.java:121)
>>  at
>> 
>>org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockDec
>>om
>> pressorStream.java:66)
>>  at
>> 
>>org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.
>>ja
>> va:74)
>>  at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>>  at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>  at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:100)
>>  at
>> 
>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:1018)
>>  at 
>>org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:966)
>>  at
>> 
>>org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.next(HFile.java:115
>>9)
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScann
>>er
>> .java:58)
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:
>>79
>> )
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:
>>23
>> 6)
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:
>>10
>> 6)
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.nextInternal(H
>>Re
>> gion.java:1915)
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.next(HRegion.j
>>av
>> a:1879)
>>  at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2500)
>>  at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2486)
>>  at
>> 
>>org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.java
>>:1
>> 733)
>>  at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>>  at
>> 
>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorI
>>mp
>> l.java:25)
>>  at java.lang.reflect.Method.invoke(Method.java:597)
>>  at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)
>>  at
>> 
>>org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915)
>>
>> 2010-11-10 00:03:57,903 DEBUG
>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>> Total=66.45012MB (69678000), Free=341.48737MB (358075472),
>>Max=407.9375MB
>> (427753472), Counts: Blocks=2147, Access=42032, Hit=39143, Miss=2889,
>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.12666654586792%, Miss
>> Ratio=6.8733349442481995%, Evicted/Run=NaN
>> 2010-11-10 00:04:57,903 DEBUG
>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>> Total=69.27812MB (72643376), Free=338.65936MB (355110096),
>>Max=407.9375MB
>> (427753472), Counts: Blocks=2192, Access=43926, Hit=40999, Miss=2927,
>> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.33652257919312%, Miss
>> Ratio=6.663479655981064%, Evicted/Run=NaN
>>
>>
>>
>> On 2010/11/09 11:59 PM, "Seraph Imalia" <se...@eisp.co.za> wrote:
>>
>>>Hi,
>>>
>>>One of our region servers keeps doing the following - it has only just
>>>started doing this since 40 minutes ago.  Our clients are able to get
>>>data
>>>from hBase, but after a short while, threads lock up and they start
>>>waiting indefinitely for data to be returned.  What is wrong? - What do
>>>we
>>>do? - I am desperate, please help as quick as you can.
>>>
>>>Regards,
>>>Seraph
>>>
>>>2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:50:57,903 DEBUG
>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>Total=25.640144MB (26885640), Free=382.29736MB (400867832),
>>>Max=407.9375MB
>>>(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
>>>Ratio=7.142169773578644%, Evicted/Run=NaN
>>>2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:51:57,903 DEBUG
>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>Total=28.028427MB (29389936), Free=379.90906MB (398363536),
>>>Max=407.9375MB
>>>(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
>>>Ratio=7.254531979560852%, Evicted/Run=NaN
>>>2010-11-09 23:52:57,903 DEBUG
>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>Total=31.233871MB (32751088), Free=376.7036MB (395002384),
>>>Max=407.9375MB
>>>(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
>>>Ratio=7.353174686431885%, Evicted/Run=NaN
>>>2010-11-09 23:53:57,903 DEBUG
>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>Total=34.532898MB (36210368), Free=373.4046MB (391543104),
>>>Max=407.9375MB
>>>(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
>>>Ratio=7.484499365091324%, Evicted/Run=NaN
>>>2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:54:57,903 DEBUG
>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
>>>(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
>>>Ratio=7.609961926937103%, Evicted/Run=NaN
>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool:
>>>Got
>>>brand-new decompressor
>>>2010-11-09 23:55:57,903 DEBUG
>>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>>Total=40.266388MB (42222368), Free=367.6711MB (385531104),
>>>Max=407.9375MB
>>>(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
>>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
>>>Ratio=7.300348579883575%, Evicted/Run=NaN
>>>
>>>
>>>
>>
>>
>>
>>




Re: org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Ryan Rawson <ry...@gmail.com>.
Looks like you are running into HDFS issues, can you check the
datanode logs for errors?

-ryan

On Tue, Nov 9, 2010 at 2:06 PM, Seraph Imalia <se...@eisp.co.za> wrote:
> Hi,
>
> Some more info: That same Region server just showed the following in the
> logs too - hope this explains it?
>
> Regards,
> Seraph
>
> 649681515:java.net.SocketTimeoutException: 60000 millis timeout while
> waiting for channel to be ready for connect. ch :
> java.nio.channels.SocketChannel[connection-pending
> remote=/192.168.2.97:50010]
>  at
> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:
> 213)
>  at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
>  at
> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFSClie
> nt.java:1848)
>  at
> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1922)
>  at org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
>  at
> org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRa
> ngeFileInputStream.java:101)
>  at
> org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRa
> ngeFileInputStream.java:88)
>  at
> org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRa
> ngeFileInputStream.java:81)
>  at
> org.apache.hadoop.io.compress.BlockDecompressorStream.rawReadInt(BlockDecom
> pressorStream.java:121)
>  at
> org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockDecom
> pressorStream.java:66)
>  at
> org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.ja
> va:74)
>  at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>  at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>  at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:100)
>  at
> org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:1018)
>  at org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:966)
>  at
> org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.next(HFile.java:1159)
>  at
> org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScanner
> .java:58)
>  at
> org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:79
> )
>  at
> org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:23
> 6)
>  at
> org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:10
> 6)
>  at
> org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.nextInternal(HRe
> gion.java:1915)
>  at
> org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.next(HRegion.jav
> a:1879)
>  at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2500)
>  at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2486)
>  at
> org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.java:1
> 733)
>  at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>  at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
> l.java:25)
>  at java.lang.reflect.Method.invoke(Method.java:597)
>  at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)
>  at
> org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915)
>
> 2010-11-10 00:03:57,903 DEBUG
> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
> Total=66.45012MB (69678000), Free=341.48737MB (358075472), Max=407.9375MB
> (427753472), Counts: Blocks=2147, Access=42032, Hit=39143, Miss=2889,
> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.12666654586792%, Miss
> Ratio=6.8733349442481995%, Evicted/Run=NaN
> 2010-11-10 00:04:57,903 DEBUG
> org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
> Total=69.27812MB (72643376), Free=338.65936MB (355110096), Max=407.9375MB
> (427753472), Counts: Blocks=2192, Access=43926, Hit=40999, Miss=2927,
> Evictions=0, Evicted=0, Ratios: Hit Ratio=93.33652257919312%, Miss
> Ratio=6.663479655981064%, Evicted/Run=NaN
>
>
>
> On 2010/11/09 11:59 PM, "Seraph Imalia" <se...@eisp.co.za> wrote:
>
>>Hi,
>>
>>One of our region servers keeps doing the following - it has only just
>>started doing this since 40 minutes ago.  Our clients are able to get data
>>from hBase, but after a short while, threads lock up and they start
>>waiting indefinitely for data to be returned.  What is wrong? - What do we
>>do? - I am desperate, please help as quick as you can.
>>
>>Regards,
>>Seraph
>>
>>2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:50:57,903 DEBUG
>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>Total=25.640144MB (26885640), Free=382.29736MB (400867832), Max=407.9375MB
>>(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
>>Ratio=7.142169773578644%, Evicted/Run=NaN
>>2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:51:57,903 DEBUG
>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>Total=28.028427MB (29389936), Free=379.90906MB (398363536), Max=407.9375MB
>>(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
>>Ratio=7.254531979560852%, Evicted/Run=NaN
>>2010-11-09 23:52:57,903 DEBUG
>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>Total=31.233871MB (32751088), Free=376.7036MB (395002384), Max=407.9375MB
>>(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
>>Ratio=7.353174686431885%, Evicted/Run=NaN
>>2010-11-09 23:53:57,903 DEBUG
>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>Total=34.532898MB (36210368), Free=373.4046MB (391543104), Max=407.9375MB
>>(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
>>Ratio=7.484499365091324%, Evicted/Run=NaN
>>2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:54:57,903 DEBUG
>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
>>(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
>>Ratio=7.609961926937103%, Evicted/Run=NaN
>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool: Got
>>brand-new decompressor
>>2010-11-09 23:55:57,903 DEBUG
>>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>>Total=40.266388MB (42222368), Free=367.6711MB (385531104), Max=407.9375MB
>>(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
>>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
>>Ratio=7.300348579883575%, Evicted/Run=NaN
>>
>>
>>
>
>
>
>

Re: org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Seraph Imalia <se...@eisp.co.za>.
Hi,

Some more info: That same Region server just showed the following in the
logs too - hope this explains it?

Regards,
Seraph

649681515:java.net.SocketTimeoutException: 60000 millis timeout while
waiting for channel to be ready for connect. ch :
java.nio.channels.SocketChannel[connection-pending
remote=/192.168.2.97:50010]
 at 
org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:
213)
 at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404)
 at 
org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFSClie
nt.java:1848)
 at 
org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1922)
 at org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
 at 
org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRa
ngeFileInputStream.java:101)
 at 
org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRa
ngeFileInputStream.java:88)
 at 
org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRa
ngeFileInputStream.java:81)
 at 
org.apache.hadoop.io.compress.BlockDecompressorStream.rawReadInt(BlockDecom
pressorStream.java:121)
 at 
org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockDecom
pressorStream.java:66)
 at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.ja
va:74)
 at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
 at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
 at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:100)
 at 
org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:1018)
 at org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:966)
 at 
org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.next(HFile.java:1159)
 at 
org.apache.hadoop.hbase.regionserver.StoreFileScanner.next(StoreFileScanner
.java:58)
 at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:79
)
 at 
org.apache.hadoop.hbase.regionserver.StoreScanner.next(StoreScanner.java:23
6)
 at 
org.apache.hadoop.hbase.regionserver.KeyValueHeap.next(KeyValueHeap.java:10
6)
 at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.nextInternal(HRe
gion.java:1915)
 at 
org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.next(HRegion.jav
a:1879)
 at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2500)
 at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:2486)
 at 
org.apache.hadoop.hbase.regionserver.HRegionServer.get(HRegionServer.java:1
733)
 at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImp
l.java:25)
 at java.lang.reflect.Method.invoke(Method.java:597)
 at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)
 at 
org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915)

2010-11-10 00:03:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=66.45012MB (69678000), Free=341.48737MB (358075472), Max=407.9375MB
(427753472), Counts: Blocks=2147, Access=42032, Hit=39143, Miss=2889,
Evictions=0, Evicted=0, Ratios: Hit Ratio=93.12666654586792%, Miss
Ratio=6.8733349442481995%, Evicted/Run=NaN
2010-11-10 00:04:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=69.27812MB (72643376), Free=338.65936MB (355110096), Max=407.9375MB
(427753472), Counts: Blocks=2192, Access=43926, Hit=40999, Miss=2927,
Evictions=0, Evicted=0, Ratios: Hit Ratio=93.33652257919312%, Miss
Ratio=6.663479655981064%, Evicted/Run=NaN



On 2010/11/09 11:59 PM, "Seraph Imalia" <se...@eisp.co.za> wrote:

>Hi,
>
>One of our region servers keeps doing the following - it has only just
>started doing this since 40 minutes ago.  Our clients are able to get data
>from hBase, but after a short while, threads lock up and they start
>waiting indefinitely for data to be returned.  What is wrong? - What do we
>do? - I am desperate, please help as quick as you can.
>
>Regards,
>Seraph
>
>2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:50:57,903 DEBUG
>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>Total=25.640144MB (26885640), Free=382.29736MB (400867832), Max=407.9375MB
>(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
>Ratio=7.142169773578644%, Evicted/Run=NaN
>2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:51:57,903 DEBUG
>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>Total=28.028427MB (29389936), Free=379.90906MB (398363536), Max=407.9375MB
>(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
>Ratio=7.254531979560852%, Evicted/Run=NaN
>2010-11-09 23:52:57,903 DEBUG
>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>Total=31.233871MB (32751088), Free=376.7036MB (395002384), Max=407.9375MB
>(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
>Ratio=7.353174686431885%, Evicted/Run=NaN
>2010-11-09 23:53:57,903 DEBUG
>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>Total=34.532898MB (36210368), Free=373.4046MB (391543104), Max=407.9375MB
>(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
>Ratio=7.484499365091324%, Evicted/Run=NaN
>2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:54:57,903 DEBUG
>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
>(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
>Ratio=7.609961926937103%, Evicted/Run=NaN
>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool: Got
>brand-new decompressor
>2010-11-09 23:55:57,903 DEBUG
>org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
>Total=40.266388MB (42222368), Free=367.6711MB (385531104), Max=407.9375MB
>(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
>Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
>Ratio=7.300348579883575%, Evicted/Run=NaN
>
>
>




org.apache.hadoop.io.compress.CodecPool: Got brand-new decompressor

Posted by Seraph Imalia <se...@eisp.co.za>.
Hi,

One of our region servers keeps doing the following - it has only just
started doing this since 40 minutes ago.  Our clients are able to get data
from hBase, but after a short while, threads lock up and they start
waiting indefinitely for data to be returned.  What is wrong? - What do we
do? - I am desperate, please help as quick as you can.

Regards,
Seraph

2010-11-09 23:49:59,102 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:49:59,159 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:49:59,224 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:49:59,226 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:00,269 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:00,730 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:01,157 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:06,916 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:06,917 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:06,918 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:09,106 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:18,271 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:20,924 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:23,151 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:33,792 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:33,793 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:44,161 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:52,489 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:50:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=25.640144MB (26885640), Free=382.29736MB (400867832), Max=407.9375MB
(427753472), Counts: Blocks=1493, Access=31181, Hit=28954, Miss=2227,
Evictions=0, Evicted=0, Ratios: Hit Ratio=92.85783171653748%, Miss
Ratio=7.142169773578644%, Evicted/Run=NaN
2010-11-09 23:50:57,996 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:51:31,922 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:51:31,923 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:51:31,924 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:51:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=28.028427MB (29389936), Free=379.90906MB (398363536), Max=407.9375MB
(427753472), Counts: Blocks=1531, Access=31277, Hit=29008, Miss=2269,
Evictions=0, Evicted=0, Ratios: Hit Ratio=92.74546504020691%, Miss
Ratio=7.254531979560852%, Evicted/Run=NaN
2010-11-09 23:52:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=31.233871MB (32751088), Free=376.7036MB (395002384), Max=407.9375MB
(427753472), Counts: Blocks=1582, Access=31483, Hit=29168, Miss=2315,
Evictions=0, Evicted=0, Ratios: Hit Ratio=92.64682531356812%, Miss
Ratio=7.353174686431885%, Evicted/Run=NaN
2010-11-09 23:53:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=34.532898MB (36210368), Free=373.4046MB (391543104), Max=407.9375MB
(427753472), Counts: Blocks=1635, Access=31612, Hit=29246, Miss=2366,
Evictions=0, Evicted=0, Ratios: Hit Ratio=92.5154983997345%, Miss
Ratio=7.484499365091324%, Evicted/Run=NaN
2010-11-09 23:54:21,831 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:54:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=37.375MB (39190528), Free=370.5625MB (388562944), Max=407.9375MB
(427753472), Counts: Blocks=1681, Access=31761, Hit=29344, Miss=2417,
Evictions=0, Evicted=0, Ratios: Hit Ratio=92.39003658294678%, Miss
Ratio=7.609961926937103%, Evicted/Run=NaN
2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:55:45,289 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:55:48,079 INFO org.apache.hadoop.io.compress.CodecPool: Got
brand-new decompressor
2010-11-09 23:55:57,903 DEBUG
org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes:
Total=40.266388MB (42222368), Free=367.6711MB (385531104), Max=407.9375MB
(427753472), Counts: Blocks=1728, Access=33834, Hit=31364, Miss=2470,
Evictions=0, Evicted=0, Ratios: Hit Ratio=92.69965291023254%, Miss
Ratio=7.300348579883575%, Evicted/Run=NaN




RE: Best way to get multiple non-sequential rows

Posted by Michael Segel <mi...@hotmail.com>.
Hi,

Non sequential rows?

Short answer... it depends.  :-)

Longer answer... how 'non-sequential' ?

If you're using a key that is hashed (SHA-1)  then your rows will be fairly random and 'non-sequential.
Here you're best bet is to fetch each row via a get().  In order to do the get you have to know the specific key so the fetch should be fairly quick and consistent regardless of the size of the database. (near linear scalability). This works great if you know your key.

If you're using some key that isn't hashed but the rows aren't sequential, you may want to do a range scan and then drop
the rows that are not needed. This may be faster in some specific situations where all of your data is within one or two regions of a large, large table.
(But its so specific, I don't know of the value in terms of a generic query.)

An extreme and bad example... suppose you want to find all of the shops along a specific street and in part of the key you include the street side but is also based on the address.
If you did a scan, you'd end up with a list where you may want every other entry.  So here it would be faster to do a sequential scan with a partial key to put a boundary on which regions to scan.  (Again this is a bad example.)
If you also write your own custom filter, you can get it to return only the rows you want.  

Again, I apologize for the bad example... it was the first thing I could think of before I finished my first cup of coffee in the morning.

HTH

-Mike


> Date: Tue, 24 Aug 2010 09:35:26 +0600
> Subject: Best way to get multiple non-sequential rows
> From: imyousuf@gmail.com
> To: user@hbase.apache.org
> 
> Hi,
> 
> I am using the HBase client API to interact with HBase. I have noticed
> that HTableInterface has operations such as put(List<Put>),
> delete(List<Delete>), but there is no similar method for Get. Using
> scan it is possible to load a range of rows, i.e. sequential rows. My
> question is -
> how would it be most efficient to load N non-sequential rows?
> 
> Currently I am using get(Get) method N times.
> 
> -- 
> Imran M Yousuf
> Blog: http://imyousuf-tech.blogs.smartitengineering.com/
> Mobile: +880-1711402557
 		 	   		  

Re: Best way to get multiple non-sequential rows

Posted by Imran M Yousuf <im...@gmail.com>.
Hi Jonathan,

On Wed, Aug 25, 2010 at 9:52 AM, Jonathan Gray <jg...@facebook.com> wrote:
> Michael,
>
> MultiGet is about performing a set of Get operations in parallel from the client.  So it buys you potential performance benefits from the concurrency/distribution of your operations.
>
> Roughly, you would bucket the gets according to their region and regionserver.  Then spawn a thread for each RS and fire off the Gets concurrently.
>
> If I have 100 Gets to perform on a random set of keys, assuming each get takes 10ms, doing them sequentially will take 1 second.  Other factors and RS concurrency aside, with MultiGet on a 10 node cluster, the total time would be reduced to 100ms. With 50 nodes, 20ms.

The MultiGet functionality does seem great and that is exactly what I
am looking for! So waiting eagerly to see it make way into the client
API.

/Imran

>
> JG
>
>
>> -----Original Message-----
>> From: Michael Segel [mailto:michael_segel@hotmail.com]
>> Sent: Tuesday, August 24, 2010 7:53 PM
>> To: user@hbase.apache.org
>> Subject: RE: Best way to get multiple non-sequential rows
>>
>>
>> Igor,
>>
>> What does this really buy you?
>>
>> I'm trying to figure out a use case that would show a benefit from just
>> fetching the rows individually. Since the rows are not contiguous, the
>> odds of the next row you want being in cache are going to slight to
>> most likely not. ;-)
>>
>> Can you give a use case where having a 'multi-get' will make life
>> easier?
>>
>> Thx
>>
>> -Mike
>>
>>
>> > Date: Wed, 25 Aug 2010 07:17:13 +0600
>> > Subject: Re: Best way to get multiple non-sequential rows
>> > From: imyousuf@gmail.com
>> > To: user@hbase.apache.org
>> >
>> > Thanks Igor, I will have a look at it.
>> >
>> > /Imran
>> >
>> > On Tue, Aug 24, 2010 at 10:36 PM, Igor Ranitovic <ir...@gmail.com>
>> wrote:
>> > > Take a look at
>> > > https://issues.apache.org/jira/browse/HBASE-1845
>> > >
>> > > As an HBase user, multi gets is something that I have been looking
>> forward
>> > > to for some time now. If there is enough interest it would be great
>> if this
>> > > becomes part of 0.90.
>> > >
>> > > Take care,
>> > > i.
>> > >
>> > > Imran M Yousuf wrote:
>> > >>
>> > >> Hi,
>> > >>
>> > >> I am using the HBase client API to interact with HBase. I have
>> noticed
>> > >> that HTableInterface has operations such as put(List<Put>),
>> > >> delete(List<Delete>), but there is no similar method for Get.
>> Using
>> > >> scan it is possible to load a range of rows, i.e. sequential rows.
>> My
>> > >> question is -
>> > >> how would it be most efficient to load N non-sequential rows?
>> > >>
>> > >> Currently I am using get(Get) method N times.
>> > >>
>> > >
>> > >
>> >
>> >
>> >
>> > --
>> > Imran M Yousuf
>> > Blog: http://imyousuf-tech.blogs.smartitengineering.com/
>> > Mobile: +880-1711402557
>>
>



-- 
Imran M Yousuf
Blog: http://imyousuf-tech.blogs.smartitengineering.com/
Mobile: +880-1711402557

RE: Best way to get multiple non-sequential rows

Posted by Jonathan Gray <jg...@facebook.com>.
Yes, something like:

List<Result> multiGet(List<Get> gets, int maxThreads)

In general, you should assume that HTable instances are not thread-safe.  Behind the scenes, HTables are sharing TCP connections to RS, but from client POV you should have one HTable per thread per table.

> -----Original Message-----
> From: Michael Segel [mailto:michael_segel@hotmail.com]
> Sent: Wednesday, August 25, 2010 3:54 AM
> To: user@hbase.apache.org
> Subject: RE: Best way to get multiple non-sequential rows
> 
> 
> Jonathan,
> 
> Ok, that makes some sense...
> So you would have some method mget(fetchKeyList,numthreads) returning
> resultList[].
> 
> So what's thread safe these days?
> 
> -Mike
> 
> > From: jgray@facebook.com
> > To: user@hbase.apache.org
> > Subject: RE: Best way to get multiple non-sequential rows
> > Date: Wed, 25 Aug 2010 03:52:38 +0000
> >
> > Michael,
> >
> > MultiGet is about performing a set of Get operations in parallel from
> the client.  So it buys you potential performance benefits from the
> concurrency/distribution of your operations.
> >
> > Roughly, you would bucket the gets according to their region and
> regionserver.  Then spawn a thread for each RS and fire off the Gets
> concurrently.
> >
> > If I have 100 Gets to perform on a random set of keys, assuming each
> get takes 10ms, doing them sequentially will take 1 second.  Other
> factors and RS concurrency aside, with MultiGet on a 10 node cluster,
> the total time would be reduced to 100ms. With 50 nodes, 20ms.
> >
> > JG
> >
> >
> > > -----Original Message-----
> > > From: Michael Segel [mailto:michael_segel@hotmail.com]
> > > Sent: Tuesday, August 24, 2010 7:53 PM
> > > To: user@hbase.apache.org
> > > Subject: RE: Best way to get multiple non-sequential rows
> > >
> > >
> > > Igor,
> > >
> > > What does this really buy you?
> > >
> > > I'm trying to figure out a use case that would show a benefit from
> just
> > > fetching the rows individually. Since the rows are not contiguous,
> the
> > > odds of the next row you want being in cache are going to slight to
> > > most likely not. ;-)
> > >
> > > Can you give a use case where having a 'multi-get' will make life
> > > easier?
> > >
> > > Thx
> > >
> > > -Mike
> > >
> > >
> > > > Date: Wed, 25 Aug 2010 07:17:13 +0600
> > > > Subject: Re: Best way to get multiple non-sequential rows
> > > > From: imyousuf@gmail.com
> > > > To: user@hbase.apache.org
> > > >
> > > > Thanks Igor, I will have a look at it.
> > > >
> > > > /Imran
> > > >
> > > > On Tue, Aug 24, 2010 at 10:36 PM, Igor Ranitovic
> <ir...@gmail.com>
> > > wrote:
> > > > > Take a look at
> > > > > https://issues.apache.org/jira/browse/HBASE-1845
> > > > >
> > > > > As an HBase user, multi gets is something that I have been
> looking
> > > forward
> > > > > to for some time now. If there is enough interest it would be
> great
> > > if this
> > > > > becomes part of 0.90.
> > > > >
> > > > > Take care,
> > > > > i.
> > > > >
> > > > > Imran M Yousuf wrote:
> > > > >>
> > > > >> Hi,
> > > > >>
> > > > >> I am using the HBase client API to interact with HBase. I have
> > > noticed
> > > > >> that HTableInterface has operations such as put(List<Put>),
> > > > >> delete(List<Delete>), but there is no similar method for Get.
> > > Using
> > > > >> scan it is possible to load a range of rows, i.e. sequential
> rows.
> > > My
> > > > >> question is -
> > > > >> how would it be most efficient to load N non-sequential rows?
> > > > >>
> > > > >> Currently I am using get(Get) method N times.
> > > > >>
> > > > >
> > > > >
> > > >
> > > >
> > > >
> > > > --
> > > > Imran M Yousuf
> > > > Blog: http://imyousuf-tech.blogs.smartitengineering.com/
> > > > Mobile: +880-1711402557
> > >
> 

RE: Best way to get multiple non-sequential rows

Posted by Michael Segel <mi...@hotmail.com>.
Jonathan,

Ok, that makes some sense...
So you would have some method mget(fetchKeyList,numthreads) returning resultList[].

So what's thread safe these days? 

-Mike

> From: jgray@facebook.com
> To: user@hbase.apache.org
> Subject: RE: Best way to get multiple non-sequential rows
> Date: Wed, 25 Aug 2010 03:52:38 +0000
> 
> Michael,
> 
> MultiGet is about performing a set of Get operations in parallel from the client.  So it buys you potential performance benefits from the concurrency/distribution of your operations.
> 
> Roughly, you would bucket the gets according to their region and regionserver.  Then spawn a thread for each RS and fire off the Gets concurrently.
> 
> If I have 100 Gets to perform on a random set of keys, assuming each get takes 10ms, doing them sequentially will take 1 second.  Other factors and RS concurrency aside, with MultiGet on a 10 node cluster, the total time would be reduced to 100ms. With 50 nodes, 20ms.
> 
> JG
> 
> 
> > -----Original Message-----
> > From: Michael Segel [mailto:michael_segel@hotmail.com]
> > Sent: Tuesday, August 24, 2010 7:53 PM
> > To: user@hbase.apache.org
> > Subject: RE: Best way to get multiple non-sequential rows
> > 
> > 
> > Igor,
> > 
> > What does this really buy you?
> > 
> > I'm trying to figure out a use case that would show a benefit from just
> > fetching the rows individually. Since the rows are not contiguous, the
> > odds of the next row you want being in cache are going to slight to
> > most likely not. ;-)
> > 
> > Can you give a use case where having a 'multi-get' will make life
> > easier?
> > 
> > Thx
> > 
> > -Mike
> > 
> > 
> > > Date: Wed, 25 Aug 2010 07:17:13 +0600
> > > Subject: Re: Best way to get multiple non-sequential rows
> > > From: imyousuf@gmail.com
> > > To: user@hbase.apache.org
> > >
> > > Thanks Igor, I will have a look at it.
> > >
> > > /Imran
> > >
> > > On Tue, Aug 24, 2010 at 10:36 PM, Igor Ranitovic <ir...@gmail.com>
> > wrote:
> > > > Take a look at
> > > > https://issues.apache.org/jira/browse/HBASE-1845
> > > >
> > > > As an HBase user, multi gets is something that I have been looking
> > forward
> > > > to for some time now. If there is enough interest it would be great
> > if this
> > > > becomes part of 0.90.
> > > >
> > > > Take care,
> > > > i.
> > > >
> > > > Imran M Yousuf wrote:
> > > >>
> > > >> Hi,
> > > >>
> > > >> I am using the HBase client API to interact with HBase. I have
> > noticed
> > > >> that HTableInterface has operations such as put(List<Put>),
> > > >> delete(List<Delete>), but there is no similar method for Get.
> > Using
> > > >> scan it is possible to load a range of rows, i.e. sequential rows.
> > My
> > > >> question is -
> > > >> how would it be most efficient to load N non-sequential rows?
> > > >>
> > > >> Currently I am using get(Get) method N times.
> > > >>
> > > >
> > > >
> > >
> > >
> > >
> > > --
> > > Imran M Yousuf
> > > Blog: http://imyousuf-tech.blogs.smartitengineering.com/
> > > Mobile: +880-1711402557
> > 
 		 	   		  

RE: Best way to get multiple non-sequential rows

Posted by Jonathan Gray <jg...@facebook.com>.
Michael,

MultiGet is about performing a set of Get operations in parallel from the client.  So it buys you potential performance benefits from the concurrency/distribution of your operations.

Roughly, you would bucket the gets according to their region and regionserver.  Then spawn a thread for each RS and fire off the Gets concurrently.

If I have 100 Gets to perform on a random set of keys, assuming each get takes 10ms, doing them sequentially will take 1 second.  Other factors and RS concurrency aside, with MultiGet on a 10 node cluster, the total time would be reduced to 100ms. With 50 nodes, 20ms.

JG


> -----Original Message-----
> From: Michael Segel [mailto:michael_segel@hotmail.com]
> Sent: Tuesday, August 24, 2010 7:53 PM
> To: user@hbase.apache.org
> Subject: RE: Best way to get multiple non-sequential rows
> 
> 
> Igor,
> 
> What does this really buy you?
> 
> I'm trying to figure out a use case that would show a benefit from just
> fetching the rows individually. Since the rows are not contiguous, the
> odds of the next row you want being in cache are going to slight to
> most likely not. ;-)
> 
> Can you give a use case where having a 'multi-get' will make life
> easier?
> 
> Thx
> 
> -Mike
> 
> 
> > Date: Wed, 25 Aug 2010 07:17:13 +0600
> > Subject: Re: Best way to get multiple non-sequential rows
> > From: imyousuf@gmail.com
> > To: user@hbase.apache.org
> >
> > Thanks Igor, I will have a look at it.
> >
> > /Imran
> >
> > On Tue, Aug 24, 2010 at 10:36 PM, Igor Ranitovic <ir...@gmail.com>
> wrote:
> > > Take a look at
> > > https://issues.apache.org/jira/browse/HBASE-1845
> > >
> > > As an HBase user, multi gets is something that I have been looking
> forward
> > > to for some time now. If there is enough interest it would be great
> if this
> > > becomes part of 0.90.
> > >
> > > Take care,
> > > i.
> > >
> > > Imran M Yousuf wrote:
> > >>
> > >> Hi,
> > >>
> > >> I am using the HBase client API to interact with HBase. I have
> noticed
> > >> that HTableInterface has operations such as put(List<Put>),
> > >> delete(List<Delete>), but there is no similar method for Get.
> Using
> > >> scan it is possible to load a range of rows, i.e. sequential rows.
> My
> > >> question is -
> > >> how would it be most efficient to load N non-sequential rows?
> > >>
> > >> Currently I am using get(Get) method N times.
> > >>
> > >
> > >
> >
> >
> >
> > --
> > Imran M Yousuf
> > Blog: http://imyousuf-tech.blogs.smartitengineering.com/
> > Mobile: +880-1711402557
> 

RE: Best way to get multiple non-sequential rows

Posted by Michael Segel <mi...@hotmail.com>.
Igor,

What does this really buy you?

I'm trying to figure out a use case that would show a benefit from just fetching the rows individually. Since the rows are not contiguous, the odds of the next row you want being in cache are going to slight to most likely not. ;-)

Can you give a use case where having a 'multi-get' will make life easier?

Thx

-Mike


> Date: Wed, 25 Aug 2010 07:17:13 +0600
> Subject: Re: Best way to get multiple non-sequential rows
> From: imyousuf@gmail.com
> To: user@hbase.apache.org
> 
> Thanks Igor, I will have a look at it.
> 
> /Imran
> 
> On Tue, Aug 24, 2010 at 10:36 PM, Igor Ranitovic <ir...@gmail.com> wrote:
> > Take a look at
> > https://issues.apache.org/jira/browse/HBASE-1845
> >
> > As an HBase user, multi gets is something that I have been looking forward
> > to for some time now. If there is enough interest it would be great if this
> > becomes part of 0.90.
> >
> > Take care,
> > i.
> >
> > Imran M Yousuf wrote:
> >>
> >> Hi,
> >>
> >> I am using the HBase client API to interact with HBase. I have noticed
> >> that HTableInterface has operations such as put(List<Put>),
> >> delete(List<Delete>), but there is no similar method for Get. Using
> >> scan it is possible to load a range of rows, i.e. sequential rows. My
> >> question is -
> >> how would it be most efficient to load N non-sequential rows?
> >>
> >> Currently I am using get(Get) method N times.
> >>
> >
> >
> 
> 
> 
> -- 
> Imran M Yousuf
> Blog: http://imyousuf-tech.blogs.smartitengineering.com/
> Mobile: +880-1711402557
 		 	   		  

Re: Best way to get multiple non-sequential rows

Posted by Imran M Yousuf <im...@gmail.com>.
Thanks Igor, I will have a look at it.

/Imran

On Tue, Aug 24, 2010 at 10:36 PM, Igor Ranitovic <ir...@gmail.com> wrote:
> Take a look at
> https://issues.apache.org/jira/browse/HBASE-1845
>
> As an HBase user, multi gets is something that I have been looking forward
> to for some time now. If there is enough interest it would be great if this
> becomes part of 0.90.
>
> Take care,
> i.
>
> Imran M Yousuf wrote:
>>
>> Hi,
>>
>> I am using the HBase client API to interact with HBase. I have noticed
>> that HTableInterface has operations such as put(List<Put>),
>> delete(List<Delete>), but there is no similar method for Get. Using
>> scan it is possible to load a range of rows, i.e. sequential rows. My
>> question is -
>> how would it be most efficient to load N non-sequential rows?
>>
>> Currently I am using get(Get) method N times.
>>
>
>



-- 
Imran M Yousuf
Blog: http://imyousuf-tech.blogs.smartitengineering.com/
Mobile: +880-1711402557

Re: Best way to get multiple non-sequential rows

Posted by Igor Ranitovic <ir...@gmail.com>.
Take a look at
https://issues.apache.org/jira/browse/HBASE-1845

As an HBase user, multi gets is something that I have been looking 
forward to for some time now. If there is enough interest it would be 
great if this becomes part of 0.90.

Take care,
i.

Imran M Yousuf wrote:
> Hi,
> 
> I am using the HBase client API to interact with HBase. I have noticed
> that HTableInterface has operations such as put(List<Put>),
> delete(List<Delete>), but there is no similar method for Get. Using
> scan it is possible to load a range of rows, i.e. sequential rows. My
> question is -
> how would it be most efficient to load N non-sequential rows?
> 
> Currently I am using get(Get) method N times.
>