You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hbase.apache.org by "Andrew Kyle Purtell (Jira)" <ji...@apache.org> on 2022/06/15 21:07:00 UTC
[jira] [Resolved] (HBASE-7402) java.io.IOException: Got error in response to OP_READ_BLOCK
[ https://issues.apache.org/jira/browse/HBASE-7402?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Andrew Kyle Purtell resolved HBASE-7402.
----------------------------------------
Resolution: Invalid
> java.io.IOException: Got error in response to OP_READ_BLOCK
> -----------------------------------------------------------
>
> Key: HBASE-7402
> URL: https://issues.apache.org/jira/browse/HBASE-7402
> Project: HBase
> Issue Type: Bug
> Components: HFile
> Affects Versions: 0.90.4, 0.94.0
> Reporter: samar
> Priority: Major
>
> Getting this error on our hbase version 0.90.4-cdh3u3
> 2012-12-18 02:35:39,082 WARN org.apache.hadoop.hdfs.DFSClient: Failed to connect to /x.x.x.x:xxxxx for file /hbase/table_x/37bea13d03ed9fa611941cc4aad6e8c2/scores/7355825801969613604 for block 3174705353677971357:java.io.IOException: Got error in response to OP_READ_BLOCK self=/x.x.x.x, remote=/x.x.x.x:xxxx for file /hbase/table_x/37bea13d03ed9fa611941cc4aad6e8c2/scores/7355825801969613604 for block 3174705353677971357_1028665
> at org.apache.hadoop.hdfs.DFSClient$RemoteBlockReader.newBlockReader(DFSClient.java:1673)
> at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.getBlockReader(DFSClient.java:2383)
> at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.fetchBlockByteRange(DFSClient.java:2272)
> at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:2438)
> at org.apache.hadoop.fs.FSDataInputStream.read(FSDataInputStream.java:46)
> at org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream.read(BoundedRangeFileInputStream.java:101)
> at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
> at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
> at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:141)
> at org.apache.hadoop.hbase.io.hfile.HFile$Reader.decompress(HFile.java:1094)
> at org.apache.hadoop.hbase.io.hfile.HFile$Reader.readBlock(HFile.java:1036)
> at org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.loadBlock(HFile.java:1446)
> at org.apache.hadoop.hbase.io.hfile.HFile$Reader$Scanner.seekTo(HFile.java:1303)
> at org.apache.hadoop.hbase.regionserver.StoreFileScanner.seekAtOrAfter(StoreFileScanner.java:136)
> at org.apache.hadoop.hbase.regionserver.StoreFileScanner.seek(StoreFileScanner.java:96)
> at org.apache.hadoop.hbase.regionserver.StoreScanner.<init>(StoreScanner.java:77)
> at org.apache.hadoop.hbase.regionserver.Store.getScanner(Store.java:1405)
> at org.apache.hadoop.hbase.regionserver.HRegion$RegionScanner.<init>(HRegion.java:2467)
> at org.apache.hadoop.hbase.regionserver.HRegion.instantiateInternalScanner(HRegion.java:1192)
> at org.apache.hadoop.hbase.regionserver.HRegion.getScanner(HRegion.java:1184)
> at org.apache.hadoop.hbase.regionserver.HRegion.getScanner(HRegion.java:1168)
> at org.apache.hadoop.hbase.regionserver.HRegion.get(HRegion.java:3215)
> this causes the HBase RS to hang and hence stops responding.
> In NameNode the block was delete before.. ( as per the timestamp)
> 2012-12-18 02:25:19,027 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* ask x.x.x.x:xxxxx to delete blk_3174705353677971357_1028665 blk_-9072685530813588257_1028824
> 2012-12-18 02:25:19,027 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* ask x.x.x.x:xxxxx to delete blk_5651962510569886604_1028711
> 2012-12-18 02:25:22,027 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* ask x.x.x.x:xxxxx to delete blk_3174705353677971357_1028665
> Looks like org.apache.hadoop.hbase.io.hfile.BoundedRangeFileInputStream is cacheing the block location and causing this issue
--
This message was sent by Atlassian Jira
(v8.20.7#820007)