You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hbase.apache.org by "Y. SREENIVASULU REDDY (Jira)" <ji...@apache.org> on 2019/09/05 06:40:00 UTC

[jira] [Updated] (HBASE-22649) FileNotFoundException shown in UI when tried to access HFILE URL of a column family name have special char (e.g #)

     [ https://issues.apache.org/jira/browse/HBASE-22649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Y. SREENIVASULU REDDY updated HBASE-22649:
------------------------------------------
    Attachment: HBASE-22649.branch-2_002.patch

> FileNotFoundException shown in UI when tried to access HFILE URL of a column family name have special char (e.g #)
> ------------------------------------------------------------------------------------------------------------------
>
>                 Key: HBASE-22649
>                 URL: https://issues.apache.org/jira/browse/HBASE-22649
>             Project: HBase
>          Issue Type: Bug
>          Components: UI
>    Affects Versions: 3.0.0, 2.1.5, 1.3.5
>            Reporter: Ashok shetty
>            Assignee: Y. SREENIVASULU REDDY
>            Priority: Major
>             Fix For: 3.0.0, 1.3.6, 2.1.7
>
>         Attachments: HBASE-22649.branch-1.002.patch, HBASE-22649.branch-1.003.patch, HBASE-22649.branch-1.patch, HBASE-22649.branch-2.patch, HBASE-22649.branch-2_002.patch, HBASE-22649.patch
>
>
> 【Test step】:
> 1. create 'specialchar' ,'#'
> 2.put 'specialchar','r1','#:cq','1000'
> 3.flush 'specialchar'
> 4.put 'specialchar','r2','#:cq','1000'
> 5.flush 'specialchar'
>  
> Once hfile is created, click the hfile link in UI.
> The following error is throwing.
> {noformat}
> java.io.FileNotFoundException: Path is not a file: /hbase/data/default/specialchar/df9d19830c562c4eeb3f8b396211d52d
>  at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:90)
>  at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:76)
>  at org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:153)
>  at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1942)
>  at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:739)
>  at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:432)
>  at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>  at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
>  at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:878)
>  at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:824)
>  at java.security.AccessController.doPrivileged(Native Method)
>  at javax.security.auth.Subject.doAs(Subject.java:422)
>  at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2684)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)