You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Antonenko Alexander (JIRA)" <ji...@apache.org> on 2015/01/24 14:16:34 UTC
[jira] [Updated] (AMBARI-9319) HBase fails to start after adding
HBase Service to a cluster that has NameNode HA already enabled
[ https://issues.apache.org/jira/browse/AMBARI-9319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Antonenko Alexander updated AMBARI-9319:
----------------------------------------
Attachment: hbase_error.txt
hbase_output.txt
hbase-site.xml
> HBase fails to start after adding HBase Service to a cluster that has NameNode HA already enabled
> -------------------------------------------------------------------------------------------------
>
> Key: AMBARI-9319
> URL: https://issues.apache.org/jira/browse/AMBARI-9319
> Project: Ambari
> Issue Type: Bug
> Components: ambari-web
> Affects Versions: 2.0.0
> Reporter: Antonenko Alexander
> Assignee: Antonenko Alexander
> Priority: Critical
> Fix For: 2.0.0
>
> Attachments: hbase-site.xml, hbase_error.txt, hbase_output.txt
>
>
> HBase fails to start when enabling HA in a 3-node cluster with Ambari 2.0.0
> (build 339) and HDP 2.2.1.0-2165
> STR:
> Install Ambari 2.0.0 with default settings
> Install HDP 2.2.1.0 on a single node with just HDFS and ZK
> Add 2 more nodes with ZK servers on all 3 nodes
> Enable HA (service name is "ha")
> Add HBase service
> {code}
> 2014-12-30 19:29:40,270 - Error while executing command 'start':
> Traceback (most recent call last):
> File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 142, in execute
> method(env)
> File "/var/lib/ambari-agent/cache/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py", line 48, in start
> self.configure(env) # for security
> File "/var/lib/ambari-agent/cache/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py", line 38, in configure
> hbase(name='master')
> File "/var/lib/ambari-agent/cache/common-services/HBASE/0.96.0.2.0/package/scripts/hbase.py", line 150, in hbase
> params.HdfsResource(None, action="execute")
> File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, in __init__
> self.env.run()
> File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 151, in run
> self.run_action(resource, action)
> File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 117, in run_action
> provider_action()
> File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 105, in action_execute
> logoutput=logoutput,
> File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, in __init__
> self.env.run()
> File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 151, in run
> self.run_action(resource, action)
> File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 117, in run_action
> provider_action()
> File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 265, in action_run
> raise ex
> Fail: Execution of 'hadoop --config /etc/hadoop/conf jar /var/lib/ambari-agent/lib/fast-hdfs-resource.jar /var/lib/ambari-agent/data/hdfs_resources.json hdfs://ha' returned 1. Creating: Resource [source=null, target=hdfs://c6404.ambari.apache.org,c6405.ambari.apache.org:8020/apps/hbase/data, type=directory, action=create, owner=hbase, group=null, mode=null, recursiveChown=false, recursiveChmod=false]
> Exception in thread "main" java.lang.IllegalArgumentException: Wrong FS: hdfs://c6404.ambari.apache.org,c6405.ambari.apache.org:8020/apps/hbase/data, expected: hdfs://ha
> at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:645)
> at org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:193)
> at org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:105)
> at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
> at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
> at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
> at org.apache.hadoop.fs.FileSystem.isFile(FileSystem.java:1426)
> at org.apache.ambari.fast_hdfs_resource.Resource.checkResourceParameters(Resource.java:152)
> at org.apache.ambari.fast_hdfs_resource.Runner.main(Runner.java:72)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)