You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Antonenko Alexander (JIRA)" <ji...@apache.org> on 2015/01/24 14:15:34 UTC

[jira] [Created] (AMBARI-9319) HBase fails to start after adding HBase Service to a cluster that has NameNode HA already enabled

Antonenko Alexander created AMBARI-9319:
-------------------------------------------

             Summary: HBase fails to start after adding HBase Service to a cluster that has NameNode HA already enabled
                 Key: AMBARI-9319
                 URL: https://issues.apache.org/jira/browse/AMBARI-9319
             Project: Ambari
          Issue Type: Bug
          Components: ambari-web
    Affects Versions: 2.0.0
            Reporter: Antonenko Alexander
            Assignee: Antonenko Alexander
            Priority: Critical
             Fix For: 2.0.0


HBase fails to start when enabling HA in a 3-node cluster with Ambari 2.0.0 
(build 339) and HDP 2.2.1.0-2165

STR:
Install Ambari 2.0.0 with default settings
Install HDP 2.2.1.0 on a single node with just HDFS and ZK
Add 2 more nodes with ZK servers on all 3 nodes
Enable HA (service name is "ha")
Add HBase service

{code}
2014-12-30 19:29:40,270 - Error while executing command 'start':
Traceback (most recent call last):
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 142, in execute
    method(env)
  File "/var/lib/ambari-agent/cache/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py", line 48, in start
    self.configure(env) # for security
  File "/var/lib/ambari-agent/cache/common-services/HBASE/0.96.0.2.0/package/scripts/hbase_master.py", line 38, in configure
    hbase(name='master')
  File "/var/lib/ambari-agent/cache/common-services/HBASE/0.96.0.2.0/package/scripts/hbase.py", line 150, in hbase
    params.HdfsResource(None, action="execute")
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, in __init__
    self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 151, in run
    self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 117, in run_action
    provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 105, in action_execute
    logoutput=logoutput,
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, in __init__
    self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 151, in run
    self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 117, in run_action
    provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 265, in action_run
    raise ex
Fail: Execution of 'hadoop --config /etc/hadoop/conf jar /var/lib/ambari-agent/lib/fast-hdfs-resource.jar /var/lib/ambari-agent/data/hdfs_resources.json hdfs://ha' returned 1. Creating: Resource [source=null, target=hdfs://c6404.ambari.apache.org,c6405.ambari.apache.org:8020/apps/hbase/data, type=directory, action=create, owner=hbase, group=null, mode=null, recursiveChown=false, recursiveChmod=false]
Exception in thread "main" java.lang.IllegalArgumentException: Wrong FS: hdfs://c6404.ambari.apache.org,c6405.ambari.apache.org:8020/apps/hbase/data, expected: hdfs://ha
	at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:645)
	at org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:193)
	at org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:105)
	at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
	at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
	at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
	at org.apache.hadoop.fs.FileSystem.isFile(FileSystem.java:1426)
	at org.apache.ambari.fast_hdfs_resource.Resource.checkResourceParameters(Resource.java:152)
	at org.apache.ambari.fast_hdfs_resource.Runner.main(Runner.java:72)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
	at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
{code}




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)