You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Hudson (JIRA)" <ji...@apache.org> on 2015/03/16 22:46:38 UTC

[jira] [Commented] (AMBARI-10095) HiveServer2 does not install/start when running w/o sudo

    [ https://issues.apache.org/jira/browse/AMBARI-10095?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14364022#comment-14364022 ] 

Hudson commented on AMBARI-10095:
---------------------------------

FAILURE: Integrated in Ambari-trunk-Commit #2050 (See [https://builds.apache.org/job/Ambari-trunk-Commit/2050/])
AMBARI-10095. HiveServer2 does not install/start when running w/o sudo (aonishuk) (aonishuk: http://git-wip-us.apache.org/repos/asf?p=ambari.git&a=commit&h=fae1eb7f385eb1924fdd98152d58bde53c5d0c1b)
* ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py


> HiveServer2 does not install/start when running w/o sudo
> --------------------------------------------------------
>
>                 Key: AMBARI-10095
>                 URL: https://issues.apache.org/jira/browse/AMBARI-10095
>             Project: Ambari
>          Issue Type: Bug
>            Reporter: Andrew Onischuk
>            Assignee: Andrew Onischuk
>             Fix For: 2.0.0
>
>
> 0) No sudo is available in env  
> 1) HDP 2.2 cluster with hdfs, yarn, tez, zk, ams  
> 2) enable kerb  
> 3) Via Add Service, add Hive. You get the following error:
>     
>     
>     
>     2015-03-15 23:45:04,329 - Error while executing command 'start':
>     Traceback (most recent call last):
>       File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 208, in execute
>         method(env)
>       File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server.py", line 61, in start
>         rolling_restart=rolling_restart )
>       File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py", line 44, in hive_service
>         check_fs_root()
>       File "/var/lib/ambari-agent/cache/common-services/HIVE/0.12.0.2.0/package/scripts/hive_service.py", line 147, in check_fs_root
>         user=params.hive_user)
>       File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, in __init__
>         self.env.run()
>       File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 152, in run
>         self.run_action(resource, action)
>       File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 118, in run_action
>         provider_action()
>       File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 276, in action_run
>         raise ex
>     Fail: Execution of 'hive --config /etc/hive/conf.server --service metatool -updateLocation hdfs://hdp2poc/apps/hive/warehouse su: warning: cannot change directory to /home/hive: No such file or directory
>     hdfs://hdp2poc/apps/hive/warehouse
>     ' returned 127. su: warning: cannot change directory to /home/hive: No such file or directory
>     Initializing HiveMetaTool..
>     Exception in thread "main" java.lang.IllegalArgumentException: java.net.URISyntaxException: Expected scheme-specific part at index 3: su:
>             at org.apache.hadoop.fs.Path.initialize(Path.java:206)
>             at org.apache.hadoop.fs.Path.<init>(Path.java:172)
>             at org.apache.hadoop.hive.metastore.tools.HiveMetaTool.main(HiveMetaTool.java:423)
>             at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>             at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>             at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>             at java.lang.reflect.Method.invoke(Method.java:606)
>             at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>             at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>     Caused by: java.net.URISyntaxException: Expected scheme-specific part at index 3: su:
>             at java.net.URI$Parser.fail(URI.java:2829)
>             at java.net.URI$Parser.failExpecting(URI.java:2835)
>             at java.net.URI$Parser.parse(URI.java:3038)
>             at java.net.URI.<init>(URI.java:753)
>             at org.apache.hadoop.fs.Path.initialize(Path.java:203)
>             ... 8 more
>     -bash: line 1: hdfs://hdp2poc/apps/hive/warehouse
>     : No such file or directory
>     



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)