You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@impala.apache.org by Philip Zeyliger <ph...@cloudera.com> on 2018/03/27 21:50:12 UTC

Updating your minicluster to Hadoop 3.0...

Hi folks,

I just sent off https://gerrit.cloudera.org/#/c/9743/ and
https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which changes
the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry 2.0,
and so on. This change *will not* be back-ported to 2.x.

When you pull that change in, you'll need to re-build your minicluster
with, e.g., ./buildall.sh -testdata -format -notests. This will pull in the
new dependencies, format your cluster, and load up all the data. As you
know, it takes 1-2 hours.

If you want to hold off, you can elso set export
IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.

Note that this choice between versions happens at build time, and CMake
depends on it. So, switching back and forth requires re-running CMake.

Please let me know if you run into any trouble. This is a big enough that
there may be some bumps on the road.

-- Philip

Re:Re:Re: Re: Updating your minicluster to Hadoop 3.0...

Posted by Quanlong Huang <hu...@126.com>.
It's hard for me to change username due to some company-specific security issues. I finally found a workaround for my problem.

The source of the impersonation invocation is in ThriftCLIService#getSessionHandle
    if (cliService.getHiveConf().getBoolVar(ConfVars.HIVE_SERVER2_ENABLE_DOAS) &&
        (userName != null)) {
      ...
      sessionHandle = cliService.openSessionWithImpersonation(...);
    } else {
      sessionHandle = cliService.openSession(...);
    }


So setting hive.server2.enable.doAs to false in hive-site.xml can avoid this invocatino. I modified my hive-site.xml templates in fe/src/test/resources for this and HiveServer2 launched successfully.
Just filed a JIRA for this workaround in case someone has an odd username like me: https://issues.apache.org/jira/browse/IMPALA-6789


Thanks,
Quanlong


在 2018-04-01 22:51:53,"Quanlong Huang" <hu...@126.com> 写道:
>Thanks, Philip!
>Since it's still unfixed, let me try to use another username.
>
>
>在 2018-04-01 12:35:42,"Philip Zeyliger" <ph...@cloudera.com> 写道:
>
>Hi Quanlong,
>
>
>You're running into https://issues.apache.org/jira/browse/HADOOP-7050. I found that JIRA via https://kb.informatica.com/solution/23/Pages/61/510035.aspx (and Google).
>
>
>What's surprising to me is that this code has been around in Hadoop for a long time. I think HiveServer2 must have changed in that it now invokes the impersonation code, whereas it must have not before.
>
>
>Here are some pointers:
>
>
>
>testdata/cluster/node_templates/common/etc/hadoop/conf/core-site.xml.tmpl:    <name>hadoop.proxyuser.${USER}.hosts</name>
>
>
>
>https://github.com/apache/hadoop/blob/dc8e3432013153ac11d31d6b462aa96f8ca2c443/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/DefaultImpersonationProvider.java#L78
>
>
>
>    String usersGroupsRegEx = prefixRegEx + "[^.]*(" +
>        Pattern.quote(CONF_USERS) + "|" + Pattern.quote(CONF_GROUPS) + ")";
>
>
>To my eye, it looks like that regular expression is wrong and is what's disallowing usernames with periods/dots in them. (Once it's loosened, any code that's parsing these may also need to be fixed.)
>
>
>
>
>-- Philip
>
>
> 
>
>
>On Fri, Mar 30, 2018 at 7:34 PM, Quanlong Huang <hu...@126.com> wrote:
>
>I failed to start the minicluster too but encountered another errors. HiveServer2 failed to launch and kept warning:
>
>
>2018-03-30T18:54:05,526  WARN [HiveServer2-Handler-Pool: Thread-49] thrift.ThriftCLIService: Error opening session:
>java.lang.RuntimeException: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: quanlong.huang is not allowed to impersonate foo
>        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:89) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_121]
>        at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_121]
>        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1962) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at com.sun.proxy.$Proxy37.open(Unknown Source) ~[?:?]
>        at org.apache.hive.service.cli.session.SessionManager.createSession(SessionManager.java:411) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.session.SessionManager.openSession(SessionManager.java:363) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.CLIService.openSessionWithImpersonation(CLIService.java:189) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.thrift.ThriftCLIService.getSessionHandle(ThriftCLIService.java:423) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:312) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.rpc.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1377) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.rpc.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1362) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_121]
>        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_121]
>        at java.lang.Thread.run(Thread.java:745) [?:1.8.0_121]
>Caused by: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: quanlong.huang is not allowed to impersonate foo
>        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:596) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
>        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
>        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
>        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
>        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        ... 21 more
>Caused by: org.apache.hadoop.ipc.RemoteException: User: quanlong.huang is not allowed to impersonate foo
>        at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1491) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.ipc.Client.call(Client.java:1437) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.ipc.Client.call(Client.java:1347) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at com.sun.proxy.$Proxy31.getFileInfo(Unknown Source) ~[?:?]
>        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:875) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
>        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
>        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
>        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
>        at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at com.sun.proxy.$Proxy32.getFileInfo(Unknown Source) ~[?:?]
>        at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1643) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
>        at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1494) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
>        at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1491) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
>        at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1506) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
>        at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1668) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>        at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:701) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:640) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:572) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
>        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
>        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
>        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
>        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>        ... 21 more
>
>
>I just rebase the codes and run ./buildall.sh -format
>Any thoughts?
>
>
>Thanks,
>Quanlong
>
>At 2018-03-30 13:08:33, "Dimitris Tsirogiannis" <dt...@cloudera.com> wrote:
>>I enabled full logging in my postgres that hosts the sentry and metastore
>>db and I don't see the table being created. If anyone has gone through the
>>process, can you: a) verify that relation SENTRY_ROLE exists in your
>>sentry_policy db, and b) tell me how many relations are in your policy_db.
>>
>>Thanks
>>Dimitris
>>
>>On Thu, Mar 29, 2018 at 9:32 PM, Dimitris Tsirogiannis <
>>dtsirogiannis@cloudera.com> wrote:
>>
>>> Good point. I used -format that in theory handles both the metastore and
>>> the sentry policy dB. The sentry_policy db is created and has some tables
>>> but not the SENTRY_ROLE.
>>>
>>> Dimitris
>>>
>>> On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:
>>>
>>>> I think I might have once fixed that using
>>>>
>>>> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
>>>>
>>>>
>>>> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
>>>> dtsirogiannis@cloudera.com> wrote:
>>>>
>>>> > I tried rebuilding my minicluster but Sentry refuses to start. I get
>>>> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
>>>> that
>>>> > ring any bells?
>>>> >
>>>> > Thanks
>>>> > Dimitris
>>>> >
>>>> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
>>>> > wrote:
>>>> >
>>>> > > Hi folks,
>>>> > >
>>>> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
>>>> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
>>>> changes
>>>> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
>>>> 2.0,
>>>> > > and so on. This change *will not* be back-ported to 2.x.
>>>> > >
>>>> > > When you pull that change in, you'll need to re-build your minicluster
>>>> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull
>>>> in
>>>> > the
>>>> > > new dependencies, format your cluster, and load up all the data. As
>>>> you
>>>> > > know, it takes 1-2 hours.
>>>> > >
>>>> > > If you want to hold off, you can elso set export
>>>> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
>>>> > >
>>>> > > Note that this choice between versions happens at build time, and
>>>> CMake
>>>> > > depends on it. So, switching back and forth requires re-running CMake.
>>>> > >
>>>> > > Please let me know if you run into any trouble. This is a big enough
>>>> that
>>>> > > there may be some bumps on the road.
>>>> > >
>>>> > > -- Philip
>>>> > >
>>>> >
>>>>
>>>
>
>

Re:Re: Re: Updating your minicluster to Hadoop 3.0...

Posted by Quanlong Huang <hu...@126.com>.
Thanks, Philip!
Since it's still unfixed, let me try to use another username.


在 2018-04-01 12:35:42,"Philip Zeyliger" <ph...@cloudera.com> 写道:

Hi Quanlong,


You're running into https://issues.apache.org/jira/browse/HADOOP-7050. I found that JIRA via https://kb.informatica.com/solution/23/Pages/61/510035.aspx (and Google).


What's surprising to me is that this code has been around in Hadoop for a long time. I think HiveServer2 must have changed in that it now invokes the impersonation code, whereas it must have not before.


Here are some pointers:



testdata/cluster/node_templates/common/etc/hadoop/conf/core-site.xml.tmpl:    <name>hadoop.proxyuser.${USER}.hosts</name>



https://github.com/apache/hadoop/blob/dc8e3432013153ac11d31d6b462aa96f8ca2c443/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/DefaultImpersonationProvider.java#L78



    String usersGroupsRegEx = prefixRegEx + "[^.]*(" +
        Pattern.quote(CONF_USERS) + "|" + Pattern.quote(CONF_GROUPS) + ")";


To my eye, it looks like that regular expression is wrong and is what's disallowing usernames with periods/dots in them. (Once it's loosened, any code that's parsing these may also need to be fixed.)




-- Philip


 


On Fri, Mar 30, 2018 at 7:34 PM, Quanlong Huang <hu...@126.com> wrote:

I failed to start the minicluster too but encountered another errors. HiveServer2 failed to launch and kept warning:


2018-03-30T18:54:05,526  WARN [HiveServer2-Handler-Pool: Thread-49] thrift.ThriftCLIService: Error opening session:
java.lang.RuntimeException: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: quanlong.huang is not allowed to impersonate foo
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:89) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_121]
        at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_121]
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1962) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at com.sun.proxy.$Proxy37.open(Unknown Source) ~[?:?]
        at org.apache.hive.service.cli.session.SessionManager.createSession(SessionManager.java:411) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.SessionManager.openSession(SessionManager.java:363) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.CLIService.openSessionWithImpersonation(CLIService.java:189) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.thrift.ThriftCLIService.getSessionHandle(ThriftCLIService.java:423) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:312) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.rpc.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1377) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.rpc.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1362) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_121]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [?:1.8.0_121]
Caused by: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: quanlong.huang is not allowed to impersonate foo
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:596) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        ... 21 more
Caused by: org.apache.hadoop.ipc.RemoteException: User: quanlong.huang is not allowed to impersonate foo
        at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1491) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.Client.call(Client.java:1437) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.Client.call(Client.java:1347) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at com.sun.proxy.$Proxy31.getFileInfo(Unknown Source) ~[?:?]
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:875) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
        at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at com.sun.proxy.$Proxy32.getFileInfo(Unknown Source) ~[?:?]
        at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1643) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1494) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1491) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1506) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1668) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:701) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:640) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:572) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        ... 21 more


I just rebase the codes and run ./buildall.sh -format
Any thoughts?


Thanks,
Quanlong

At 2018-03-30 13:08:33, "Dimitris Tsirogiannis" <dt...@cloudera.com> wrote:
>I enabled full logging in my postgres that hosts the sentry and metastore
>db and I don't see the table being created. If anyone has gone through the
>process, can you: a) verify that relation SENTRY_ROLE exists in your
>sentry_policy db, and b) tell me how many relations are in your policy_db.
>
>Thanks
>Dimitris
>
>On Thu, Mar 29, 2018 at 9:32 PM, Dimitris Tsirogiannis <
>dtsirogiannis@cloudera.com> wrote:
>
>> Good point. I used -format that in theory handles both the metastore and
>> the sentry policy dB. The sentry_policy db is created and has some tables
>> but not the SENTRY_ROLE.
>>
>> Dimitris
>>
>> On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:
>>
>>> I think I might have once fixed that using
>>>
>>> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
>>>
>>>
>>> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
>>> dtsirogiannis@cloudera.com> wrote:
>>>
>>> > I tried rebuilding my minicluster but Sentry refuses to start. I get
>>> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
>>> that
>>> > ring any bells?
>>> >
>>> > Thanks
>>> > Dimitris
>>> >
>>> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
>>> > wrote:
>>> >
>>> > > Hi folks,
>>> > >
>>> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
>>> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
>>> changes
>>> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
>>> 2.0,
>>> > > and so on. This change *will not* be back-ported to 2.x.
>>> > >
>>> > > When you pull that change in, you'll need to re-build your minicluster
>>> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull
>>> in
>>> > the
>>> > > new dependencies, format your cluster, and load up all the data. As
>>> you
>>> > > know, it takes 1-2 hours.
>>> > >
>>> > > If you want to hold off, you can elso set export
>>> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
>>> > >
>>> > > Note that this choice between versions happens at build time, and
>>> CMake
>>> > > depends on it. So, switching back and forth requires re-running CMake.
>>> > >
>>> > > Please let me know if you run into any trouble. This is a big enough
>>> that
>>> > > there may be some bumps on the road.
>>> > >
>>> > > -- Philip
>>> > >
>>> >
>>>
>>



Re: Re: Updating your minicluster to Hadoop 3.0...

Posted by Philip Zeyliger <ph...@cloudera.com>.
Hi Quanlong,

You're running into https://issues.apache.org/jira/browse/HADOOP-7050. I
found that JIRA via
https://kb.informatica.com/solution/23/Pages/61/510035.aspx (and Google).

What's surprising to me is that this code has been around in Hadoop for a
long time. I think HiveServer2 must have changed in that it now invokes the
impersonation code, whereas it must have not before.

Here are some pointers:

testdata/cluster/node_templates/common/etc/hadoop/conf/core-site.xml.tmpl:
  <name>hadoop.proxyuser.${USER}.hosts</name>

https://github.com/apache/hadoop/blob/dc8e3432013153ac11d31d6b462aa96f8ca2c443/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/DefaultImpersonationProvider.java#L78

    String usersGroupsRegEx = prefixRegEx + "[^.]*(" +
        Pattern.quote(CONF_USERS) + "|" + Pattern.quote(CONF_GROUPS) + ")";

To my eye, it looks like that regular expression is wrong and is what's
disallowing usernames with periods/dots in them. (Once it's loosened, any
code that's parsing these may also need to be fixed.)


-- Philip



On Fri, Mar 30, 2018 at 7:34 PM, Quanlong Huang <hu...@126.com>
wrote:

> I failed to start the minicluster too but encountered another errors.
> HiveServer2 failed to launch and kept warning:
>
> 2018-03-30T18:54:05,526  WARN [HiveServer2-Handler-Pool: Thread-49]
> thrift.ThriftCLIService: Error opening session:
> java.lang.RuntimeException: java.lang.RuntimeException:
> org.apache.hadoop.ipc.RemoteException(org.apache.
> hadoop.security.authorize.AuthorizationException): User: quanlong.huang
> is not allowed to impersonate foo
>         at org.apache.hive.service.cli.session.HiveSessionProxy.
> invoke(HiveSessionProxy.java:89) ~[hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.session.HiveSessionProxy.
> access$000(HiveSessionProxy.java:36) ~[hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
> ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at java.security.AccessController.doPrivileged(Native Method)
> ~[?:1.8.0_121]
>         at javax.security.auth.Subject.doAs(Subject.java:422)
> ~[?:1.8.0_121]
>         at org.apache.hadoop.security.UserGroupInformation.doAs(
> UserGroupInformation.java:1962) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at org.apache.hive.service.cli.session.HiveSessionProxy.
> invoke(HiveSessionProxy.java:59) ~[hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at com.sun.proxy.$Proxy37.open(Unknown Source) ~[?:?]
>         at org.apache.hive.service.cli.session.SessionManager.
> createSession(SessionManager.java:411) [hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.session.SessionManager.
> openSession(SessionManager.java:363) [hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.CLIService.
> openSessionWithImpersonation(CLIService.java:189)
> [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.thrift.ThriftCLIService.
> getSessionHandle(ThriftCLIService.java:423) [hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.thrift.ThriftCLIService.
> OpenSession(ThriftCLIService.java:312) [hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.rpc.thrift.TCLIService$Processor$
> OpenSession.getResult(TCLIService.java:1377) [hive-exec-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.rpc.thrift.TCLIService$Processor$
> OpenSession.getResult(TCLIService.java:1362) [hive-exec-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
> [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
> [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.auth.TSetIpAddressProcessor.process(
> TSetIpAddressProcessor.java:56) [hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286)
> [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [?:1.8.0_121]
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [?:1.8.0_121]
>         at java.lang.Thread.run(Thread.java:745) [?:1.8.0_121]
> Caused by: java.lang.RuntimeException: org.apache.hadoop.ipc.
> RemoteException(org.apache.hadoop.security.authorize.AuthorizationException):
> User: quanlong.huang is not allowed to impersonate foo
>         at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:596)
> ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539)
> ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169)
> ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> ~[?:1.8.0_121]
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
>         at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
>         at org.apache.hive.service.cli.session.HiveSessionProxy.
> invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         ... 21 more
> Caused by: org.apache.hadoop.ipc.RemoteException: User: quanlong.huang is
> not allowed to impersonate foo
>         at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1491)
> ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>         at org.apache.hadoop.ipc.Client.call(Client.java:1437)
> ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>         at org.apache.hadoop.ipc.Client.call(Client.java:1347)
> ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.
> invoke(ProtobufRpcEngine.java:228) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.
> invoke(ProtobufRpcEngine.java:116) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at com.sun.proxy.$Proxy31.getFileInfo(Unknown Source) ~[?:?]
>         at org.apache.hadoop.hdfs.protocolPB.
> ClientNamenodeProtocolTranslatorPB.getFileInfo(
> ClientNamenodeProtocolTranslatorPB.java:875) ~[hadoop-hdfs-client-3.0.0-
> cdh6.x-20180302.192732-2.jar:?]
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> ~[?:1.8.0_121]
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
>         at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
>         at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(
> RetryInvocationHandler.java:422) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.
> invokeMethod(RetryInvocationHandler.java:165)
> ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>         at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.
> invoke(RetryInvocationHandler.java:157) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.
> invokeOnce(RetryInvocationHandler.java:95) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(
> RetryInvocationHandler.java:359) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at com.sun.proxy.$Proxy32.getFileInfo(Unknown Source) ~[?:?]
>         at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1643)
> ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
>         at org.apache.hadoop.hdfs.DistributedFileSystem$29.
> doCall(DistributedFileSystem.java:1494) ~[hadoop-hdfs-client-3.0.0-
> cdh6.x-20180302.192732-2.jar:?]
>         at org.apache.hadoop.hdfs.DistributedFileSystem$29.
> doCall(DistributedFileSystem.java:1491) ~[hadoop-hdfs-client-3.0.0-
> cdh6.x-20180302.192732-2.jar:?]
>         at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(
> FileSystemLinkResolver.java:81) ~[hadoop-common-3.0.0-cdh6.x-
> 20180302.191654-1.jar:?]
>         at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(
> DistributedFileSystem.java:1506) ~[hadoop-hdfs-client-3.0.0-
> cdh6.x-20180302.192732-2.jar:?]
>         at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1668)
> ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
>         at org.apache.hadoop.hive.ql.session.SessionState.
> createRootHDFSDir(SessionState.java:701) ~[hive-exec-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hadoop.hive.ql.session.SessionState.
> createSessionDirs(SessionState.java:640) ~[hive-exec-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:572)
> ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539)
> ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169)
> ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> ~[?:1.8.0_121]
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
>         at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
>         at org.apache.hive.service.cli.session.HiveSessionProxy.
> invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-
> SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
>         ... 21 more
>
> I just rebase the codes and run ./buildall.sh -format
> Any thoughts?
>
> Thanks,
> Quanlong
>
>
> At 2018-03-30 13:08:33, "Dimitris Tsirogiannis" <dt...@cloudera.com> wrote:
> >I enabled full logging in my postgres that hosts the sentry and metastore
> >db and I don't see the table being created. If anyone has gone through the
> >process, can you: a) verify that relation SENTRY_ROLE exists in your
> >sentry_policy db, and b) tell me how many relations are in your policy_db.
> >
> >Thanks
> >Dimitris
> >
> >On Thu, Mar 29, 2018 at 9:32 PM, Dimitris Tsirogiannis <
> >dtsirogiannis@cloudera.com> wrote:
> >
> >> Good point. I used -format that in theory handles both the metastore and
> >> the sentry policy dB. The sentry_policy db is created and has some tables
> >> but not the SENTRY_ROLE.
> >>
> >> Dimitris
> >>
> >> On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:
> >>
> >>> I think I might have once fixed that using
> >>>
> >>> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
> >>>
> >>>
> >>> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
> >>> dtsirogiannis@cloudera.com> wrote:
> >>>
> >>> > I tried rebuilding my minicluster but Sentry refuses to start. I get
> >>> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
> >>> that
> >>> > ring any bells?
> >>> >
> >>> > Thanks
> >>> > Dimitris
> >>> >
> >>> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
> >>> > wrote:
> >>> >
> >>> > > Hi folks,
> >>> > >
> >>> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
> >>> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
> >>> changes
> >>> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
> >>> 2.0,
> >>> > > and so on. This change *will not* be back-ported to 2.x.
> >>> > >
> >>> > > When you pull that change in, you'll need to re-build your minicluster
> >>> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull
> >>> in
> >>> > the
> >>> > > new dependencies, format your cluster, and load up all the data. As
> >>> you
> >>> > > know, it takes 1-2 hours.
> >>> > >
> >>> > > If you want to hold off, you can elso set export
> >>> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
> >>> > >
> >>> > > Note that this choice between versions happens at build time, and
> >>> CMake
> >>> > > depends on it. So, switching back and forth requires re-running CMake.
> >>> > >
> >>> > > Please let me know if you run into any trouble. This is a big enough
> >>> that
> >>> > > there may be some bumps on the road.
> >>> > >
> >>> > > -- Philip
> >>> > >
> >>> >
> >>>
> >>
>
>

Re:Re: Updating your minicluster to Hadoop 3.0...

Posted by Quanlong Huang <hu...@126.com>.
I failed to start the minicluster too but encountered another errors. HiveServer2 failed to launch and kept warning:


2018-03-30T18:54:05,526  WARN [HiveServer2-Handler-Pool: Thread-49] thrift.ThriftCLIService: Error opening session:
java.lang.RuntimeException: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: quanlong.huang is not allowed to impersonate foo
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:89) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_121]
        at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_121]
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1962) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at com.sun.proxy.$Proxy37.open(Unknown Source) ~[?:?]
        at org.apache.hive.service.cli.session.SessionManager.createSession(SessionManager.java:411) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.SessionManager.openSession(SessionManager.java:363) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.CLIService.openSessionWithImpersonation(CLIService.java:189) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.thrift.ThriftCLIService.getSessionHandle(ThriftCLIService.java:423) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:312) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.rpc.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1377) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.rpc.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1362) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56) [hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286) [hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_121]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_121]
        at java.lang.Thread.run(Thread.java:745) [?:1.8.0_121]
Caused by: java.lang.RuntimeException: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.authorize.AuthorizationException): User: quanlong.huang is not allowed to impersonate foo
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:596) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        ... 21 more
Caused by: org.apache.hadoop.ipc.RemoteException: User: quanlong.huang is not allowed to impersonate foo
        at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1491) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.Client.call(Client.java:1437) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.Client.call(Client.java:1347) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at com.sun.proxy.$Proxy31.getFileInfo(Unknown Source) ~[?:?]
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:875) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
        at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at com.sun.proxy.$Proxy32.getFileInfo(Unknown Source) ~[?:?]
        at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1643) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1494) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.hdfs.DistributedFileSystem$29.doCall(DistributedFileSystem.java:1491) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1506) ~[hadoop-hdfs-client-3.0.0-cdh6.x-20180302.192732-2.jar:?]
        at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1668) ~[hadoop-common-3.0.0-cdh6.x-20180302.191654-1.jar:?]
        at org.apache.hadoop.hive.ql.session.SessionState.createRootHDFSDir(SessionState.java:701) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:640) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:572) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:539) ~[hive-exec-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at org.apache.hive.service.cli.session.HiveSessionImpl.open(HiveSessionImpl.java:169) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_121]
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_121]
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_121]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_121]
        at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) ~[hive-service-2.1.1-cdh6.x-SNAPSHOT.jar:2.1.1-cdh6.x-SNAPSHOT]
        ... 21 more


I just rebase the codes and run ./buildall.sh -format
Any thoughts?


Thanks,
Quanlong

At 2018-03-30 13:08:33, "Dimitris Tsirogiannis" <dt...@cloudera.com> wrote:
>I enabled full logging in my postgres that hosts the sentry and metastore
>db and I don't see the table being created. If anyone has gone through the
>process, can you: a) verify that relation SENTRY_ROLE exists in your
>sentry_policy db, and b) tell me how many relations are in your policy_db.
>
>Thanks
>Dimitris
>
>On Thu, Mar 29, 2018 at 9:32 PM, Dimitris Tsirogiannis <
>dtsirogiannis@cloudera.com> wrote:
>
>> Good point. I used -format that in theory handles both the metastore and
>> the sentry policy dB. The sentry_policy db is created and has some tables
>> but not the SENTRY_ROLE.
>>
>> Dimitris
>>
>> On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:
>>
>>> I think I might have once fixed that using
>>>
>>> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
>>>
>>>
>>> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
>>> dtsirogiannis@cloudera.com> wrote:
>>>
>>> > I tried rebuilding my minicluster but Sentry refuses to start. I get
>>> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
>>> that
>>> > ring any bells?
>>> >
>>> > Thanks
>>> > Dimitris
>>> >
>>> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
>>> > wrote:
>>> >
>>> > > Hi folks,
>>> > >
>>> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
>>> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
>>> changes
>>> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
>>> 2.0,
>>> > > and so on. This change *will not* be back-ported to 2.x.
>>> > >
>>> > > When you pull that change in, you'll need to re-build your minicluster
>>> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull
>>> in
>>> > the
>>> > > new dependencies, format your cluster, and load up all the data. As
>>> you
>>> > > know, it takes 1-2 hours.
>>> > >
>>> > > If you want to hold off, you can elso set export
>>> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
>>> > >
>>> > > Note that this choice between versions happens at build time, and
>>> CMake
>>> > > depends on it. So, switching back and forth requires re-running CMake.
>>> > >
>>> > > Please let me know if you run into any trouble. This is a big enough
>>> that
>>> > > there may be some bumps on the road.
>>> > >
>>> > > -- Philip
>>> > >
>>> >
>>>
>>

Re: Updating your minicluster to Hadoop 3.0...

Posted by Philip Zeyliger <ph...@cloudera.com>.
Hi Dimitris,

Here's a listing from my machine after I ran "./buildall.sh
-start_minicluster -start_impala_cluster -format -testdata -notests".

-- Philip

$sudo -u postgres psql
psql (9.5.12)
Type "help" for help.

postgres=# \c sentry_policy
You are now connected to database "sentry_policy" as user "postgres".
sentry_policy=# select * from "SENTRY_ROLE";
 ROLE_ID | CREATE_TIME | ROLE_NAME
---------+-------------+-----------
(0 rows)

sentry_policy=# select * from "SENTRY_VERSION";
 VER_ID | SCHEMA_VERSION |        VERSION_COMMENT
--------+----------------+-------------------------------
      1 | 2.0.0          | Schema version set implicitly
(1 row)

sentry_policy=# select * from "SENTRY_USER";
 USER_ID | CREATE_TIME | USER_NAME
---------+-------------+-----------
(0 rows)

sentry_policy=# \dt
                    List of relations
 Schema |             Name             | Type  |  Owner
--------+------------------------------+-------+----------
 public | AUTHZ_PATH                   | table | hiveuser
 public | AUTHZ_PATHS_MAPPING          | table | hiveuser
 public | AUTHZ_PATHS_SNAPSHOT_ID      | table | hiveuser
 public | SENTRY_DB_PRIVILEGE          | table | hiveuser
 public | SENTRY_GM_PRIVILEGE          | table | hiveuser
 public | SENTRY_GROUP                 | table | hiveuser
 public | SENTRY_HMS_NOTIFICATION_ID   | table | hiveuser
 public | SENTRY_PATH_CHANGE           | table | hiveuser
 public | SENTRY_PERM_CHANGE           | table | hiveuser
 public | SENTRY_ROLE                  | table | hiveuser
 public | SENTRY_ROLE_DB_PRIVILEGE_MAP | table | hiveuser
 public | SENTRY_ROLE_GM_PRIVILEGE_MAP | table | hiveuser
 public | SENTRY_ROLE_GROUP_MAP        | table | hiveuser
 public | SENTRY_ROLE_USER_MAP         | table | hiveuser
 public | SENTRY_USER                  | table | hiveuser
 public | SENTRY_VERSION               | table | hiveuser
 public | SEQUENCE_TABLE               | table | hiveuser
(17 rows)



On Thu, Mar 29, 2018 at 10:08 PM, Dimitris Tsirogiannis <
dtsirogiannis@cloudera.com> wrote:

> I enabled full logging in my postgres that hosts the sentry and metastore
> db and I don't see the table being created. If anyone has gone through the
> process, can you: a) verify that relation SENTRY_ROLE exists in your
> sentry_policy db, and b) tell me how many relations are in your policy_db.
>
> Thanks
> Dimitris
>
> On Thu, Mar 29, 2018 at 9:32 PM, Dimitris Tsirogiannis <
> dtsirogiannis@cloudera.com> wrote:
>
> > Good point. I used -format that in theory handles both the metastore and
> > the sentry policy dB. The sentry_policy db is created and has some tables
> > but not the SENTRY_ROLE.
> >
> > Dimitris
> >
> > On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:
> >
> >> I think I might have once fixed that using
> >>
> >> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
> >>
> >>
> >> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
> >> dtsirogiannis@cloudera.com> wrote:
> >>
> >> > I tried rebuilding my minicluster but Sentry refuses to start. I get
> >> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
> >> that
> >> > ring any bells?
> >> >
> >> > Thanks
> >> > Dimitris
> >> >
> >> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <philip@cloudera.com
> >
> >> > wrote:
> >> >
> >> > > Hi folks,
> >> > >
> >> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
> >> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
> >> changes
> >> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
> >> 2.0,
> >> > > and so on. This change *will not* be back-ported to 2.x.
> >> > >
> >> > > When you pull that change in, you'll need to re-build your
> minicluster
> >> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull
> >> in
> >> > the
> >> > > new dependencies, format your cluster, and load up all the data. As
> >> you
> >> > > know, it takes 1-2 hours.
> >> > >
> >> > > If you want to hold off, you can elso set export
> >> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
> >> > >
> >> > > Note that this choice between versions happens at build time, and
> >> CMake
> >> > > depends on it. So, switching back and forth requires re-running
> CMake.
> >> > >
> >> > > Please let me know if you run into any trouble. This is a big enough
> >> that
> >> > > there may be some bumps on the road.
> >> > >
> >> > > -- Philip
> >> > >
> >> >
> >>
> >
>

Re: Updating your minicluster to Hadoop 3.0...

Posted by Dimitris Tsirogiannis <dt...@cloudera.com>.
I enabled full logging in my postgres that hosts the sentry and metastore
db and I don't see the table being created. If anyone has gone through the
process, can you: a) verify that relation SENTRY_ROLE exists in your
sentry_policy db, and b) tell me how many relations are in your policy_db.

Thanks
Dimitris

On Thu, Mar 29, 2018 at 9:32 PM, Dimitris Tsirogiannis <
dtsirogiannis@cloudera.com> wrote:

> Good point. I used -format that in theory handles both the metastore and
> the sentry policy dB. The sentry_policy db is created and has some tables
> but not the SENTRY_ROLE.
>
> Dimitris
>
> On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:
>
>> I think I might have once fixed that using
>>
>> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
>>
>>
>> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
>> dtsirogiannis@cloudera.com> wrote:
>>
>> > I tried rebuilding my minicluster but Sentry refuses to start. I get
>> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
>> that
>> > ring any bells?
>> >
>> > Thanks
>> > Dimitris
>> >
>> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
>> > wrote:
>> >
>> > > Hi folks,
>> > >
>> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
>> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
>> changes
>> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
>> 2.0,
>> > > and so on. This change *will not* be back-ported to 2.x.
>> > >
>> > > When you pull that change in, you'll need to re-build your minicluster
>> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull
>> in
>> > the
>> > > new dependencies, format your cluster, and load up all the data. As
>> you
>> > > know, it takes 1-2 hours.
>> > >
>> > > If you want to hold off, you can elso set export
>> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
>> > >
>> > > Note that this choice between versions happens at build time, and
>> CMake
>> > > depends on it. So, switching back and forth requires re-running CMake.
>> > >
>> > > Please let me know if you run into any trouble. This is a big enough
>> that
>> > > there may be some bumps on the road.
>> > >
>> > > -- Philip
>> > >
>> >
>>
>

Re: Updating your minicluster to Hadoop 3.0...

Posted by Dimitris Tsirogiannis <dt...@cloudera.com>.
Good point. I used -format that in theory handles both the metastore and
the sentry policy dB. The sentry_policy db is created and has some tables
but not the SENTRY_ROLE.

Dimitris

On Thu, Mar 29, 2018 at 6:29 PM Jim Apple <jb...@cloudera.com> wrote:

> I think I might have once fixed that using
>
> ./buildall.sh -notests -format_metastore -format_sentry_policy_db
>
>
> On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
> dtsirogiannis@cloudera.com> wrote:
>
> > I tried rebuilding my minicluster but Sentry refuses to start. I get
> > "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does
> that
> > ring any bells?
> >
> > Thanks
> > Dimitris
> >
> > On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
> > wrote:
> >
> > > Hi folks,
> > >
> > > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
> > > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which
> changes
> > > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry
> 2.0,
> > > and so on. This change *will not* be back-ported to 2.x.
> > >
> > > When you pull that change in, you'll need to re-build your minicluster
> > > with, e.g., ./buildall.sh -testdata -format -notests. This will pull in
> > the
> > > new dependencies, format your cluster, and load up all the data. As you
> > > know, it takes 1-2 hours.
> > >
> > > If you want to hold off, you can elso set export
> > > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
> > >
> > > Note that this choice between versions happens at build time, and CMake
> > > depends on it. So, switching back and forth requires re-running CMake.
> > >
> > > Please let me know if you run into any trouble. This is a big enough
> that
> > > there may be some bumps on the road.
> > >
> > > -- Philip
> > >
> >
>

Re: Updating your minicluster to Hadoop 3.0...

Posted by Jim Apple <jb...@cloudera.com>.
I think I might have once fixed that using

./buildall.sh -notests -format_metastore -format_sentry_policy_db


On Thu, Mar 29, 2018 at 6:15 PM, Dimitris Tsirogiannis <
dtsirogiannis@cloudera.com> wrote:

> I tried rebuilding my minicluster but Sentry refuses to start. I get
> "ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does that
> ring any bells?
>
> Thanks
> Dimitris
>
> On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
> wrote:
>
> > Hi folks,
> >
> > I just sent off https://gerrit.cloudera.org/#/c/9743/ and
> > https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which changes
> > the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry 2.0,
> > and so on. This change *will not* be back-ported to 2.x.
> >
> > When you pull that change in, you'll need to re-build your minicluster
> > with, e.g., ./buildall.sh -testdata -format -notests. This will pull in
> the
> > new dependencies, format your cluster, and load up all the data. As you
> > know, it takes 1-2 hours.
> >
> > If you want to hold off, you can elso set export
> > IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
> >
> > Note that this choice between versions happens at build time, and CMake
> > depends on it. So, switching back and forth requires re-running CMake.
> >
> > Please let me know if you run into any trouble. This is a big enough that
> > there may be some bumps on the road.
> >
> > -- Philip
> >
>

Re: Updating your minicluster to Hadoop 3.0...

Posted by Dimitris Tsirogiannis <dt...@cloudera.com>.
I tried rebuilding my minicluster but Sentry refuses to start. I get
"ERROR: relation "SENTRY_ROLE" does not exist in the sentry logs. Does that
ring any bells?

Thanks
Dimitris

On Tue, Mar 27, 2018 at 2:50 PM, Philip Zeyliger <ph...@cloudera.com>
wrote:

> Hi folks,
>
> I just sent off https://gerrit.cloudera.org/#/c/9743/ and
> https://issues.apache.org/jira/browse/IMPALA-4277 for GVD, which changes
> the default minicluster to be based on Hadoop 3.0, Hive 2.1, Sentry 2.0,
> and so on. This change *will not* be back-ported to 2.x.
>
> When you pull that change in, you'll need to re-build your minicluster
> with, e.g., ./buildall.sh -testdata -format -notests. This will pull in the
> new dependencies, format your cluster, and load up all the data. As you
> know, it takes 1-2 hours.
>
> If you want to hold off, you can elso set export
> IMPALA_MINICLUSTER_PROFILE_OVERRIDE=2 in your environment.
>
> Note that this choice between versions happens at build time, and CMake
> depends on it. So, switching back and forth requires re-running CMake.
>
> Please let me know if you run into any trouble. This is a big enough that
> there may be some bumps on the road.
>
> -- Philip
>