You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@zeppelin.apache.org by "lim se yoon (Jira)" <ji...@apache.org> on 2020/09/26 14:03:00 UTC

[jira] [Created] (ZEPPELIN-5072) zeppelin on kubernetes hive connection bug (zeppelin0.9.0-preview2)

lim se yoon created ZEPPELIN-5072:
-------------------------------------

             Summary: zeppelin on kubernetes hive connection bug (zeppelin0.9.0-preview2)
                 Key: ZEPPELIN-5072
                 URL: https://issues.apache.org/jira/browse/ZEPPELIN-5072
             Project: Zeppelin
          Issue Type: Bug
          Components: Interpreters
         Environment: zeppelin0.9.0-preview2

 

hadoop 2.6 (kerberos)

hive 1.1.0 (kerberos)
            Reporter: lim se yoon
         Attachments: zeppelin--56e7561bff11_not_running_on_kubernetes.log, zeppelin--zeppelin-server_running_on_kubernetes.log, zeppelin-interpreter-hive-shared_process--56e7561bff11_not_running_on_kubernetes.log, zeppelin-interpreter-hive-shared_process--jdbc-kppvww_running_on_kubernetes.log

Hi, I'm having a trouble when I'm connecting to the hive 1.1.0 with zeppelin0.9.0-preview2

 

KERBEROS authentication is required for connecting my hive.

 

First of all, there was a problem with the code below to connect with hive 1.1.0.
 jdbc/src/main/java/org/apache/zeppelin/jdbc/JDBCInterpreter.java
 lines (749 to 753)
{code:java}
if (getJDBCConfiguration).getPropertyMap (dbPrefix)getProperty (URL_KEY)
    .startsWith ("jdbc:hive2://") {
  HiveUtils.startHiveMonitorThread(statement, context,
      Boolean.parseBolean ("hive.log.display", "true"));
}{code}
 

 

Due to this code, there was a problem with the connection with hive 1.1.0, but I annotated and built the code to create a new docker file, and when I executed the server through bin/zeppelin.sh in the docker image(not running kubernetes just docker local env), it was connected with kerberos hive well.

 

*However*, there is a problem when operating the image as *running kubernetes mode*.

There was a problem when running on running kubernetes just like the docker environment, and I wanted to find the problem through the debugging mode.

The suspicious part is that running kubernetes says the following logs:

'sun.nio.ch.EPollSelectorImp' was able to confirm that no further updates were made.

 

here is the zeppelin--*server.log

*running on kubernetes mode*
{code:java}
DEBUG [2020-09-26 10:42:30,150] ({SchedulerFactory2} RemoteInterpreterUtils.java[checkIfRemoteEndpointAccessible]:127) - Remote endpoint 'jdbc-kppvww.default.svc:12321' is not accessible (might be initializing): jdbc-kppvww.default.svc
DEBUG [2020-09-26 10:42:31,151] ({SchedulerFactory2} RemoteInterpreterUtils.java[checkIfRemoteEndpointAccessible]:127) - Remote endpoint 'jdbc-kppvww.default.svc:12321' is not accessible (might be initializing): jdbc-kppvww.default.svc
DEBUG [2020-09-26 10:42:32,151] ({SchedulerFactory2} RemoteInterpreterUtils.java[checkIfRemoteEndpointAccessible]:127) - Remote endpoint 'jdbc-kppvww.default.svc:12321' is not accessible (might be initializing): jdbc-kppvww.default.svc
DEBUG [2020-09-26 10:42:33,152] ({SchedulerFactory2} RemoteInterpreterUtils.java[checkIfRemoteEndpointAccessible]:127) - Remote endpoint 'jdbc-kppvww.default.svc:12321' is not accessible (might be initializing): jdbc-kppvww.default.svc
{code}
 

Zeppelin-interpreter/src/main/java/org/apache/zeppelin/interpreter/remote/RemoteInterpreterUtils.java

A problem occurred with the checkIfRemoteEnpointAccessibe function of above java code and I don't think it's updated in 'sun.nio.ch.EPollSelectorImp'.

 

*not running on kubernetes mode (local mode (same docker image))*
{code:java}
DEBUG [2020-09-26 10:30:57,418] ({qtp1412925683-13} QueuedThreadPool.java[run]:940) - ran CEP:SocketChannelEndPoint@2754cfd7{/172.17.0.1:60076<->/172.17.0.2:8080,OPEN,fill=FI,flush=-,to=0/300000}{io=0/1,kio=0,kro=1}->WebSocketServerConnection@69f97a84[s=ConnectionState@75a8b010[OPENED],f=org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection$Flusher@5756b226[IDLE][queueSize=0,aggregateSize=-1,terminated=null],g=Generator[SERVER,validating,+rsv1],p=Parser@311d31c8[ExtensionStack,s=START,c=0,len=8,f=null]]:runFillable:BLOCKING in QueuedThreadPool[qtp1412925683]@543788f3{STARTED,8<=8<=400,i=1,r=8,q=0}[ReservedThreadExecutor@53fe15ff{s=1/8,p=0}]DEBUG [2020-09-26 10:30:57,418] ({qtp1412925683-18} ManagedSelector.java[select]:476) - Selector sun.nio.ch.EPollSelectorImpl@615ea420 woken with none selected
DEBUG [2020-09-26 10:30:57,418] ({qtp1412925683-18} ManagedSelector.java[select]:485) - Selector sun.nio.ch.EPollSelectorImpl@615ea420 woken up from select, 0/0/1 selected
DEBUG [2020-09-26 10:30:57,418] ({qtp1412925683-18} ManagedSelector.java[select]:498) - Selector sun.nio.ch.EPollSelectorImpl@615ea420 processing 0 keys, 1 update{code}
 

'sun.nio.ch.EPollSelectorImpl' shows that the update continues.

*As a result*, there seems to be a difference between interpreter_process.log in the docker not running kubernetes environment and running kubernetes environment.

 

And Here is the zeppelin-interpreter-*_process-jdbc.log

*running on kubernetes mode* 
{code:java}
ERROR [2020-09-26 10:42:40,800] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[open]:225) - zeppelin will be ignored. driver.zeppelin and zeppelin.url is mandatory.DEBUG [2020-09-26 10:42:40,800] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[open]:235) - JDBC PropertiesMap: {default={url=jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@*.HADOOP, completer.schemaFilters=, user=PERSONAL_INFO@PERSONAL_INFO.HADOOP, statementPrecode=, splitQueries=true, proxy.user.property=hive.server2.proxy.user, password=, driver=org.apache.hive.jdbc.HiveDriver, completer.ttlInSeconds=120, precode=}, common={max_count=1000}}DEBUG [2020-09-26 10:42:40,802] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:212) - key: zeppelin.jdbc.maxRows, value: 1000DEBUG [2020-09-26 10:42:40,805] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: default.precode, value: DEBUG [2020-09-26 10:42:40,807] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: common.precode, value: nullDEBUG [2020-09-26 10:42:40,813] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[jobRun]:775) - Script after hooks: show databases;DEBUG [2020-09-26 10:42:40,814] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:212) - key: zeppelin.jdbc.interpolation, value: falseDEBUG [2020-09-26 10:42:40,816] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[internalInterpret]:877) - Run SQL command 'show databases;'DEBUG [2020-09-26 10:42:40,816] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[internalInterpret]:879) - DBPrefix: default, SQL command: 'show databases;'DEBUG [2020-09-26 10:42:40,819] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.auth.type, value: KERBEROSDEBUG [2020-09-26 10:42:41,124] ({ParallelScheduler-Worker-1} AbstractScheduler.java[runJob]:143) - Job Error, paragraph_1601116938240_457009787, %text java.lang.NoClassDefFoundError: Could not initialize class org.apache.hadoop.security.UserGroupInformation  at org.apache.zeppelin.jdbc.security.JDBCSecurityImpl.createSecureConfiguration(JDBCSecurityImpl.java:48)  at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:512)  at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:706)  at org.apache.zeppelin.jdbc.JDBCInterpreter.internalInterpret(JDBCInterpreter.java:881)  at org.apache.zeppelin.interpreter.AbstractInterpreter.interpret(AbstractInterpreter.java:47)  at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:110)  at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:776)  at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)  at org.apache.zeppelin.scheduler.Job.run(Job.java:172)  at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)  at org.apache.zeppelin.scheduler.ParallelScheduler.lambda$runJobInScheduler$0(ParallelScheduler.java:39)  at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)  at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)  at java.lang.Thread.run(Thread.java:748)
 INFO [2020-09-26 10:42:41,126] ({ParallelScheduler-Worker-1} AbstractScheduler.java[runJob]:152) - Job paragraph_1601116938240_457009787 finished by scheduler org.apache.zeppelin.jdbc.JDBCInterpreter176877981DEBUG [2020-09-26 10:42:41,287] ({pool-2-thread-3} RemoteInterpreterServer.java[resourcePoolGetAll]:1112) - Request resourcePoolGetAll from ZeppelinServer
{code}
 

*not running on kubernetes mode (local mode (same docker image))*
{code:java}
ERROR [2020-09-26 10:30:58,754] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[open]:225) - zeppelin will be ignored. driver.zeppelin and zeppelin.url is mandatory.ERROR [2020-09-26 10:30:58,754] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[open]:225) - zeppelin will be ignored. driver.zeppelin and zeppelin.url is mandatory.DEBUG [2020-09-26 10:30:58,754] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[open]:235) - JDBC PropertiesMap: {default={url=jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOP, completer.schemaFilters=, user=PERSONAL_INFO@PERSONAL_INFO.HADOOP, statementPrecode=, splitQueries=true, proxy.user.property=hive.server2.proxy.user, password=, driver=org.apache.hive.jdbc.HiveDriver, completer.ttlInSeconds=120, precode=}, common={max_count=1000}}DEBUG [2020-09-26 10:30:58,755] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:212) - key: zeppelin.jdbc.maxRows, value: 1000DEBUG [2020-09-26 10:30:58,757] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: default.precode, value: DEBUG [2020-09-26 10:30:58,758] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: common.precode, value: nullDEBUG [2020-09-26 10:30:58,760] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[jobRun]:775) - Script after hooks: show databases;DEBUG [2020-09-26 10:30:58,761] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:212) - key: zeppelin.jdbc.interpolation, value: falseDEBUG [2020-09-26 10:30:58,761] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[internalInterpret]:877) - Run SQL command 'show databases;'DEBUG [2020-09-26 10:30:58,761] ({ParallelScheduler-Worker-1} JDBCInterpreter.java[internalInterpret]:879) - DBPrefix: default, SQL command: 'show databases;'DEBUG [2020-09-26 10:30:58,763] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.auth.type, value: KERBEROSDEBUG [2020-09-26 10:30:58,764] ({pool-2-thread-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.concurrent.use, value: trueDEBUG [2020-09-26 10:30:58,764] ({pool-2-thread-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.concurrent.max_connection, value: 10DEBUG [2020-09-26 10:30:58,781] ({pool-3-thread-1} MutableMetricsFactory.java[newForField]:42) - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of successful kerberos logins and latency (milliseconds)])DEBUG [2020-09-26 10:30:58,793] ({pool-3-thread-1} MutableMetricsFactory.java[newForField]:42) - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of failed kerberos logins and latency (milliseconds)])DEBUG [2020-09-26 10:30:58,793] ({pool-3-thread-1} MutableMetricsFactory.java[newForField]:42) - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[GetGroups])DEBUG [2020-09-26 10:30:58,795] ({pool-3-thread-1} MetricsSystemImpl.java[register]:231) - UgiMetrics, User and group related metricsDEBUG [2020-09-26 10:30:58,984] ({pool-3-thread-1} Groups.java[getUserToGroupsMappingService]:278) -  Creating new Groups objectDEBUG [2020-09-26 10:30:58,987] ({pool-3-thread-1} NativeCodeLoader.java[<clinit>]:46) - Trying to load the custom-built native-hadoop library...DEBUG [2020-09-26 10:30:58,987] ({pool-3-thread-1} NativeCodeLoader.java[<clinit>]:55) - Failed to load native-hadoop with error: java.lang.UnsatisfiedLinkError: no hadoop in java.library.pathDEBUG [2020-09-26 10:30:58,988] ({pool-3-thread-1} NativeCodeLoader.java[<clinit>]:56) - java.library.path=/usr/java/packages/lib/amd64:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib WARN [2020-09-26 10:30:58,988] ({pool-3-thread-1} NativeCodeLoader.java[<clinit>]:62) - Unable to load native-hadoop library for your platform... using builtin-java classes where applicableDEBUG [2020-09-26 10:30:58,988] ({pool-3-thread-1} JniBasedUnixGroupsMappingWithFallback.java[<init>]:41) - Falling back to shell basedDEBUG [2020-09-26 10:30:58,989] ({pool-3-thread-1} JniBasedUnixGroupsMappingWithFallback.java[<init>]:45) - Group mapping impl=org.apache.hadoop.security.ShellBasedUnixGroupsMappingDEBUG [2020-09-26 10:30:59,052] ({pool-3-thread-1} Shell.java[checkHadoopHome]:320) - Failed to detect a valid hadoop home directoryjava.io.IOException: HADOOP_HOME or hadoop.home.dir are not set. at org.apache.hadoop.util.Shell.checkHadoopHome(Shell.java:302) at org.apache.hadoop.util.Shell.<clinit>(Shell.java:327) at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:79) at org.apache.hadoop.security.Groups.parseStaticMapping(Groups.java:104) at org.apache.hadoop.security.Groups.<init>(Groups.java:86) at org.apache.hadoop.security.Groups.<init>(Groups.java:66) at org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:280) at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:271) at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:248) at org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:763) at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:748) at org.apache.hadoop.security.UserGroupInformation.isLoginKeytabBased(UserGroupInformation.java:1142) at org.apache.zeppelin.jdbc.JDBCInterpreter.runKerberosLogin(JDBCInterpreter.java:187) at org.apache.zeppelin.interpreter.KerberosInterpreter$1.call(KerberosInterpreter.java:135) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748)DEBUG [2020-09-26 10:30:59,065] ({pool-3-thread-1} Shell.java[isSetsidSupported]:396) - setsid exited with exit code 0DEBUG [2020-09-26 10:30:59,065] ({pool-3-thread-1} Groups.java[<init>]:91) - Group mapping impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback; cacheTimeout=300000; warningDeltaMs=5000DEBUG [2020-09-26 10:30:59,070] ({pool-3-thread-1} UserGroupInformation.java[login]:209) - hadoop loginDEBUG [2020-09-26 10:30:59,071] ({pool-3-thread-1} UserGroupInformation.java[commit]:144) - hadoop login commitDEBUG [2020-09-26 10:30:59,074] ({pool-3-thread-1} UserGroupInformation.java[commit]:174) - using local user:UnixPrincipal: rootDEBUG [2020-09-26 10:30:59,074] ({pool-3-thread-1} UserGroupInformation.java[commit]:180) - Using user: "UnixPrincipal: root" with name rootDEBUG [2020-09-26 10:30:59,074] ({pool-3-thread-1} UserGroupInformation.java[commit]:190) - User entry: "root"DEBUG [2020-09-26 10:30:59,075] ({pool-3-thread-1} UserGroupInformation.java[loginUserFromSubject]:799) - UGI loginUser:root (auth:SIMPLE) INFO [2020-09-26 10:30:59,075] ({pool-3-thread-1} KerberosInterpreter.java[call]:143) - runKerberosLogin failed for 1 time(s).DEBUG [2020-09-26 10:31:00,661] ({ParallelScheduler-Worker-1} UserGroupInformation.java[login]:209) - hadoop loginDEBUG [2020-09-26 10:31:00,663] ({ParallelScheduler-Worker-1} UserGroupInformation.java[commit]:144) - hadoop login commitDEBUG [2020-09-26 10:31:00,663] ({ParallelScheduler-Worker-1} UserGroupInformation.java[commit]:158) - using kerberos user:PERSONAL_INFO@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:00,663] ({ParallelScheduler-Worker-1} UserGroupInformation.java[commit]:180) - Using user: "PERSONAL_INFO@PERSONAL_INFO.HADOOP" with name PERSONAL_INFO@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:00,663] ({ParallelScheduler-Worker-1} UserGroupInformation.java[commit]:190) - User entry: "PERSONAL_INFO@PERSONAL_INFO.HADOOP" INFO [2020-09-26 10:31:00,664] ({ParallelScheduler-Worker-1} UserGroupInformation.java[loginUserFromKeytab]:938) - Login successful for user PERSONAL_INFO@PERSONAL_INFO.HADOOP using keytab file /zeppelin/PERSONAL_INFO.PERSONAL_INFO.HADOOP.keytabDEBUG [2020-09-26 10:31:00,664] ({pool-3-thread-1} UserGroupInformation.java[reloginFromTicketCache]:1054) - Initiating logout for PERSONAL_INFO@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:00,664] ({pool-3-thread-1} UserGroupInformation.java[logout]:217) - hadoop logoutDEBUG [2020-09-26 10:31:00,664] ({pool-3-thread-1} UserGroupInformation.java[reloginFromTicketCache]:1066) - Initiating re-login for PERSONAL_INFO@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:00,666] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.auth.kerberos.proxy.enable, value: nullDEBUG [2020-09-26 10:31:00,666] ({pool-3-thread-1} UserGroupInformation.java[login]:209) - hadoop loginDEBUG [2020-09-26 10:31:00,667] ({pool-3-thread-1} UserGroupInformation.java[commit]:144) - hadoop login commitDEBUG [2020-09-26 10:31:00,667] ({pool-3-thread-1} UserGroupInformation.java[commit]:149) - using existing subject:[PERSONAL_INFO@PERSONAL_INFO.HADOOP, UnixPrincipal: root, UnixNumericUserPrincipal: 0, UnixNumericGroupPrincipal [Primary Group]: 0] INFO [2020-09-26 10:31:00,667] ({pool-3-thread-1} KerberosInterpreter.java[call]:136) - Ran runKerberosLogin command successfully.DEBUG [2020-09-26 10:31:00,708] ({ParallelScheduler-Worker-1} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.maxConnLifetime, value: -1 INFO [2020-09-26 10:31:00,740] ({ParallelScheduler-Worker-1} Utils.java[parseURL]:285) - Supplied authorities: PERSONAL_INFO.io:10000 INFO [2020-09-26 10:31:00,740] ({ParallelScheduler-Worker-1} Utils.java[parseURL]:372) - Resolved authority: PERSONAL_INFO.io:10000DEBUG [2020-09-26 10:31:00,753] ({ParallelScheduler-Worker-1} HadoopThriftAuthBridge.java[loginUserHasCurrentAuthMethod]:155) - Current authMethod = KERBEROSDEBUG [2020-09-26 10:31:00,754] ({ParallelScheduler-Worker-1} HadoopThriftAuthBridge.java[createClientWithConf]:90) - Not setting UGI conf as passed-in authMethod of kerberos = current. INFO [2020-09-26 10:31:00,774] ({ParallelScheduler-Worker-1} HiveConnection.java[openTransport]:189) - Will try to open client transport with JDBC Uri: jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:00,775] ({ParallelScheduler-Worker-1} UserGroupInformation.java[logPrivilegedAction]:1652) - PrivilegedAction as:PERSONAL_INFO@PERSONAL_INFO.HADOOP (auth:KERBEROS) from:org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49)DEBUG [2020-09-26 10:31:00,776] ({ParallelScheduler-Worker-1} TSaslTransport.java[open]:261) - opening transport org.apache.thrift.transport.TSaslClientTransport@c2d98ERROR [2020-09-26 10:31:00,897] ({ParallelScheduler-Worker-1} TSaslTransport.java[open]:315) - SASL negotiation failurejavax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)] at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:211) at org.apache.thrift.transport.TSaslClientTransport.handleSaslStartMessage(TSaslClientTransport.java:94) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:271) at org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49) at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:190) at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:163) at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:208) at org.apache.commons.dbcp2.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:79) at org.apache.commons.dbcp2.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:205) at org.apache.commons.pool2.impl.GenericObjectPool.create(GenericObjectPool.java:836) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:434) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:361) at org.apache.commons.dbcp2.PoolingDriver.connect(PoolingDriver.java:129) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnectionFromPool(JDBCInterpreter.java:487) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:520) at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:706) at org.apache.zeppelin.jdbc.JDBCInterpreter.internalInterpret(JDBCInterpreter.java:881) at org.apache.zeppelin.interpreter.AbstractInterpreter.interpret(AbstractInterpreter.java:47) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:110) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:776) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668) at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130) at org.apache.zeppelin.scheduler.ParallelScheduler.lambda$runJobInScheduler$0(ParallelScheduler.java:39) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748)Caused by: GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt) at sun.security.jgss.krb5.Krb5InitCredential.getInstance(Krb5InitCredential.java:162) at sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement(Krb5MechFactory.java:122) at sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext(Krb5MechFactory.java:189) at sun.security.jgss.GSSManagerImpl.getMechanismContext(GSSManagerImpl.java:224) at sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:212) at sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:179) at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:192) ... 36 moreDEBUG [2020-09-26 10:31:00,900] ({ParallelScheduler-Worker-1} TSaslTransport.java[sendSaslMessage]:162) - CLIENT: Writing message with status BAD and payload length 19 INFO [2020-09-26 10:31:00,901] ({ParallelScheduler-Worker-1} HiveConnection.java[openTransport]:194) - Could not open client transport with JDBC Uri: jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:00,905] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onUpdate]:957) - Output Update for index 0: DEBUG [2020-09-26 10:31:00,930] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append: java.sql.SQLException: Could not open client transport with JDBC Uri: jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOP: GSS initiate failed
DEBUG [2020-09-26 10:31:00,953] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:215)
DEBUG [2020-09-26 10:31:00,953] ({pool-2-thread-2} RemoteInterpreterServer.java[resourcePoolGetAll]:1112) - Request resourcePoolGetAll from ZeppelinServerDEBUG [2020-09-26 10:31:00,955] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:163)
DEBUG [2020-09-26 10:31:00,956] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105)
DEBUG [2020-09-26 10:31:00,956] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.sql.DriverManager.getConnection(DriverManager.java:664)
DEBUG [2020-09-26 10:31:00,957] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.sql.DriverManager.getConnection(DriverManager.java:208)
DEBUG [2020-09-26 10:31:00,959] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.commons.dbcp2.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:79)
DEBUG [2020-09-26 10:31:00,960] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.commons.dbcp2.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:205)
DEBUG [2020-09-26 10:31:00,960] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.commons.pool2.impl.GenericObjectPool.create(GenericObjectPool.java:836)
DEBUG [2020-09-26 10:31:00,961] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:434)
DEBUG [2020-09-26 10:31:00,961] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:361)
DEBUG [2020-09-26 10:31:00,962] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.commons.dbcp2.PoolingDriver.connect(PoolingDriver.java:129)
DEBUG [2020-09-26 10:31:00,963] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.sql.DriverManager.getConnection(DriverManager.java:664)
DEBUG [2020-09-26 10:31:00,964] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.sql.DriverManager.getConnection(DriverManager.java:270)
DEBUG [2020-09-26 10:31:00,964] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnectionFromPool(JDBCInterpreter.java:487)
DEBUG [2020-09-26 10:31:00,965] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:520)
DEBUG [2020-09-26 10:31:00,966] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:706)
DEBUG [2020-09-26 10:31:00,967] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.jdbc.JDBCInterpreter.internalInterpret(JDBCInterpreter.java:881)
DEBUG [2020-09-26 10:31:00,968] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.interpreter.AbstractInterpreter.interpret(AbstractInterpreter.java:47)
DEBUG [2020-09-26 10:31:00,968] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:110)
DEBUG [2020-09-26 10:31:00,969] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:776)
DEBUG [2020-09-26 10:31:00,969] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
DEBUG [2020-09-26 10:31:00,970] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
DEBUG [2020-09-26 10:31:00,970] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
DEBUG [2020-09-26 10:31:00,971] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.zeppelin.scheduler.ParallelScheduler.lambda$runJobInScheduler$0(ParallelScheduler.java:39)
DEBUG [2020-09-26 10:31:00,971] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
DEBUG [2020-09-26 10:31:00,971] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
DEBUG [2020-09-26 10:31:00,972] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.lang.Thread.run(Thread.java:748)
DEBUG [2020-09-26 10:31:00,972] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append: Caused by: org.apache.thrift.transport.TTransportException: GSS initiate failed
DEBUG [2020-09-26 10:31:00,973] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232)
DEBUG [2020-09-26 10:31:00,973] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:316)
DEBUG [2020-09-26 10:31:00,974] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37)
DEBUG [2020-09-26 10:31:00,974] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52)
DEBUG [2020-09-26 10:31:00,975] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49)
DEBUG [2020-09-26 10:31:00,975] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at java.security.AccessController.doPrivileged(Native Method)
DEBUG [2020-09-26 10:31:00,975] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at javax.security.auth.Subject.doAs(Subject.java:422)
DEBUG [2020-09-26 10:31:00,976] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
DEBUG [2020-09-26 10:31:00,976] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49)
DEBUG [2020-09-26 10:31:00,977] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:190)
DEBUG [2020-09-26 10:31:00,977] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[onAppend]:947) - Output Append:  ... 26 more
DEBUG [2020-09-26 10:31:00,980] ({ParallelScheduler-Worker-1} RemoteInterpreterServer.java[jobRun]:795) - InterpreterResultMessage: %text java.sql.SQLException: Could not open client transport with JDBC Uri: jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOP: GSS initiate failed at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:215) at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:163) at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:208) at org.apache.commons.dbcp2.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:79) at org.apache.commons.dbcp2.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:205) at org.apache.commons.pool2.impl.GenericObjectPool.create(GenericObjectPool.java:836) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:434) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:361) at org.apache.commons.dbcp2.PoolingDriver.connect(PoolingDriver.java:129) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnectionFromPool(JDBCInterpreter.java:487) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:520) at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:706) at org.apache.zeppelin.jdbc.JDBCInterpreter.internalInterpret(JDBCInterpreter.java:881) at org.apache.zeppelin.interpreter.AbstractInterpreter.interpret(AbstractInterpreter.java:47) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:110) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:776) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668) at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130) at org.apache.zeppelin.scheduler.ParallelScheduler.lambda$runJobInScheduler$0(ParallelScheduler.java:39) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748)Caused by: org.apache.thrift.transport.TTransportException: GSS initiate failed at org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:316) at org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49) at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:190) ... 26 more
DEBUG [2020-09-26 10:31:00,981] ({ParallelScheduler-Worker-1} AbstractScheduler.java[runJob]:143) - Job Error, paragraph_1601116251411_1220021722, %text java.sql.SQLException: Could not open client transport with JDBC Uri: jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOP: GSS initiate failed at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:215) at org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:163) at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:105) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:208) at org.apache.commons.dbcp2.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:79) at org.apache.commons.dbcp2.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:205) at org.apache.commons.pool2.impl.GenericObjectPool.create(GenericObjectPool.java:836) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:434) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:361) at org.apache.commons.dbcp2.PoolingDriver.connect(PoolingDriver.java:129) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnectionFromPool(JDBCInterpreter.java:487) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:520) at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:706) at org.apache.zeppelin.jdbc.JDBCInterpreter.internalInterpret(JDBCInterpreter.java:881) at org.apache.zeppelin.interpreter.AbstractInterpreter.interpret(AbstractInterpreter.java:47) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:110) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:776) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668) at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130) at org.apache.zeppelin.scheduler.ParallelScheduler.lambda$runJobInScheduler$0(ParallelScheduler.java:39) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748)Caused by: org.apache.thrift.transport.TTransportException: GSS initiate failed at org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:316) at org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49) at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:190) ... 26 more
 INFO [2020-09-26 10:31:00,981] ({ParallelScheduler-Worker-1} AbstractScheduler.java[runJob]:152) - Job paragraph_1601116251411_1220021722 finished by scheduler org.apache.zeppelin.jdbc.JDBCInterpreter2082164389DEBUG [2020-09-26 10:31:01,027] ({pool-2-thread-1} RemoteInterpreterServer.java[resourcePoolGetAll]:1112) - Request resourcePoolGetAll from ZeppelinServerDEBUG [2020-09-26 10:31:02,483] ({pool-2-thread-1} RemoteInterpreterServer.java[interpret]:593) - st:show databases;DEBUG [2020-09-26 10:31:02,484] ({pool-2-thread-1} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.concurrent.use, value: trueDEBUG [2020-09-26 10:31:02,485] ({pool-2-thread-1} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.concurrent.max_connection, value: 10 INFO [2020-09-26 10:31:02,498] ({ParallelScheduler-Worker-2} AbstractScheduler.java[runJob]:125) - Job paragraph_1601116251411_1220021722 started by scheduler org.apache.zeppelin.jdbc.JDBCInterpreter2082164389DEBUG [2020-09-26 10:31:02,509] ({ParallelScheduler-Worker-2} RemoteInterpreterServer.java[jobRun]:775) - Script after hooks: show databases;DEBUG [2020-09-26 10:31:02,509] ({ParallelScheduler-Worker-2} Interpreter.java[getProperty]:212) - key: zeppelin.jdbc.interpolation, value: falseDEBUG [2020-09-26 10:31:02,511] ({ParallelScheduler-Worker-2} JDBCInterpreter.java[internalInterpret]:877) - Run SQL command 'show databases;'DEBUG [2020-09-26 10:31:02,511] ({ParallelScheduler-Worker-2} JDBCInterpreter.java[internalInterpret]:879) - DBPrefix: default, SQL command: 'show databases;'DEBUG [2020-09-26 10:31:02,511] ({ParallelScheduler-Worker-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.auth.type, value: KERBEROSDEBUG [2020-09-26 10:31:02,583] ({pool-2-thread-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.concurrent.use, value: trueDEBUG [2020-09-26 10:31:02,584] ({pool-2-thread-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.concurrent.max_connection, value: 10DEBUG [2020-09-26 10:31:02,724] ({ParallelScheduler-Worker-2} UserGroupInformation.java[login]:209) - hadoop loginDEBUG [2020-09-26 10:31:02,725] ({ParallelScheduler-Worker-2} UserGroupInformation.java[commit]:144) - hadoop login commitDEBUG [2020-09-26 10:31:02,725] ({ParallelScheduler-Worker-2} UserGroupInformation.java[commit]:158) - using kerberos user:PERSONAL_INFO@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:02,725] ({ParallelScheduler-Worker-2} UserGroupInformation.java[commit]:180) - Using user: "PERSONAL_INFO@PERSONAL_INFO.HADOOP" with name PERSONAL_INFO@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:02,725] ({ParallelScheduler-Worker-2} UserGroupInformation.java[commit]:190) - User entry: "PERSONAL_INFO@PERSONAL_INFO.HADOOP" INFO [2020-09-26 10:31:02,726] ({ParallelScheduler-Worker-2} UserGroupInformation.java[loginUserFromKeytab]:938) - Login successful for user PERSONAL_INFO@PERSONAL_INFO.HADOOP using keytab file /zeppelin/PERSONAL_INFO.PERSONAL_INFO.HADOOP.keytabDEBUG [2020-09-26 10:31:02,726] ({ParallelScheduler-Worker-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.auth.kerberos.proxy.enable, value: nullDEBUG [2020-09-26 10:31:02,727] ({ParallelScheduler-Worker-2} Interpreter.java[getProperty]:205) - key: zeppelin.jdbc.maxConnLifetime, value: -1 INFO [2020-09-26 10:31:02,728] ({ParallelScheduler-Worker-2} Utils.java[parseURL]:285) - Supplied authorities: PERSONAL_INFO.io:10000 INFO [2020-09-26 10:31:02,729] ({ParallelScheduler-Worker-2} Utils.java[parseURL]:372) - Resolved authority: PERSONAL_INFO.io:10000DEBUG [2020-09-26 10:31:02,729] ({ParallelScheduler-Worker-2} HadoopThriftAuthBridge.java[loginUserHasCurrentAuthMethod]:155) - Current authMethod = KERBEROSDEBUG [2020-09-26 10:31:02,729] ({ParallelScheduler-Worker-2} HadoopThriftAuthBridge.java[createClientWithConf]:90) - Not setting UGI conf as passed-in authMethod of kerberos = current. INFO [2020-09-26 10:31:02,730] ({ParallelScheduler-Worker-2} HiveConnection.java[openTransport]:189) - Will try to open client transport with JDBC Uri: jdbc:hive2://PERSONAL_INFO.io:10000/default;principal=hive/PERSONAL_INFO.io@PERSONAL_INFO.HADOOPDEBUG [2020-09-26 10:31:02,730] ({ParallelScheduler-Worker-2} UserGroupInformation.java[logPrivilegedAction]:1652) - PrivilegedAction as:PERSONAL_INFO@PERSONAL_INFO.HADOOP (auth:KERBEROS) from:org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49)DEBUG [2020-09-26 10:31:02,730] ({ParallelScheduler-Worker-2} TSaslTransport.java[open]:261) - opening transport org.apache.thrift.transport.TSaslClientTransport@11d7cb45DEBUG [2020-09-26 10:31:05,746] ({ParallelScheduler-Worker-2} TSaslClientTransport.java[handleSaslStartMessage]:96) - Sending mechanism name GSSAPI and initial response of length 585DEBUG [2020-09-26 10:31:05,746] ({ParallelScheduler-Worker-2} TSaslTransport.java[sendSaslMessage]:162) - CLIENT: Writing message with status START and payload length 6DEBUG [2020-09-26 10:31:05,746] ({ParallelScheduler-Worker-2} TSaslTransport.java[sendSaslMessage]:162) - CLIENT: Writing message with status OK and payload length 585DEBUG [2020-09-26 10:31:05,746] ({ParallelScheduler-Worker-2} TSaslTransport.java[open]:273) - CLIENT: Start message handledDEBUG [2020-09-26 10:31:05,931] ({ParallelScheduler-Worker-2} TSaslTransport.java[receiveSaslMessage]:206) - CLIENT: Received message with status OK and payload length 108DEBUG [2020-09-26 10:31:05,934] ({ParallelScheduler-Worker-2} TSaslTransport.java[sendSaslMessage]:162) - CLIENT: Writing message with status OK and payload length 0DEBUG [2020-09-26 10:31:06,112] ({ParallelScheduler-Worker-2} TSaslTransport.java[receiveSaslMessage]:206) - CLIENT: Received message with status OK and payload length 32DEBUG [2020-09-26 10:31:06,114] ({ParallelScheduler-Worker-2} TSaslTransport.java[sendSaslMessage]:162) - CLIENT: Writing message with status COMPLETE and payload length 32DEBUG [2020-09-26 10:31:06,115] ({ParallelScheduler-Worker-2} TSaslTransport.java[open]:296) - CLIENT: Main negotiation loop completeDEBUG [2020-09-26 10:31:06,115] ({ParallelScheduler-Worker-2} TSaslTransport.java[open]:306) - CLIENT: SASL Client receiving last messageDEBUG [2020-09-26 10:31:06,225] ({ParallelScheduler-Worker-2} TSaslTransport.java[receiveSaslMessage]:206) - CLIENT: Received message with status COMPLETE and payload length 0DEBUG [2020-09-26 10:31:06,244] ({ParallelScheduler-Worker-2} TSaslTransport.java[flush]:498) - writing data length: 71DEBUG [2020-09-26 10:31:06,619] ({ParallelScheduler-Worker-2} TSaslTransport.java[readFrame]:459) - CLIENT: reading data length: 109DEBUG [2020-09-26 10:31:06,643] ({ParallelScheduler-Worker-2} Interpreter.java[getProperty]:205) - key: default.statementPrecode, value: DEBUG [2020-09-26 10:31:06,649] ({ParallelScheduler-Worker-2} TSaslTransport.java[flush]:498) - writing data length: 121DEBUG [2020-09-26 10:31:06,719] ({ParallelScheduler-Worker-2} TSaslTransport.java[readFrame]:459) - CLIENT: reading data length: 109DEBUG [2020-09-26 10:31:06,733] ({ParallelScheduler-Worker-2} TSaslTransport.java[flush]:498) - writing data length: 100DEBUG [2020-09-26 10:31:06,776] ({ParallelScheduler-Worker-2} TSaslTransport.java[readFrame]:459) - CLIENT: reading data length: 321DEBUG [2020-09-26 10:31:06,790] ({ParallelScheduler-Worker-2} TSaslTransport.java[flush]:498) - writing data length: 102DEBUG [2020-09-26 10:31:06,925] ({ParallelScheduler-Worker-2} TSaslTransport.java[readFrame]:459) - CLIENT: reading data length: 136DEBUG [2020-09-26 10:31:06,963] ({ParallelScheduler-Worker-2} TSaslTransport.java[flush]:498) - writing data length: 112DEBUG [2020-09-26 10:31:07,121] ({ParallelScheduler-Worker-2} TSaslTransport.java[readFrame]:459) - CLIENT: reading data length: 3283DEBUG [2020-09-26 10:31:07,149] ({ParallelScheduler-Worker-2} HiveQueryResultSet.java[next]:381) - Fetched row string:
{code}
 

At the not running on kubernetes mode, I finally found out that the main negotiation loop was successfully completed at TSaslTransport.java and received the QueryResultSet from the hive.

 

I'd appreciate your help. Thank you for reading it.

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)