You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@zeppelin.apache.org by Pradeep Reddy <pr...@gmail.com> on 2016/09/01 01:36:53 UTC

Re: Zeppelin Kerberos error

No luck, even after copying the hive-site.xml in interpreter/spark/dep....I
also tried, downloading 0.6.1 src  vs 0.5.6. for the same steps I followed,
0.5.6 is able to talk to my hive metastore, where as the other recent
builds are not, they just show one "default" database.

Thanks,
Pradeep

On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jo...@gmail.com> wrote:

> I think it's related to https://issues.apache.org/
> jira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin
> launches interpreter. Could you please check your hive-site.xml is included
> in your interpreter process? It looks like a configuration issue because
> you can see the default database. If it doesn't exists, you should copy
> your xml into interpreter/spark/dep/
>
> Regards,
> JL
>
> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy <
> pradeepreddy.alla@gmail.com> wrote:
>
>> Hi Jongyoul- I followed the exact same steps for compiling and setting up
>> the new build from source as 0.5.6 (only difference is, I acquired the
>> source for latest build using "git clone")
>>
>> hive-site.xml was copied to conf directory. But, the spark interpreter is
>> not talking to the hive metastore. Both the 0.5.6 & the latest builds are
>> running in the same machine. In 0.5.6 when i run the below command, I see
>> 116 databases listed, as per my expectations and I'm able to run my
>> notebooks built on those databases.
>>
>> [image: Inline image 1]
>>
>> Thanks,
>> Pradeep
>>
>>
>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jo...@gmail.com> wrote:
>>
>>> Hello,
>>>
>>> Do you copy your hive-site.xml in a proper position?
>>>
>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <
>>> pradeepreddy.alla@gmail.com> wrote:
>>>
>>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds
>>>> stabilize.
>>>>
>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zj...@gmail.com> wrote:
>>>>
>>>>> Then I guess maybe you are connecting to different database. Why not
>>>>> using  'z.show(sql("databases"))' to display the databases ? Then you
>>>>> will get a hint what's going on.
>>>>>
>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>
>>>>>> Yes...I didn't wish to show the names of the databases that we have
>>>>>> in our data lake on that screen shot. so thats why I chose to display the
>>>>>> count. The latest zeppelin build just shows 1 count which is "default"
>>>>>> database.
>>>>>>
>>>>>> Thanks,
>>>>>> Pradeep
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zj...@gmail.com> wrote:
>>>>>>
>>>>>>> 116 is the databases count number. Do you expect a list of database
>>>>>>> ? then you need to use 'z.show(sql("databases"))'
>>>>>>>
>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>
>>>>>>>> Here it is Jeff
>>>>>>>>
>>>>>>>> [image: Inline image 1]
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi Pradeep,
>>>>>>>>>
>>>>>>>>> I don't see the databases on your screenshot (second one for
>>>>>>>>> 0.5.6). I think the output is correct.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin
>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk to the
>>>>>>>>>> shared hive meta store are defects in the current build.
>>>>>>>>>>
>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Jeff-
>>>>>>>>>>>
>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases &
>>>>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>>>>
>>>>>>>>>>> Zeppelin Latest Build
>>>>>>>>>>>
>>>>>>>>>>> [image: Inline image 3]
>>>>>>>>>>>
>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my
>>>>>>>>>>> databases and tables.
>>>>>>>>>>>
>>>>>>>>>>> [image: Inline image 4]
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> > the spark interpreter is not showing my tables & databases,
>>>>>>>>>>>> may be its running in an isolated mode... I'm just getting empty list, so I
>>>>>>>>>>>> attempted to do kerberos authentication to workaround that issue, and
>>>>>>>>>>>> bumped into this road block.
>>>>>>>>>>>>
>>>>>>>>>>>> kerberos would not help here, actually I think it would make
>>>>>>>>>>>> the problem more complicated.  You need to first check the log why you get
>>>>>>>>>>>> empty list.
>>>>>>>>>>>>
>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>>>>>> running kdestroy, with the below command and was able to get to my hive
>>>>>>>>>>>>> tables.
>>>>>>>>>>>>>
>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode
>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but
>>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark interpreter is not
>>>>>>>>>>>>>> showing my tables & databases, may be its running in an isolated mode...
>>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos authentication
>>>>>>>>>>>>>> to workaround that issue, and bumped into this road block.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file and
>>>>>>>>>>>>>> its working fine.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master
>>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>>>>       "properties": {
>>>>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>>>>         "args": "",
>>>>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,f
>>>>>>>>>>>>>> alse;"
>>>>>>>>>>>>>>       },
>>>>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>> rkInterpreter",
>>>>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Dep
>>>>>>>>>>>>>> Interpreter",
>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>       ],
>>>>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>>>>       "option": {
>>>>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>>>>         "users": []
>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjffdu@gmail.com
>>>>>>>>>>>>>> > wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you mind
>>>>>>>>>>>>>>> to share your configuration ? Or you can first try to run spark-shell using
>>>>>>>>>>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added "spark.yarn.keytab" &
>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But I keep
>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the notebook
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>> kInterpreter335845091
>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>> setRules(KerberosName.java:386)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati
>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia
>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica
>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn
>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269)
>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro
>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820)
>>>>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre
>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ
>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>> Job.run(Job.java:176)
>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>
>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best Regards
>>>>>>>>>
>>>>>>>>> Jeff Zhang
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best Regards
>>>>>>>
>>>>>>> Jeff Zhang
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best Regards
>>>>>
>>>>> Jeff Zhang
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> 이종열, Jongyoul Lee, 李宗烈
>>> http://madeng.net
>>>
>>
>>
>
>
> --
> 이종열, Jongyoul Lee, 李宗烈
> http://madeng.net
>

Re: Zeppelin Kerberos error

Posted by Pradeep Reddy <pr...@gmail.com>.
Jeff- Thanks!! I figured the issue, I didn't need to copy the hive-site.xml
to spark_home/conf.

All I needed to do was set the SPARK_HOME environment variable in
"zeppelin-env.sh". That has made the local mode to work as well.

export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark

In 0.5.6, even though I didn't set the SPARK_HOME in "zeppelin-env.sh", the
embedded spark was able to bootstrap itself and talk to my hive metastore.

Thanks & Regards,
Pradeep




On Wed, Aug 31, 2016 at 9:09 PM, Jeff Zhang <zj...@gmail.com> wrote:

> AFAIK, kerberos should not related here. ZEPPELIN-1175 just remove
> ZEPPELIN_HOME/conf from the classpath of interpreter process. I guess you
> put hive-site.xml under ZEPPELIN_HOME/conf, can you try to put it
> under SPARK_HOME/conf ?
>
> On Thu, Sep 1, 2016 at 10:01 AM, Pradeep Reddy <
> pradeepreddy.alla@gmail.com> wrote:
>
>> I just tried enabling kerberos on 0.6.1 and its now able to talk to my
>> hive metastore, I see all my databases and tables. However the moment I
>> takeout kerberos and run zeppelin 0.6.1 in local mode by resetting the
>> "master" & removing spark variables, i see just one default database.
>>
>> I can live with this, as enabling kerberos integration is able to make
>> zeppelin talk to my hive metastore.
>>
>> Thanks,
>> Pradeep
>>
>> On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy <
>> pradeepreddy.alla@gmail.com> wrote:
>>
>>> No luck, even after copying the hive-site.xml in
>>> interpreter/spark/dep....I also tried, downloading 0.6.1 src  vs 0.5.6. for
>>> the same steps I followed, 0.5.6 is able to talk to my hive metastore,
>>> where as the other recent builds are not, they just show one "default"
>>> database.
>>>
>>> Thanks,
>>> Pradeep
>>>
>>> On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jo...@gmail.com>
>>> wrote:
>>>
>>>> I think it's related to https://issues.apache.org/j
>>>> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin
>>>> launches interpreter. Could you please check your hive-site.xml is included
>>>> in your interpreter process? It looks like a configuration issue because
>>>> you can see the default database. If it doesn't exists, you should copy
>>>> your xml into interpreter/spark/dep/
>>>>
>>>> Regards,
>>>> JL
>>>>
>>>> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy <
>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>
>>>>> Hi Jongyoul- I followed the exact same steps for compiling and setting
>>>>> up the new build from source as 0.5.6 (only difference is, I acquired the
>>>>> source for latest build using "git clone")
>>>>>
>>>>> hive-site.xml was copied to conf directory. But, the spark interpreter
>>>>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds
>>>>> are running in the same machine. In 0.5.6 when i run the below command, I
>>>>> see 116 databases listed, as per my expectations and I'm able to run my
>>>>> notebooks built on those databases.
>>>>>
>>>>> [image: Inline image 1]
>>>>>
>>>>> Thanks,
>>>>> Pradeep
>>>>>
>>>>>
>>>>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jo...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hello,
>>>>>>
>>>>>> Do you copy your hive-site.xml in a proper position?
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <
>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>
>>>>>>> nothing obvious. I will stick to 0.5.6 build, until the latest
>>>>>>> builds stabilize.
>>>>>>>
>>>>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Then I guess maybe you are connecting to different database. Why
>>>>>>>> not using  'z.show(sql("databases"))' to display the databases ?
>>>>>>>> Then you will get a hint what's going on.
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Yes...I didn't wish to show the names of the databases that we
>>>>>>>>> have in our data lake on that screen shot. so thats why I chose to display
>>>>>>>>> the count. The latest zeppelin build just shows 1 count which is "default"
>>>>>>>>> database.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Pradeep
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> 116 is the databases count number. Do you expect a list of
>>>>>>>>>> database ? then you need to use 'z.show(sql("databases"))'
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Here it is Jeff
>>>>>>>>>>>
>>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Pradeep,
>>>>>>>>>>>>
>>>>>>>>>>>> I don't see the databases on your screenshot (second one for
>>>>>>>>>>>> 0.5.6). I think the output is correct.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin
>>>>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk to the
>>>>>>>>>>>>> shared hive meta store are defects in the current build.
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Jeff-
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases &
>>>>>>>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Zeppelin Latest Build
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [image: Inline image 3]
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show
>>>>>>>>>>>>>> my databases and tables.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> [image: Inline image 4]
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjffdu@gmail.com
>>>>>>>>>>>>>> > wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> > the spark interpreter is not showing my tables &
>>>>>>>>>>>>>>> databases, may be its running in an isolated mode... I'm just getting empty
>>>>>>>>>>>>>>> list, so I attempted to do kerberos authentication to workaround that
>>>>>>>>>>>>>>> issue, and bumped into this road block.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> kerberos would not help here, actually I think it would make
>>>>>>>>>>>>>>> the problem more complicated.  You need to first check the log why you get
>>>>>>>>>>>>>>> empty list.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell,
>>>>>>>>>>>>>>>> after running kdestroy, with the below command and was able to get to my
>>>>>>>>>>>>>>>> hive tables.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode
>>>>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode,
>>>>>>>>>>>>>>>>> but when I migrated from 0.5.6 to this version, the spark interpreter is
>>>>>>>>>>>>>>>>> not showing my tables & databases, may be its running in an isolated
>>>>>>>>>>>>>>>>> mode... I'm just getting empty list, so I attempted to do kerberos
>>>>>>>>>>>>>>>>> authentication to workaround that issue, and bumped into this road block.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file
>>>>>>>>>>>>>>>>> and its working fine.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client
>>>>>>>>>>>>>>>>> --master yarn --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>>>>>>>       "properties": {
>>>>>>>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>>>>>>>         "args": "",
>>>>>>>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>>>>>>>> "spark-packages,http://dl.bint
>>>>>>>>>>>>>>>>> ray.com/spark-packages/maven,false;"
>>>>>>>>>>>>>>>>>       },
>>>>>>>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>>>> rkInterpreter",
>>>>>>>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Dep
>>>>>>>>>>>>>>>>> Interpreter",
>>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>>>       ],
>>>>>>>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>>>>>>>       "option": {
>>>>>>>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>>>>>>>         "users": []
>>>>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <
>>>>>>>>>>>>>>>>> zjffdu@gmail.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you
>>>>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to run spark-shell
>>>>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added "spark.yarn.keytab" &
>>>>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But I keep
>>>>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the notebook
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> 3536728 started by scheduler
>>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.
>>>>>>>>>>>>>>>>>>> COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>>>> setRules(KerberosName.java:386)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati
>>>>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>>> rGroupInformation.initialize(U
>>>>>>>>>>>>>>>>>>> serGroupInformation.java:227)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia
>>>>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica
>>>>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn
>>>>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269)
>>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro
>>>>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820)
>>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre
>>>>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ
>>>>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>>>> Job.run(Job.java:176)
>>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledT
>>>>>>>>>>>>>>>>>>> hreadPoolExecutor.java:180)
>>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>
>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best Regards
>>>>>>>>>>
>>>>>>>>>> Jeff Zhang
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best Regards
>>>>>>>>
>>>>>>>> Jeff Zhang
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> 이종열, Jongyoul Lee, 李宗烈
>>>>>> http://madeng.net
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> 이종열, Jongyoul Lee, 李宗烈
>>>> http://madeng.net
>>>>
>>>
>>>
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>

Re: Zeppelin Kerberos error

Posted by Jeff Zhang <zj...@gmail.com>.
AFAIK, kerberos should not related here. ZEPPELIN-1175 just remove
ZEPPELIN_HOME/conf from the classpath of interpreter process. I guess you
put hive-site.xml under ZEPPELIN_HOME/conf, can you try to put it
under SPARK_HOME/conf ?

On Thu, Sep 1, 2016 at 10:01 AM, Pradeep Reddy <pr...@gmail.com>
wrote:

> I just tried enabling kerberos on 0.6.1 and its now able to talk to my
> hive metastore, I see all my databases and tables. However the moment I
> takeout kerberos and run zeppelin 0.6.1 in local mode by resetting the
> "master" & removing spark variables, i see just one default database.
>
> I can live with this, as enabling kerberos integration is able to make
> zeppelin talk to my hive metastore.
>
> Thanks,
> Pradeep
>
> On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy <
> pradeepreddy.alla@gmail.com> wrote:
>
>> No luck, even after copying the hive-site.xml in
>> interpreter/spark/dep....I also tried, downloading 0.6.1 src  vs 0.5.6. for
>> the same steps I followed, 0.5.6 is able to talk to my hive metastore,
>> where as the other recent builds are not, they just show one "default"
>> database.
>>
>> Thanks,
>> Pradeep
>>
>> On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jo...@gmail.com>
>> wrote:
>>
>>> I think it's related to https://issues.apache.org/j
>>> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin
>>> launches interpreter. Could you please check your hive-site.xml is included
>>> in your interpreter process? It looks like a configuration issue because
>>> you can see the default database. If it doesn't exists, you should copy
>>> your xml into interpreter/spark/dep/
>>>
>>> Regards,
>>> JL
>>>
>>> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy <
>>> pradeepreddy.alla@gmail.com> wrote:
>>>
>>>> Hi Jongyoul- I followed the exact same steps for compiling and setting
>>>> up the new build from source as 0.5.6 (only difference is, I acquired the
>>>> source for latest build using "git clone")
>>>>
>>>> hive-site.xml was copied to conf directory. But, the spark interpreter
>>>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds
>>>> are running in the same machine. In 0.5.6 when i run the below command, I
>>>> see 116 databases listed, as per my expectations and I'm able to run my
>>>> notebooks built on those databases.
>>>>
>>>> [image: Inline image 1]
>>>>
>>>> Thanks,
>>>> Pradeep
>>>>
>>>>
>>>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jo...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> Do you copy your hive-site.xml in a proper position?
>>>>>
>>>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <
>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>
>>>>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds
>>>>>> stabilize.
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zj...@gmail.com> wrote:
>>>>>>
>>>>>>> Then I guess maybe you are connecting to different database. Why not
>>>>>>> using  'z.show(sql("databases"))' to display the databases ? Then
>>>>>>> you will get a hint what's going on.
>>>>>>>
>>>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>
>>>>>>>> Yes...I didn't wish to show the names of the databases that we have
>>>>>>>> in our data lake on that screen shot. so thats why I chose to display the
>>>>>>>> count. The latest zeppelin build just shows 1 count which is "default"
>>>>>>>> database.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Pradeep
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> 116 is the databases count number. Do you expect a list of
>>>>>>>>> database ? then you need to use 'z.show(sql("databases"))'
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Here it is Jeff
>>>>>>>>>>
>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Pradeep,
>>>>>>>>>>>
>>>>>>>>>>> I don't see the databases on your screenshot (second one for
>>>>>>>>>>> 0.5.6). I think the output is correct.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin
>>>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk to the
>>>>>>>>>>>> shared hive meta store are defects in the current build.
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Jeff-
>>>>>>>>>>>>>
>>>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases &
>>>>>>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>>>>>>
>>>>>>>>>>>>> Zeppelin Latest Build
>>>>>>>>>>>>>
>>>>>>>>>>>>> [image: Inline image 3]
>>>>>>>>>>>>>
>>>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show
>>>>>>>>>>>>> my databases and tables.
>>>>>>>>>>>>>
>>>>>>>>>>>>> [image: Inline image 4]
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> > the spark interpreter is not showing my tables &
>>>>>>>>>>>>>> databases, may be its running in an isolated mode... I'm just getting empty
>>>>>>>>>>>>>> list, so I attempted to do kerberos authentication to workaround that
>>>>>>>>>>>>>> issue, and bumped into this road block.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> kerberos would not help here, actually I think it would make
>>>>>>>>>>>>>> the problem more complicated.  You need to first check the log why you get
>>>>>>>>>>>>>> empty list.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>>>>>>>> running kdestroy, with the below command and was able to get to my hive
>>>>>>>>>>>>>>> tables.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode
>>>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but
>>>>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark interpreter is not
>>>>>>>>>>>>>>>> showing my tables & databases, may be its running in an isolated mode...
>>>>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos authentication
>>>>>>>>>>>>>>>> to workaround that issue, and bumped into this road block.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file
>>>>>>>>>>>>>>>> and its working fine.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master
>>>>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>>>>>>       "properties": {
>>>>>>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>>>>>>         "args": "",
>>>>>>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>>>>>>> "spark-packages,http://dl.bint
>>>>>>>>>>>>>>>> ray.com/spark-packages/maven,false;"
>>>>>>>>>>>>>>>>       },
>>>>>>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>>> rkInterpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Dep
>>>>>>>>>>>>>>>> Interpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>>       ],
>>>>>>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>>>>>>       "option": {
>>>>>>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>>>>>>         "users": []
>>>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <
>>>>>>>>>>>>>>>> zjffdu@gmail.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you
>>>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to run spark-shell
>>>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added "spark.yarn.keytab" &
>>>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But I keep
>>>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the notebook
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> 3536728 started by scheduler
>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>>> setRules(KerberosName.java:386)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati
>>>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.initialize(U
>>>>>>>>>>>>>>>>>> serGroupInformation.java:227)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia
>>>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica
>>>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn
>>>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro
>>>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre
>>>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ
>>>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>>> Job.run(Job.java:176)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Best Regards
>>>>>>>>>>>
>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best Regards
>>>>>>>>>
>>>>>>>>> Jeff Zhang
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best Regards
>>>>>>>
>>>>>>> Jeff Zhang
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> 이종열, Jongyoul Lee, 李宗烈
>>>>> http://madeng.net
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> 이종열, Jongyoul Lee, 李宗烈
>>> http://madeng.net
>>>
>>
>>
>


-- 
Best Regards

Jeff Zhang

Re: Zeppelin Kerberos error

Posted by Pradeep Reddy <pr...@gmail.com>.
I just tried enabling kerberos on 0.6.1 and its now able to talk to my hive
metastore, I see all my databases and tables. However the moment I takeout
kerberos and run zeppelin 0.6.1 in local mode by resetting the "master" &
removing spark variables, i see just one default database.

I can live with this, as enabling kerberos integration is able to make
zeppelin talk to my hive metastore.

Thanks,
Pradeep

On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy <pr...@gmail.com>
wrote:

> No luck, even after copying the hive-site.xml in
> interpreter/spark/dep....I also tried, downloading 0.6.1 src  vs 0.5.6. for
> the same steps I followed, 0.5.6 is able to talk to my hive metastore,
> where as the other recent builds are not, they just show one "default"
> database.
>
> Thanks,
> Pradeep
>
> On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jo...@gmail.com> wrote:
>
>> I think it's related to https://issues.apache.org/j
>> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin
>> launches interpreter. Could you please check your hive-site.xml is included
>> in your interpreter process? It looks like a configuration issue because
>> you can see the default database. If it doesn't exists, you should copy
>> your xml into interpreter/spark/dep/
>>
>> Regards,
>> JL
>>
>> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy <
>> pradeepreddy.alla@gmail.com> wrote:
>>
>>> Hi Jongyoul- I followed the exact same steps for compiling and setting
>>> up the new build from source as 0.5.6 (only difference is, I acquired the
>>> source for latest build using "git clone")
>>>
>>> hive-site.xml was copied to conf directory. But, the spark interpreter
>>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds
>>> are running in the same machine. In 0.5.6 when i run the below command, I
>>> see 116 databases listed, as per my expectations and I'm able to run my
>>> notebooks built on those databases.
>>>
>>> [image: Inline image 1]
>>>
>>> Thanks,
>>> Pradeep
>>>
>>>
>>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jo...@gmail.com>
>>> wrote:
>>>
>>>> Hello,
>>>>
>>>> Do you copy your hive-site.xml in a proper position?
>>>>
>>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <
>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>
>>>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds
>>>>> stabilize.
>>>>>
>>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zj...@gmail.com> wrote:
>>>>>
>>>>>> Then I guess maybe you are connecting to different database. Why not
>>>>>> using  'z.show(sql("databases"))' to display the databases ? Then
>>>>>> you will get a hint what's going on.
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>
>>>>>>> Yes...I didn't wish to show the names of the databases that we have
>>>>>>> in our data lake on that screen shot. so thats why I chose to display the
>>>>>>> count. The latest zeppelin build just shows 1 count which is "default"
>>>>>>> database.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Pradeep
>>>>>>>
>>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> 116 is the databases count number. Do you expect a list of database
>>>>>>>> ? then you need to use 'z.show(sql("databases"))'
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Here it is Jeff
>>>>>>>>>
>>>>>>>>> [image: Inline image 1]
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Pradeep,
>>>>>>>>>>
>>>>>>>>>> I don't see the databases on your screenshot (second one for
>>>>>>>>>> 0.5.6). I think the output is correct.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin
>>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk to the
>>>>>>>>>>> shared hive meta store are defects in the current build.
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Jeff-
>>>>>>>>>>>>
>>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases &
>>>>>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>>>>>
>>>>>>>>>>>> Zeppelin Latest Build
>>>>>>>>>>>>
>>>>>>>>>>>> [image: Inline image 3]
>>>>>>>>>>>>
>>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my
>>>>>>>>>>>> databases and tables.
>>>>>>>>>>>>
>>>>>>>>>>>> [image: Inline image 4]
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zj...@gmail.com>
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> > the spark interpreter is not showing my tables & databases,
>>>>>>>>>>>>> may be its running in an isolated mode... I'm just getting empty list, so I
>>>>>>>>>>>>> attempted to do kerberos authentication to workaround that issue, and
>>>>>>>>>>>>> bumped into this road block.
>>>>>>>>>>>>>
>>>>>>>>>>>>> kerberos would not help here, actually I think it would make
>>>>>>>>>>>>> the problem more complicated.  You need to first check the log why you get
>>>>>>>>>>>>> empty list.
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>>>>>>> running kdestroy, with the below command and was able to get to my hive
>>>>>>>>>>>>>> tables.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode
>>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but
>>>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark interpreter is not
>>>>>>>>>>>>>>> showing my tables & databases, may be its running in an isolated mode...
>>>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos authentication
>>>>>>>>>>>>>>> to workaround that issue, and bumped into this road block.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file and
>>>>>>>>>>>>>>> its working fine.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master
>>>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>>>>>       "properties": {
>>>>>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>>>>>         "args": "",
>>>>>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,f
>>>>>>>>>>>>>>> alse;"
>>>>>>>>>>>>>>>       },
>>>>>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>> rkInterpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Dep
>>>>>>>>>>>>>>> Interpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>       ],
>>>>>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>>>>>       "option": {
>>>>>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>>>>>         "users": []
>>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <
>>>>>>>>>>>>>>> zjffdu@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you
>>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to run spark-shell
>>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>>>>>> pradeepreddy.alla@gmail.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added "spark.yarn.keytab" &
>>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But I keep
>>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the notebook
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>> kInterpreter335845091
>>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>> setRules(KerberosName.java:386)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati
>>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.initialize(U
>>>>>>>>>>>>>>>>> serGroupInformation.java:227)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia
>>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica
>>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn
>>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269)
>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro
>>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre
>>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ
>>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>> Job.run(Job.java:176)
>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>
>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best Regards
>>>>>>>>>>
>>>>>>>>>> Jeff Zhang
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best Regards
>>>>>>>>
>>>>>>>> Jeff Zhang
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best Regards
>>>>>>
>>>>>> Jeff Zhang
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> 이종열, Jongyoul Lee, 李宗烈
>>>> http://madeng.net
>>>>
>>>
>>>
>>
>>
>> --
>> 이종열, Jongyoul Lee, 李宗烈
>> http://madeng.net
>>
>
>