You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@toree.apache.org by "Harpreet Sawhney (JIRA)" <ji...@apache.org> on 2016/12/23 05:38:58 UTC

[jira] [Commented] (TOREE-353) Spark defaulting to internal metastore instead of hive

    [ https://issues.apache.org/jira/browse/TOREE-353?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15771997#comment-15771997 ] 

Harpreet Sawhney commented on TOREE-353:
----------------------------------------

I am facing similar issue with spark 1.6.3 

> Spark defaulting to internal metastore instead of hive
> ------------------------------------------------------
>
>                 Key: TOREE-353
>                 URL: https://issues.apache.org/jira/browse/TOREE-353
>             Project: TOREE
>          Issue Type: Bug
>    Affects Versions: 0.1.0
>         Environment: Vanilla Hadoop 2.7.3, Hive 2.1.0, Spark 2.0.2 with Hadoop 2.7 binary distribution, Toree build from git repo and then installed using pip from torte-pip folder
>            Reporter: Abhijit Bhole
>
> I cannot see Hive tables inside Toree kernel. 
> For testing I have created a table in Hive and I execute the query "show tables from default" in spark. Some more details -
> - I have added the hive-submit.conf in SPARK_HOME/conf directory. 
> - pyspark and spark-submit scripts work file in both yarn and local mode without any additional arguments. The pyspark shell also works fine and returns the table name. 
> - On the driver UI page (on 4040 port) I can see  spark.sql.catalogImplementation set to hive when running pyspark shell. But the variable is missing when running Toree kernel.
> - I have tried also giving --driver-class-path hive-site.xml --files /opt/hive-2.1.0/conf/hive-site.xml unsuccessfully
> I have tried everything I could think of hence filing a bug report. I have missed something please help.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)