You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hive.apache.org by شجاع الرحمن بیگ <sh...@gmail.com> on 2016/08/19 11:53:02 UTC

javax.jdo.JDODataStoreException: Required table missing : "VERSION" in Catalog "" Schema "". DataNucleus requires this table to perform its persistence operations. Either your MetaData is incorrect, or you need to enable "datanucleus.schema.autoCreateTables"

Hey,

Could you please help me resolving this error?
version= hive 2.1.0

ERROR StatusLogger No log4j2 configuration file found. Using default
configuration: logging only errors to the console.
13:41:37.848 [main] ERROR hive.ql.metadata.Hive - Cannot initialize
metastore due to autoCreate error
javax.jdo.JDODataStoreException: Required table missing : "VERSION" in
Catalog "" Schema "". DataNucleus requires this table to perform its
persistence operations. Either your MetaData is incorrect, or you need to
enable "datanucleus.schema.autoCreateTables"
    at
org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:553)
~[datanucleus-api-jdo-4.2.1.jar:?]
    at
org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:720)
~[datanucleus-api-jdo-4.2.1.jar:?]
    at
org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:740)
~[datanucleus-api-jdo-4.2.1.jar:?]
    at
org.apache.hadoop.hive.metastore.ObjectStore.setMetaStoreSchemaVersion(ObjectStore.java:7763)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.ObjectStore.checkSchema(ObjectStore.java:7657)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.ObjectStore.verifySchema(ObjectStore.java:7632)
~[hive-exec-2.1.0.jar:2.1.0]
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.7.0_79]
    at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
~[?:1.7.0_79]
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.7.0_79]
    at java.lang.reflect.Method.invoke(Method.java:606) ~[?:1.7.0_79]
    at
org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:101)
~[hive-exec-2.1.0.jar:2.1.0]
    at com.sun.proxy.$Proxy11.verifySchema(Unknown Source) ~[?:?]
    at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:547)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:612)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:398)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6396)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:236)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:70)
~[hive-exec-2.1.0.jar:2.1.0]
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[?:1.7.0_79]
    at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
~[?:1.7.0_79]
    at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[?:1.7.0_79]
    at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
~[?:1.7.0_79]
    at
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1625)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:80)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:130)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:101)
~[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3317)
~[hive-exec-2.1.0.jar:2.1.0]
    at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3356)
[hive-exec-2.1.0.jar:2.1.0]
    at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3336)
[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3590)
[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:236)
[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:221)
[hive-exec-2.1.0.jar:2.1.0]
    at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:366)
[hive-exec-2.1.0.jar:2.1.0]
    at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:310)
[hive-exec-2.1.0.jar:2.1.0]
    at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:290)
[hive-exec-2.1.0.jar:2.1.0]
    at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:266)
[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:545)
[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:513)
[hive-exec-2.1.0.jar:2.1.0]
    at
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:116)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
    at
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.7.0_79]
    at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
~[?:1.7.0_79]
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.7.0_79]
    at java.lang.reflect.Method.invoke(Method.java:606) ~[?:1.7.0_79]
    at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
    at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
    at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
    at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
    at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
[spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
Caused by: org.datanucleus.store.rdbms.exceptions.MissingTableException:
Required table missing : "VERSION" in Catalog "" Schema "". DataNucleus
requires this table to perform its persistence operations. Either your
MetaData is incorrect, or you need to enable
"datanucleus.schema.autoCreateTables"
    at
org.datanucleus.store.rdbms.table.AbstractTable.exists(AbstractTable.java:606)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.rdbms.RDBMSStoreManager$ClassAdder.performTablesValidation(RDBMSStoreManager.java:3365)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.rdbms.RDBMSStoreManager$ClassAdder.run(RDBMSStoreManager.java:2877)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.rdbms.AbstractSchemaTransaction.execute(AbstractSchemaTransaction.java:119)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.rdbms.RDBMSStoreManager.manageClasses(RDBMSStoreManager.java:1608)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.rdbms.RDBMSStoreManager.getDatastoreClass(RDBMSStoreManager.java:671)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.rdbms.RDBMSStoreManager.getPropertiesForGenerator(RDBMSStoreManager.java:2069)
~[datanucleus-rdbms-4.1.7.jar:?]
    at
org.datanucleus.store.AbstractStoreManager.getStrategyValue(AbstractStoreManager.java:1271)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.ExecutionContextImpl.newObjectId(ExecutionContextImpl.java:3759)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.state.StateManagerImpl.setIdentity(StateManagerImpl.java:2267)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.state.StateManagerImpl.initialiseForPersistentNew(StateManagerImpl.java:484)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.state.StateManagerImpl.initialiseForPersistentNew(StateManagerImpl.java:120)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.state.ObjectProviderFactoryImpl.newForPersistentNew(ObjectProviderFactoryImpl.java:218)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.ExecutionContextImpl.persistObjectInternal(ExecutionContextImpl.java:2078)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.ExecutionContextImpl.persistObjectWork(ExecutionContextImpl.java:1922)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.ExecutionContextImpl.persistObject(ExecutionContextImpl.java:1777)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.ExecutionContextThreadedImpl.persistObject(ExecutionContextThreadedImpl.java:217)
~[datanucleus-core-4.1.6.jar:?]
    at
org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:715)
~[datanucleus-api-jdo-4.2.1.jar:?]
    ... 49 more
Exception in thread "main" java.lang.RuntimeException:
org.apache.hadoop.hive.ql.metadata.HiveException:
org.apache.hadoop.hive.ql.metadata.HiveException:
MetaException(message:Hive metastore database is not initialized. Please
use schematool (e.g. ./schematool -initSchema -dbType ...) to create the
schema. If needed, don't forget to include the option to auto-create the
underlying database in your JDBC connection string (e.g.
?createDatabaseIfNotExist=true for mysql))
    at
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:578)
    at
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:513)
    at
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:116)
    at
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
    at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
    at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
    at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
    at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
    at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
org.apache.hadoop.hive.ql.metadata.HiveException:
MetaException(message:Hive metastore database is not initialized. Please
use schematool (e.g. ./schematool -initSchema -dbType ...) to create the
schema. If needed, don't forget to include the option to auto-create the
underlying database in your JDBC connection string (e.g.
?createDatabaseIfNotExist=true for mysql))
    at
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:226)
    at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:366)
    at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:310)
    at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:290)
    at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:266)
    at
org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:545)
    ... 12 more
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
MetaException(message:Hive metastore database is not initialized. Please
use schematool (e.g. ./schematool -initSchema -dbType ...) to create the
schema. If needed, don't forget to include the option to auto-create the
underlying database in your JDBC connection string (e.g.
?createDatabaseIfNotExist=true for mysql))
    at
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3593)
    at
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:236)
    at
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:221)
    ... 17 more
Caused by: MetaException(message:Hive metastore database is not
initialized. Please use schematool (e.g. ./schematool -initSchema -dbType
...) to create the schema. If needed, don't forget to include the option to
auto-create the underlying database in your JDBC connection string (e.g.
?createDatabaseIfNotExist=true for mysql))
    at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3364)
    at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3336)
    at
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3590)
    ... 19 more


Thanks
Shuja

-- 
Regards
Shuja-ur-Rehman Baig
<http://pk.linkedin.com/in/shujamughal>

Re: javax.jdo.JDODataStoreException: Required table missing : "VERSION" in Catalog "" Schema "". DataNucleus requires this table to perform its persistence operations. Either your MetaData is incorrect, or you need to enable "datanucleus.schema.autoCreateTables"

Posted by Thejas Nair <th...@gmail.com>.
As the error message indicated, please use "schematool -initSchema
-dbType.. " command to create the proper metadata schema in metastore.


On Fri, Aug 19, 2016 at 4:53 AM, شجاع الرحمن بیگ <sh...@gmail.com>
wrote:

> Hey,
>
> Could you please help me resolving this error?
> version= hive 2.1.0
>
> ERROR StatusLogger No log4j2 configuration file found. Using default
> configuration: logging only errors to the console.
> 13:41:37.848 [main] ERROR hive.ql.metadata.Hive - Cannot initialize
> metastore due to autoCreate error
> javax.jdo.JDODataStoreException: Required table missing : "VERSION" in
> Catalog "" Schema "". DataNucleus requires this table to perform its
> persistence operations. Either your MetaData is incorrect, or you need to
> enable "datanucleus.schema.autoCreateTables"
>     at org.datanucleus.api.jdo.NucleusJDOHelper.
> getJDOExceptionForNucleusException(NucleusJDOHelper.java:553)
> ~[datanucleus-api-jdo-4.2.1.jar:?]
>     at org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(
> JDOPersistenceManager.java:720) ~[datanucleus-api-jdo-4.2.1.jar:?]
>     at org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(
> JDOPersistenceManager.java:740) ~[datanucleus-api-jdo-4.2.1.jar:?]
>     at org.apache.hadoop.hive.metastore.ObjectStore.
> setMetaStoreSchemaVersion(ObjectStore.java:7763)
> ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.ObjectStore.
> checkSchema(ObjectStore.java:7657) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.ObjectStore.
> verifySchema(ObjectStore.java:7632) ~[hive-exec-2.1.0.jar:2.1.0]
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> ~[?:1.7.0_79]
>     at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57) ~[?:1.7.0_79]
>     at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43) ~[?:1.7.0_79]
>     at java.lang.reflect.Method.invoke(Method.java:606) ~[?:1.7.0_79]
>     at org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:101)
> ~[hive-exec-2.1.0.jar:2.1.0]
>     at com.sun.proxy.$Proxy11.verifySchema(Unknown Source) ~[?:?]
>     at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:547)
> ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.HiveMetaStore$
> HMSHandler.createDefaultDB(HiveMetaStore.java:612)
> ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.HiveMetaStore$
> HMSHandler.init(HiveMetaStore.java:398) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<
> init>(RetryingHMSHandler.java:78) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.RetryingHMSHandler.
> getProxy(RetryingHMSHandler.java:84) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.HiveMetaStore.
> newRetryingHMSHandler(HiveMetaStore.java:6396)
> ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.
> <init>(HiveMetaStoreClient.java:236) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<
> init>(SessionHiveMetaStoreClient.java:70) ~[hive-exec-2.1.0.jar:2.1.0]
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method) ~[?:1.7.0_79]
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance(
> NativeConstructorAccessorImpl.java:57) ~[?:1.7.0_79]
>     at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> DelegatingConstructorAccessorImpl.java:45) ~[?:1.7.0_79]
>     at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> ~[?:1.7.0_79]
>     at org.apache.hadoop.hive.metastore.MetaStoreUtils.
> newInstance(MetaStoreUtils.java:1625) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>
> (RetryingMetaStoreClient.java:80) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(
> RetryingMetaStoreClient.java:130) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(
> RetryingMetaStoreClient.java:101) ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3317)
> ~[hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3356)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3336)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3590)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:236)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.
> registerAllFunctionsOnce(Hive.java:221) [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:366)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:310)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:290)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:266)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:545)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:513)
> [hive-exec-2.1.0.jar:2.1.0]
>     at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:116)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
>     at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> ~[?:1.7.0_79]
>     at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57) ~[?:1.7.0_79]
>     at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43) ~[?:1.7.0_79]
>     at java.lang.reflect.Method.invoke(Method.java:606) ~[?:1.7.0_79]
>     at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
>     at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
>     at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> [spark-assembly-1.6.1-hadoop2.6.0.jar:1.6.1]
> Caused by: org.datanucleus.store.rdbms.exceptions.MissingTableException:
> Required table missing : "VERSION" in Catalog "" Schema "". DataNucleus
> requires this table to perform its persistence operations. Either your
> MetaData is incorrect, or you need to enable "datanucleus.schema.
> autoCreateTables"
>     at org.datanucleus.store.rdbms.table.AbstractTable.exists(AbstractTable.java:606)
> ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.rdbms.RDBMSStoreManager$ClassAdder.
> performTablesValidation(RDBMSStoreManager.java:3365)
> ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.rdbms.RDBMSStoreManager$ClassAdder.
> run(RDBMSStoreManager.java:2877) ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.rdbms.AbstractSchemaTransaction.execute(
> AbstractSchemaTransaction.java:119) ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.rdbms.RDBMSStoreManager.manageClasses(RDBMSStoreManager.java:1608)
> ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.rdbms.RDBMSStoreManager.getDatastoreClass(RDBMSStoreManager.java:671)
> ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.rdbms.RDBMSStoreManager.
> getPropertiesForGenerator(RDBMSStoreManager.java:2069)
> ~[datanucleus-rdbms-4.1.7.jar:?]
>     at org.datanucleus.store.AbstractStoreManager.getStrategyValue(
> AbstractStoreManager.java:1271) ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.ExecutionContextImpl.newObjectId(
> ExecutionContextImpl.java:3759) ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.state.StateManagerImpl.setIdentity(StateManagerImpl.java:2267)
> ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.state.StateManagerImpl.initialiseForPersistentNew(StateManagerImpl.java:484)
> ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.state.StateManagerImpl.initialiseForPersistentNew(StateManagerImpl.java:120)
> ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.state.ObjectProviderFactoryImpl.
> newForPersistentNew(ObjectProviderFactoryImpl.java:218)
> ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.ExecutionContextImpl.persistObjectInternal(
> ExecutionContextImpl.java:2078) ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.ExecutionContextImpl.persistObjectWork(
> ExecutionContextImpl.java:1922) ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.ExecutionContextImpl.persistObject(
> ExecutionContextImpl.java:1777) ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.ExecutionContextThreadedImpl.persistObject(
> ExecutionContextThreadedImpl.java:217) ~[datanucleus-core-4.1.6.jar:?]
>     at org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(
> JDOPersistenceManager.java:715) ~[datanucleus-api-jdo-4.2.1.jar:?]
>     ... 49 more
> Exception in thread "main" java.lang.RuntimeException:
> org.apache.hadoop.hive.ql.metadata.HiveException:
> org.apache.hadoop.hive.ql.metadata.HiveException:
> MetaException(message:Hive metastore database is not initialized. Please
> use schematool (e.g. ./schematool -initSchema -dbType ...) to create the
> schema. If needed, don't forget to include the option to auto-create the
> underlying database in your JDBC connection string (e.g.
> ?createDatabaseIfNotExist=true for mysql))
>     at org.apache.hadoop.hive.ql.session.SessionState.start(
> SessionState.java:578)
>     at org.apache.hadoop.hive.ql.session.SessionState.start(
> SessionState.java:513)
>     at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(
> SparkSQLCLIDriver.scala:116)
>     at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(
> SparkSQLCLIDriver.scala)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
>     at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:606)
>     at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
>     at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:181)
>     at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
> org.apache.hadoop.hive.ql.metadata.HiveException:
> MetaException(message:Hive metastore database is not initialized. Please
> use schematool (e.g. ./schematool -initSchema -dbType ...) to create the
> schema. If needed, don't forget to include the option to auto-create the
> underlying database in your JDBC connection string (e.g.
> ?createDatabaseIfNotExist=true for mysql))
>     at org.apache.hadoop.hive.ql.metadata.Hive.
> registerAllFunctionsOnce(Hive.java:226)
>     at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:366)
>     at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:310)
>     at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:290)
>     at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:266)
>     at org.apache.hadoop.hive.ql.session.SessionState.start(
> SessionState.java:545)
>     ... 12 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
> MetaException(message:Hive metastore database is not initialized. Please
> use schematool (e.g. ./schematool -initSchema -dbType ...) to create the
> schema. If needed, don't forget to include the option to auto-create the
> underlying database in your JDBC connection string (e.g.
> ?createDatabaseIfNotExist=true for mysql))
>     at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(
> Hive.java:3593)
>     at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(
> Hive.java:236)
>     at org.apache.hadoop.hive.ql.metadata.Hive.
> registerAllFunctionsOnce(Hive.java:221)
>     ... 17 more
> Caused by: MetaException(message:Hive metastore database is not
> initialized. Please use schematool (e.g. ./schematool -initSchema -dbType
> ...) to create the schema. If needed, don't forget to include the option to
> auto-create the underlying database in your JDBC connection string (e.g.
> ?createDatabaseIfNotExist=true for mysql))
>     at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3364)
>     at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3336)
>     at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(
> Hive.java:3590)
>     ... 19 more
>
>
> Thanks
> Shuja
>
> --
> Regards
> Shuja-ur-Rehman Baig
> <http://pk.linkedin.com/in/shujamughal>
>