You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Wei Cui (JIRA)" <ji...@apache.org> on 2014/05/19 23:21:37 UTC

[jira] [Updated] (SPARK-1884) Shark failed to start

     [ https://issues.apache.org/jira/browse/SPARK-1884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Wei Cui updated SPARK-1884:
---------------------------

    Summary: Shark failed to start  (was: Shark )

> Shark failed to start
> ---------------------
>
>                 Key: SPARK-1884
>                 URL: https://issues.apache.org/jira/browse/SPARK-1884
>             Project: Spark
>          Issue Type: Bug
>    Affects Versions: 0.9.1
>         Environment: ubuntu 14.04, spark 0.9.1, hive 0.13.0, hadoop 2.4.0 (stand alone), scala 2.11.0
>            Reporter: Wei Cui
>            Priority: Blocker
>
> the hadoop, hive, spark works fine.
> when start the shark, it failed with the following messages:
> Starting the Shark Command Line Client
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.input.dir.recursive is deprecated. Instead, use mapreduce.input.fileinputformat.input.dir.recursive
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.max.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.min.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.min.split.size.per.rack is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize.per.rack
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.min.split.size.per.node is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize.per.node
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces
> 14/05/19 16:47:21 INFO Configuration.deprecation: mapred.reduce.tasks.speculative.execution is deprecated. Instead, use mapreduce.reduce.speculative
> 14/05/19 16:47:22 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@48c724c:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval;  Ignoring.
> 14/05/19 16:47:22 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@48c724c:an attempt to override final parameter: mapreduce.cluster.local.dir;  Ignoring.
> 14/05/19 16:47:22 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@48c724c:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts;  Ignoring.
> 14/05/19 16:47:22 WARN conf.Configuration: org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@48c724c:an attempt to override final parameter: mapreduce.cluster.temp.dir;  Ignoring.
> Logging initialized using configuration in jar:file:/usr/local/shark/lib_managed/jars/edu.berkeley.cs.shark/hive-common/hive-common-0.11.0-shark-0.9.1.jar!/hive-log4j.properties
> Hive history file=/tmp/root/hive_job_log_root_14857@ubuntu_201405191647_897494215.txt
> 6.004: [GC 279616K->18440K(1013632K), 0.0438980 secs]
> 6.445: [Full GC 59125K->7949K(1013632K), 0.0685160 secs]
> Reloading cached RDDs from previous Shark sessions... (use -skipRddReload flag to skip reloading)
> 7.535: [Full GC 104136K->13059K(1013632K), 0.0885820 secs]
> 8.459: [Full GC 61237K->18031K(1013632K), 0.0820400 secs]
> 8.662: [Full GC 29832K->8958K(1013632K), 0.0869700 secs]
> 8.751: [Full GC 13433K->8998K(1013632K), 0.0856520 secs]
> 10.435: [Full GC 72246K->14140K(1013632K), 0.1797530 secs]
> Exception in thread "main" org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient
> 	at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1072)
> 	at shark.memstore2.TableRecovery$.reloadRdds(TableRecovery.scala:49)
> 	at shark.SharkCliDriver.<init>(SharkCliDriver.scala:283)
> 	at shark.SharkCliDriver$.main(SharkCliDriver.scala:162)
> 	at shark.SharkCliDriver.main(SharkCliDriver.scala)
> Caused by: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient
> 	at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1139)
> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:51)
> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:61)
> 	at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:2288)
> 	at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:2299)
> 	at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1070)
> 	... 4 more
> Caused by: java.lang.reflect.InvocationTargetException
> 	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> 	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> 	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> 	at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1137)
> 	... 9 more
> Caused by: java.lang.VerifyError: class org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$SetOwnerRequestProto overrides final method getUnknownFields.()Lcom/google/protobuf/UnknownFieldSet;
> 	at java.lang.ClassLoader.defineClass1(Native Method)
> 	at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
> 	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
> 	at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
> 	at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> 	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
> 	at java.lang.Class.getDeclaredMethods0(Native Method)
> 	at java.lang.Class.privateGetDeclaredMethods(Class.java:2531)
> 	at java.lang.Class.privateGetPublicMethods(Class.java:2651)
> 	at java.lang.Class.privateGetPublicMethods(Class.java:2661)
> 	at java.lang.Class.getMethods(Class.java:1467)
> 	at sun.misc.ProxyGenerator.generateClassFile(ProxyGenerator.java:426)
> 	at sun.misc.ProxyGenerator.generateProxyClass(ProxyGenerator.java:323)
> 	at java.lang.reflect.Proxy.getProxyClass0(Proxy.java:636)
> 	at java.lang.reflect.Proxy.newProxyInstance(Proxy.java:722)
> 	at org.apache.hadoop.ipc.ProtobufRpcEngine.getProxy(ProtobufRpcEngine.java:92)
> 	at org.apache.hadoop.ipc.RPC.getProtocolProxy(RPC.java:537)
> 	at org.apache.hadoop.hdfs.NameNodeProxies.createNNProxyWithClientProtocol(NameNodeProxies.java:334)
> 	at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:241)
> 	at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:141)
> 	at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:576)
> 	at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:521)
> 	at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:146)
> 	at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2397)
> 	at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89)
> 	at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2431)
> 	at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2413)
> 	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:368)
> 	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:167)
> 	at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:352)
> 	at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
> 	at org.apache.hadoop.hive.metastore.Warehouse.getFs(Warehouse.java:105)
> 	at org.apache.hadoop.hive.metastore.Warehouse.getDnsPath(Warehouse.java:137)
> 	at org.apache.hadoop.hive.metastore.Warehouse.getWhRoot(Warehouse.java:152)
> 	at org.apache.hadoop.hive.metastore.Warehouse.getDefaultDatabasePath(Warehouse.java:170)
> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB_core(HiveMetaStore.java:420)
> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:439)
> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:325)
> 	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.<init>(HiveMetaStore.java:285)
> 	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:54)
> 	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:59)
> 	at org.apache.hadoop.hive.metastore.HiveMetaStore.newHMSHandler(HiveMetaStore.java:4102)
> 	at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:121)
> 	... 14 more



--
This message was sent by Atlassian JIRA
(v6.2#6252)