You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@phoenix.apache.org by Raghavendra Channarayappa <ra...@myntra.com> on 2018/12/04 12:17:16 UTC

JDBC Connection to Apache Phoenix failing

Can someone help with this? TIA
I am trying to connect to Apache Phoenix from the spark-shell, and the
following stack trace is getting thown.

val con =
DriverManager.getConnection("jdbc:phoenix:ec2-xx.xx.xx.xx.ap-southeast-1.compute.amazonaws.com:2181
:/hbase-unsecure")

2018-12-04 17:32:16 WARN  ConnectionImplementation:529 - Retrieve cluster
id failed
java.util.concurrent.ExecutionException:
org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException$NoNodeException:
KeeperErrorCode = NoNode for /hbase-unsecure/hbaseid
    at
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
    at
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
    at
org.apache.hadoop.hbase.client.ConnectionImplementation.retrieveClusterId(ConnectionImplementation.java:527)
    at
org.apache.hadoop.hbase.client.ConnectionImplementation.<init>(ConnectionImplementation.java:287)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    at
org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:219)
    at
org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:114)
    at
org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
    at
org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:430)
    at
org.apache.phoenix.query.ConnectionQueryServicesImpl.access$400(ConnectionQueryServicesImpl.java:272)
    at
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2556)
    at
org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2532)
    at
org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)
    at
org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2532)
    at
org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255)
    at
org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150)
    at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221)
    at java.sql.DriverManager.getConnection(DriverManager.java:664)
    at java.sql.DriverManager.getConnection(DriverManager.java:270)
    at $line17.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:24)
    at $line17.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:29)
    at $line17.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:31)
    at $line17.$read$$iw$$iw$$iw$$iw$$iw.<init>(<console>:33)
    at $line17.$read$$iw$$iw$$iw$$iw.<init>(<console>:35)
    at $line17.$read$$iw$$iw$$iw.<init>(<console>:37)
    at $line17.$read$$iw$$iw.<init>(<console>:39)
    at $line17.$read$$iw.<init>(<console>:41)
    at $line17.$read.<init>(<console>:43)
    at $line17.$read$.<init>(<console>:47)
    at $line17.$read$.<clinit>(<console>)
    at $line17.$eval$.$print$lzycompute(<console>:7)
    at $line17.$eval$.$print(<console>:6)
    at $line17.$eval.$print(<console>)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786)
    at
scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047)
    at
scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638)
    at
scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637)
    at
scala.reflect.internal.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:31)
    at
scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19)
    at
scala.tools.nsc.interpreter.IMain$WrappedRequest.loadAndRunReq(IMain.scala:637)
    at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569)
    at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)
    at
scala.tools.nsc.interpreter.ILoop.interpretStartingWith(ILoop.scala:807)
    at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)
    at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395)
    at scala.tools.nsc.interpreter.ILoop.loop(ILoop.scala:415)
    at
scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply$mcZ$sp(ILoop.scala:923)
    at
scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
    at
scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
    at
scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:97)
    at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)
    at org.apache.spark.repl.Main$.doMain(Main.scala:76)
    at org.apache.spark.repl.Main$.main(Main.scala:56)
    at org.apache.spark.repl.Main.main(Main.scala)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
    at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:879)
    at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:197)
    at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:227)
    at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:136)
    at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by:
org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException$NoNodeException:
KeeperErrorCode = NoNode for /hbase-unsecure/hbaseid
    at
org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
    at
org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
    at
org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$ZKTask$1.exec(ReadOnlyZKClient.java:168)
    at
org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient.run(ReadOnlyZKClient.java:323)
    at java.lang.Thread.run(Thread.java:748)

Re: JDBC Connection to Apache Phoenix failing

Posted by Ankit Singhal <an...@gmail.com>.
Is connecting and running some commands through HBase shell working? As per
the stack trace, It seems your HBase is not up  , Look at the master and
regionserver logs for errors.

On Tue, Dec 4, 2018 at 4:17 AM Raghavendra Channarayappa <
raghavendra.channarayappa@myntra.com> wrote:

> Can someone help with this? TIA
> I am trying to connect to Apache Phoenix from the spark-shell, and the
> following stack trace is getting thown.
>
> val con =
> DriverManager.getConnection("jdbc:phoenix:ec2-xx.xx.xx.xx.ap-southeast-1.compute.amazonaws.com:2181
> :/hbase-unsecure")
>
> 2018-12-04 17:32:16 WARN  ConnectionImplementation:529 - Retrieve cluster
> id failed
> java.util.concurrent.ExecutionException:
> org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException$NoNodeException:
> KeeperErrorCode = NoNode for /hbase-unsecure/hbaseid
>     at
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>     at
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
>     at
> org.apache.hadoop.hbase.client.ConnectionImplementation.retrieveClusterId(ConnectionImplementation.java:527)
>     at
> org.apache.hadoop.hbase.client.ConnectionImplementation.<init>(ConnectionImplementation.java:287)
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>     at
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>     at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>     at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>     at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:219)
>     at
> org.apache.hadoop.hbase.client.ConnectionFactory.createConnection(ConnectionFactory.java:114)
>     at
> org.apache.phoenix.query.HConnectionFactory$HConnectionFactoryImpl.createConnection(HConnectionFactory.java:47)
>     at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.openConnection(ConnectionQueryServicesImpl.java:430)
>     at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.access$400(ConnectionQueryServicesImpl.java:272)
>     at
> org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2556)
>     at
> org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2532)
>     at
> org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)
>     at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2532)
>     at
> org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255)
>     at
> org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150)
>     at
> org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221)
>     at java.sql.DriverManager.getConnection(DriverManager.java:664)
>     at java.sql.DriverManager.getConnection(DriverManager.java:270)
>     at $line17.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:24)
>     at $line17.$read$$iw$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:29)
>     at $line17.$read$$iw$$iw$$iw$$iw$$iw$$iw.<init>(<console>:31)
>     at $line17.$read$$iw$$iw$$iw$$iw$$iw.<init>(<console>:33)
>     at $line17.$read$$iw$$iw$$iw$$iw.<init>(<console>:35)
>     at $line17.$read$$iw$$iw$$iw.<init>(<console>:37)
>     at $line17.$read$$iw$$iw.<init>(<console>:39)
>     at $line17.$read$$iw.<init>(<console>:41)
>     at $line17.$read.<init>(<console>:43)
>     at $line17.$read$.<init>(<console>:47)
>     at $line17.$read$.<clinit>(<console>)
>     at $line17.$eval$.$print$lzycompute(<console>:7)
>     at $line17.$eval$.$print(<console>:6)
>     at $line17.$eval.$print(<console>)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:498)
>     at
> scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786)
>     at
> scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047)
>     at
> scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638)
>     at
> scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637)
>     at
> scala.reflect.internal.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:31)
>     at
> scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19)
>     at
> scala.tools.nsc.interpreter.IMain$WrappedRequest.loadAndRunReq(IMain.scala:637)
>     at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569)
>     at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)
>     at
> scala.tools.nsc.interpreter.ILoop.interpretStartingWith(ILoop.scala:807)
>     at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)
>     at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395)
>     at scala.tools.nsc.interpreter.ILoop.loop(ILoop.scala:415)
>     at
> scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply$mcZ$sp(ILoop.scala:923)
>     at
> scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
>     at
> scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)
>     at
> scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:97)
>     at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)
>     at org.apache.spark.repl.Main$.doMain(Main.scala:76)
>     at org.apache.spark.repl.Main$.main(Main.scala:56)
>     at org.apache.spark.repl.Main.main(Main.scala)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:498)
>     at
> org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
>     at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:879)
>     at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:197)
>     at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:227)
>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:136)
>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by:
> org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException$NoNodeException:
> KeeperErrorCode = NoNode for /hbase-unsecure/hbaseid
>     at
> org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
>     at
> org.apache.phoenix.shaded.org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
>     at
> org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$ZKTask$1.exec(ReadOnlyZKClient.java:168)
>     at
> org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient.run(ReadOnlyZKClient.java:323)
>     at java.lang.Thread.run(Thread.java:748)
>