You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@phoenix.apache.org by "Nico Pappagianis (JIRA)" <ji...@apache.org> on 2016/12/15 20:15:58 UTC

[jira] [Comment Edited] (PHOENIX-3532) Enable DataFrames and RDDs to read from a tenant-specific table

    [ https://issues.apache.org/jira/browse/PHOENIX-3532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15752394#comment-15752394 ] 

Nico Pappagianis edited comment on PHOENIX-3532 at 12/15/16 8:15 PM:
---------------------------------------------------------------------

I'm getting the same exception. I'm on a MBP OS X El Capitan.

{code:title=Bar.java|borderStyle=solid}
Formatting using clusterid: testClusterID
1    [ScalaTest-3] ERROR org.apache.hadoop.hdfs.MiniDFSCluster  - IOE creating namenodes. Permissions dump:
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/data': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/data
	permissions: ----
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis
	permissions: drwx
path '/Users/nico.pappagianis/git': 
	absolute:/Users/nico.pappagianis/git
	permissions: drwx
path '/Users/nico.pappagianis': 
	absolute:/Users/nico.pappagianis
	permissions: drwx
path '/Users': 
	absolute:/Users
	permissions: dr-x
path '/': 
	absolute:/
	permissions: dr-x

java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176)
	at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:742)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:585)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:982)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:863)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:845)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:832)
	at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:588)
	at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:509)
	at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:483)
	at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:561)
	at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:557)
	at org.apache.phoenix.end2end.BaseHBaseManagedTimeIT.doSetup(BaseHBaseManagedTimeIT.java:57)
	at org.apache.phoenix.spark.PhoenixSparkITHelper$.doSetup(AbstractPhoenixSparkIT.scala:33)
	at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:88)
	at org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)
	at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:44)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
	at org.apache.phoenix.spark.AbstractPhoenixSparkIT.run(AbstractPhoenixSparkIT.scala:44)
	at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Exception encountered when invoking run on a nested suite - java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current *** ABORTED ***
  java.lang.RuntimeException: java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current
  at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:591)
  at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:509)
  at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:483)
  at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:561)
  at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:557)
  at org.apache.phoenix.end2end.BaseHBaseManagedTimeIT.doSetup(BaseHBaseManagedTimeIT.java:57)
  at org.apache.phoenix.spark.PhoenixSparkITHelper$.doSetup(AbstractPhoenixSparkIT.scala:33)
  at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:88)
  at org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)
  at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:44)
  ...
  Cause: java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current
  at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337)
  at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548)
  at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569)
  at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161)
  at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
  at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342)
  at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176)
  at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973)
  at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811)
  at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:742)
  ...
{code}


was (Author: nico.pappagianis):
I'm getting the same exception. I'm on a MBP OS X El Capitan.

{{Formatting using clusterid: testClusterID
1    [ScalaTest-3] ERROR org.apache.hadoop.hdfs.MiniDFSCluster  - IOE creating namenodes. Permissions dump:
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/data': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/data
	permissions: ----
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis/phoenix': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis/phoenix
	permissions: drwx
path '/Users/nico.pappagianis/git/nico-pappagianis': 
	absolute:/Users/nico.pappagianis/git/nico-pappagianis
	permissions: drwx
path '/Users/nico.pappagianis/git': 
	absolute:/Users/nico.pappagianis/git
	permissions: drwx
path '/Users/nico.pappagianis': 
	absolute:/Users/nico.pappagianis
	permissions: drwx
path '/Users': 
	absolute:/Users
	permissions: dr-x
path '/': 
	absolute:/
	permissions: dr-x

java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176)
	at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:742)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:585)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:982)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:863)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:845)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:832)
	at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:588)
	at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:509)
	at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:483)
	at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:561)
	at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:557)
	at org.apache.phoenix.end2end.BaseHBaseManagedTimeIT.doSetup(BaseHBaseManagedTimeIT.java:57)
	at org.apache.phoenix.spark.PhoenixSparkITHelper$.doSetup(AbstractPhoenixSparkIT.scala:33)
	at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:88)
	at org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)
	at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:44)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
	at org.apache.phoenix.spark.AbstractPhoenixSparkIT.run(AbstractPhoenixSparkIT.scala:44)
	at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)
Exception encountered when invoking run on a nested suite - java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current *** ABORTED ***
  java.lang.RuntimeException: java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current
  at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:591)
  at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:509)
  at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:483)
  at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:561)
  at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:557)
  at org.apache.phoenix.end2end.BaseHBaseManagedTimeIT.doSetup(BaseHBaseManagedTimeIT.java:57)
  at org.apache.phoenix.spark.PhoenixSparkITHelper$.doSetup(AbstractPhoenixSparkIT.scala:33)
  at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:88)
  at org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)
  at org.apache.phoenix.spark.AbstractPhoenixSparkIT.beforeAll(AbstractPhoenixSparkIT.scala:44)
  ...
  Cause: java.io.IOException: Cannot create directory /Users/nico.pappagianis/git/nico-pappagianis/phoenix/phoenix-spark/target/test-data/cbc82c32-ae51-49c9-81a8-07f5ff4c8e47/dfscluster_d42892a5-e09d-42b7-a3b7-43a4499310ad/dfs/name1/current
  at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:337)
  at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:548)
  at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:569)
  at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:161)
  at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:991)
  at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:342)
  at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:176)
  at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:973)
  at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:811)
  at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:742)
  ...
}}

> Enable DataFrames and RDDs to read from a tenant-specific table
> ---------------------------------------------------------------
>
>                 Key: PHOENIX-3532
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-3532
>             Project: Phoenix
>          Issue Type: Bug
>            Reporter: Nico Pappagianis
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> Currently the method phoenixTableAsDataFrame in SparkSqlContextFunctions
> and phoenixTableAsRDD in SparkContextFunctions do not pass the tenantId parameter along to the PhoenixRDD constructor. The tenantId parameter was added as part of PHOENIX-3427 but was not properly implemented (by me). This JIRA will fix this issue and add tests around reading tenant-specific tables as both DataFrames and RDDs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)