You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@carbondata.apache.org by yixu2001 <yi...@163.com> on 2018/04/12 06:17:16 UTC
carbon java.lang.NullPointerException
dev
spark 2.1.1 carbondate 1.1.1
scala> cc.sql("select count(*) from public.prod_offer_inst_cab").show;
18/04/12 11:47:17 AUDIT CarbonMetaStoreFactory: [hdd340][ip_crm][Thread-1]File based carbon metastore is enabled
java.lang.NullPointerException
at org.apache.carbondata.core.mutate.CarbonUpdateUtil.getSegmentBlockNameKey(CarbonUpdateUtil.java:801)
at org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager.populateMap(SegmentUpdateStatusManager.java:117)
at org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager.<init>(SegmentUpdateStatusManager.java:106)
at org.apache.spark.sql.execution.strategy.CarbonLateDecodeStrategy.driverSideCountStar(CarbonLateDecodeStrategy.scala:96)
at org.apache.spark.sql.execution.strategy.CarbonLateDecodeStrategy.apply(CarbonLateDecodeStrategy.scala:81)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:62)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:62)
at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:92)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:74)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:66)
at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:92)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:74)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:66)
at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:92)
at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(QueryExecution.scala:84)
at org.apache.spark.sql.execution.QueryExecution.sparkPlan(QueryExecution.scala:80)
at org.apache.spark.sql.execution.QueryExecution.executedPlan$lzycompute(QueryExecution.scala:89)
at org.apache.spark.sql.execution.QueryExecution.executedPlan(QueryExecution.scala:89)
at org.apache.spark.sql.Dataset.withTypedCallback(Dataset.scala:2814)
at org.apache.spark.sql.Dataset.head(Dataset.scala:2127)
at org.apache.spark.sql.Dataset.take(Dataset.scala:2342)
at org.apache.spark.sql.Dataset.showString(Dataset.scala:248)
at org.apache.spark.sql.Dataset.show(Dataset.scala:638)
at org.apache.spark.sql.Dataset.show(Dataset.scala:597)
at org.apache.spark.sql.Dataset.show(Dataset.scala:606)
... 50 elided
SegmentUpdateStatusManager.java
String blockIdentifier = CarbonUpdateUtil
.getSegmentBlockNameKey(blockDetails.getSegmentName(), blockDetails.getActualBlockName()); return null
yixu2001
Re: carbon java.lang.NullPointerException
Posted by Kumar Vishal <ku...@gmail.com>.
Hi Yixu2001,
Can u please provide the detailed step to reproduce this issue.
-Regards
Kumar Vishal
On Thu, Apr 12, 2018 at 9:16 PM, yixu2001 <yi...@163.com> wrote:
> dev
>
> version spark2.1.1 carbondata 1.3.1
>
>
> yixu2001
>
> From: yixu2001
> Date: 2018-04-12 14:17
> To: dev
> Subject: carbon java.lang.NullPointerException
> dev
> spark 2.1.1 carbondate 1.1.1
>
> scala> cc.sql("select count(*) from public.prod_offer_inst_cab").show;
> 18/04/12 11:47:17 AUDIT CarbonMetaStoreFactory:
> [hdd340][ip_crm][Thread-1]File based carbon metastore is enabled
> java.lang.NullPointerException
> at org.apache.carbondata.core.mutate.CarbonUpdateUtil.
> getSegmentBlockNameKey(CarbonUpdateUtil.java:801)
> at org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager.
> populateMap(SegmentUpdateStatusManager.java:117)
> at org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager.<
> init>(SegmentUpdateStatusManager.java:106)
> at org.apache.spark.sql.execution.strategy.CarbonLateDecodeStrategy.
> driverSideCountStar(CarbonLateDecodeStrategy.scala:96)
> at org.apache.spark.sql.execution.strategy.CarbonLateDecodeStrategy.
> apply(CarbonLateDecodeStrategy.scala:81)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$1.apply(QueryPlanner.scala:62)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$1.apply(QueryPlanner.scala:62)
> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(
> QueryPlanner.scala:92)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(
> TraversableOnce.scala:157)
> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(
> TraversableOnce.scala:157)
> at scala.collection.Iterator$class.foreach(Iterator.scala:893)
> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
> at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.
> scala:157)
> at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2.apply(QueryPlanner.scala:74)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2.apply(QueryPlanner.scala:66)
> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(
> QueryPlanner.scala:92)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(
> TraversableOnce.scala:157)
> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(
> TraversableOnce.scala:157)
> at scala.collection.Iterator$class.foreach(Iterator.scala:893)
> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
> at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.
> scala:157)
> at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2.apply(QueryPlanner.scala:74)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$
> anonfun$2.apply(QueryPlanner.scala:66)
> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(
> QueryPlanner.scala:92)
> at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(
> QueryExecution.scala:84)
> at org.apache.spark.sql.execution.QueryExecution.
> sparkPlan(QueryExecution.scala:80)
> at org.apache.spark.sql.execution.QueryExecution.
> executedPlan$lzycompute(QueryExecution.scala:89)
> at org.apache.spark.sql.execution.QueryExecution.
> executedPlan(QueryExecution.scala:89)
> at org.apache.spark.sql.Dataset.withTypedCallback(Dataset.scala:2814)
> at org.apache.spark.sql.Dataset.head(Dataset.scala:2127)
> at org.apache.spark.sql.Dataset.take(Dataset.scala:2342)
> at org.apache.spark.sql.Dataset.showString(Dataset.scala:248)
> at org.apache.spark.sql.Dataset.show(Dataset.scala:638)
> at org.apache.spark.sql.Dataset.show(Dataset.scala:597)
> at org.apache.spark.sql.Dataset.show(Dataset.scala:606)
> ... 50 elided
>
>
>
>
> SegmentUpdateStatusManager.java
> String blockIdentifier = CarbonUpdateUtil
> .getSegmentBlockNameKey(blockDetails.getSegmentName(),
> blockDetails.getActualBlockName()); return null
>
> yixu2001
>
Re: carbon java.lang.NullPointerException
Posted by yixu2001 <yi...@163.com>.
dev
version spark2.1.1 carbondata 1.3.1
yixu2001
From: yixu2001
Date: 2018-04-12 14:17
To: dev
Subject: carbon java.lang.NullPointerException
dev
spark 2.1.1 carbondate 1.1.1
scala> cc.sql("select count(*) from public.prod_offer_inst_cab").show;
18/04/12 11:47:17 AUDIT CarbonMetaStoreFactory: [hdd340][ip_crm][Thread-1]File based carbon metastore is enabled
java.lang.NullPointerException
at org.apache.carbondata.core.mutate.CarbonUpdateUtil.getSegmentBlockNameKey(CarbonUpdateUtil.java:801)
at org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager.populateMap(SegmentUpdateStatusManager.java:117)
at org.apache.carbondata.core.statusmanager.SegmentUpdateStatusManager.<init>(SegmentUpdateStatusManager.java:106)
at org.apache.spark.sql.execution.strategy.CarbonLateDecodeStrategy.driverSideCountStar(CarbonLateDecodeStrategy.scala:96)
at org.apache.spark.sql.execution.strategy.CarbonLateDecodeStrategy.apply(CarbonLateDecodeStrategy.scala:81)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:62)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:62)
at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:92)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:74)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:66)
at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:92)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:77)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:74)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:74)
at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:66)
at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:92)
at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(QueryExecution.scala:84)
at org.apache.spark.sql.execution.QueryExecution.sparkPlan(QueryExecution.scala:80)
at org.apache.spark.sql.execution.QueryExecution.executedPlan$lzycompute(QueryExecution.scala:89)
at org.apache.spark.sql.execution.QueryExecution.executedPlan(QueryExecution.scala:89)
at org.apache.spark.sql.Dataset.withTypedCallback(Dataset.scala:2814)
at org.apache.spark.sql.Dataset.head(Dataset.scala:2127)
at org.apache.spark.sql.Dataset.take(Dataset.scala:2342)
at org.apache.spark.sql.Dataset.showString(Dataset.scala:248)
at org.apache.spark.sql.Dataset.show(Dataset.scala:638)
at org.apache.spark.sql.Dataset.show(Dataset.scala:597)
at org.apache.spark.sql.Dataset.show(Dataset.scala:606)
... 50 elided
SegmentUpdateStatusManager.java
String blockIdentifier = CarbonUpdateUtil
.getSegmentBlockNameKey(blockDetails.getSegmentName(), blockDetails.getActualBlockName()); return null
yixu2001