You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@carbondata.apache.org by "Ravindra Pesala (JIRA)" <ji...@apache.org> on 2016/07/22 05:24:20 UTC

[jira] [Commented] (CARBONDATA-82) NullPointerException by ColumnSchemaDetailsWrapper.(ColumnSchemaDetailsWrapper.java:75)

    [ https://issues.apache.org/jira/browse/CARBONDATA-82?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15388939#comment-15388939 ] 

Ravindra Pesala commented on CARBONDATA-82:
-------------------------------------------

[~xiaoyesoso] Actually you used numeric datatypes for all columns, so carbondata treats all columns as measure but no dimension. That is why it is thrown error.  It is an issue as it supposed throw error at the time of creating the table it self.
Carbon requires atleast one dimension to work, so for this type scenario please use either DICTIONARY_INCLUDE or DICTIONARY_EXCLUDE for atleast one column to include that column as dimension.
You can use as follows.
{code}
create table if not exists store_sales
(
ss_sold_date_sk int,
ss_sold_time_sk int,
ss_item_sk int,
ss_customer_sk int,
ss_cdemo_sk int,
ss_hdemo_sk int,
ss_addr_sk int,
ss_store_sk int,
ss_promo_sk int,
ss_ticket_number int,
ss_quantity int,
ss_wholesale_cost double,
ss_list_price double,
ss_sales_price double,
ss_ext_discount_amt double,
ss_ext_sales_price double,
ss_ext_wholesale_cost double,
ss_ext_list_price double,
ss_ext_tax double,
ss_coupon_amt double,
ss_net_paid double,
ss_net_paid_inc_tax double,
ss_net_profit double
)
STORED BY 'org.apache.carbondata.format'
TBLPROPERTIES ('DICTIONARY_INCLUDE'='ss_sold_date_sk')
{code}

> NullPointerException by ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> ---------------------------------------------------------------------------------------------
>
>                 Key: CARBONDATA-82
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-82
>             Project: CarbonData
>          Issue Type: Bug
>          Components: carbon-spark
>    Affects Versions: Apache CarbonData 0.2.0-incubating
>            Reporter: Shoujie Zhuo
>             Fix For: Apache CarbonData 0.2.0-incubating
>
>
> csv file:
> ss_sold_date_sk,ss_sold_time_sk,ss_item_sk,ss_customer_sk,ss_cdemo_sk,ss_hdemo_sk,ss_addr_sk,ss_store_sk,ss_promo_sk,ss_ticket_number,ss_quantity,ss_wholesale_cost,ss_list_price,ss_sales_price,ss_ext_discount_amt,ss_ext_sales_price,ss_ext_wholesale_cost,ss_ext_list_price,ss_ext_tax,ss_coupon_amt,ss_net_paid,ss_net_paid_inc_tax,ss_net_profit
> 2451813,65495,7649,79006,591617,3428,44839,10,5,1,79,11.41,18.71,2.80,99.54,221.20,901.39,1478.09,6.08,99.54,121.66,127.74,-779.73
> DDL:
> create table if not exists  store_sales
> (
>     ss_sold_date_sk           int,
>     ss_sold_time_sk           int,
>     ss_item_sk                int,
>     ss_customer_sk            int,
>     ss_cdemo_sk               int,
>     ss_hdemo_sk               int,
>     ss_addr_sk                int,
>     ss_store_sk               int,
>     ss_promo_sk               int,
>     ss_ticket_number          int,
>     ss_quantity               int,
>     ss_wholesale_cost         double,
>     ss_list_price             double,
>     ss_sales_price            double,
>     ss_ext_discount_amt       double,
>     ss_ext_sales_price        double,
>     ss_ext_wholesale_cost     double,
>     ss_ext_list_price         double,
>     ss_ext_tax                double,
>     ss_coupon_amt             double,
>     ss_net_paid               double,
>     ss_net_paid_inc_tax       double,
>     ss_net_profit             double
> )
> STORED BY 'org.apache.carbondata.format';
> Log:
> > LOAD DATA  inpath 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table store_sales;
> INFO  20-07 13:43:39,249 - main Query [LOAD DATA  INPATH 'HDFS://HOLODESK01/USER/CARBON-SPARK-SQL/TPCDS/2/STORE_SALES' INTO TABLE STORE_SALES]
> INFO  20-07 13:43:39,307 - Successfully able to get the table metadata file lock
> INFO  20-07 13:43:39,324 - main Initiating Direct Load for the Table : (tpcds_carbon_2.store_sales)
> INFO  20-07 13:43:39,331 - [Block Distribution]
> INFO  20-07 13:43:39,332 - totalInputSpaceConsumed : 778266079 , defaultParallelism : 24
> INFO  20-07 13:43:39,332 - mapreduce.input.fileinputformat.split.maxsize : 32427753
> INFO  20-07 13:43:39,392 - Block broadcast_8 stored as values in memory (estimated size 264.0 KB, free 573.6 KB)
> INFO  20-07 13:43:39,465 - Block broadcast_8_piece0 stored as bytes in memory (estimated size 23.9 KB, free 597.4 KB)
> INFO  20-07 13:43:39,467 - Added broadcast_8_piece0 in memory on localhost:50762 (size: 23.9 KB, free: 511.4 MB)
> INFO  20-07 13:43:39,468 - Created broadcast 8 from NewHadoopRDD at CarbonTextFile.scala:45
> INFO  20-07 13:43:39,478 - Total input paths to process : 1
> INFO  20-07 13:43:39,493 - Starting job: take at CarbonCsvRelation.scala:175
> INFO  20-07 13:43:39,494 - Got job 5 (take at CarbonCsvRelation.scala:175) with 1 output partitions
> INFO  20-07 13:43:39,494 - Final stage: ResultStage 6 (take at CarbonCsvRelation.scala:175)
> INFO  20-07 13:43:39,494 - Parents of final stage: List()
> INFO  20-07 13:43:39,495 - Missing parents: List()
> INFO  20-07 13:43:39,496 - Submitting ResultStage 6 (MapPartitionsRDD[23] at map at CarbonTextFile.scala:55), which has no missing parents
> INFO  20-07 13:43:39,499 - Block broadcast_9 stored as values in memory (estimated size 2.6 KB, free 600.0 KB)
> INFO  20-07 13:43:39,511 - Block broadcast_9_piece0 stored as bytes in memory (estimated size 1600.0 B, free 601.5 KB)
> INFO  20-07 13:43:39,512 - Added broadcast_9_piece0 in memory on localhost:50762 (size: 1600.0 B, free: 511.4 MB)
> INFO  20-07 13:43:39,513 - Created broadcast 9 from broadcast at DAGScheduler.scala:1006
> INFO  20-07 13:43:39,514 - Submitting 1 missing tasks from ResultStage 6 (MapPartitionsRDD[23] at map at CarbonTextFile.scala:55)
> INFO  20-07 13:43:39,514 - Adding task set 6.0 with 1 tasks
> INFO  20-07 13:43:39,517 - Starting task 0.0 in stage 6.0 (TID 9, localhost, partition 0,ANY, 2302 bytes)
> INFO  20-07 13:43:39,518 - Running task 0.0 in stage 6.0 (TID 9)
> INFO  20-07 13:43:39,523 - Input split: hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales/data-m-00001.csv:0+32427753
> INFO  20-07 13:43:39,545 - Finished task 0.0 in stage 6.0 (TID 9). 3580 bytes result sent to driver
> INFO  20-07 13:43:39,558 - Finished task 0.0 in stage 6.0 (TID 9) in 42 ms on localhost (1/1)
> INFO  20-07 13:43:39,558 - ResultStage 6 (take at CarbonCsvRelation.scala:175) finished in 0.042 s
> INFO  20-07 13:43:39,558 - Removed TaskSet 6.0, whose tasks have all completed, from pool 
> INFO  20-07 13:43:39,558 - Job 5 finished: take at CarbonCsvRelation.scala:175, took 0.065209 s
> INFO  20-07 13:43:39,558 - Finished stage: org.apache.spark.scheduler.StageInfo@6c7379d3
> INFO  20-07 13:43:39,561 - task runtime:(count: 1, mean: 42.000000, stdev: 0.000000, max: 42.000000, min: 42.000000)
> INFO  20-07 13:43:39,561 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,561 - 	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms
> INFO  20-07 13:43:39,563 - task result size:(count: 1, mean: 3580.000000, stdev: 0.000000, max: 3580.000000, min: 3580.000000)
> INFO  20-07 13:43:39,563 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,563 - 	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB
> INFO  20-07 13:43:39,564 - have no column need to generate global dictionary
> AUDIT 20-07 13:43:39,564 - [holodesk01][hdfs][Thread-1]Data load request has been received for table tpcds_carbon_2.store_sales
> INFO  20-07 13:43:39,565 - executor (non-fetch) time pct: (count: 1, mean: 26.190476, stdev: 0.000000, max: 26.190476, min: 26.190476)
> INFO  20-07 13:43:39,565 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,565 - 	26 %	26 %	26 %	26 %	26 %	26 %	26 %	26 %	26 %
> INFO  20-07 13:43:39,567 - other time pct: (count: 1, mean: 73.809524, stdev: 0.000000, max: 73.809524, min: 73.809524)
> INFO  20-07 13:43:39,567 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,568 - 	74 %	74 %	74 %	74 %	74 %	74 %	74 %	74 %	74 %
> INFO  20-07 13:43:39,582 - main compaction need status is false
> INFO  20-07 13:43:39,583 - [Block Distribution]
> INFO  20-07 13:43:39,584 - totalInputSpaceConsumed : 778266079 , defaultParallelism : 24
> INFO  20-07 13:43:39,584 - mapreduce.input.fileinputformat.split.maxsize : 32427753
> INFO  20-07 13:43:39,586 - Total input paths to process : 1
> INFO  20-07 13:43:39,599 - Total no of blocks : 24, No.of Nodes : 4
> INFO  20-07 13:43:39,599 - #Node: holodesk02 no.of.blocks: 6
> #Node: holodesk01 no.of.blocks: 6
> #Node: holodesk04 no.of.blocks: 6
> #Node: holodesk03 no.of.blocks: 6
> INFO  20-07 13:43:40,605 - Starting job: collect at CarbonDataRDDFactory.scala:717
> INFO  20-07 13:43:40,606 - Got job 6 (collect at CarbonDataRDDFactory.scala:717) with 4 output partitions
> INFO  20-07 13:43:40,606 - Final stage: ResultStage 7 (collect at CarbonDataRDDFactory.scala:717)
> INFO  20-07 13:43:40,607 - Parents of final stage: List()
> INFO  20-07 13:43:40,607 - Missing parents: List()
> INFO  20-07 13:43:40,607 - Submitting ResultStage 7 (CarbonDataLoadRDD[24] at RDD at CarbonDataLoadRDD.scala:94), which has no missing parents
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk02
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk01
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk04
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk03
> INFO  20-07 13:43:40,613 - Block broadcast_10 stored as values in memory (estimated size 15.8 KB, free 617.3 KB)
> INFO  20-07 13:43:40,625 - Block broadcast_10_piece0 stored as bytes in memory (estimated size 5.9 KB, free 623.2 KB)
> INFO  20-07 13:43:40,627 - Added broadcast_10_piece0 in memory on localhost:50762 (size: 5.9 KB, free: 511.4 MB)
> INFO  20-07 13:43:40,627 - Created broadcast 10 from broadcast at DAGScheduler.scala:1006
> INFO  20-07 13:43:40,628 - Submitting 4 missing tasks from ResultStage 7 (CarbonDataLoadRDD[24] at RDD at CarbonDataLoadRDD.scala:94)
> INFO  20-07 13:43:40,628 - Adding task set 7.0 with 4 tasks
> INFO  20-07 13:43:40,631 - Starting task 0.0 in stage 7.0 (TID 10, localhost, partition 0,ANY, 2892 bytes)
> INFO  20-07 13:43:40,632 - Starting task 1.0 in stage 7.0 (TID 11, localhost, partition 1,ANY, 2892 bytes)
> INFO  20-07 13:43:40,633 - Starting task 2.0 in stage 7.0 (TID 12, localhost, partition 2,ANY, 2892 bytes)
> INFO  20-07 13:43:40,634 - Starting task 3.0 in stage 7.0 (TID 13, localhost, partition 3,ANY, 2892 bytes)
> INFO  20-07 13:43:40,634 - Running task 0.0 in stage 7.0 (TID 10)
> INFO  20-07 13:43:40,635 - Running task 1.0 in stage 7.0 (TID 11)
> INFO  20-07 13:43:40,635 - Running task 2.0 in stage 7.0 (TID 12)
> INFO  20-07 13:43:40,635 - Running task 3.0 in stage 7.0 (TID 13)
> INFO  20-07 13:43:40,648 - Input split: holodesk04
> INFO  20-07 13:43:40,648 - The Block Count in this node :6
> INFO  20-07 13:43:40,649 - Input split: holodesk01
> INFO  20-07 13:43:40,649 - The Block Count in this node :6
> INFO  20-07 13:43:40,649 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - Input split: holodesk03
> INFO  20-07 13:43:40,650 - The Block Count in this node :6
> INFO  20-07 13:43:40,650 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - Input split: holodesk02
> INFO  20-07 13:43:40,651 - The Block Count in this node :6
> INFO  20-07 13:43:40,651 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,701 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Kettle environment initialized
> INFO  20-07 13:43:40,706 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Kettle environment initialized
> INFO  20-07 13:43:40,707 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Kettle environment initialized
> INFO  20-07 13:43:40,713 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Kettle environment initialized
> INFO  20-07 13:43:40,751 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] ** Using csv file **
> INFO  20-07 13:43:40,756 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] ** Using csv file **
> INFO  20-07 13:43:40,764 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,774 - [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,788 - [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,795 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Graph execution is started /mnt/disk1/spark/438978154880668/3/etl/tpcds_carbon_2/store_sales/0/3/store_sales.ktr
> INFO  20-07 13:43:40,798 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,809 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Graph execution is started /mnt/disk1/spark/438978153902729/1/etl/tpcds_carbon_2/store_sales/0/1/store_sales.ktr
> INFO  20-07 13:43:40,813 - [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,814 - [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2] *****************started csv reading by thread***********
> ERROR 20-07 13:43:40,819 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO  20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,820 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] ** Using csv file **
> ERROR 20-07 13:43:40,821 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk1/spark/438978154880668/3/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/3
> INFO  20-07 13:43:40,841 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] ** Using csv file **
> INFO  20-07 13:43:40,854 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Graph execution is started /mnt/disk2/spark/438978155737218/0/etl/tpcds_carbon_2/store_sales/0/0/store_sales.ktr
> ERROR 20-07 13:43:40,854 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> ERROR 20-07 13:43:40,855 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk1/spark/438978153902729/1/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/1
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,856 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO  20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,867 - [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,869 - [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,872 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,878 - [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,881 - [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,886 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Graph execution is started /mnt/disk1/spark/438978153678637/2/etl/tpcds_carbon_2/store_sales/0/2/store_sales.ktr
> ERROR 20-07 13:43:40,898 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,899 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk2/spark/438978155737218/0/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/0
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,900 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> ERROR 20-07 13:43:40,904 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,906 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,906 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,907 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk1/spark/438978153678637/2/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/2
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> INFO  20-07 13:43:41,464 - Cleaned accumulator 18
> INFO  20-07 13:43:41,492 - Removed broadcast_8_piece0 on localhost:50762 in memory (size: 23.9 KB, free: 511.5 MB)
> INFO  20-07 13:43:41,497 - Removed broadcast_7_piece0 on localhost:50762 in memory (size: 23.9 KB, free: 511.5 MB)
> INFO  20-07 13:43:41,499 - Removed broadcast_9_piece0 on localhost:50762 in memory (size: 1600.0 B, free: 511.5 MB)
> INFO  20-07 13:43:49,599 - [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:49,855 - [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:49,957 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:49,957 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Graph execution is finished.
> ERROR 20-07 13:43:49,957 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Graph Execution had errors
> ERROR 20-07 13:43:49,957 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:49,958 - DataLoad failure
> INFO  20-07 13:43:49,969 - Finished task 1.0 in stage 7.0 (TID 11). 952 bytes result sent to driver
> INFO  20-07 13:43:49,982 - Finished task 1.0 in stage 7.0 (TID 11) in 9350 ms on localhost (1/4)
> INFO  20-07 13:43:50,482 - [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:50,943 - [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:51,270 - [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:51,408 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:51,408 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Graph execution is finished.
> ERROR 20-07 13:43:51,409 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Graph Execution had errors
> ERROR 20-07 13:43:51,409 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:51,409 - DataLoad failure
> INFO  20-07 13:43:51,420 - Finished task 3.0 in stage 7.0 (TID 13). 952 bytes result sent to driver
> INFO  20-07 13:43:51,434 - Finished task 3.0 in stage 7.0 (TID 13) in 10800 ms on localhost (2/4)
> INFO  20-07 13:43:51,435 - [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,466 - [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,588 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:52,590 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Graph execution is finished.
> ERROR 20-07 13:43:52,590 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Graph Execution had errors
> ERROR 20-07 13:43:52,590 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:52,591 - DataLoad failure
> INFO  20-07 13:43:52,603 - Finished task 0.0 in stage 7.0 (TID 10). 952 bytes result sent to driver
> INFO  20-07 13:43:52,614 - Finished task 0.0 in stage 7.0 (TID 10) in 11984 ms on localhost (3/4)
> INFO  20-07 13:43:52,638 - [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,824 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:52,824 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Graph execution is finished.
> ERROR 20-07 13:43:52,825 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Graph Execution had errors
> ERROR 20-07 13:43:52,825 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:52,825 - DataLoad failure
> INFO  20-07 13:43:52,837 - Finished task 2.0 in stage 7.0 (TID 12). 952 bytes result sent to driver
> INFO  20-07 13:43:52,849 - Finished task 2.0 in stage 7.0 (TID 12) in 12216 ms on localhost (4/4)
> INFO  20-07 13:43:52,849 - ResultStage 7 (collect at CarbonDataRDDFactory.scala:717) finished in 12.219 s
> INFO  20-07 13:43:52,849 - Removed TaskSet 7.0, whose tasks have all completed, from pool 
> INFO  20-07 13:43:52,849 - Finished stage: org.apache.spark.scheduler.StageInfo@46ffcf8b
> INFO  20-07 13:43:52,849 - Job 6 finished: collect at CarbonDataRDDFactory.scala:717, took 12.244086 s
> INFO  20-07 13:43:52,850 - ********starting clean up**********
> INFO  20-07 13:43:52,851 - task runtime:(count: 4, mean: 11087.500000, stdev: 1137.847419, max: 12216.000000, min: 9350.000000)
> INFO  20-07 13:43:52,851 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,851 - 	9.4 s	9.4 s	9.4 s	10.8 s	12.0 s	12.2 s	12.2 s	12.2 s	12.2 s
> INFO  20-07 13:43:52,853 - task result size:(count: 4, mean: 952.000000, stdev: 0.000000, max: 952.000000, min: 952.000000)
> INFO  20-07 13:43:52,853 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,853 - 	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B
> INFO  20-07 13:43:52,855 - executor (non-fetch) time pct: (count: 4, mean: 99.639701, stdev: 0.042276, max: 99.688933, min: 99.572193)
> INFO  20-07 13:43:52,855 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,855 - 	100 %	100 %	100 %	100 %	100 %	100 %	100 %	100 %	100 %
> INFO  20-07 13:43:52,857 - other time pct: (count: 4, mean: 0.360299, stdev: 0.042276, max: 0.427807, min: 0.311067)
> INFO  20-07 13:43:52,857 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,857 - 	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %
> INFO  20-07 13:43:53,079 - ********clean up done**********
> AUDIT 20-07 13:43:53,079 - [holodesk01][hdfs][Thread-1]Data load is failed for tpcds_carbon_2.store_sales
> WARN  20-07 13:43:53,080 - Unable to write load metadata file
> ERROR 20-07 13:43:53,080 - main 
> java.lang.Exception: Dataload failure
> 	at org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> 	at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> 	at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> 	at org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> 	at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> 	at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> AUDIT 20-07 13:43:53,081 - [holodesk01][hdfs][Thread-1]Dataload failure for tpcds_carbon_2.store_sales. Please check the logs
> INFO  20-07 13:43:53,083 - Table MetaData Unlocked Successfully after data load
> ERROR 20-07 13:43:53,083 - Failed in [LOAD DATA  inpath 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table store_sales]
> java.lang.Exception: Dataload failure
> 	at org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> 	at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> 	at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> 	at org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> 	at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> 	at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> java.lang.Exception: Dataload failure
> 	at org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> 	at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> 	at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> 	at org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> 	at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> 	at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> ```



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)