You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@carbondata.apache.org by "Shoujie Zhuo (JIRA)" <ji...@apache.org> on 2016/07/20 06:35:20 UTC

[jira] [Issue Comment Deleted] (CARBONDATA-82) Dataload Failure then the csv file is big

     [ https://issues.apache.org/jira/browse/CARBONDATA-82?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Shoujie Zhuo updated CARBONDATA-82:
-----------------------------------
    Comment: was deleted

(was: This issue happens only on big file.
sql command:load data inpath 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' into table store_sales;
The file size is 742.2123709MB  with 5760750 rows
)

> Dataload Failure then the csv file is big
> -----------------------------------------
>
>                 Key: CARBONDATA-82
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-82
>             Project: CarbonData
>          Issue Type: Improvement
>          Components: carbon-spark
>    Affects Versions: Apache CarbonData 0.2.0-incubating
>            Reporter: Shoujie Zhuo
>             Fix For: Apache CarbonData 0.2.0-incubating
>
>
> > LOAD DATA  inpath 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table store_sales;
> INFO  20-07 13:43:39,249 - main Query [LOAD DATA  INPATH 'HDFS://HOLODESK01/USER/CARBON-SPARK-SQL/TPCDS/2/STORE_SALES' INTO TABLE STORE_SALES]
> INFO  20-07 13:43:39,307 - Successfully able to get the table metadata file lock
> INFO  20-07 13:43:39,324 - main Initiating Direct Load for the Table : (tpcds_carbon_2.store_sales)
> INFO  20-07 13:43:39,331 - [Block Distribution]
> INFO  20-07 13:43:39,332 - totalInputSpaceConsumed : 778266079 , defaultParallelism : 24
> INFO  20-07 13:43:39,332 - mapreduce.input.fileinputformat.split.maxsize : 32427753
> INFO  20-07 13:43:39,392 - Block broadcast_8 stored as values in memory (estimated size 264.0 KB, free 573.6 KB)
> INFO  20-07 13:43:39,465 - Block broadcast_8_piece0 stored as bytes in memory (estimated size 23.9 KB, free 597.4 KB)
> INFO  20-07 13:43:39,467 - Added broadcast_8_piece0 in memory on localhost:50762 (size: 23.9 KB, free: 511.4 MB)
> INFO  20-07 13:43:39,468 - Created broadcast 8 from NewHadoopRDD at CarbonTextFile.scala:45
> INFO  20-07 13:43:39,478 - Total input paths to process : 1
> INFO  20-07 13:43:39,493 - Starting job: take at CarbonCsvRelation.scala:175
> INFO  20-07 13:43:39,494 - Got job 5 (take at CarbonCsvRelation.scala:175) with 1 output partitions
> INFO  20-07 13:43:39,494 - Final stage: ResultStage 6 (take at CarbonCsvRelation.scala:175)
> INFO  20-07 13:43:39,494 - Parents of final stage: List()
> INFO  20-07 13:43:39,495 - Missing parents: List()
> INFO  20-07 13:43:39,496 - Submitting ResultStage 6 (MapPartitionsRDD[23] at map at CarbonTextFile.scala:55), which has no missing parents
> INFO  20-07 13:43:39,499 - Block broadcast_9 stored as values in memory (estimated size 2.6 KB, free 600.0 KB)
> INFO  20-07 13:43:39,511 - Block broadcast_9_piece0 stored as bytes in memory (estimated size 1600.0 B, free 601.5 KB)
> INFO  20-07 13:43:39,512 - Added broadcast_9_piece0 in memory on localhost:50762 (size: 1600.0 B, free: 511.4 MB)
> INFO  20-07 13:43:39,513 - Created broadcast 9 from broadcast at DAGScheduler.scala:1006
> INFO  20-07 13:43:39,514 - Submitting 1 missing tasks from ResultStage 6 (MapPartitionsRDD[23] at map at CarbonTextFile.scala:55)
> INFO  20-07 13:43:39,514 - Adding task set 6.0 with 1 tasks
> INFO  20-07 13:43:39,517 - Starting task 0.0 in stage 6.0 (TID 9, localhost, partition 0,ANY, 2302 bytes)
> INFO  20-07 13:43:39,518 - Running task 0.0 in stage 6.0 (TID 9)
> INFO  20-07 13:43:39,523 - Input split: hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales/data-m-00001.csv:0+32427753
> INFO  20-07 13:43:39,545 - Finished task 0.0 in stage 6.0 (TID 9). 3580 bytes result sent to driver
> INFO  20-07 13:43:39,558 - Finished task 0.0 in stage 6.0 (TID 9) in 42 ms on localhost (1/1)
> INFO  20-07 13:43:39,558 - ResultStage 6 (take at CarbonCsvRelation.scala:175) finished in 0.042 s
> INFO  20-07 13:43:39,558 - Removed TaskSet 6.0, whose tasks have all completed, from pool 
> INFO  20-07 13:43:39,558 - Job 5 finished: take at CarbonCsvRelation.scala:175, took 0.065209 s
> INFO  20-07 13:43:39,558 - Finished stage: org.apache.spark.scheduler.StageInfo@6c7379d3
> INFO  20-07 13:43:39,561 - task runtime:(count: 1, mean: 42.000000, stdev: 0.000000, max: 42.000000, min: 42.000000)
> INFO  20-07 13:43:39,561 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,561 - 	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms	42.0 ms
> INFO  20-07 13:43:39,563 - task result size:(count: 1, mean: 3580.000000, stdev: 0.000000, max: 3580.000000, min: 3580.000000)
> INFO  20-07 13:43:39,563 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,563 - 	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB	3.5 KB
> INFO  20-07 13:43:39,564 - have no column need to generate global dictionary
> AUDIT 20-07 13:43:39,564 - [holodesk01][hdfs][Thread-1]Data load request has been received for table tpcds_carbon_2.store_sales
> INFO  20-07 13:43:39,565 - executor (non-fetch) time pct: (count: 1, mean: 26.190476, stdev: 0.000000, max: 26.190476, min: 26.190476)
> INFO  20-07 13:43:39,565 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,565 - 	26 %	26 %	26 %	26 %	26 %	26 %	26 %	26 %	26 %
> INFO  20-07 13:43:39,567 - other time pct: (count: 1, mean: 73.809524, stdev: 0.000000, max: 73.809524, min: 73.809524)
> INFO  20-07 13:43:39,567 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:39,568 - 	74 %	74 %	74 %	74 %	74 %	74 %	74 %	74 %	74 %
> INFO  20-07 13:43:39,582 - main compaction need status is false
> INFO  20-07 13:43:39,583 - [Block Distribution]
> INFO  20-07 13:43:39,584 - totalInputSpaceConsumed : 778266079 , defaultParallelism : 24
> INFO  20-07 13:43:39,584 - mapreduce.input.fileinputformat.split.maxsize : 32427753
> INFO  20-07 13:43:39,586 - Total input paths to process : 1
> INFO  20-07 13:43:39,599 - Total no of blocks : 24, No.of Nodes : 4
> INFO  20-07 13:43:39,599 - #Node: holodesk02 no.of.blocks: 6
> #Node: holodesk01 no.of.blocks: 6
> #Node: holodesk04 no.of.blocks: 6
> #Node: holodesk03 no.of.blocks: 6
> INFO  20-07 13:43:40,605 - Starting job: collect at CarbonDataRDDFactory.scala:717
> INFO  20-07 13:43:40,606 - Got job 6 (collect at CarbonDataRDDFactory.scala:717) with 4 output partitions
> INFO  20-07 13:43:40,606 - Final stage: ResultStage 7 (collect at CarbonDataRDDFactory.scala:717)
> INFO  20-07 13:43:40,607 - Parents of final stage: List()
> INFO  20-07 13:43:40,607 - Missing parents: List()
> INFO  20-07 13:43:40,607 - Submitting ResultStage 7 (CarbonDataLoadRDD[24] at RDD at CarbonDataLoadRDD.scala:94), which has no missing parents
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk02
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk01
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk04
> INFO  20-07 13:43:40,608 - Prefered Location for split : holodesk03
> INFO  20-07 13:43:40,613 - Block broadcast_10 stored as values in memory (estimated size 15.8 KB, free 617.3 KB)
> INFO  20-07 13:43:40,625 - Block broadcast_10_piece0 stored as bytes in memory (estimated size 5.9 KB, free 623.2 KB)
> INFO  20-07 13:43:40,627 - Added broadcast_10_piece0 in memory on localhost:50762 (size: 5.9 KB, free: 511.4 MB)
> INFO  20-07 13:43:40,627 - Created broadcast 10 from broadcast at DAGScheduler.scala:1006
> INFO  20-07 13:43:40,628 - Submitting 4 missing tasks from ResultStage 7 (CarbonDataLoadRDD[24] at RDD at CarbonDataLoadRDD.scala:94)
> INFO  20-07 13:43:40,628 - Adding task set 7.0 with 4 tasks
> INFO  20-07 13:43:40,631 - Starting task 0.0 in stage 7.0 (TID 10, localhost, partition 0,ANY, 2892 bytes)
> INFO  20-07 13:43:40,632 - Starting task 1.0 in stage 7.0 (TID 11, localhost, partition 1,ANY, 2892 bytes)
> INFO  20-07 13:43:40,633 - Starting task 2.0 in stage 7.0 (TID 12, localhost, partition 2,ANY, 2892 bytes)
> INFO  20-07 13:43:40,634 - Starting task 3.0 in stage 7.0 (TID 13, localhost, partition 3,ANY, 2892 bytes)
> INFO  20-07 13:43:40,634 - Running task 0.0 in stage 7.0 (TID 10)
> INFO  20-07 13:43:40,635 - Running task 1.0 in stage 7.0 (TID 11)
> INFO  20-07 13:43:40,635 - Running task 2.0 in stage 7.0 (TID 12)
> INFO  20-07 13:43:40,635 - Running task 3.0 in stage 7.0 (TID 13)
> INFO  20-07 13:43:40,648 - Input split: holodesk04
> INFO  20-07 13:43:40,648 - The Block Count in this node :6
> INFO  20-07 13:43:40,649 - Input split: holodesk01
> INFO  20-07 13:43:40,649 - The Block Count in this node :6
> INFO  20-07 13:43:40,649 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - Input split: holodesk03
> INFO  20-07 13:43:40,650 - The Block Count in this node :6
> INFO  20-07 13:43:40,650 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,649 - Input split: holodesk02
> INFO  20-07 13:43:40,651 - The Block Count in this node :6
> INFO  20-07 13:43:40,651 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] ************* Is Columnar Storagetrue
> INFO  20-07 13:43:40,701 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Kettle environment initialized
> INFO  20-07 13:43:40,706 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Kettle environment initialized
> INFO  20-07 13:43:40,707 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Kettle environment initialized
> INFO  20-07 13:43:40,713 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Kettle environment initialized
> INFO  20-07 13:43:40,751 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] ** Using csv file **
> INFO  20-07 13:43:40,756 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] ** Using csv file **
> INFO  20-07 13:43:40,764 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,774 - [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,788 - [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,795 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Graph execution is started /mnt/disk1/spark/438978154880668/3/etl/tpcds_carbon_2/store_sales/0/3/store_sales.ktr
> INFO  20-07 13:43:40,798 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,809 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Graph execution is started /mnt/disk1/spark/438978153902729/1/etl/tpcds_carbon_2/store_sales/0/1/store_sales.ktr
> INFO  20-07 13:43:40,813 - [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,814 - [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2] *****************started csv reading by thread***********
> ERROR 20-07 13:43:40,819 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,819 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO  20-07 13:43:40,820 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,820 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] ** Using csv file **
> ERROR 20-07 13:43:40,821 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk1/spark/438978154880668/3/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/3
> INFO  20-07 13:43:40,841 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] ** Using csv file **
> INFO  20-07 13:43:40,854 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Graph execution is started /mnt/disk2/spark/438978155737218/0/etl/tpcds_carbon_2/store_sales/0/0/store_sales.ktr
> ERROR 20-07 13:43:40,854 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> ERROR 20-07 13:43:40,855 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk1/spark/438978153902729/1/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/1
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,855 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,856 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> INFO  20-07 13:43:40,857 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,867 - [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,869 - [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,872 - store_sales: Graph - CSV Input *****************Started all csv reading***********
> INFO  20-07 13:43:40,878 - [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,881 - [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2] *****************started csv reading by thread***********
> INFO  20-07 13:43:40,886 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Graph execution is started /mnt/disk1/spark/438978153678637/2/etl/tpcds_carbon_2/store_sales/0/2/store_sales.ktr
> ERROR 20-07 13:43:40,898 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,899 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk2/spark/438978155737218/0/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/0
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,899 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,900 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> ERROR 20-07 13:43:40,904 - [store_sales: Graph - Carbon Surrogate Key Generator][partitionID:0] 
> java.lang.NullPointerException
> 	at org.carbondata.processing.schema.metadata.ColumnSchemaDetailsWrapper.<init>(ColumnSchemaDetailsWrapper.java:75)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenMeta.initialize(CarbonCSVBasedSeqGenMeta.java:787)
> 	at org.carbondata.processing.surrogatekeysgenerator.csvbased.CarbonCSVBasedSeqGenStep.processRow(CarbonCSVBasedSeqGenStep.java:294)
> 	at org.pentaho.di.trans.step.RunThread.run(RunThread.java:50)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:40,906 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Record Processed For table: store_sales
> INFO  20-07 13:43:40,906 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Record Procerssed For table: store_sales
> ERROR 20-07 13:43:40,907 - [store_sales: Graph - MDKeyGenstore_sales][partitionID:0] Local data load folder location does not exist: /mnt/disk1/spark/438978153678637/2/tpcds_carbon_2/store_sales/Fact/Part0/Segment_0/2
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Number of Records was Zero
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Carbon Slice Mergerstore_sales][partitionID:sales] Summary: Carbon Slice Merger Step: Read: 0: Write: 0
> INFO  20-07 13:43:40,907 - [store_sales: Graph - Sort Key: Sort keysstore_sales][partitionID:0] Summary: Carbon Sort Key Step: Read: 0: Write: 0
> INFO  20-07 13:43:41,464 - Cleaned accumulator 18
> INFO  20-07 13:43:41,492 - Removed broadcast_8_piece0 on localhost:50762 in memory (size: 23.9 KB, free: 511.5 MB)
> INFO  20-07 13:43:41,497 - Removed broadcast_7_piece0 on localhost:50762 in memory (size: 23.9 KB, free: 511.5 MB)
> INFO  20-07 13:43:41,499 - Removed broadcast_9_piece0 on localhost:50762 in memory (size: 1600.0 B, free: 511.5 MB)
> INFO  20-07 13:43:49,599 - [pool-41-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:49,855 - [pool-41-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-41-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:49,957 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:49,957 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Graph execution is finished.
> ERROR 20-07 13:43:49,957 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] Graph Execution had errors
> ERROR 20-07 13:43:49,957 - [Executor task launch worker-6][partitionID:tpcds_carbon_2_store_sales_6302551d-dc77-4440-a26e-cbafb9d22c8c] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:49,958 - DataLoad failure
> INFO  20-07 13:43:49,969 - Finished task 1.0 in stage 7.0 (TID 11). 952 bytes result sent to driver
> INFO  20-07 13:43:49,982 - Finished task 1.0 in stage 7.0 (TID 11) in 9350 ms on localhost (1/4)
> INFO  20-07 13:43:50,482 - [pool-40-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:50,943 - [pool-42-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:51,270 - [pool-40-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-40-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:51,408 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:51,408 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Graph execution is finished.
> ERROR 20-07 13:43:51,409 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] Graph Execution had errors
> ERROR 20-07 13:43:51,409 - [Executor task launch worker-8][partitionID:tpcds_carbon_2_store_sales_94282d67-f4de-42dd-b61c-af8483cf3d21] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:51,409 - DataLoad failure
> INFO  20-07 13:43:51,420 - Finished task 3.0 in stage 7.0 (TID 13). 952 bytes result sent to driver
> INFO  20-07 13:43:51,434 - Finished task 3.0 in stage 7.0 (TID 13) in 10800 ms on localhost (2/4)
> INFO  20-07 13:43:51,435 - [pool-43-thread-2][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-2] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,466 - [pool-42-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-42-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,588 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:52,590 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Graph execution is finished.
> ERROR 20-07 13:43:52,590 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] Graph Execution had errors
> ERROR 20-07 13:43:52,590 - [Executor task launch worker-5][partitionID:tpcds_carbon_2_store_sales_3e4ba964-bcdc-4196-8d81-c590f2c67605] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:52,591 - DataLoad failure
> INFO  20-07 13:43:52,603 - Finished task 0.0 in stage 7.0 (TID 10). 952 bytes result sent to driver
> INFO  20-07 13:43:52,614 - Finished task 0.0 in stage 7.0 (TID 10) in 11984 ms on localhost (3/4)
> INFO  20-07 13:43:52,638 - [pool-43-thread-1][partitionID:PROCESS_BLOCKS;queryID:pool-43-thread-1] *****************Completed csv reading by thread***********
> INFO  20-07 13:43:52,824 - store_sales: Graph - CSV Input *****************Completed all csv reading***********
> INFO  20-07 13:43:52,824 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Graph execution is finished.
> ERROR 20-07 13:43:52,825 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] Graph Execution had errors
> ERROR 20-07 13:43:52,825 - [Executor task launch worker-7][partitionID:tpcds_carbon_2_store_sales_00be80d1-400a-425d-9c7f-4acf3b3a7bb3] 
> org.carbondata.processing.etl.DataLoadingException: Internal Errors
> 	at org.carbondata.processing.csvload.DataGraphExecuter.execute(DataGraphExecuter.java:253)
> 	at org.carbondata.processing.csvload.DataGraphExecuter.executeGraph(DataGraphExecuter.java:168)
> 	at org.carbondata.spark.load.CarbonLoaderUtil.executeGraph(CarbonLoaderUtil.java:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD$$anon$1.<init>(CarbonDataLoadRDD.scala:189)
> 	at org.carbondata.spark.rdd.CarbonDataLoadRDD.compute(CarbonDataLoadRDD.scala:148)
> 	at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
> 	at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
> 	at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> 	at org.apache.spark.scheduler.Task.run(Task.scala:89)
> 	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> INFO  20-07 13:43:52,825 - DataLoad failure
> INFO  20-07 13:43:52,837 - Finished task 2.0 in stage 7.0 (TID 12). 952 bytes result sent to driver
> INFO  20-07 13:43:52,849 - Finished task 2.0 in stage 7.0 (TID 12) in 12216 ms on localhost (4/4)
> INFO  20-07 13:43:52,849 - ResultStage 7 (collect at CarbonDataRDDFactory.scala:717) finished in 12.219 s
> INFO  20-07 13:43:52,849 - Removed TaskSet 7.0, whose tasks have all completed, from pool 
> INFO  20-07 13:43:52,849 - Finished stage: org.apache.spark.scheduler.StageInfo@46ffcf8b
> INFO  20-07 13:43:52,849 - Job 6 finished: collect at CarbonDataRDDFactory.scala:717, took 12.244086 s
> INFO  20-07 13:43:52,850 - ********starting clean up**********
> INFO  20-07 13:43:52,851 - task runtime:(count: 4, mean: 11087.500000, stdev: 1137.847419, max: 12216.000000, min: 9350.000000)
> INFO  20-07 13:43:52,851 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,851 - 	9.4 s	9.4 s	9.4 s	10.8 s	12.0 s	12.2 s	12.2 s	12.2 s	12.2 s
> INFO  20-07 13:43:52,853 - task result size:(count: 4, mean: 952.000000, stdev: 0.000000, max: 952.000000, min: 952.000000)
> INFO  20-07 13:43:52,853 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,853 - 	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B	952.0 B
> INFO  20-07 13:43:52,855 - executor (non-fetch) time pct: (count: 4, mean: 99.639701, stdev: 0.042276, max: 99.688933, min: 99.572193)
> INFO  20-07 13:43:52,855 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,855 - 	100 %	100 %	100 %	100 %	100 %	100 %	100 %	100 %	100 %
> INFO  20-07 13:43:52,857 - other time pct: (count: 4, mean: 0.360299, stdev: 0.042276, max: 0.427807, min: 0.311067)
> INFO  20-07 13:43:52,857 - 	0%	5%	10%	25%	50%	75%	90%	95%	100%
> INFO  20-07 13:43:52,857 - 	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %	 0 %
> INFO  20-07 13:43:53,079 - ********clean up done**********
> AUDIT 20-07 13:43:53,079 - [holodesk01][hdfs][Thread-1]Data load is failed for tpcds_carbon_2.store_sales
> WARN  20-07 13:43:53,080 - Unable to write load metadata file
> ERROR 20-07 13:43:53,080 - main 
> java.lang.Exception: Dataload failure
> 	at org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> 	at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> 	at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> 	at org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> 	at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> 	at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> AUDIT 20-07 13:43:53,081 - [holodesk01][hdfs][Thread-1]Dataload failure for tpcds_carbon_2.store_sales. Please check the logs
> INFO  20-07 13:43:53,083 - Table MetaData Unlocked Successfully after data load
> ERROR 20-07 13:43:53,083 - Failed in [LOAD DATA  inpath 'hdfs://holodesk01/user/carbon-spark-sql/tpcds/2/store_sales' INTO table store_sales]
> java.lang.Exception: Dataload failure
> 	at org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> 	at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> 	at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> 	at org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> 	at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> 	at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> java.lang.Exception: Dataload failure
> 	at org.carbondata.spark.rdd.CarbonDataRDDFactory$.loadCarbonData(CarbonDataRDDFactory.scala:779)
> 	at org.apache.spark.sql.execution.command.LoadTable.run(carbonTableSchema.scala:1146)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:58)
> 	at org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:56)
> 	at org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:70)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:132)
> 	at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:130)
> 	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> 	at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:130)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:55)
> 	at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:55)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:145)
> 	at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:130)
> 	at org.carbondata.spark.rdd.CarbonDataFrameRDD.<init>(CarbonDataFrameRDD.scala:23)
> 	at org.apache.spark.sql.CarbonContext.sql(CarbonContext.scala:109)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:311)
> 	at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
> 	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver$.main(CarbonSQLCLIDriver.scala:40)
> 	at org.apache.spark.sql.hive.cli.CarbonSQLCLIDriver.main(CarbonSQLCLIDriver.scala)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
> 	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> ```



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)