You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@carbondata.apache.org by "Vandana Yadav (JIRA)" <ji...@apache.org> on 2018/01/25 09:39:00 UTC
[jira] [Created] (CARBONDATA-2079) Error displays while executing
minor compaction on the cluster
Vandana Yadav created CARBONDATA-2079:
-----------------------------------------
Summary: Error displays while executing minor compaction on the cluster
Key: CARBONDATA-2079
URL: https://issues.apache.org/jira/browse/CARBONDATA-2079
Project: CarbonData
Issue Type: Bug
Components: data-load
Affects Versions: 1.3.0
Environment: spark 2.1
Reporter: Vandana Yadav
Attachments: 7000_UniqData.csv
Error displays while executing minor compaction on the cluster:
Steps to Reproduce:
1) Create Table:
CREATE TABLE uniqdata_batchsort_compact (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'carbondata' TBLPROPERTIES('SORT_SCOPE'='BATCH_SORT')
2) Load Data :
LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/7000_UniqData.csv' into table uniqdata_batchsort_compact OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','batch_sort_size_inmb'='1')
LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/7000_UniqData.csv' into table uniqdata_batchsort_compact OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','batch_sort_size_inmb'='1')
LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/7000_UniqData.csv' into table uniqdata_batchsort_compact OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','batch_sort_size_inmb'='1')
LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/uniqdata/7000_UniqData.csv' into table uniqdata_batchsort_compact OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1','batch_sort_size_inmb'='1')
3)Execute Queries:
alter table uniqdata_batchsort_compact compact 'minor'
output:
Compaction failed. Please check logs for more info. Exception in compaction / by zero
logs:
[exec] 18/01/24 16:47:33 INFO SelectQuery: Executing Query: alter table uniqdata_batchsort_compact compact 'minor'
[exec] 18/01/24 16:47:33 INFO CarbonSparkSqlParser: Parsing command: alter table uniqdata_batchsort_compact compact 'minor'
[exec] 18/01/24 16:47:33 INFO CarbonLateDecodeRule: main skip CarbonOptimizer
[exec] 18/01/24 16:47:33 INFO CarbonLateDecodeRule: main Skip CarbonOptimizer
[exec] 18/01/24 16:47:33 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_batchsort_compact
[exec] 18/01/24 16:47:33 INFO audit: ugi=root ip=unknown-ip-addr cmd=get_table : db=default tbl=uniqdata_batchsort_compact
[exec] 18/01/24 16:47:33 INFO CatalystSqlParser: Parsing command: array<string>
[exec] 18/01/24 16:47:33 INFO HiveMetaStore: 0: get_table : db=default tbl=uniqdata_batchsort_compact
[exec] 18/01/24 16:47:33 INFO audit: ugi=root ip=unknown-ip-addr cmd=get_table : db=default tbl=uniqdata_batchsort_compact
[exec] 18/01/24 16:47:33 INFO CatalystSqlParser: Parsing command: array<string>
[exec] 18/01/24 16:47:33 AUDIT CarbonAlterTableCompactionCommand: [hadoop-master][root][Thread-1]Compaction request received for table default.uniqdata_batchsort_compact
[exec] 18/01/24 16:47:33 INFO HdfsFileLock: main HDFS lock path:hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/compaction.lock
[exec] 18/01/24 16:47:34 INFO CarbonAlterTableCompactionCommand: main Acquired the compaction lock for table default.uniqdata_batchsort_compact
[exec] 18/01/24 16:47:34 INFO CarbonTableCompactor: main loads identified for merge is 0
[exec] 18/01/24 16:47:34 INFO CarbonTableCompactor: main loads identified for merge is 1
[exec] 18/01/24 16:47:34 INFO CarbonTableCompactor: main loads identified for merge is 2
[exec] 18/01/24 16:47:34 INFO CarbonTableCompactor: main loads identified for merge is 3
[exec] 18/01/24 16:47:34 INFO CarbonTableCompactor: main spark.executor.instances property is set to = 3
[exec] 18/01/24 16:47:34 INFO TableInfo: main Table block size not specified for default_uniqdata_batchsort_compact. Therefore considering the default value 1024 MB
[exec] 18/01/24 16:47:34 INFO BlockletDataMap: main Time taken to load blocklet datamap from file : hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/Fact/Part0/Segment_0/0_batchno0-0-1516792651041.carbonindexis 0
[exec] 18/01/24 16:47:34 INFO BlockletDataMap: main Time taken to load blocklet datamap from file : hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/Fact/Part0/Segment_1/0_batchno0-0-1516792651780.carbonindexis 1
[exec] 18/01/24 16:47:34 INFO BlockletDataMap: main Time taken to load blocklet datamap from file : hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/Fact/Part0/Segment_2/0_batchno0-0-1516792652481.carbonindexis 0
[exec] 18/01/24 16:47:34 INFO BlockletDataMap: main Time taken to load blocklet datamap from file : hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/Fact/Part0/Segment_3/0_batchno0-0-1516792653232.carbonindexis 0
[exec] 18/01/24 16:47:34 ERROR CarbonTableCompactor: main Exception in compaction thread / by zero
[exec] java.lang.ArithmeticException: / by zero
[exec] at org.apache.carbondata.processing.util.CarbonLoaderUtil.nodeBlockMapping(CarbonLoaderUtil.java:524)
[exec] at org.apache.carbondata.processing.util.CarbonLoaderUtil.nodeBlockMapping(CarbonLoaderUtil.java:453)
[exec] at org.apache.carbondata.spark.rdd.CarbonMergerRDD.getPartitions(CarbonMergerRDD.scala:400)
[exec] at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:252)
[exec] at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:250)
[exec] at scala.Option.getOrElse(Option.scala:121)
[exec] at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
[exec] at org.apache.spark.SparkContext.runJob(SparkContext.scala:1958)
[exec] at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:935)
[exec] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
[exec] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
[exec] at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
[exec] at org.apache.spark.rdd.RDD.collect(RDD.scala:934)
[exec] at org.apache.carbondata.spark.rdd.CarbonTableCompactor.triggerCompaction(CarbonTableCompactor.scala:211)
[exec] at org.apache.carbondata.spark.rdd.CarbonTableCompactor.scanSegmentsAndSubmitJob(CarbonTableCompactor.scala:120)
[exec] at org.apache.carbondata.spark.rdd.CarbonTableCompactor.executeCompaction(CarbonTableCompactor.scala:71)
[exec] at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$$anon$2.run(CarbonDataRDDFactory.scala:182)
[exec] at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.startCompactionThreads(CarbonDataRDDFactory.scala:269)
[exec] at org.apache.spark.sql.execution.command.management.CarbonAlterTableCompactionCommand.alterTableForCompaction(CarbonAlterTableCompactionCommand.scala:258)
[exec] at org.apache.spark.sql.execution.command.management.CarbonAlterTableCompactionCommand.processData(CarbonAlterTableCompactionCommand.scala:111)
[exec] at org.apache.spark.sql.execution.command.DataCommand.run(package.scala:71)
[exec] at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
[exec] at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
[exec] at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
[exec] at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
[exec] at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
[exec] at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
[exec] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
[exec] at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
[exec] at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
[exec] at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
[exec] at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
[exec] at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
[exec] at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
[exec] at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
[exec] at com.huawei.spark.SessionManager.sql(SessionManager.java:42)
[exec] at com.huawei.querymanagement.QueryManagement.sql(QueryManagement.java:62)
[exec] at com.huawei.querymanagement.SelectQuery.testQuery(SelectQuery.java:70)
[exec] at sun.reflect.GeneratedMethodAccessor66.invoke(Unknown Source)
[exec] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[exec] at java.lang.reflect.Method.invoke(Method.java:498)
[exec] at org.junit.internal.runners.TestMethod.invoke(TestMethod.java:59)
[exec] at org.junit.internal.runners.MethodRoadie.runTestMethod(MethodRoadie.java:98)
[exec] at org.junit.internal.runners.MethodRoadie$2.run(MethodRoadie.java:79)
[exec] at org.junit.internal.runners.MethodRoadie.runBeforesThenTestThenAfters(MethodRoadie.java:87)
[exec] at org.junit.internal.runners.MethodRoadie.runTest(MethodRoadie.java:77)
[exec] at org.junit.internal.runners.MethodRoadie.run(MethodRoadie.java:42)
[exec] at org.junit.internal.runners.JUnit4ClassRunner.invokeTestMethod(JUnit4ClassRunner.java:88)
[exec] at org.junit.internal.runners.JUnit4ClassRunner.runMethods(JUnit4ClassRunner.java:51)
[exec] at org.junit.runners.Parameterized$TestClassRunnerForParameters.run(Parameterized.java:98)
[exec] at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
[exec] at org.junit.runners.Parameterized.access$000(Parameterized.java:55)
[exec] at org.junit.runners.Parameterized$1.run(Parameterized.java:131)
[exec] at org.junit.internal.runners.ClassRoadie.runUnprotected(ClassRoadie.java:27)
[exec] at org.junit.internal.runners.ClassRoadie.runProtected(ClassRoadie.java:37)
[exec] at org.junit.runners.Parameterized.run(Parameterized.java:129)
[exec] at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
[exec] at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:28)
[exec] at org.junit.runner.JUnitCore.run(JUnitCore.java:130)
[exec] at org.junit.runner.JUnitCore.run(JUnitCore.java:109)
[exec] at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
[exec] at org.junit.runner.JUnitCore.runClasses(JUnitCore.java:60)
[exec] at com.huawei.querymanagement.SelectQuerySuite.main(SelectQuerySuite.java:18)
[exec] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[exec] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[exec] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[exec] at java.lang.reflect.Method.invoke(Method.java:498)
[exec] at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
[exec] at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
[exec] at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
[exec] at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
[exec] at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
[exec] 18/01/24 16:47:34 ERROR CarbonDataRDDFactory$: main Exception in compaction thread / by zero
[exec] 18/01/24 16:47:34 INFO HdfsFileLock: main Deleted the lock file hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/compaction.lock
[exec] 18/01/24 16:47:34 ERROR CarbonAlterTableCompactionCommand: main Exception in start compaction thread. Exception in compaction / by zero
[exec] 18/01/24 16:47:34 ERROR HdfsFileLock: main Not able to delete the lock file because it is not existed in location hdfs://hadoop-master:54311//opt/CarbonStore/default/uniqdata_batchsort_compact/compaction.lock
[exec] 18/01/24 16:47:34 ERROR SelectQuery: An exception has occurred:
[exec] org.apache.spark.sql.AnalysisException: Compaction failed. Please check logs for more info. Exception in compaction / by zero;
[exec] at org.apache.spark.sql.util.CarbonException$.analysisException(CarbonException.scala:23)
[exec] at org.apache.spark.sql.execution.command.management.CarbonAlterTableCompactionCommand.processData(CarbonAlterTableCompactionCommand.scala:120)
[exec] at org.apache.spark.sql.execution.command.DataCommand.run(package.scala:71)
[exec] at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
[exec] at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
[exec] at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
[exec] at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
[exec] at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
[exec] at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
[exec] at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
[exec] at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
[exec] at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
[exec] at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
[exec] at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:87)
[exec] at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
[exec] at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
[exec] at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592)
[exec] at com.huawei.spark.SessionManager.sql(SessionManager.java:42)
[exec] at com.huawei.querymanagement.QueryManagement.sql(QueryManagement.java:62)
[exec] at com.huawei.querymanagement.SelectQuery.testQuery(SelectQuery.java:70)
[exec] at sun.reflect.GeneratedMethodAccessor66.invoke(Unknown Source)
[exec] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[exec] at java.lang.reflect.Method.invoke(Method.java:498)
[exec] at org.junit.internal.runners.TestMethod.invoke(TestMethod.java:59)
[exec] at org.junit.internal.runners.MethodRoadie.runTestMethod(MethodRoadie.java:98)
[exec] at org.junit.internal.runners.MethodRoadie$2.run(MethodRoadie.java:79)
[exec] at org.junit.internal.runners.MethodRoadie.runBeforesThenTestThenAfters(MethodRoadie.java:87)
[exec] at org.junit.internal.runners.MethodRoadie.runTest(MethodRoadie.java:77)
[exec] at org.junit.internal.runners.MethodRoadie.run(MethodRoadie.java:42)
[exec] at org.junit.internal.runners.JUnit4ClassRunner.invokeTestMethod(JUnit4ClassRunner.java:88)
[exec] at org.junit.internal.runners.JUnit4ClassRunner.runMethods(JUnit4ClassRunner.java:51)
[exec] at org.junit.runners.Parameterized$TestClassRunnerForParameters.run(Parameterized.java:98)
[exec] at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
[exec] at org.junit.runners.Parameterized.access$000(Parameterized.java:55)
[exec] at org.junit.runners.Parameterized$1.run(Parameterized.java:131)
[exec] at org.junit.internal.runners.ClassRoadie.runUnprotected(ClassRoadie.java:27)
[exec] at org.junit.internal.runners.ClassRoadie.runProtected(ClassRoadie.java:37)
[exec] at org.junit.runners.Parameterized.run(Parameterized.java:129)
[exec] at org.junit.internal.runners.CompositeRunner.runChildren(CompositeRunner.java:33)
[exec] at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:28)
[exec] at org.junit.runner.JUnitCore.run(JUnitCore.java:130)
[exec] at org.junit.runner.JUnitCore.run(JUnitCore.java:109)
[exec] at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
[exec] at org.junit.runner.JUnitCore.runClasses(JUnitCore.java:60)
[exec] at com.huawei.querymanagement.SelectQuerySuite.main(SelectQuerySuite.java:18)
[exec] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
[exec] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
[exec] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
[exec] at java.lang.reflect.Method.invoke(Method.java:498)
[exec] at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
[exec] at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
[exec] at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
[exec] at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
[exec] at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)