You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@carbondata.apache.org by aj...@apache.org on 2019/11/25 08:44:02 UTC

[carbondata] branch master updated: [CARBONDATA-3589]: Adding NULL segments check and empty segments check before prepriming

This is an automated email from the ASF dual-hosted git repository.

ajantha pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/carbondata.git


The following commit(s) were added to refs/heads/master by this push:
     new ca13a2e  [CARBONDATA-3589]: Adding NULL segments check and empty segments check before prepriming
ca13a2e is described below

commit ca13a2ef7c50f2c6cf235121d9e7b04418c42b55
Author: Vikram Ahuja <vi...@gmail.com>
AuthorDate: Mon Nov 18 15:16:50 2019 +0530

    [CARBONDATA-3589]: Adding NULL segments check and empty segments check before prepriming
    
    Insert into select from hive table into carbon table having partition
    fails with index server running because of the fact that empty segments
    were being sent for prepriming. Added a check for the same.
    
    This closes #3468
---
 .../scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala   | 3 ++-
 .../spark/sql/execution/command/management/CarbonLoadDataCommand.scala | 2 +-
 2 files changed, 3 insertions(+), 2 deletions(-)

diff --git a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
index fa7b9f5..6bb7e76 100644
--- a/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
+++ b/integration/spark2/src/main/scala/org/apache/carbondata/spark/rdd/CarbonDataRDDFactory.scala
@@ -28,6 +28,7 @@ import scala.collection.mutable.ListBuffer
 import scala.util.Random
 import scala.util.control.Breaks._
 
+import org.apache.commons.lang3.StringUtils
 import org.apache.hadoop.conf.Configuration
 import org.apache.hadoop.fs.Path
 import org.apache.hadoop.io.NullWritable
@@ -594,7 +595,7 @@ object CarbonDataRDDFactory {
 
       // code to handle Pre-Priming cache for loading
 
-      if (carbonLoadModel.getSegmentId != null) {
+      if (!StringUtils.isEmpty(carbonLoadModel.getSegmentId)) {
         DistributedRDDUtils.triggerPrepriming(sqlContext.sparkSession, carbonTable, Seq(),
           operationContext, hadoopConf, List(carbonLoadModel.getSegmentId))
       }
diff --git a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
index 72340f4..2227094 100644
--- a/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
+++ b/integration/spark2/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala
@@ -869,7 +869,7 @@ case class CarbonLoadDataCommand(
       }
 
       // Prepriming for Partition table here
-      if (carbonLoadModel.getSegmentId != null) {
+      if (!StringUtils.isEmpty(carbonLoadModel.getSegmentId)) {
         DistributedRDDUtils.triggerPrepriming(sparkSession,
           table,
           Seq(),