You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by kg...@apache.org on 2019/01/23 09:50:09 UTC

[2/2] hive git commit: HIVE-21138: Fix some of the alerts raised by lgtm.com (Malcolm Taylor via Zoltan Haindrich)

HIVE-21138: Fix some of the alerts raised by lgtm.com (Malcolm Taylor via Zoltan Haindrich)

Signed-off-by: Zoltan Haindrich <ki...@rxd.hu>


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/eba9646b
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/eba9646b
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/eba9646b

Branch: refs/heads/master
Commit: eba9646b41cc1306b7c5b0f67dcf2853e0b8171d
Parents: 461d8a0
Author: Malcolm Taylor <ma...@semmle.com>
Authored: Wed Jan 23 10:41:48 2019 +0100
Committer: Zoltan Haindrich <ki...@rxd.hu>
Committed: Wed Jan 23 10:41:48 2019 +0100

----------------------------------------------------------------------
 .../hadoop/hive/llap/cli/LlapServiceDriver.java |  2 +-
 .../hadoop/hive/ql/exec/repl/ReplDumpTask.java  |  2 +-
 .../ql/exec/tez/KillTriggerActionHandler.java   | 34 ++++++++++----------
 .../ql/exec/vector/VectorizedBatchUtil.java     | 12 +++----
 .../hive/ql/metadata/CheckConstraint.java       |  2 --
 .../ql/optimizer/lineage/OpProcFactory.java     |  4 +--
 .../hive/ql/optimizer/physical/Vectorizer.java  |  1 -
 .../stats/annotation/StatsRulesProcFactory.java |  5 +--
 .../hadoop/hive/ql/parse/CalcitePlanner.java    |  4 +--
 .../repl/dump/events/AddPartitionHandler.java   |  4 +--
 .../repl/dump/events/CreateTableHandler.java    |  2 +-
 .../hadoop/hive/ql/plan/AggregationDesc.java    |  6 +---
 .../hive/ql/stats/TestStatsUpdaterThread.java   |  6 ++++
 .../hive/metastore/HiveMetaStoreClient.java     |  2 +-
 .../hadoop/hive/metastore/ObjectStore.java      |  4 +--
 .../hadoop/hive/metastore/txn/TxnHandler.java   |  2 +-
 16 files changed, 43 insertions(+), 49 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
----------------------------------------------------------------------
diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java b/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
index ffdd340..4bc2431 100644
--- a/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
+++ b/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
@@ -159,7 +159,7 @@ public class LlapServiceDriver {
         conf.set(key, (String) props.getValue());
       } else {
         if (key.startsWith(HiveConf.PREFIX_LLAP) || key.startsWith(HiveConf.PREFIX_HIVE_LLAP)) {
-          LOG.warn("Adding key [{}] even though it is not in the set of known llap-server keys");
+          LOG.warn("Adding key [{}] even though it is not in the set of known llap-server keys", key);
           conf.set(key, (String) props.getValue());
         } else {
           LOG.warn("Ignoring unknown llap server parameter: [{}]", key);

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
index 497e103..a5b944b 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
@@ -263,7 +263,7 @@ public class ReplDumpTask extends Task<ReplDumpWork> implements Serializable {
                 && TableType.EXTERNAL_TABLE.equals(tableTuple.object.getTableType())
                 && !conf.getBoolVar(HiveConf.ConfVars.REPL_DUMP_METADATA_ONLY);
             if (shouldWriteExternalTableLocationInfo) {
-              LOG.debug("adding table {} to external tables list");
+              LOG.debug("adding table {} to external tables list", tblName);
               writer.dataLocationDump(tableTuple.object);
             }
             dumpTable(dbName, tblName, validTxnList, dbRoot, bootDumpBeginReplId, hiveDb,

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
index ee539ba..06e9ff6 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
@@ -37,25 +37,25 @@ public class KillTriggerActionHandler implements TriggerActionHandler<TezSession
   public void applyAction(final Map<TezSessionState, Trigger> queriesViolated) {
     for (Map.Entry<TezSessionState, Trigger> entry : queriesViolated.entrySet()) {
       switch (entry.getValue().getAction().getType()) {
-        case KILL_QUERY:
-          TezSessionState sessionState = entry.getKey();
-          String queryId = sessionState.getWmContext().getQueryId();
-          try {
-            SessionState ss = new SessionState(new HiveConf());
-            ss.setIsHiveServerQuery(true);
-            SessionState.start(ss);
-            KillQuery killQuery = sessionState.getKillQuery();
-            // if kill query is null then session might have been released to pool or closed already
-            if (killQuery != null) {
-              sessionState.getKillQuery().killQuery(queryId, entry.getValue().getViolationMsg(),
+      case KILL_QUERY:
+        TezSessionState sessionState = entry.getKey();
+        String queryId = sessionState.getWmContext().getQueryId();
+        try {
+          SessionState ss = new SessionState(new HiveConf());
+          ss.setIsHiveServerQuery(true);
+          SessionState.start(ss);
+          KillQuery killQuery = sessionState.getKillQuery();
+          // if kill query is null then session might have been released to pool or closed already
+          if (killQuery != null) {
+            sessionState.getKillQuery().killQuery(queryId, entry.getValue().getViolationMsg(),
                       sessionState.getConf());
-            }
-          } catch (HiveException e) {
-            LOG.warn("Unable to kill query {} for trigger violation");
           }
-          break;
-        default:
-          throw new RuntimeException("Unsupported action: " + entry.getValue());
+        } catch (HiveException e) {
+          LOG.warn("Unable to kill query {} for trigger violation", queryId);
+        }
+        break;
+      default:
+        throw new RuntimeException("Unsupported action: " + entry.getValue());
       }
     }
   }

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
index 211f452..617cbf1 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
@@ -576,7 +576,12 @@ public class VectorizedBatchUtil {
 
   public static ColumnVector makeLikeColumnVector(ColumnVector source
                                         ) throws HiveException{
-    if (source instanceof LongColumnVector) {
+    if (source instanceof Decimal64ColumnVector) {
+      Decimal64ColumnVector dec64ColVector = (Decimal64ColumnVector) source;
+      return new DecimalColumnVector(dec64ColVector.vector.length,
+          dec64ColVector.precision,
+          dec64ColVector.scale);
+    } else if (source instanceof LongColumnVector) {
       return new LongColumnVector(((LongColumnVector) source).vector.length);
     } else if (source instanceof DoubleColumnVector) {
       return new DoubleColumnVector(((DoubleColumnVector) source).vector.length);
@@ -587,11 +592,6 @@ public class VectorizedBatchUtil {
       return new DecimalColumnVector(decColVector.vector.length,
           decColVector.precision,
           decColVector.scale);
-    } else if (source instanceof Decimal64ColumnVector) {
-        Decimal64ColumnVector dec64ColVector = (Decimal64ColumnVector) source;
-        return new DecimalColumnVector(dec64ColVector.vector.length,
-            dec64ColVector.precision,
-            dec64ColVector.scale);
     } else if (source instanceof TimestampColumnVector) {
       return new TimestampColumnVector(((TimestampColumnVector) source).getLength());
     } else if (source instanceof IntervalDayTimeColumnVector) {

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
index db11310..af45788 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
@@ -56,8 +56,6 @@ public class CheckConstraint implements Serializable {
   public CheckConstraint() {}
 
   public CheckConstraint(List<SQLCheckConstraint> checkConstraintsList) {
-    this.tableName = tableName;
-    this.databaseName = databaseName;
     checkConstraints = new TreeMap<String, List<CheckConstraintCol>>();
     checkExpressionList = new ArrayList<>();
     if (checkConstraintsList == null) {

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
index 2360032..825ece6 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
@@ -271,9 +271,9 @@ public class OpProcFactory {
         }
         int left = conds[i].getLeft();
         int right = conds[i].getRight();
-        if (joinKeys.length < left
+        if (joinKeys.length <= left
             || joinKeys[left].length == 0
-            || joinKeys.length < right
+            || joinKeys.length <= right
             || joinKeys[right].length == 0
             || parents < left
             || parents < right) {

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
index 5023f2f..1fe0a79 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
@@ -2404,7 +2404,6 @@ public class Vectorizer implements PhysicalPlanResolver {
   @Override
   public PhysicalContext resolve(PhysicalContext physicalContext) throws SemanticException {
 
-    physicalContext = physicalContext;
     hiveConf = physicalContext.getConf();
     planMapper = physicalContext.getContext().getPlanMapper();
 

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
index 1bd2511..b19c7de 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
@@ -31,7 +31,6 @@ import java.util.Optional;
 import java.util.Set;
 import java.util.Stack;
 
-import org.apache.commons.collections.CollectionUtils;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
 import org.apache.hadoop.hive.ql.Context;
@@ -91,7 +90,6 @@ import org.apache.hadoop.hive.ql.udf.generic.GenericUDFIn;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDFInBloomFilter;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPAnd;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqual;
-import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqualNS;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqualOrGreaterThan;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqualOrLessThan;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPGreaterThan;
@@ -1095,8 +1093,7 @@ public class StatsRulesProcFactory {
         ExprNodeGenericFuncDesc genFunc = (ExprNodeGenericFuncDesc) child;
         GenericUDF udf = genFunc.getGenericUDF();
 
-        if (udf instanceof GenericUDFOPEqual ||
-            udf instanceof GenericUDFOPEqualNS) {
+        if (udf instanceof GenericUDFOPEqual) {
           String colName = null;
           boolean isConst = false;
           Object prevConst = null;

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
index aacf9e3..646ce09 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
@@ -281,7 +281,6 @@ import org.joda.time.Interval;
 import java.io.IOException;
 import java.lang.reflect.Field;
 import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.UndeclaredThrowableException;
 import java.math.BigDecimal;
 import java.util.AbstractMap.SimpleEntry;
 import java.util.ArrayDeque;
@@ -1655,8 +1654,7 @@ public class CalcitePlanner extends SemanticAnalyzer {
   }
 
   private boolean isUselessCause(Throwable t) {
-    return t instanceof RuntimeException || t instanceof InvocationTargetException
-        || t instanceof UndeclaredThrowableException;
+    return t instanceof RuntimeException || t instanceof InvocationTargetException;
   }
 
   private RowResolver genRowResolver(Operator op, QB qb) {

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
index 5c16887..0b7f910 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
@@ -54,7 +54,7 @@ class AddPartitionHandler extends AbstractEventHandler {
     AddPartitionMessage apm = (AddPartitionMessage) eventMessage;
     org.apache.hadoop.hive.metastore.api.Table tobj = apm.getTableObj();
     if (tobj == null) {
-      LOG.debug("Event#{} was a ADD_PTN_EVENT with no table listed");
+      LOG.debug("Event#{} was a ADD_PTN_EVENT with no table listed", fromEventId());
       return;
     }
 
@@ -65,7 +65,7 @@ class AddPartitionHandler extends AbstractEventHandler {
 
     Iterable<org.apache.hadoop.hive.metastore.api.Partition> ptns = apm.getPartitionObjs();
     if ((ptns == null) || (!ptns.iterator().hasNext())) {
-      LOG.debug("Event#{} was an ADD_PTN_EVENT with no partitions");
+      LOG.debug("Event#{} was an ADD_PTN_EVENT with no partitions", fromEventId());
       return;
     }
 

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
index 550a82d..5870876 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
@@ -47,7 +47,7 @@ class CreateTableHandler extends AbstractEventHandler<CreateTableMessage> {
     org.apache.hadoop.hive.metastore.api.Table tobj = eventMessage.getTableObj();
 
     if (tobj == null) {
-      LOG.debug("Event#{} was a CREATE_TABLE_EVENT with no table listed");
+      LOG.debug("Event#{} was a CREATE_TABLE_EVENT with no table listed", fromEventId());
       return;
     }
 

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
index 4b2ef1f..87a7f4d 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
@@ -18,11 +18,8 @@
 
 package org.apache.hadoop.hive.ql.plan;
 
-import java.io.Externalizable;
 import java.io.Serializable;
 
-import org.apache.hadoop.hive.common.JavaUtils;
-import org.apache.hadoop.hive.ql.exec.PTFUtils;
 import org.apache.hadoop.hive.ql.exec.Utilities;
 import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFEvaluator;
 import org.apache.hadoop.util.ReflectionUtils;
@@ -72,8 +69,7 @@ public class AggregationDesc implements java.io.Serializable {
   public void setGenericUDAFEvaluator(
       final GenericUDAFEvaluator genericUDAFEvaluator) {
     this.genericUDAFEvaluator = genericUDAFEvaluator;
-    if (genericUDAFEvaluator instanceof Serializable ||
-        genericUDAFEvaluator instanceof Externalizable) {
+    if (genericUDAFEvaluator instanceof Serializable) {
       this.genericUDAFWritableEvaluator = genericUDAFEvaluator;
     } else {
       this.genericUDAFEvaluatorClassName = genericUDAFEvaluator.getClass().getName();

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
----------------------------------------------------------------------
diff --git a/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java b/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
index 2512579..24acd6d 100644
--- a/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
+++ b/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
@@ -274,6 +274,12 @@ public class TestStatsUpdaterThread {
         Lists.newArrayList("s"), currentWriteIds);
     assertEquals(1, stats.size());
 
+    // Test with null list of partNames
+    stats = msClient.getPartitionColumnStatistics(
+        dbName, tblName, null,
+        Lists.newArrayList("s"), currentWriteIds);
+    assertEquals(0, stats.size());
+
     // New reader.
     currentWriteIds = msClient.getValidWriteIds(fqName).toString();
     stats = msClient.getPartitionColumnStatistics(

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
index 30edc56..3d4467d 100644
--- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
+++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
@@ -924,7 +924,7 @@ public class HiveMetaStoreClient implements IMetaStoreClient, AutoCloseable {
       List<String> colNames, String validWriteIdList)
       throws NoSuchObjectException, MetaException, TException {
     PartitionsStatsRequest rqst = new PartitionsStatsRequest(dbName, tableName, colNames,
-        partNames);
+        partNames == null ? new ArrayList<String>() : partNames);
     rqst.setCatName(catName);
     rqst.setValidWriteIdList(validWriteIdList);
     return client.get_partitions_statistics_req(rqst).getPartStats();

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
index ad60af7..65fea87 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
@@ -8810,7 +8810,7 @@ public class ObjectStore implements RawStore, Configurable {
       List<String> partNames, List<String> colNames,
       String writeIdList)
       throws MetaException, NoSuchObjectException {
-    if (partNames == null && partNames.isEmpty()) {
+    if (partNames == null || partNames.isEmpty()) {
       return null;
     }
     List<ColumnStatistics> allStats = getPartitionColumnStatisticsInternal(
@@ -8897,7 +8897,7 @@ public class ObjectStore implements RawStore, Configurable {
     // If the current stats in the metastore doesn't comply with
     // the isolation level of the query, return null.
     if (writeIdList != null) {
-      if (partNames == null && partNames.isEmpty()) {
+      if (partNames == null || partNames.isEmpty()) {
         return null;
       }
 

http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
index 91a9ab4..a17abfc 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
@@ -2028,7 +2028,7 @@ abstract class TxnHandler implements TxnStore, TxnStore.MutexAPI {
         ValidWriteIdList tblValidWriteIdList =
             validReaderWriteIdList.getTableValidWriteIdList(fullyQualifiedName);
         if (tblValidWriteIdList == null) {
-          LOG.warn("ValidWriteIdList for table {} not present in creation metadata, this should not happen");
+          LOG.warn("ValidWriteIdList for table {} not present in creation metadata, this should not happen", fullyQualifiedName);
           return null;
         }
         query.append(" AND (ctc_writeid > " + tblValidWriteIdList.getHighWatermark());