You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by kg...@apache.org on 2019/01/23 09:50:09 UTC
[2/2] hive git commit: HIVE-21138: Fix some of the alerts raised by
lgtm.com (Malcolm Taylor via Zoltan Haindrich)
HIVE-21138: Fix some of the alerts raised by lgtm.com (Malcolm Taylor via Zoltan Haindrich)
Signed-off-by: Zoltan Haindrich <ki...@rxd.hu>
Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/eba9646b
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/eba9646b
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/eba9646b
Branch: refs/heads/master
Commit: eba9646b41cc1306b7c5b0f67dcf2853e0b8171d
Parents: 461d8a0
Author: Malcolm Taylor <ma...@semmle.com>
Authored: Wed Jan 23 10:41:48 2019 +0100
Committer: Zoltan Haindrich <ki...@rxd.hu>
Committed: Wed Jan 23 10:41:48 2019 +0100
----------------------------------------------------------------------
.../hadoop/hive/llap/cli/LlapServiceDriver.java | 2 +-
.../hadoop/hive/ql/exec/repl/ReplDumpTask.java | 2 +-
.../ql/exec/tez/KillTriggerActionHandler.java | 34 ++++++++++----------
.../ql/exec/vector/VectorizedBatchUtil.java | 12 +++----
.../hive/ql/metadata/CheckConstraint.java | 2 --
.../ql/optimizer/lineage/OpProcFactory.java | 4 +--
.../hive/ql/optimizer/physical/Vectorizer.java | 1 -
.../stats/annotation/StatsRulesProcFactory.java | 5 +--
.../hadoop/hive/ql/parse/CalcitePlanner.java | 4 +--
.../repl/dump/events/AddPartitionHandler.java | 4 +--
.../repl/dump/events/CreateTableHandler.java | 2 +-
.../hadoop/hive/ql/plan/AggregationDesc.java | 6 +---
.../hive/ql/stats/TestStatsUpdaterThread.java | 6 ++++
.../hive/metastore/HiveMetaStoreClient.java | 2 +-
.../hadoop/hive/metastore/ObjectStore.java | 4 +--
.../hadoop/hive/metastore/txn/TxnHandler.java | 2 +-
16 files changed, 43 insertions(+), 49 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
----------------------------------------------------------------------
diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java b/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
index ffdd340..4bc2431 100644
--- a/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
+++ b/llap-server/src/java/org/apache/hadoop/hive/llap/cli/LlapServiceDriver.java
@@ -159,7 +159,7 @@ public class LlapServiceDriver {
conf.set(key, (String) props.getValue());
} else {
if (key.startsWith(HiveConf.PREFIX_LLAP) || key.startsWith(HiveConf.PREFIX_HIVE_LLAP)) {
- LOG.warn("Adding key [{}] even though it is not in the set of known llap-server keys");
+ LOG.warn("Adding key [{}] even though it is not in the set of known llap-server keys", key);
conf.set(key, (String) props.getValue());
} else {
LOG.warn("Ignoring unknown llap server parameter: [{}]", key);
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
index 497e103..a5b944b 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/repl/ReplDumpTask.java
@@ -263,7 +263,7 @@ public class ReplDumpTask extends Task<ReplDumpWork> implements Serializable {
&& TableType.EXTERNAL_TABLE.equals(tableTuple.object.getTableType())
&& !conf.getBoolVar(HiveConf.ConfVars.REPL_DUMP_METADATA_ONLY);
if (shouldWriteExternalTableLocationInfo) {
- LOG.debug("adding table {} to external tables list");
+ LOG.debug("adding table {} to external tables list", tblName);
writer.dataLocationDump(tableTuple.object);
}
dumpTable(dbName, tblName, validTxnList, dbRoot, bootDumpBeginReplId, hiveDb,
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
index ee539ba..06e9ff6 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/KillTriggerActionHandler.java
@@ -37,25 +37,25 @@ public class KillTriggerActionHandler implements TriggerActionHandler<TezSession
public void applyAction(final Map<TezSessionState, Trigger> queriesViolated) {
for (Map.Entry<TezSessionState, Trigger> entry : queriesViolated.entrySet()) {
switch (entry.getValue().getAction().getType()) {
- case KILL_QUERY:
- TezSessionState sessionState = entry.getKey();
- String queryId = sessionState.getWmContext().getQueryId();
- try {
- SessionState ss = new SessionState(new HiveConf());
- ss.setIsHiveServerQuery(true);
- SessionState.start(ss);
- KillQuery killQuery = sessionState.getKillQuery();
- // if kill query is null then session might have been released to pool or closed already
- if (killQuery != null) {
- sessionState.getKillQuery().killQuery(queryId, entry.getValue().getViolationMsg(),
+ case KILL_QUERY:
+ TezSessionState sessionState = entry.getKey();
+ String queryId = sessionState.getWmContext().getQueryId();
+ try {
+ SessionState ss = new SessionState(new HiveConf());
+ ss.setIsHiveServerQuery(true);
+ SessionState.start(ss);
+ KillQuery killQuery = sessionState.getKillQuery();
+ // if kill query is null then session might have been released to pool or closed already
+ if (killQuery != null) {
+ sessionState.getKillQuery().killQuery(queryId, entry.getValue().getViolationMsg(),
sessionState.getConf());
- }
- } catch (HiveException e) {
- LOG.warn("Unable to kill query {} for trigger violation");
}
- break;
- default:
- throw new RuntimeException("Unsupported action: " + entry.getValue());
+ } catch (HiveException e) {
+ LOG.warn("Unable to kill query {} for trigger violation", queryId);
+ }
+ break;
+ default:
+ throw new RuntimeException("Unsupported action: " + entry.getValue());
}
}
}
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
index 211f452..617cbf1 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorizedBatchUtil.java
@@ -576,7 +576,12 @@ public class VectorizedBatchUtil {
public static ColumnVector makeLikeColumnVector(ColumnVector source
) throws HiveException{
- if (source instanceof LongColumnVector) {
+ if (source instanceof Decimal64ColumnVector) {
+ Decimal64ColumnVector dec64ColVector = (Decimal64ColumnVector) source;
+ return new DecimalColumnVector(dec64ColVector.vector.length,
+ dec64ColVector.precision,
+ dec64ColVector.scale);
+ } else if (source instanceof LongColumnVector) {
return new LongColumnVector(((LongColumnVector) source).vector.length);
} else if (source instanceof DoubleColumnVector) {
return new DoubleColumnVector(((DoubleColumnVector) source).vector.length);
@@ -587,11 +592,6 @@ public class VectorizedBatchUtil {
return new DecimalColumnVector(decColVector.vector.length,
decColVector.precision,
decColVector.scale);
- } else if (source instanceof Decimal64ColumnVector) {
- Decimal64ColumnVector dec64ColVector = (Decimal64ColumnVector) source;
- return new DecimalColumnVector(dec64ColVector.vector.length,
- dec64ColVector.precision,
- dec64ColVector.scale);
} else if (source instanceof TimestampColumnVector) {
return new TimestampColumnVector(((TimestampColumnVector) source).getLength());
} else if (source instanceof IntervalDayTimeColumnVector) {
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
index db11310..af45788 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/CheckConstraint.java
@@ -56,8 +56,6 @@ public class CheckConstraint implements Serializable {
public CheckConstraint() {}
public CheckConstraint(List<SQLCheckConstraint> checkConstraintsList) {
- this.tableName = tableName;
- this.databaseName = databaseName;
checkConstraints = new TreeMap<String, List<CheckConstraintCol>>();
checkExpressionList = new ArrayList<>();
if (checkConstraintsList == null) {
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
index 2360032..825ece6 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/lineage/OpProcFactory.java
@@ -271,9 +271,9 @@ public class OpProcFactory {
}
int left = conds[i].getLeft();
int right = conds[i].getRight();
- if (joinKeys.length < left
+ if (joinKeys.length <= left
|| joinKeys[left].length == 0
- || joinKeys.length < right
+ || joinKeys.length <= right
|| joinKeys[right].length == 0
|| parents < left
|| parents < right) {
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
index 5023f2f..1fe0a79 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java
@@ -2404,7 +2404,6 @@ public class Vectorizer implements PhysicalPlanResolver {
@Override
public PhysicalContext resolve(PhysicalContext physicalContext) throws SemanticException {
- physicalContext = physicalContext;
hiveConf = physicalContext.getConf();
planMapper = physicalContext.getContext().getPlanMapper();
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
index 1bd2511..b19c7de 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/optimizer/stats/annotation/StatsRulesProcFactory.java
@@ -31,7 +31,6 @@ import java.util.Optional;
import java.util.Set;
import java.util.Stack;
-import org.apache.commons.collections.CollectionUtils;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
import org.apache.hadoop.hive.ql.Context;
@@ -91,7 +90,6 @@ import org.apache.hadoop.hive.ql.udf.generic.GenericUDFIn;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDFInBloomFilter;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPAnd;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqual;
-import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqualNS;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqualOrGreaterThan;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPEqualOrLessThan;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPGreaterThan;
@@ -1095,8 +1093,7 @@ public class StatsRulesProcFactory {
ExprNodeGenericFuncDesc genFunc = (ExprNodeGenericFuncDesc) child;
GenericUDF udf = genFunc.getGenericUDF();
- if (udf instanceof GenericUDFOPEqual ||
- udf instanceof GenericUDFOPEqualNS) {
+ if (udf instanceof GenericUDFOPEqual) {
String colName = null;
boolean isConst = false;
Object prevConst = null;
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
index aacf9e3..646ce09 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java
@@ -281,7 +281,6 @@ import org.joda.time.Interval;
import java.io.IOException;
import java.lang.reflect.Field;
import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.UndeclaredThrowableException;
import java.math.BigDecimal;
import java.util.AbstractMap.SimpleEntry;
import java.util.ArrayDeque;
@@ -1655,8 +1654,7 @@ public class CalcitePlanner extends SemanticAnalyzer {
}
private boolean isUselessCause(Throwable t) {
- return t instanceof RuntimeException || t instanceof InvocationTargetException
- || t instanceof UndeclaredThrowableException;
+ return t instanceof RuntimeException || t instanceof InvocationTargetException;
}
private RowResolver genRowResolver(Operator op, QB qb) {
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
index 5c16887..0b7f910 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/AddPartitionHandler.java
@@ -54,7 +54,7 @@ class AddPartitionHandler extends AbstractEventHandler {
AddPartitionMessage apm = (AddPartitionMessage) eventMessage;
org.apache.hadoop.hive.metastore.api.Table tobj = apm.getTableObj();
if (tobj == null) {
- LOG.debug("Event#{} was a ADD_PTN_EVENT with no table listed");
+ LOG.debug("Event#{} was a ADD_PTN_EVENT with no table listed", fromEventId());
return;
}
@@ -65,7 +65,7 @@ class AddPartitionHandler extends AbstractEventHandler {
Iterable<org.apache.hadoop.hive.metastore.api.Partition> ptns = apm.getPartitionObjs();
if ((ptns == null) || (!ptns.iterator().hasNext())) {
- LOG.debug("Event#{} was an ADD_PTN_EVENT with no partitions");
+ LOG.debug("Event#{} was an ADD_PTN_EVENT with no partitions", fromEventId());
return;
}
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
index 550a82d..5870876 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/repl/dump/events/CreateTableHandler.java
@@ -47,7 +47,7 @@ class CreateTableHandler extends AbstractEventHandler<CreateTableMessage> {
org.apache.hadoop.hive.metastore.api.Table tobj = eventMessage.getTableObj();
if (tobj == null) {
- LOG.debug("Event#{} was a CREATE_TABLE_EVENT with no table listed");
+ LOG.debug("Event#{} was a CREATE_TABLE_EVENT with no table listed", fromEventId());
return;
}
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
index 4b2ef1f..87a7f4d 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/AggregationDesc.java
@@ -18,11 +18,8 @@
package org.apache.hadoop.hive.ql.plan;
-import java.io.Externalizable;
import java.io.Serializable;
-import org.apache.hadoop.hive.common.JavaUtils;
-import org.apache.hadoop.hive.ql.exec.PTFUtils;
import org.apache.hadoop.hive.ql.exec.Utilities;
import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFEvaluator;
import org.apache.hadoop.util.ReflectionUtils;
@@ -72,8 +69,7 @@ public class AggregationDesc implements java.io.Serializable {
public void setGenericUDAFEvaluator(
final GenericUDAFEvaluator genericUDAFEvaluator) {
this.genericUDAFEvaluator = genericUDAFEvaluator;
- if (genericUDAFEvaluator instanceof Serializable ||
- genericUDAFEvaluator instanceof Externalizable) {
+ if (genericUDAFEvaluator instanceof Serializable) {
this.genericUDAFWritableEvaluator = genericUDAFEvaluator;
} else {
this.genericUDAFEvaluatorClassName = genericUDAFEvaluator.getClass().getName();
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
----------------------------------------------------------------------
diff --git a/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java b/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
index 2512579..24acd6d 100644
--- a/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
+++ b/ql/src/test/org/apache/hadoop/hive/ql/stats/TestStatsUpdaterThread.java
@@ -274,6 +274,12 @@ public class TestStatsUpdaterThread {
Lists.newArrayList("s"), currentWriteIds);
assertEquals(1, stats.size());
+ // Test with null list of partNames
+ stats = msClient.getPartitionColumnStatistics(
+ dbName, tblName, null,
+ Lists.newArrayList("s"), currentWriteIds);
+ assertEquals(0, stats.size());
+
// New reader.
currentWriteIds = msClient.getValidWriteIds(fqName).toString();
stats = msClient.getPartitionColumnStatistics(
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
index 30edc56..3d4467d 100644
--- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
+++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java
@@ -924,7 +924,7 @@ public class HiveMetaStoreClient implements IMetaStoreClient, AutoCloseable {
List<String> colNames, String validWriteIdList)
throws NoSuchObjectException, MetaException, TException {
PartitionsStatsRequest rqst = new PartitionsStatsRequest(dbName, tableName, colNames,
- partNames);
+ partNames == null ? new ArrayList<String>() : partNames);
rqst.setCatName(catName);
rqst.setValidWriteIdList(validWriteIdList);
return client.get_partitions_statistics_req(rqst).getPartStats();
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
index ad60af7..65fea87 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/ObjectStore.java
@@ -8810,7 +8810,7 @@ public class ObjectStore implements RawStore, Configurable {
List<String> partNames, List<String> colNames,
String writeIdList)
throws MetaException, NoSuchObjectException {
- if (partNames == null && partNames.isEmpty()) {
+ if (partNames == null || partNames.isEmpty()) {
return null;
}
List<ColumnStatistics> allStats = getPartitionColumnStatisticsInternal(
@@ -8897,7 +8897,7 @@ public class ObjectStore implements RawStore, Configurable {
// If the current stats in the metastore doesn't comply with
// the isolation level of the query, return null.
if (writeIdList != null) {
- if (partNames == null && partNames.isEmpty()) {
+ if (partNames == null || partNames.isEmpty()) {
return null;
}
http://git-wip-us.apache.org/repos/asf/hive/blob/eba9646b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
----------------------------------------------------------------------
diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
index 91a9ab4..a17abfc 100644
--- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
+++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/txn/TxnHandler.java
@@ -2028,7 +2028,7 @@ abstract class TxnHandler implements TxnStore, TxnStore.MutexAPI {
ValidWriteIdList tblValidWriteIdList =
validReaderWriteIdList.getTableValidWriteIdList(fullyQualifiedName);
if (tblValidWriteIdList == null) {
- LOG.warn("ValidWriteIdList for table {} not present in creation metadata, this should not happen");
+ LOG.warn("ValidWriteIdList for table {} not present in creation metadata, this should not happen", fullyQualifiedName);
return null;
}
query.append(" AND (ctc_writeid > " + tblValidWriteIdList.getHighWatermark());