You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by mg...@apache.org on 2020/05/04 09:27:13 UTC
[hive] branch master updated: HIVE-22028 Clean up Add Partition
ADDENDUM - fix typo (Miklos Gergely, reviewed by Zoltan Haindrich)
This is an automated email from the ASF dual-hosted git repository.
mgergely pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push:
new 342f8fb HIVE-22028 Clean up Add Partition ADDENDUM - fix typo (Miklos Gergely, reviewed by Zoltan Haindrich)
342f8fb is described below
commit 342f8fbd521ce5728098887cd1749fc7931e370f
Author: miklosgergely <mg...@cloudera.com>
AuthorDate: Sat May 2 16:42:30 2020 +0200
HIVE-22028 Clean up Add Partition ADDENDUM - fix typo (Miklos Gergely, reviewed by Zoltan Haindrich)
---
.../table/partition/add/AlterTableAddPartitionOperation.java | 10 +++++-----
ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java | 4 +---
2 files changed, 6 insertions(+), 8 deletions(-)
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java
index 6910e10..ddc47a4 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java
@@ -157,7 +157,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
List<Partition> partitions) throws HiveException {
// TODO: normally, the result is not necessary; might make sense to pass false
List<org.apache.hadoop.hive.ql.metadata.Partition> outPartitions = new ArrayList<>();
- for (Partition outPart : context.getDb().addPartition(partitions, desc.isIfNotExists(), true)) {
+ for (Partition outPart : context.getDb().addPartitions(partitions, desc.isIfNotExists(), true)) {
outPartitions.add(new org.apache.hadoop.hive.ql.metadata.Partition(table, outPart));
}
return outPartitions;
@@ -170,7 +170,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
// no choice but to iterate over the partitions here.
List<Partition> partitionsToAdd = new ArrayList<>();
- List<Partition> partitionssToAlter = new ArrayList<>();
+ List<Partition> partitionsToAlter = new ArrayList<>();
List<String> partitionNames = new ArrayList<>();
for (Partition partition : partitions){
partitionNames.add(getPartitionName(table, partition));
@@ -178,7 +178,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
Partition p = context.getDb().getPartition(desc.getDbName(), desc.getTableName(), partition.getValues());
if (desc.getReplicationSpec().allowReplacementInto(p.getParameters())){
ReplicationSpec.copyLastReplId(p.getParameters(), partition.getParameters());
- partitionssToAlter.add(partition);
+ partitionsToAlter.add(partition);
} // else ptn already exists, but we do nothing with it.
} catch (HiveException e){
if (e.getCause() instanceof NoSuchObjectException) {
@@ -191,7 +191,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
}
List<org.apache.hadoop.hive.ql.metadata.Partition> outPartitions = new ArrayList<>();
- for (Partition outPartition : context.getDb().addPartition(partitionsToAdd, desc.isIfNotExists(), true)) {
+ for (Partition outPartition : context.getDb().addPartitions(partitionsToAdd, desc.isIfNotExists(), true)) {
outPartitions.add(new org.apache.hadoop.hive.ql.metadata.Partition(table, outPartition));
}
@@ -199,7 +199,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
EnvironmentContext ec = new EnvironmentContext();
ec.putToProperties(StatsSetupConst.DO_NOT_UPDATE_STATS, StatsSetupConst.TRUE);
String validWriteIdList = getValidWriteIdList(table, writeId);
- context.getDb().alterPartitions(desc.getDbName(), desc.getTableName(), partitionssToAlter, ec, validWriteIdList,
+ context.getDb().alterPartitions(desc.getDbName(), desc.getTableName(), partitionsToAlter, ec, validWriteIdList,
writeId);
for (Partition outPartition : context.getDb().getPartitionsByNames(desc.getDbName(), desc.getTableName(),
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
index bd1fae1..f6a5207 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
@@ -40,7 +40,6 @@ import java.io.PrintStream;
import java.net.InetAddress;
import java.net.UnknownHostException;
import java.nio.ByteBuffer;
-import java.sql.SQLIntegrityConstraintViolationException;
import java.util.ArrayList;
import java.util.Arrays;
import java.util.BitSet;
@@ -80,7 +79,6 @@ import org.apache.calcite.rel.core.TableScan;
import org.apache.calcite.rex.RexBuilder;
import org.apache.commons.io.FilenameUtils;
import org.apache.commons.lang3.ObjectUtils;
-import org.apache.commons.lang3.exception.ExceptionUtils;
import org.apache.commons.lang3.tuple.Pair;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileChecksum;
@@ -3220,7 +3218,7 @@ private void constructOneLBLocationMap(FileStatus fSta,
}
}
- public List<org.apache.hadoop.hive.metastore.api.Partition> addPartition(
+ public List<org.apache.hadoop.hive.metastore.api.Partition> addPartitions(
List<org.apache.hadoop.hive.metastore.api.Partition> partitions, boolean ifNotExists, boolean needResults)
throws HiveException {
try {