You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by mg...@apache.org on 2020/05/04 09:27:13 UTC

[hive] branch master updated: HIVE-22028 Clean up Add Partition ADDENDUM - fix typo (Miklos Gergely, reviewed by Zoltan Haindrich)

This is an automated email from the ASF dual-hosted git repository.

mgergely pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hive.git


The following commit(s) were added to refs/heads/master by this push:
     new 342f8fb  HIVE-22028 Clean up Add Partition ADDENDUM - fix typo (Miklos Gergely, reviewed by Zoltan Haindrich)
342f8fb is described below

commit 342f8fbd521ce5728098887cd1749fc7931e370f
Author: miklosgergely <mg...@cloudera.com>
AuthorDate: Sat May 2 16:42:30 2020 +0200

    HIVE-22028 Clean up Add Partition ADDENDUM - fix typo (Miklos Gergely, reviewed by Zoltan Haindrich)
---
 .../table/partition/add/AlterTableAddPartitionOperation.java   | 10 +++++-----
 ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java       |  4 +---
 2 files changed, 6 insertions(+), 8 deletions(-)

diff --git a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java
index 6910e10..ddc47a4 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/ddl/table/partition/add/AlterTableAddPartitionOperation.java
@@ -157,7 +157,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
       List<Partition> partitions) throws HiveException {
     // TODO: normally, the result is not necessary; might make sense to pass false
     List<org.apache.hadoop.hive.ql.metadata.Partition> outPartitions = new ArrayList<>();
-    for (Partition outPart : context.getDb().addPartition(partitions, desc.isIfNotExists(), true)) {
+    for (Partition outPart : context.getDb().addPartitions(partitions, desc.isIfNotExists(), true)) {
       outPartitions.add(new org.apache.hadoop.hive.ql.metadata.Partition(table, outPart));
     }
     return outPartitions;
@@ -170,7 +170,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
     // no choice but to iterate over the partitions here.
 
     List<Partition> partitionsToAdd = new ArrayList<>();
-    List<Partition> partitionssToAlter = new ArrayList<>();
+    List<Partition> partitionsToAlter = new ArrayList<>();
     List<String> partitionNames = new ArrayList<>();
     for (Partition partition : partitions){
       partitionNames.add(getPartitionName(table, partition));
@@ -178,7 +178,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
         Partition p = context.getDb().getPartition(desc.getDbName(), desc.getTableName(), partition.getValues());
         if (desc.getReplicationSpec().allowReplacementInto(p.getParameters())){
           ReplicationSpec.copyLastReplId(p.getParameters(), partition.getParameters());
-          partitionssToAlter.add(partition);
+          partitionsToAlter.add(partition);
         } // else ptn already exists, but we do nothing with it.
       } catch (HiveException e){
         if (e.getCause() instanceof NoSuchObjectException) {
@@ -191,7 +191,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
     }
 
     List<org.apache.hadoop.hive.ql.metadata.Partition> outPartitions = new ArrayList<>();
-    for (Partition outPartition : context.getDb().addPartition(partitionsToAdd, desc.isIfNotExists(), true)) {
+    for (Partition outPartition : context.getDb().addPartitions(partitionsToAdd, desc.isIfNotExists(), true)) {
       outPartitions.add(new org.apache.hadoop.hive.ql.metadata.Partition(table, outPartition));
     }
 
@@ -199,7 +199,7 @@ public class AlterTableAddPartitionOperation extends DDLOperation<AlterTableAddP
     EnvironmentContext ec = new EnvironmentContext();
     ec.putToProperties(StatsSetupConst.DO_NOT_UPDATE_STATS, StatsSetupConst.TRUE);
     String validWriteIdList = getValidWriteIdList(table, writeId);
-    context.getDb().alterPartitions(desc.getDbName(), desc.getTableName(), partitionssToAlter, ec, validWriteIdList,
+    context.getDb().alterPartitions(desc.getDbName(), desc.getTableName(), partitionsToAlter, ec, validWriteIdList,
         writeId);
 
     for (Partition outPartition : context.getDb().getPartitionsByNames(desc.getDbName(), desc.getTableName(),
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
index bd1fae1..f6a5207 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
@@ -40,7 +40,6 @@ import java.io.PrintStream;
 import java.net.InetAddress;
 import java.net.UnknownHostException;
 import java.nio.ByteBuffer;
-import java.sql.SQLIntegrityConstraintViolationException;
 import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.BitSet;
@@ -80,7 +79,6 @@ import org.apache.calcite.rel.core.TableScan;
 import org.apache.calcite.rex.RexBuilder;
 import org.apache.commons.io.FilenameUtils;
 import org.apache.commons.lang3.ObjectUtils;
-import org.apache.commons.lang3.exception.ExceptionUtils;
 import org.apache.commons.lang3.tuple.Pair;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileChecksum;
@@ -3220,7 +3218,7 @@ private void constructOneLBLocationMap(FileStatus fSta,
     }
   }
 
-  public List<org.apache.hadoop.hive.metastore.api.Partition> addPartition(
+  public List<org.apache.hadoop.hive.metastore.api.Partition> addPartitions(
       List<org.apache.hadoop.hive.metastore.api.Partition> partitions, boolean ifNotExists, boolean needResults)
           throws HiveException {
     try {