You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by an...@apache.org on 2015/05/25 18:42:19 UTC

hbase git commit: HBASE-13760 Cleanup Findbugs keySet iterator warnings. (Gabor Liptak)

Repository: hbase
Updated Branches:
  refs/heads/master d45e0a7d4 -> 9ee184b67


HBASE-13760 Cleanup Findbugs keySet iterator warnings. (Gabor Liptak)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/9ee184b6
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/9ee184b6
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/9ee184b6

Branch: refs/heads/master
Commit: 9ee184b6733d38e2ae764781dda427b7fd9f165f
Parents: d45e0a7
Author: anoopsjohn <an...@gmail.com>
Authored: Mon May 25 22:11:49 2015 +0530
Committer: anoopsjohn <an...@gmail.com>
Committed: Mon May 25 22:11:49 2015 +0530

----------------------------------------------------------------------
 .../apache/hadoop/hbase/HColumnDescriptor.java  |  8 +++---
 .../apache/hadoop/hbase/HTableDescriptor.java   | 14 +++++------
 .../hadoop/hbase/master/ServerManager.java      |  4 +--
 .../apache/hadoop/hbase/wal/WALSplitter.java    | 26 +++++++++++---------
 4 files changed, 27 insertions(+), 25 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/9ee184b6/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
index 2cb587b..20cfbef 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
@@ -920,13 +920,13 @@ public class HColumnDescriptor implements Comparable<HColumnDescriptor> {
     boolean hasConfigKeys = false;
 
     // print all reserved keys first
-    for (Bytes k : values.keySet()) {
-      if (!RESERVED_KEYWORDS.contains(k)) {
+    for (Map.Entry<Bytes, Bytes> entry : values.entrySet()) {
+      if (!RESERVED_KEYWORDS.contains(entry.getKey())) {
         hasConfigKeys = true;
         continue;
       }
-      String key = Bytes.toString(k.get());
-      String value = Bytes.toStringBinary(values.get(k).get());
+      String key = Bytes.toString(entry.getKey().get());
+      String value = Bytes.toStringBinary(entry.getValue().get());
       if (printDefaults
           || !DEFAULT_VALUES.containsKey(key)
           || !DEFAULT_VALUES.get(key).equalsIgnoreCase(value)) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/9ee184b6/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
index a0a3c79..f2f0077 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -886,16 +886,16 @@ public class HTableDescriptor implements Comparable<HTableDescriptor> {
     // step 1: set partitioning and pruning
     Set<Bytes> reservedKeys = new TreeSet<Bytes>();
     Set<Bytes> userKeys = new TreeSet<Bytes>();
-    for (Bytes k : values.keySet()) {
-      if (k == null || k.get() == null) continue;
-      String key = Bytes.toString(k.get());
+    for (Map.Entry<Bytes, Bytes> entry : values.entrySet()) {
+      if (entry.getKey() == null || entry.getKey().get() == null) continue;
+      String key = Bytes.toString(entry.getKey().get());
       // in this section, print out reserved keywords + coprocessor info
-      if (!RESERVED_KEYWORDS.contains(k) && !key.startsWith("coprocessor$")) {
-        userKeys.add(k);
+      if (!RESERVED_KEYWORDS.contains(entry.getKey()) && !key.startsWith("coprocessor$")) {
+        userKeys.add(entry.getKey());
         continue;
       }
       // only print out IS_ROOT/IS_META if true
-      String value = Bytes.toString(values.get(k).get());
+      String value = Bytes.toString(entry.getValue().get());
       if (key.equalsIgnoreCase(IS_ROOT) || key.equalsIgnoreCase(IS_META)) {
         if (Boolean.valueOf(value) == false) continue;
       }
@@ -903,7 +903,7 @@ public class HTableDescriptor implements Comparable<HTableDescriptor> {
       if (printDefaults
           || !DEFAULT_VALUES.containsKey(key)
           || !DEFAULT_VALUES.get(key).equalsIgnoreCase(value)) {
-        reservedKeys.add(k);
+        reservedKeys.add(entry.getKey());
       }
     }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/9ee184b6/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
index 1ed2514..fa99a92 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
@@ -677,8 +677,8 @@ public class ServerManager {
       LOG.info("AssignmentManager hasn't finished failover cleanup; waiting");
     }
 
-    for(ServerName tmpServerName : requeuedDeadServers.keySet()){
-      processDeadServer(tmpServerName, requeuedDeadServers.get(tmpServerName));
+    for (Map.Entry<ServerName, Boolean> entry : requeuedDeadServers.entrySet()) {
+      processDeadServer(entry.getKey(), entry.getValue());
     }
     requeuedDeadServers.clear();
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/9ee184b6/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java
index 2c68a68..b9b8486 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/wal/WALSplitter.java
@@ -70,7 +70,6 @@ import org.apache.hadoop.hbase.ServerName;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.TableNotFoundException;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.client.Connection;
 import org.apache.hadoop.hbase.client.ConnectionFactory;
 import org.apache.hadoop.hbase.client.ConnectionUtils;
 import org.apache.hadoop.hbase.client.Delete;
@@ -1673,12 +1672,13 @@ public class WALSplitter {
       int maxSize = 0;
       List<Pair<HRegionLocation, Entry>> maxQueue = null;
       synchronized (this.serverToBufferQueueMap) {
-        for (String key : this.serverToBufferQueueMap.keySet()) {
-          List<Pair<HRegionLocation, Entry>> curQueue = this.serverToBufferQueueMap.get(key);
+        for (Map.Entry<String, List<Pair<HRegionLocation, Entry>>> entry :
+            this.serverToBufferQueueMap.entrySet()) {
+          List<Pair<HRegionLocation, Entry>> curQueue = entry.getValue();
           if (curQueue.size() > maxSize) {
             maxSize = curQueue.size();
             maxQueue = curQueue;
-            maxLocKey = key;
+            maxLocKey = entry.getKey();
           }
         }
         if (maxSize < minBatchSize
@@ -1969,11 +1969,12 @@ public class WALSplitter {
       int curSize = 0;
       List<Pair<HRegionLocation, Entry>> curQueue = null;
       synchronized (this.serverToBufferQueueMap) {
-        for (String locationKey : this.serverToBufferQueueMap.keySet()) {
-          curQueue = this.serverToBufferQueueMap.get(locationKey);
+        for (Map.Entry<String, List<Pair<HRegionLocation, Entry>>> entry :
+                this.serverToBufferQueueMap.entrySet()) {
+          curQueue = entry.getValue();
           if (!curQueue.isEmpty()) {
             curSize = curQueue.size();
-            curLoc = locationKey;
+            curLoc = entry.getKey();
             break;
           }
         }
@@ -2043,12 +2044,12 @@ public class WALSplitter {
           }
         } finally {
           synchronized (writers) {
-            for (String locationKey : writers.keySet()) {
-              RegionServerWriter tmpW = writers.get(locationKey);
+            for (Map.Entry<String, RegionServerWriter> entry : writers.entrySet()) {
+              RegionServerWriter tmpW = entry.getValue();
               try {
                 tmpW.close();
               } catch (IOException ioe) {
-                LOG.error("Couldn't close writer for region server:" + locationKey, ioe);
+                LOG.error("Couldn't close writer for region server:" + entry.getKey(), ioe);
                 result.add(ioe);
               }
             }
@@ -2056,8 +2057,9 @@ public class WALSplitter {
 
           // close connections
           synchronized (this.tableNameToHConnectionMap) {
-            for (TableName tableName : this.tableNameToHConnectionMap.keySet()) {
-              HConnection hconn = this.tableNameToHConnectionMap.get(tableName);
+            for (Map.Entry<TableName,HConnection> entry :
+                    this.tableNameToHConnectionMap.entrySet()) {
+              HConnection hconn = entry.getValue();
               try {
                 hconn.clearRegionCache();
                 hconn.close();