You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by sz...@apache.org on 2009/05/19 03:05:43 UTC
svn commit: r776146 - in /hadoop/core/trunk: ./
src/hdfs/org/apache/hadoop/hdfs/server/common/
src/hdfs/org/apache/hadoop/hdfs/server/datanode/
src/hdfs/org/apache/hadoop/hdfs/server/namenode/
Author: szetszwo
Date: Tue May 19 01:05:43 2009
New Revision: 776146
URL: http://svn.apache.org/viewvc?rev=776146&view=rev
Log:
HADOOP-5829. Fix javac warnings found in ReplicationTargetChooser, FSImage, Checkpointer, SecondaryNameNode and a few other hdfs classes. Contributed by Suresh Srinivas
Modified:
hadoop/core/trunk/CHANGES.txt
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/InconsistentFSStateException.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/IncorrectVersionException.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/datanode/DataBlockScanner.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/FSImage.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/LeaseExpiredException.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/NotReplicatedYetException.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/ReplicationTargetChooser.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SafeModeException.java
hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
Modified: hadoop/core/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/CHANGES.txt?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/CHANGES.txt (original)
+++ hadoop/core/trunk/CHANGES.txt Tue May 19 01:05:43 2009
@@ -617,6 +617,10 @@
HADOOP-5582. Fixes a problem in Hadoop Vaidya to do with reading
counters from job history files. (Suhas Gogate via ddas)
+ HADOOP-5829. Fix javac warnings found in ReplicationTargetChooser,
+ FSImage, Checkpointer, SecondaryNameNode and a few other hdfs classes.
+ (Suresh Srinivas via szetszwo)
+
Release 0.20.1 - Unreleased
INCOMPATIBLE CHANGES
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/InconsistentFSStateException.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/InconsistentFSStateException.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/InconsistentFSStateException.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/InconsistentFSStateException.java Tue May 19 01:05:43 2009
@@ -27,6 +27,7 @@
*
*/
public class InconsistentFSStateException extends IOException {
+ private static final long serialVersionUID = 1L;
public InconsistentFSStateException(File dir, String descr) {
super("Directory " + getFilePath(dir)
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/IncorrectVersionException.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/IncorrectVersionException.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/IncorrectVersionException.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/common/IncorrectVersionException.java Tue May 19 01:05:43 2009
@@ -23,10 +23,11 @@
/**
* The exception is thrown when external version does not match
- * current version of the appication.
+ * current version of the application.
*
*/
public class IncorrectVersionException extends IOException {
+ private static final long serialVersionUID = 1L;
public IncorrectVersionException(int versionReported, String ofWhat) {
this(versionReported, ofWhat, FSConstants.LAYOUT_VERSION);
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/datanode/DataBlockScanner.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/datanode/DataBlockScanner.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/datanode/DataBlockScanner.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/datanode/DataBlockScanner.java Tue May 19 01:05:43 2009
@@ -957,6 +957,8 @@
}
public static class Servlet extends HttpServlet {
+ private static final long serialVersionUID = 1L;
+
public void doGet(HttpServletRequest request,
HttpServletResponse response) throws IOException {
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/Checkpointer.java Tue May 19 01:05:43 2009
@@ -21,6 +21,7 @@
import java.io.File;
import java.net.InetAddress;
import java.net.InetSocketAddress;
+import java.util.Collection;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
@@ -28,6 +29,8 @@
import org.apache.hadoop.hdfs.protocol.FSConstants;
import org.apache.hadoop.hdfs.server.common.HdfsConstants.NamenodeRole;
import org.apache.hadoop.hdfs.server.namenode.FSImage.CheckpointStates;
+import org.apache.hadoop.hdfs.server.namenode.FSImage.NameNodeDirType;
+import org.apache.hadoop.hdfs.server.namenode.FSImage.NameNodeFile;
import org.apache.hadoop.hdfs.server.protocol.CheckpointCommand;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol;
import org.apache.hadoop.hdfs.server.protocol.NamenodeCommand;
@@ -165,7 +168,9 @@
private void downloadCheckpoint(CheckpointSignature sig) throws IOException {
// Retrieve image file
String fileid = "getimage=1";
- File[] files = getFSImage().getImageFiles();
+ Collection<File> list = getFSImage().getFiles(NameNodeFile.IMAGE,
+ NameNodeDirType.IMAGE);
+ File[] files = list.toArray(new File[list.size()]);
assert files.length > 0 : "No checkpoint targets.";
String nnHttpAddr = backupNode.nnHttpAddress;
TransferFsImage.getFileClient(nnHttpAddr, fileid, files);
@@ -174,7 +179,8 @@
// Retrieve edits file
fileid = "getedit=1";
- files = getFSImage().getEditsFiles();
+ list = getFSImage().getFiles(NameNodeFile.EDITS, NameNodeDirType.EDITS);
+ files = list.toArray(new File[list.size()]);
assert files.length > 0 : "No checkpoint targets.";
TransferFsImage.getFileClient(nnHttpAddr, fileid, files);
LOG.info("Downloaded file " + files[0].getName() + " size " +
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/FSImage.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/FSImage.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/FSImage.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/FSImage.java Tue May 19 01:05:43 2009
@@ -256,27 +256,6 @@
return list;
}
- @Deprecated // Use getFiles() instead
- File[] getFileNames(NameNodeFile type, NameNodeDirType dirType) {
- Collection<File> list = getFiles(type, dirType);
- return list.toArray(new File[list.size()]);
- }
-
- @Deprecated
- File[] getImageFiles() {
- return getFileNames(NameNodeFile.IMAGE, NameNodeDirType.IMAGE);
- }
-
- @Deprecated
- File[] getEditsFiles() {
- return getFileNames(NameNodeFile.EDITS, NameNodeDirType.EDITS);
- }
-
- @Deprecated // should be removed
- File[] getTimeFiles() {
- return getFileNames(NameNodeFile.TIME, null);
- }
-
Collection<File> getDirectories(NameNodeDirType dirType) {
ArrayList<File> list = new ArrayList<File>();
Iterator<StorageDirectory> it = (dirType == null) ? dirIterator() :
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/LeaseExpiredException.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/LeaseExpiredException.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/LeaseExpiredException.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/LeaseExpiredException.java Tue May 19 01:05:43 2009
@@ -24,6 +24,8 @@
* The lease that was being used to create this file has expired.
*/
public class LeaseExpiredException extends IOException {
+ private static final long serialVersionUID = 1L;
+
public LeaseExpiredException(String msg) {
super(msg);
}
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/NotReplicatedYetException.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/NotReplicatedYetException.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/NotReplicatedYetException.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/NotReplicatedYetException.java Tue May 19 01:05:43 2009
@@ -24,6 +24,8 @@
* The file has not finished being written to enough datanodes yet.
*/
public class NotReplicatedYetException extends IOException {
+ private static final long serialVersionUID = 1L;
+
public NotReplicatedYetException(String msg) {
super(msg);
}
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/ReplicationTargetChooser.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/ReplicationTargetChooser.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/ReplicationTargetChooser.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/ReplicationTargetChooser.java Tue May 19 01:05:43 2009
@@ -47,6 +47,8 @@
}
private static class NotEnoughReplicasException extends Exception {
+ private static final long serialVersionUID = 1L;
+
NotEnoughReplicasException(String msg) {
super(msg);
}
@@ -147,24 +149,25 @@
int numOfResults = results.size();
boolean newBlock = (numOfResults==0);
if (writer == null && !newBlock) {
- writer = (DatanodeDescriptor)results.get(0);
+ writer = results.get(0);
}
try {
- switch(numOfResults) {
- case 0:
+ if (numOfResults == 0) {
writer = chooseLocalNode(writer, excludedNodes,
blocksize, maxNodesPerRack, results);
if (--numOfReplicas == 0) {
- break;
+ return writer;
}
- case 1:
+ }
+ if (numOfResults <= 1) {
chooseRemoteRack(1, results.get(0), excludedNodes,
blocksize, maxNodesPerRack, results);
if (--numOfReplicas == 0) {
- break;
+ return writer;
}
- case 2:
+ }
+ if (numOfResults <= 2) {
if (clusterMap.isOnSameRack(results.get(0), results.get(1))) {
chooseRemoteRack(1, results.get(0), excludedNodes,
blocksize, maxNodesPerRack, results);
@@ -176,12 +179,11 @@
maxNodesPerRack, results);
}
if (--numOfReplicas == 0) {
- break;
+ return writer;
}
- default:
- chooseRandom(numOfReplicas, NodeBase.ROOT, excludedNodes,
- blocksize, maxNodesPerRack, results);
}
+ chooseRandom(numOfReplicas, NodeBase.ROOT, excludedNodes,
+ blocksize, maxNodesPerRack, results);
} catch (NotEnoughReplicasException e) {
FSNamesystem.LOG.warn("Not able to place enough replicas, still in need of "
+ numOfReplicas);
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SafeModeException.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SafeModeException.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SafeModeException.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SafeModeException.java Tue May 19 01:05:43 2009
@@ -26,6 +26,7 @@
*
*/
public class SafeModeException extends IOException {
+ private static final long serialVersionUID = 1L;
public SafeModeException(String text, FSNamesystem.SafeModeInfo mode ) {
super(text + ". Name node is in safe mode.\n" + mode.getTurnOffTip());
Modified: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java?rev=776146&r1=776145&r2=776146&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/namenode/SecondaryNameNode.java Tue May 19 01:05:43 2009
@@ -34,6 +34,8 @@
import org.apache.hadoop.hdfs.DFSUtil.ErrorSimulator;
import org.apache.hadoop.hdfs.server.common.HdfsConstants;
import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException;
+import org.apache.hadoop.hdfs.server.namenode.FSImage.NameNodeDirType;
+import org.apache.hadoop.hdfs.server.namenode.FSImage.NameNodeFile;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol;
import org.apache.hadoop.http.HttpServer;
import org.apache.hadoop.ipc.RPC;
@@ -237,7 +239,9 @@
// get fsimage
String fileid = "getimage=1";
- File[] srcNames = checkpointImage.getImageFiles();
+ Collection<File> list = checkpointImage.getFiles(NameNodeFile.IMAGE,
+ NameNodeDirType.IMAGE);
+ File[] srcNames = list.toArray(new File[list.size()]);
assert srcNames.length > 0 : "No checkpoint targets.";
TransferFsImage.getFileClient(fsName, fileid, srcNames);
LOG.info("Downloaded file " + srcNames[0].getName() + " size " +
@@ -245,7 +249,8 @@
// get edits file
fileid = "getedit=1";
- srcNames = checkpointImage.getEditsFiles();
+ list = getFSImage().getFiles(NameNodeFile.EDITS, NameNodeDirType.EDITS);
+ srcNames = list.toArray(new File[list.size()]);;
assert srcNames.length > 0 : "No checkpoint targets.";
TransferFsImage.getFileClient(fsName, fileid, srcNames);
LOG.info("Downloaded file " + srcNames[0].getName() + " size " +