You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by at...@apache.org on 2011/06/26 07:39:50 UTC
svn commit: r1139715 - in /hadoop/common/trunk/hdfs: CHANGES.txt
src/test/hdfs/org/apache/hadoop/hdfs/MiniDFSCluster.java
src/test/hdfs/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java
src/test/hdfs/org/apache/hadoop/hdfs/UpgradeUtilities.java
Author: atm
Date: Sun Jun 26 05:39:50 2011
New Revision: 1139715
URL: http://svn.apache.org/viewvc?rev=1139715&view=rev
Log:
HDFS-1381. HDFS javadocs hard-code references to dfs.namenode.name.dir and dfs.datanode.data.dir parameters (Jim Plush via atm)
Modified:
hadoop/common/trunk/hdfs/CHANGES.txt
hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/MiniDFSCluster.java
hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java
hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/UpgradeUtilities.java
Modified: hadoop/common/trunk/hdfs/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hdfs/CHANGES.txt?rev=1139715&r1=1139714&r2=1139715&view=diff
==============================================================================
--- hadoop/common/trunk/hdfs/CHANGES.txt (original)
+++ hadoop/common/trunk/hdfs/CHANGES.txt Sun Jun 26 05:39:50 2011
@@ -780,6 +780,9 @@ Trunk (unreleased changes)
HDFS-1321. If service port and main port are the same, there is no clear
log message explaining the issue. (Jim Plush via atm)
+ HDFS-1381. HDFS javadocs hard-code references to dfs.namenode.name.dir and
+ dfs.datanode.data.dir parameters (Jim Plush via atm)
+
Release 0.22.0 - Unreleased
INCOMPATIBLE CHANGES
Modified: hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/MiniDFSCluster.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/MiniDFSCluster.java?rev=1139715&r1=1139714&r2=1139715&view=diff
==============================================================================
--- hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/MiniDFSCluster.java (original)
+++ hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/MiniDFSCluster.java Sun Jun 26 05:39:50 2011
@@ -309,7 +309,8 @@ public class MiniDFSCluster {
* Servers will be started on free ports.
* <p>
* The caller must manage the creation of NameNode and DataNode directories
- * and have already set dfs.namenode.name.dir and dfs.datanode.data.dir in the given conf.
+ * and have already set {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY} and
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} in the given conf.
*
* @param conf the base configuration to use in starting the servers. This
* will be modified as necessary.
@@ -370,7 +371,8 @@ public class MiniDFSCluster {
/**
* NOTE: if possible, the other constructors that don't have nameNode port
- * parameter should be used as they will ensure that the servers use free ports.
+ * parameter should be used as they will ensure that the servers use free
+ * ports.
* <p>
* Modify the config and start up the servers.
*
@@ -379,9 +381,12 @@ public class MiniDFSCluster {
* @param conf the base configuration to use in starting the servers. This
* will be modified as necessary.
* @param numDataNodes Number of DataNodes to start; may be zero
- * @param format if true, format the NameNode and DataNodes before starting up
+ * @param format if true, format the NameNode and DataNodes before starting
+ * up
* @param manageDfsDirs if true, the data directories for servers will be
- * created and dfs.namenode.name.dir and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY} and
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be set in
+ * the conf
* @param operation the operation with which to start the servers. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -411,7 +416,9 @@ public class MiniDFSCluster {
* @param numDataNodes Number of DataNodes to start; may be zero
* @param format if true, format the NameNode and DataNodes before starting up
* @param manageDfsDirs if true, the data directories for servers will be
- * created and dfs.namenode.name.dir and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY} and
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be set in
+ * the conf
* @param operation the operation with which to start the servers. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -443,9 +450,12 @@ public class MiniDFSCluster {
* @param numDataNodes Number of DataNodes to start; may be zero
* @param format if true, format the NameNode and DataNodes before starting up
* @param manageNameDfsDirs if true, the data directories for servers will be
- * created and dfs.namenode.name.dir and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY} and
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be set in
+ * the conf
* @param manageDataDfsDirs if true, the data directories for datanodes will
- * be created and dfs.datanode.data.dir set to same in the conf
+ * be created and {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY}
+ * set to same in the conf
* @param operation the operation with which to start the servers. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -710,7 +720,8 @@ public class MiniDFSCluster {
* will be modified as necessary.
* @param numDataNodes Number of DataNodes to start; may be zero
* @param manageDfsDirs if true, the data directories for DataNodes will be
- * created and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be set
+ * in the conf
* @param operation the operation with which to start the DataNodes. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -741,7 +752,8 @@ public class MiniDFSCluster {
* will be modified as necessary.
* @param numDataNodes Number of DataNodes to start; may be zero
* @param manageDfsDirs if true, the data directories for DataNodes will be
- * created and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be
+ * set in the conf
* @param operation the operation with which to start the DataNodes. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -774,7 +786,8 @@ public class MiniDFSCluster {
* will be modified as necessary.
* @param numDataNodes Number of DataNodes to start; may be zero
* @param manageDfsDirs if true, the data directories for DataNodes will be
- * created and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be
+ * set in the conf
* @param operation the operation with which to start the DataNodes. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -900,7 +913,8 @@ public class MiniDFSCluster {
* will be modified as necessary.
* @param numDataNodes Number of DataNodes to start; may be zero
* @param manageDfsDirs if true, the data directories for DataNodes will be
- * created and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will be
+ * set in the conf
* @param operation the operation with which to start the DataNodes. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
@@ -930,7 +944,8 @@ public class MiniDFSCluster {
* will be modified as necessary.
* @param numDataNodes Number of DataNodes to start; may be zero
* @param manageDfsDirs if true, the data directories for DataNodes will be
- * created and dfs.datanode.data.dir will be set in the conf
+ * created and {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} will
+ * be set in the conf
* @param operation the operation with which to start the DataNodes. If null
* or StartupOption.FORMAT, then StartupOption.REGULAR will be used.
* @param racks array of strings indicating the rack that each DataNode is on
Modified: hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java?rev=1139715&r1=1139714&r2=1139715&view=diff
==============================================================================
--- hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java (original)
+++ hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java Sun Jun 26 05:39:50 2011
@@ -110,7 +110,8 @@ public class TestDFSStorageStateRecovery
/**
* Sets up the storage directories for namenode as defined by
- * dfs.namenode.name.dir. For each element in dfs.namenode.name.dir, the subdirectories
+ * {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY}. For each element
+ * in {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY}, the subdirectories
* represented by the first four elements of the <code>state</code> array
* will be created and populated.
*
@@ -139,7 +140,8 @@ public class TestDFSStorageStateRecovery
/**
* Sets up the storage directories for a datanode under
- * dfs.datanode.data.dir. For each element in dfs.datanode.data.dir, the subdirectories
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY}. For each element in
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY}, the subdirectories
* represented by the first four elements of the <code>state</code> array
* will be created and populated.
* See {@link UpgradeUtilities#createDataNodeStorageDirs()}
@@ -167,7 +169,8 @@ public class TestDFSStorageStateRecovery
/**
* Sets up the storage directories for a block pool under
- * dfs.datanode.data.dir. For each element in dfs.datanode.data.dir, the subdirectories
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY}. For each element
+ * in {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY}, the subdirectories
* represented by the first four elements of the <code>state</code> array
* will be created and populated.
* See {@link UpgradeUtilities#createBlockPoolStorageDirs()}
Modified: hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/UpgradeUtilities.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/UpgradeUtilities.java?rev=1139715&r1=1139714&r2=1139715&view=diff
==============================================================================
--- hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/UpgradeUtilities.java (original)
+++ hadoop/common/trunk/hdfs/src/test/hdfs/org/apache/hadoop/hdfs/UpgradeUtilities.java Sun Jun 26 05:39:50 2011
@@ -181,8 +181,9 @@ public class UpgradeUtilities {
}
/**
- * Initialize dfs.namenode.name.dir and dfs.datanode.data.dir with the specified number of
- * directory entries. Also initialize dfs.blockreport.intervalMsec.
+ * Initialize {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY} and
+ * {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} with the specified
+ * number of directory entries. Also initialize dfs.blockreport.intervalMsec.
*/
public static Configuration initializeStorageStateConf(int numDirs,
Configuration conf) {
@@ -306,7 +307,8 @@ public class UpgradeUtilities {
}
/**
- * Simulate the <code>dfs.namenode.name.dir</code> of a populated DFS filesystem.
+ * Simulate the {@link DFSConfigKeys#DFS_NAMENODE_NAME_DIR_KEY} of a populated
+ * DFS filesystem.
* This method populates for each parent directory, <code>parent/dirName</code>
* with the content of namenode storage directory that comes from a singleton
* namenode master (that contains edits, fsimage, version and time files).
@@ -333,7 +335,8 @@ public class UpgradeUtilities {
}
/**
- * Simulate the <code>dfs.datanode.data.dir</code> of a populated DFS filesystem.
+ * Simulate the {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} of a
+ * populated DFS filesystem.
* This method populates for each parent directory, <code>parent/dirName</code>
* with the content of datanode storage directory that comes from a singleton
* datanode master (that contains version and block files). If the destination
@@ -360,7 +363,8 @@ public class UpgradeUtilities {
}
/**
- * Simulate the <code>dfs.datanode.data.dir</code> of a populated DFS filesystem.
+ * Simulate the {@link DFSConfigKeys#DFS_DATANODE_DATA_DIR_KEY} of a
+ * populated DFS filesystem.
* This method populates for each parent directory, <code>parent/dirName</code>
* with the content of block pool storage directory that comes from a singleton
* datanode master (that contains version and block files). If the destination