You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by br...@apache.org on 2013/10/08 02:39:56 UTC
svn commit: r1530119 - in
/hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs: ./
src/main/java/org/apache/hadoop/hdfs/server/namenode/
src/test/java/org/apache/hadoop/hdfs/server/datanode/
Author: brandonli
Date: Tue Oct 8 00:39:56 2013
New Revision: 1530119
URL: http://svn.apache.org/r1530119
Log:
HDFS-5317. Merging change 1530117 from branch-2
Modified:
hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeJspHelper.java
hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java
Modified: hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt?rev=1530119&r1=1530118&r2=1530119&view=diff
==============================================================================
--- hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt (original)
+++ hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt Tue Oct 8 00:39:56 2013
@@ -67,6 +67,9 @@ Release 2.2.0 - 2013-10-13
HDFS-5307. Support both HTTP and HTTPS in jsp pages (Haohui Mai via
branconli)
+ HDFS-5317. Go back to DFS Home link does not work on datanode webUI
+ (Haohui Mai via brandonli)
+
Release 2.1.1-beta - 2013-09-23
INCOMPATIBLE CHANGES
Modified: hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeJspHelper.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeJspHelper.java?rev=1530119&r1=1530118&r2=1530119&view=diff
==============================================================================
--- hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeJspHelper.java (original)
+++ hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NamenodeJspHelper.java Tue Oct 8 00:39:56 2013
@@ -23,6 +23,7 @@ import java.io.IOException;
import java.lang.management.ManagementFactory;
import java.lang.management.MemoryMXBean;
import java.lang.management.MemoryUsage;
+import java.net.InetAddress;
import java.net.InetSocketAddress;
import java.net.URI;
import java.net.URLEncoder;
@@ -60,6 +61,7 @@ import org.apache.hadoop.hdfs.server.nam
import org.apache.hadoop.hdfs.server.namenode.startupprogress.Step;
import org.apache.hadoop.hdfs.server.namenode.startupprogress.StepType;
import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
+import org.apache.hadoop.http.HttpConfig;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.net.NodeBase;
import org.apache.hadoop.security.UserGroupInformation;
@@ -724,11 +726,11 @@ class NamenodeJspHelper {
}
private void generateNodeDataHeader(JspWriter out, DatanodeDescriptor d,
- String suffix, boolean alive, int nnHttpPort, String nnaddr, String scheme)
+ String suffix, boolean alive, int nnInfoPort, String nnaddr, String scheme)
throws IOException {
// from nn_browsedfscontent.jsp:
String url = "///" + JspHelper.Url.authority(scheme, d)
- + "/browseDirectory.jsp?namenodeInfoPort=" + nnHttpPort + "&dir="
+ + "/browseDirectory.jsp?namenodeInfoPort=" + nnInfoPort + "&dir="
+ URLEncoder.encode("/", "UTF-8")
+ JspHelper.getUrlParam(JspHelper.NAMENODE_ADDRESS, nnaddr);
@@ -745,9 +747,9 @@ class NamenodeJspHelper {
}
void generateDecommissioningNodeData(JspWriter out, DatanodeDescriptor d,
- String suffix, boolean alive, int nnHttpPort, String nnaddr, String scheme)
+ String suffix, boolean alive, int nnInfoPort, String nnaddr, String scheme)
throws IOException {
- generateNodeDataHeader(out, d, suffix, alive, nnHttpPort, nnaddr, scheme);
+ generateNodeDataHeader(out, d, suffix, alive, nnInfoPort, nnaddr, scheme);
if (!alive) {
return;
}
@@ -771,7 +773,7 @@ class NamenodeJspHelper {
}
void generateNodeData(JspWriter out, DatanodeDescriptor d, String suffix,
- boolean alive, int nnHttpPort, String nnaddr, String scheme) throws IOException {
+ boolean alive, int nnInfoPort, String nnaddr, String scheme) throws IOException {
/*
* Say the datanode is dn1.hadoop.apache.org with ip 192.168.0.5 we use:
* 1) d.getHostName():d.getPort() to display. Domain and port are stripped
@@ -783,7 +785,7 @@ class NamenodeJspHelper {
* interact with datanodes.
*/
- generateNodeDataHeader(out, d, suffix, alive, nnHttpPort, nnaddr, scheme);
+ generateNodeDataHeader(out, d, suffix, alive, nnInfoPort, nnaddr, scheme);
if (!alive) {
out.print("<td class=\"decommissioned\"> " +
d.isDecommissioned() + "\n");
@@ -849,11 +851,8 @@ class NamenodeJspHelper {
final List<DatanodeDescriptor> dead = new ArrayList<DatanodeDescriptor>();
dm.fetchDatanodes(live, dead, true);
- InetSocketAddress nnSocketAddress =
- (InetSocketAddress)context.getAttribute(
- NameNodeHttpServer.NAMENODE_ADDRESS_ATTRIBUTE_KEY);
- String nnaddr = nnSocketAddress.getAddress().getHostAddress() + ":"
- + nnSocketAddress.getPort();
+ String nnaddr = nn.getServiceRpcAddress().getAddress().getHostName() + ":"
+ + nn.getServiceRpcAddress().getPort();
whatNodes = request.getParameter("whatNodes"); // show only live or only
// dead nodes
@@ -889,16 +888,11 @@ class NamenodeJspHelper {
counterReset();
- try {
- Thread.sleep(1000);
- } catch (InterruptedException e) {
- }
-
if (live.isEmpty() && dead.isEmpty()) {
out.print("There are no datanodes in the cluster");
} else {
- int nnHttpPort = nn.getHttpAddress().getPort();
+ int nnInfoPort = request.getServerPort();
out.print("<div id=\"dfsnodetable\"> ");
if (whatNodes.equals("LIVE")) {
out.print("<a name=\"LiveNodes\" id=\"title\">" + "Live Datanodes : "
@@ -940,7 +934,7 @@ class NamenodeJspHelper {
JspHelper.sortNodeList(live, sorterField, sorterOrder);
for (int i = 0; i < live.size(); i++) {
- generateNodeData(out, live.get(i), port_suffix, true, nnHttpPort,
+ generateNodeData(out, live.get(i), port_suffix, true, nnInfoPort,
nnaddr, request.getScheme());
}
}
@@ -961,7 +955,7 @@ class NamenodeJspHelper {
JspHelper.sortNodeList(dead, sorterField, sorterOrder);
for (int i = 0; i < dead.size(); i++) {
generateNodeData(out, dead.get(i), port_suffix, false,
- nnHttpPort, nnaddr, request.getScheme());
+ nnInfoPort, nnaddr, request.getScheme());
}
out.print("</table>\n");
@@ -992,7 +986,7 @@ class NamenodeJspHelper {
JspHelper.sortNodeList(decommissioning, "name", "ASC");
for (int i = 0; i < decommissioning.size(); i++) {
generateDecommissioningNodeData(out, decommissioning.get(i),
- port_suffix, true, nnHttpPort, nnaddr, request.getScheme());
+ port_suffix, true, nnInfoPort, nnaddr, request.getScheme());
}
out.print("</table>\n");
}
Modified: hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java?rev=1530119&r1=1530118&r2=1530119&view=diff
==============================================================================
--- hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java (original)
+++ hadoop/common/branches/branch-2.2.0/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestDatanodeJsp.java Tue Oct 8 00:39:56 2013
@@ -61,9 +61,10 @@ public class TestDatanodeJsp {
InetSocketAddress nnIpcAddress = cluster.getNameNode().getNameNodeAddress();
InetSocketAddress nnHttpAddress = cluster.getNameNode().getHttpAddress();
- int dnInfoPort = cluster.getDataNodes().get(0).getInfoPort();
-
- URL url = new URL("http://localhost:" + dnInfoPort + "/"
+ String base = JspHelper.Url.url("http", cluster.getDataNodes().get(0)
+ .getDatanodeId());
+
+ URL url = new URL(base + "/"
+ "browseDirectory.jsp" + JspHelper.getUrlParam("dir",
URLEncoder.encode(testPath.toString(), "UTF-8"), true)
+ JspHelper.getUrlParam("namenodeInfoPort", Integer