You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by ha...@apache.org on 2012/07/16 05:07:53 UTC
svn commit: r1361895 [3/3] - in
/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs: ./
src/main/java/org/apache/hadoop/hdfs/
src/main/java/org/apache/hadoop/hdfs/protocol/
src/main/java/org/apache/hadoop/hdfs/protocolPB/ src/main/java/org...
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/NNThroughputBenchmark.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/NNThroughputBenchmark.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/NNThroughputBenchmark.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/NNThroughputBenchmark.java Mon Jul 16 03:07:49 2012
@@ -392,6 +392,7 @@ public class NNThroughputBenchmark {
setName(toString());
}
+ @Override
public void run() {
localNumOpsExecuted = 0;
localCumulativeTime = 0;
@@ -404,6 +405,7 @@ public class NNThroughputBenchmark {
}
}
+ @Override
public String toString() {
return "StatsDaemon-" + daemonId;
}
@@ -446,16 +448,19 @@ public class NNThroughputBenchmark {
keepResults = true;
}
+ @Override
String getOpName() {
return OP_CLEAN_NAME;
}
+ @Override
void parseArguments(List<String> args) {
boolean ignoreUnrelatedOptions = verifyOpArgument(args);
if(args.size() > 2 && !ignoreUnrelatedOptions)
printUsage();
}
+ @Override
void generateInputs(int[] opsPerThread) throws IOException {
// do nothing
}
@@ -463,6 +468,7 @@ public class NNThroughputBenchmark {
/**
* Does not require the argument
*/
+ @Override
String getExecutionArgument(int daemonId) {
return null;
}
@@ -470,6 +476,7 @@ public class NNThroughputBenchmark {
/**
* Remove entire benchmark directory.
*/
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore)
throws IOException {
nameNodeProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE);
@@ -479,6 +486,7 @@ public class NNThroughputBenchmark {
return end-start;
}
+ @Override
void printResults() {
LOG.info("--- " + getOpName() + " inputs ---");
LOG.info("Remove directory " + BASE_DIR_NAME);
@@ -508,10 +516,12 @@ public class NNThroughputBenchmark {
parseArguments(args);
}
+ @Override
String getOpName() {
return OP_CREATE_NAME;
}
+ @Override
void parseArguments(List<String> args) {
boolean ignoreUnrelatedOptions = verifyOpArgument(args);
int nrFilesPerDir = 4;
@@ -534,6 +544,7 @@ public class NNThroughputBenchmark {
nameGenerator = new FileNameGenerator(getBaseDir(), nrFilesPerDir);
}
+ @Override
void generateInputs(int[] opsPerThread) throws IOException {
assert opsPerThread.length == numThreads : "Error opsPerThread.length";
nameNodeProto.setSafeMode(HdfsConstants.SafeModeAction.SAFEMODE_LEAVE);
@@ -557,6 +568,7 @@ public class NNThroughputBenchmark {
/**
* returns client name
*/
+ @Override
String getExecutionArgument(int daemonId) {
return getClientName(daemonId);
}
@@ -564,6 +576,7 @@ public class NNThroughputBenchmark {
/**
* Do file create.
*/
+ @Override
long executeOp(int daemonId, int inputIdx, String clientName)
throws IOException {
long start = Time.now();
@@ -578,6 +591,7 @@ public class NNThroughputBenchmark {
return end-start;
}
+ @Override
void printResults() {
LOG.info("--- " + getOpName() + " inputs ---");
LOG.info("nrFiles = " + numOpsRequired);
@@ -607,10 +621,12 @@ public class NNThroughputBenchmark {
super(args);
}
+ @Override
String getOpName() {
return OP_OPEN_NAME;
}
+ @Override
void parseArguments(List<String> args) {
int ueIndex = args.indexOf("-useExisting");
useExisting = (ueIndex >= 0);
@@ -620,6 +636,7 @@ public class NNThroughputBenchmark {
super.parseArguments(args);
}
+ @Override
void generateInputs(int[] opsPerThread) throws IOException {
// create files using opsPerThread
String[] createArgs = new String[] {
@@ -652,6 +669,7 @@ public class NNThroughputBenchmark {
/**
* Do file open.
*/
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore)
throws IOException {
long start = Time.now();
@@ -676,10 +694,12 @@ public class NNThroughputBenchmark {
super(args);
}
+ @Override
String getOpName() {
return OP_DELETE_NAME;
}
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore)
throws IOException {
long start = Time.now();
@@ -704,10 +724,12 @@ public class NNThroughputBenchmark {
super(args);
}
+ @Override
String getOpName() {
return OP_FILE_STATUS_NAME;
}
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore)
throws IOException {
long start = Time.now();
@@ -734,10 +756,12 @@ public class NNThroughputBenchmark {
super(args);
}
+ @Override
String getOpName() {
return OP_RENAME_NAME;
}
+ @Override
void generateInputs(int[] opsPerThread) throws IOException {
super.generateInputs(opsPerThread);
destNames = new String[fileNames.length][];
@@ -749,6 +773,7 @@ public class NNThroughputBenchmark {
}
}
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore)
throws IOException {
long start = Time.now();
@@ -793,6 +818,7 @@ public class NNThroughputBenchmark {
this.nrBlocks = 0;
}
+ @Override
public String toString() {
return dnRegistration.toString();
}
@@ -868,6 +894,7 @@ public class NNThroughputBenchmark {
return blockReportList;
}
+ @Override
public int compareTo(String xferAddr) {
return getXferAddr().compareTo(xferAddr);
}
@@ -959,10 +986,12 @@ public class NNThroughputBenchmark {
return numThreads;
}
+ @Override
String getOpName() {
return OP_BLOCK_REPORT_NAME;
}
+ @Override
void parseArguments(List<String> args) {
boolean ignoreUnrelatedOptions = verifyOpArgument(args);
for (int i = 2; i < args.size(); i++) { // parse command line
@@ -983,6 +1012,7 @@ public class NNThroughputBenchmark {
}
}
+ @Override
void generateInputs(int[] ignore) throws IOException {
int nrDatanodes = getNumDatanodes();
int nrBlocks = (int)Math.ceil((double)blocksPerReport * nrDatanodes
@@ -1044,10 +1074,12 @@ public class NNThroughputBenchmark {
/**
* Does not require the argument
*/
+ @Override
String getExecutionArgument(int daemonId) {
return null;
}
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore) throws IOException {
assert daemonId < numThreads : "Wrong daemonId.";
TinyDatanode dn = datanodes[daemonId];
@@ -1060,6 +1092,7 @@ public class NNThroughputBenchmark {
return end-start;
}
+ @Override
void printResults() {
String blockDistribution = "";
String delim = "(";
@@ -1120,10 +1153,12 @@ public class NNThroughputBenchmark {
numPendingBlocks = 0;
}
+ @Override
String getOpName() {
return OP_REPLICATION_NAME;
}
+ @Override
void parseArguments(List<String> args) {
boolean ignoreUnrelatedOptions = verifyOpArgument(args);
for (int i = 2; i < args.size(); i++) { // parse command line
@@ -1147,6 +1182,7 @@ public class NNThroughputBenchmark {
}
}
+ @Override
void generateInputs(int[] ignore) throws IOException {
final FSNamesystem namesystem = nameNode.getNamesystem();
@@ -1193,10 +1229,12 @@ public class NNThroughputBenchmark {
/**
* Does not require the argument
*/
+ @Override
String getExecutionArgument(int daemonId) {
return null;
}
+ @Override
long executeOp(int daemonId, int inputIdx, String ignore) throws IOException {
assert daemonId < numThreads : "Wrong daemonId.";
long start = Time.now();
@@ -1210,6 +1248,7 @@ public class NNThroughputBenchmark {
return end-start;
}
+ @Override
void printResults() {
String blockDistribution = "";
String delim = "(";
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/OfflineEditsViewerHelper.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/OfflineEditsViewerHelper.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/OfflineEditsViewerHelper.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/OfflineEditsViewerHelper.java Mon Jul 16 03:07:49 2012
@@ -205,6 +205,7 @@ public class OfflineEditsViewerHelper {
"JobTracker/foo.com@FOO.COM");
try {
longUgi.doAs(new PrivilegedExceptionAction<Object>() {
+ @Override
public Object run() throws IOException, InterruptedException {
token.renew(config);
token.cancel(config);
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCheckpoint.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCheckpoint.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCheckpoint.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCheckpoint.java Mon Jul 16 03:07:49 2012
@@ -1386,6 +1386,7 @@ public class TestCheckpoint extends Test
final Answer<Object> delegator = new GenericTestUtils.DelegateAnswer(origNN);
NamenodeProtocol spyNN = Mockito.mock(NamenodeProtocol.class, delegator);
DelayAnswer delayer = new DelayAnswer(LOG) {
+ @Override
protected Object passThrough(InvocationOnMock invocation) throws Throwable {
return delegator.answer(invocation);
}
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java Mon Jul 16 03:07:49 2012
@@ -131,6 +131,7 @@ public class TestEditLog extends TestCas
}
// add a bunch of transactions.
+ @Override
public void run() {
PermissionStatus p = namesystem.createFsOwnerPermissions(
new FsPermission((short)0777));
@@ -365,6 +366,7 @@ public class TestEditLog extends TestCas
final String filename) throws Exception
{
exec.submit(new Callable<Void>() {
+ @Override
public Void call() {
log.logSetReplication(filename, (short)1);
return null;
@@ -376,6 +378,7 @@ public class TestEditLog extends TestCas
throws Exception
{
exec.submit(new Callable<Void>() {
+ @Override
public Void call() {
log.logSync();
return null;
@@ -387,6 +390,7 @@ public class TestEditLog extends TestCas
throws Exception
{
exec.submit(new Callable<Void>() {
+ @Override
public Void call() throws Exception {
log.logSyncAll();
return null;
@@ -1138,6 +1142,7 @@ public class TestEditLog extends TestCas
final long endGapTxId = 2*TXNS_PER_ROLL;
File[] files = new File(f1, "current").listFiles(new FilenameFilter() {
+ @Override
public boolean accept(File dir, String name) {
if (name.startsWith(NNStorage.getFinalizedEditsFileName(startGapTxId,
endGapTxId))) {
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogRace.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogRace.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogRace.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLogRace.java Mon Jul 16 03:07:49 2012
@@ -112,6 +112,7 @@ public class TestEditLogRace {
}
// add a bunch of transactions.
+ @Override
public void run() {
thr = Thread.currentThread();
PermissionStatus p = namesystem.createFsOwnerPermissions(
@@ -368,6 +369,7 @@ public class TestEditLogRace {
final CountDownLatch waitToEnterFlush = new CountDownLatch(1);
final Thread doAnEditThread = new Thread() {
+ @Override
public void run() {
try {
LOG.info("Starting mkdirs");
@@ -463,6 +465,7 @@ public class TestEditLogRace {
final CountDownLatch waitToEnterSync = new CountDownLatch(1);
final Thread doAnEditThread = new Thread() {
+ @Override
public void run() {
try {
LOG.info("Starting mkdirs");
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileJournalManager.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileJournalManager.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileJournalManager.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFileJournalManager.java Mon Jul 16 03:07:49 2012
@@ -307,6 +307,7 @@ public class TestFileJournalManager {
final long startGapTxId = 3*TXNS_PER_ROLL + 1;
final long endGapTxId = 4*TXNS_PER_ROLL;
File[] files = new File(f, "current").listFiles(new FilenameFilter() {
+ @Override
public boolean accept(File dir, String name) {
if (name.startsWith(NNStorage.getFinalizedEditsFileName(startGapTxId, endGapTxId))) {
return true;
@@ -338,6 +339,7 @@ public class TestFileJournalManager {
StorageDirectory sd = storage.dirIterator(NameNodeDirType.EDITS).next();
File[] files = new File(f, "current").listFiles(new FilenameFilter() {
+ @Override
public boolean accept(File dir, String name) {
if (name.startsWith("edits_inprogress")) {
return true;
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeRecovery.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeRecovery.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeRecovery.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeRecovery.java Mon Jul 16 03:07:49 2012
@@ -181,6 +181,7 @@ public class TestNameNodeRecovery {
final private long BAD_TXID = 4;
final private long MAX_TXID = 10;
+ @Override
public void addTransactionsToLog(EditLogOutputStream elos,
OpInstanceCache cache) throws IOException {
for (long txid = 1; txid <= MAX_TXID; txid++) {
@@ -199,10 +200,12 @@ public class TestNameNodeRecovery {
}
}
+ @Override
public long getLastValidTxId() {
return BAD_TXID - 1;
}
+ @Override
public Set<Long> getValidTxIds() {
return Sets.newHashSet(1L , 2L, 3L, 5L, 6L, 7L, 8L, 9L, 10L);
}
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSaveNamespace.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSaveNamespace.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSaveNamespace.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSaveNamespace.java Mon Jul 16 03:07:49 2012
@@ -90,6 +90,7 @@ public class TestSaveNamespace {
this.throwRTE = throwRTE;
}
+ @Override
public Void answer(InvocationOnMock invocation) throws Throwable {
Object[] args = invocation.getArguments();
StorageDirectory sd = (StorageDirectory)args[1];
@@ -564,6 +565,7 @@ public class TestSaveNamespace {
delayer.waitForCall();
// then cancel the saveNamespace
Future<Void> cancelFuture = pool.submit(new Callable<Void>() {
+ @Override
public Void call() throws Exception {
canceler.cancel("cancelled");
return null;
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSecurityTokenEditLog.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSecurityTokenEditLog.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSecurityTokenEditLog.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestSecurityTokenEditLog.java Mon Jul 16 03:07:49 2012
@@ -62,6 +62,7 @@ public class TestSecurityTokenEditLog ex
}
// add a bunch of transactions.
+ @Override
public void run() {
FSEditLog editLog = namesystem.getEditLog();
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java Mon Jul 16 03:07:49 2012
@@ -88,6 +88,7 @@ public class TestStartup extends TestCas
}
+ @Override
protected void setUp() throws Exception {
config = new HdfsConfiguration();
hdfsDir = new File(MiniDFSCluster.getBaseDirectory());
@@ -114,6 +115,7 @@ public class TestStartup extends TestCas
/**
* clean up
*/
+ @Override
public void tearDown() throws Exception {
if ( hdfsDir.exists() && !FileUtil.fullyDelete(hdfsDir) ) {
throw new IOException("Could not delete hdfs directory in tearDown '" + hdfsDir + "'");
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java Mon Jul 16 03:07:49 2012
@@ -53,21 +53,26 @@ import org.mortbay.jetty.InclusiveByteRa
*/
class MockFSInputStream extends FSInputStream {
long currentPos = 0;
+ @Override
public int read() throws IOException {
return (int)(currentPos++);
}
+ @Override
public void close() throws IOException {
}
+ @Override
public void seek(long pos) throws IOException {
currentPos = pos;
}
+ @Override
public long getPos() throws IOException {
return currentPos;
}
+ @Override
public boolean seekToNewSource(long targetPos) throws IOException {
return false;
}
@@ -87,6 +92,7 @@ public class TestStreamFile {
StreamFile sfile = new StreamFile() {
private static final long serialVersionUID = -5513776238875189473L;
+ @Override
public ServletContext getServletContext() {
return mockServletContext;
}
@@ -216,6 +222,7 @@ public class TestStreamFile {
private static final long serialVersionUID = 7715590481809562722L;
+ @Override
public ServletContext getServletContext() {
return mockServletContext;
}
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDNFencingWithReplication.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDNFencingWithReplication.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDNFencingWithReplication.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDNFencingWithReplication.java Mon Jul 16 03:07:49 2012
@@ -98,6 +98,7 @@ public class TestDNFencingWithReplicatio
}
}
+ @Override
public String toString() {
return "Toggler for " + path;
}
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDelegationTokensWithHA.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDelegationTokensWithHA.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDelegationTokensWithHA.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestDelegationTokensWithHA.java Mon Jul 16 03:07:49 2012
@@ -159,6 +159,7 @@ public class TestDelegationTokensWithHA
final UserGroupInformation shortUgi = UserGroupInformation
.createRemoteUser("JobTracker");
longUgi.doAs(new PrivilegedExceptionAction<Void>() {
+ @Override
public Void run() throws Exception {
DistributedFileSystem dfs = (DistributedFileSystem)
HATestUtil.configureFailoverFs(cluster, conf);
@@ -168,6 +169,7 @@ public class TestDelegationTokensWithHA
}
});
shortUgi.doAs(new PrivilegedExceptionAction<Void>() {
+ @Override
public Void run() throws Exception {
DistributedFileSystem dfs = (DistributedFileSystem)
HATestUtil.configureFailoverFs(cluster, conf);
@@ -176,6 +178,7 @@ public class TestDelegationTokensWithHA
}
});
longUgi.doAs(new PrivilegedExceptionAction<Void>() {
+ @Override
public Void run() throws Exception {
DistributedFileSystem dfs = (DistributedFileSystem)
HATestUtil.configureFailoverFs(cluster, conf);
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPipelinesFailover.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPipelinesFailover.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPipelinesFailover.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPipelinesFailover.java Mon Jul 16 03:07:49 2012
@@ -88,12 +88,14 @@ public class TestPipelinesFailover {
enum TestScenario {
GRACEFUL_FAILOVER {
+ @Override
void run(MiniDFSCluster cluster) throws IOException {
cluster.transitionToStandby(0);
cluster.transitionToActive(1);
}
},
ORIGINAL_ACTIVE_CRASHED {
+ @Override
void run(MiniDFSCluster cluster) throws IOException {
cluster.restartNameNode(0);
cluster.transitionToActive(1);
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java Mon Jul 16 03:07:49 2012
@@ -96,6 +96,7 @@ public class TestFSMainOperationsWebHdfs
return defaultWorkingDirectory;
}
+ @Override
@Test
public void testMkdirsFailsForSubdirectoryOfExistingFile() throws Exception {
Path testDir = getTestRootPath(fSys, "test/hadoop");
Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tools/TestJMXGet.java
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tools/TestJMXGet.java?rev=1361895&r1=1361894&r2=1361895&view=diff
==============================================================================
--- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tools/TestJMXGet.java (original)
+++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tools/TestJMXGet.java Mon Jul 16 03:07:49 2012
@@ -62,6 +62,7 @@ public class TestJMXGet extends TestCase
}
+ @Override
protected void setUp() throws Exception {
config = new HdfsConfiguration();
}
@@ -69,6 +70,7 @@ public class TestJMXGet extends TestCase
/**
* clean up
*/
+ @Override
public void tearDown() throws Exception {
if(cluster.isClusterUp())
cluster.shutdown();