You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2011/08/15 19:46:26 UTC

Hadoop-Hdfs-trunk - Build # 750 - Still Failing

See https://builds.apache.org/job/Hadoop-Hdfs-trunk/750/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 924 lines...]
A         buildMR-279Branch.sh
AU        hudsonBuildHadoopPatch.sh
AU        hudsonBuildHadoopRelease.sh
AU        processHadoopPatchEmailRemote.sh
AU        hudsonPatchQueueAdmin.sh
AU        processHadoopPatchEmail.sh
A         README.txt
A         test-patch
A         test-patch/test-patch.sh
At revision 1157923
no change for http://svn.apache.org/repos/asf/hadoop/nightly since the previous build
no change for https://svn.apache.org/repos/asf/hadoop/common/trunk/common/src/test/bin since the previous build
No emails were triggered.
[Hadoop-Hdfs-trunk] $ /bin/bash /tmp/hudson5719509509992190546.sh
1024


======================================================================
======================================================================
CLEAN: cleaning workspace
======================================================================
======================================================================


/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/nightly/commitBuild.sh: line 43: /home/hudson/tools/ant/latest/bin/ant: No such file or directory


======================================================================
======================================================================
BUILD: ant clean binary findbugs -Dtest.junit.output.format=xml -Dtest.output=yes -Dcompile.c++=true -Dcompile.native=true -Dfindbugs.home=$FINDBUGS_HOME -Dforrest.home=$FORREST_HOME -Dclover.home=$CLOVER_HOME -Declipse.home=$ECLIPSE_HOME
======================================================================
======================================================================


/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/nightly/commitBuild.sh: line 20: /home/hudson/tools/ant/latest/bin/ant: No such file or directory


======================================================================
======================================================================
STORE: saving artifacts
======================================================================
======================================================================


mv: cannot stat `build/*.tar.gz': No such file or directory
mv: cannot stat `build/*.jar': No such file or directory
mv: cannot stat `build/test/findbugs': No such file or directory
mv: cannot stat `build/docs/api': No such file or directory
Build Failed
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Updating HDFS-73
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 760 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/760/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 8950 lines...]
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.468 sec
Running org.apache.hadoop.hdfs.TestFSInputChecker
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.281 sec
Running org.apache.hadoop.fs.TestGlobPaths
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.635 sec
Running org.apache.hadoop.hdfs.TestModTime
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.986 sec
Running org.apache.hadoop.hdfs.TestBlockMissingException
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.21 sec
Running org.apache.hadoop.hdfs.TestReplication
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.209 sec
Running org.apache.hadoop.fs.TestFcHdfsCreateMkdir
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.514 sec
Running org.apache.hadoop.hdfs.protocol.TestCorruptFileBlocks
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.07 sec
Running org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.559 sec
Running org.apache.hadoop.hdfs.server.datanode.TestRefreshNamenodes
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.019 sec
Running org.apache.hadoop.fs.viewfs.TestViewFileSystemHdfs
Tests run: 37, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.047 sec
Running org.apache.hadoop.cli.TestHDFSCLI
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 71.875 sec

Results :

Failed tests: 

Tests in error: 

Tests run: 833, Failures: 4, Errors: 1, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 1:11:32.340s
[INFO] Finished at: Mon Aug 22 18:53:59 UTC 2011
[INFO] Final Memory: 9M/117M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.6:test (default-test) on project hadoop-hdfs: There are test failures.
[ERROR] 
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Recording test results
Publishing Javadoc
Recording fingerprints
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 759 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/759/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
Started by timer
Building remotely on hadoop6
Location 'http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs' does not exist
Cleaning workspace /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk
Checking out http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs
ERROR: Failed to check out http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs
org.tmatesoft.svn.core.SVNException: svn: URL 'http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs' doesn't exist
	at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:64)
	at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:51)
	at org.tmatesoft.svn.core.wc.SVNUpdateClient.doCheckout(SVNUpdateClient.java:910)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:83)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:135)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:726)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:707)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:691)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1994)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:287)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
	at java.util.concurrent.FutureTask.run(FutureTask.java:138)
	at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
	at java.lang.Thread.run(Thread.java:662)
Caused by: org.tmatesoft.svn.core.SVNErrorMessage: svn: URL 'http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs' doesn't exist
	at org.tmatesoft.svn.core.SVNErrorMessage.create(SVNErrorMessage.java:163)
	at org.tmatesoft.svn.core.SVNErrorMessage.create(SVNErrorMessage.java:118)
	at org.tmatesoft.svn.core.wc.SVNUpdateClient.doCheckout(SVNUpdateClient.java:909)
	... 15 more
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 758 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/758/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
Started by timer
Building remotely on hadoop2
Location 'http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs' does not exist
Cleaning workspace /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk
Checking out http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs
ERROR: Failed to check out http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs
org.tmatesoft.svn.core.SVNException: svn: URL 'http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs' doesn't exist
	at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:64)
	at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:51)
	at org.tmatesoft.svn.core.wc.SVNUpdateClient.doCheckout(SVNUpdateClient.java:910)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:83)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:135)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:726)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:707)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:691)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1994)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:270)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
	at java.util.concurrent.FutureTask.run(FutureTask.java:138)
	at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
	at java.lang.Thread.run(Thread.java:662)
Caused by: org.tmatesoft.svn.core.SVNErrorMessage: svn: URL 'http://svn.apache.org/repos/asf/hadoop/common/trunk/hdfs' doesn't exist
	at org.tmatesoft.svn.core.SVNErrorMessage.create(SVNErrorMessage.java:163)
	at org.tmatesoft.svn.core.SVNErrorMessage.create(SVNErrorMessage.java:118)
	at org.tmatesoft.svn.core.wc.SVNUpdateClient.doCheckout(SVNUpdateClient.java:909)
	... 15 more
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 757 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/757/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1250 lines...]

clean-fi:

clean-sign:

clean:
   [delete] Deleting directory /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build

BUILD SUCCESSFUL
Total time: 0 seconds


======================================================================
======================================================================
ANALYSIS: ant -Drun.clover=true clover checkstyle test generate-clover-reports -Dtest.junit.output.format=xml -Dtest.output=yes -Dcompile.c++=true -Dcompile.native=true -Dfindbugs.home=$FINDBUGS_HOME -Dforrest.home=$FORREST_HOME -Dclover.home=$CLOVER_HOME -Declipse.home=$ECLIPSE_HOME
======================================================================
======================================================================


Buildfile: build.xml

clover.setup:
    [mkdir] Created dir: /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/clover/db
[clover-setup] Clover Version 3.1.0, built on May 31 2011 (build-821)
[clover-setup] Loaded from: /home/jenkins/tools/clover/latest/lib/clover.jar

BUILD FAILED
java.lang.RuntimeException: Clover upgrades for your license ended December 14 2010, and this version of Clover was built May 31 2011. Please visit http://www.atlassian.com/clover/renew for information on upgrading your license.
	at com.cenqua.clover.CloverStartup.loadLicense(CloverStartup.java:103)
	at com.cenqua.clover.CloverStartup.loadLicense(CloverStartup.java:25)
	at com.cenqua.clover.tasks.AbstractCloverTask.execute(AbstractCloverTask.java:52)
	at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:288)
	at sun.reflect.GeneratedMethodAccessor1.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
	at org.apache.tools.ant.Task.perform(Task.java:348)
	at org.apache.tools.ant.Target.execute(Target.java:357)
	at org.apache.tools.ant.Target.performTasks(Target.java:385)
	at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1337)
	at org.apache.tools.ant.Project.executeTarget(Project.java:1306)
	at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
	at org.apache.tools.ant.Project.executeTargets(Project.java:1189)
	at org.apache.tools.ant.Main.runBuild(Main.java:758)
	at org.apache.tools.ant.Main.startAnt(Main.java:217)
	at org.apache.tools.ant.launch.Launcher.run(Launcher.java:257)
	at org.apache.tools.ant.launch.Launcher.main(Launcher.java:104)

Total time: 0 seconds
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 756 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/756/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1471 lines...]
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:331: cannot find symbol
    [javac] symbol  : class UnresolvedLinkException
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]       throws IOException, UnresolvedLinkException {
    [javac]                           ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:337: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst) 
    [javac]                              ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:337: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst) 
    [javac]                                        ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:338: cannot find symbol
    [javac] symbol  : class UnresolvedLinkException
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]     throws IOException, UnresolvedLinkException {
    [javac]                         ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:343: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst, boolean overwrite)
    [javac]                              ^
    [javac] Note: Some input files use or override a deprecated API.
    [javac] Note: Recompile with -Xlint:deprecation for details.
    [javac] 100 errors

BUILD FAILED
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build.xml:370: Compile failed; see the compiler error output for details.

Total time: 19 seconds


======================================================================
======================================================================
STORE: saving artifacts
======================================================================
======================================================================


mv: cannot stat `build/*.tar.gz': No such file or directory
mv: cannot stat `build/*.jar': No such file or directory
mv: cannot stat `build/test/findbugs': No such file or directory
mv: cannot stat `build/docs/api': No such file or directory
Build Failed
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Updating HDFS-2265
Updating HDFS-2260
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 755 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/755/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1467 lines...]
    [javac]   public FSDataInputStream open(Path f, int bufferSize) 
    [javac]          ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:331: cannot find symbol
    [javac] symbol  : class UnresolvedLinkException
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]       throws IOException, UnresolvedLinkException {
    [javac]                           ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:337: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst) 
    [javac]                              ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:337: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst) 
    [javac]                                        ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:338: cannot find symbol
    [javac] symbol  : class UnresolvedLinkException
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]     throws IOException, UnresolvedLinkException {
    [javac]                         ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:343: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst, boolean overwrite)
    [javac]                              ^
    [javac] Note: Some input files use or override a deprecated API.
    [javac] Note: Recompile with -Xlint:deprecation for details.
    [javac] 100 errors

BUILD FAILED
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build.xml:370: Compile failed; see the compiler error output for details.

Total time: 17 seconds


======================================================================
======================================================================
STORE: saving artifacts
======================================================================
======================================================================


mv: cannot stat `build/*.tar.gz': No such file or directory
mv: cannot stat `build/*.jar': No such file or directory
mv: cannot stat `build/test/findbugs': No such file or directory
mv: cannot stat `build/docs/api': No such file or directory
Build Failed
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 754 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/754/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1243 lines...]
     [echo] contrib: fuse-dfs

clean-fi:

clean-sign:

clean:
   [delete] Deleting directory /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build

BUILD SUCCESSFUL
Total time: 0 seconds


======================================================================
======================================================================
ANALYSIS: ant -Drun.clover=true clover checkstyle test generate-clover-reports -Dtest.junit.output.format=xml -Dtest.output=yes -Dcompile.c++=true -Dcompile.native=true -Dfindbugs.home=$FINDBUGS_HOME -Dforrest.home=$FORREST_HOME -Dclover.home=$CLOVER_HOME -Declipse.home=$ECLIPSE_HOME
======================================================================
======================================================================


Buildfile: build.xml

clover.setup:
    [mkdir] Created dir: /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/clover/db
[clover-setup] Clover Version 3.1.0, built on May 31 2011 (build-821)
[clover-setup] Loaded from: /home/jenkins/tools/clover/latest/lib/clover.jar

BUILD FAILED
java.lang.RuntimeException: Clover upgrades for your license ended December 14 2010, and this version of Clover was built May 31 2011. Please visit http://www.atlassian.com/clover/renew for information on upgrading your license.
	at com.cenqua.clover.CloverStartup.loadLicense(CloverStartup.java:103)
	at com.cenqua.clover.CloverStartup.loadLicense(CloverStartup.java:25)
	at com.cenqua.clover.tasks.AbstractCloverTask.execute(AbstractCloverTask.java:52)
	at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:288)
	at sun.reflect.GeneratedMethodAccessor1.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.apache.tools.ant.dispatch.DispatchUtils.execute(DispatchUtils.java:106)
	at org.apache.tools.ant.Task.perform(Task.java:348)
	at org.apache.tools.ant.Target.execute(Target.java:357)
	at org.apache.tools.ant.Target.performTasks(Target.java:385)
	at org.apache.tools.ant.Project.executeSortedTargets(Project.java:1337)
	at org.apache.tools.ant.Project.executeTarget(Project.java:1306)
	at org.apache.tools.ant.helper.DefaultExecutor.executeTargets(DefaultExecutor.java:41)
	at org.apache.tools.ant.Project.executeTargets(Project.java:1189)
	at org.apache.tools.ant.Main.runBuild(Main.java:758)
	at org.apache.tools.ant.Main.startAnt(Main.java:217)
	at org.apache.tools.ant.launch.Launcher.run(Launcher.java:257)
	at org.apache.tools.ant.launch.Launcher.main(Launcher.java:104)

Total time: 0 seconds
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 753 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/753/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1402551 lines...]
    [junit] 2011-08-16 04:27:18,730 INFO  datanode.FSDatasetAsyncDiskService (FSDatasetAsyncDiskService.java:shutdown(142)) - All async disk service threads have been shut down.
    [junit] 2011-08-16 04:27:18,731 INFO  mortbay.log (Slf4jLog.java:info(67)) - Stopped SelectChannelConnector@localhost:0
    [junit] 2011-08-16 04:27:18,832 WARN  blockmanagement.BlockManager (BlockManager.java:run(2614)) - ReplicationMonitor thread received InterruptedException.
    [junit] java.lang.InterruptedException: sleep interrupted
    [junit] 	at java.lang.Thread.sleep(Native Method)
    [junit] 	at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor.run(BlockManager.java:2612)
    [junit] 	at java.lang.Thread.run(Thread.java:662)
    [junit] 2011-08-16 04:27:18,832 WARN  blockmanagement.DecommissionManager (DecommissionManager.java:run(75)) - Monitor interrupted: java.lang.InterruptedException: sleep interrupted
    [junit] 2011-08-16 04:27:18,832 INFO  namenode.FSEditLog (FSEditLog.java:endCurrentLogSegment(859)) - Ending log segment 1
    [junit] 2011-08-16 04:27:18,843 INFO  namenode.FSEditLog (FSEditLog.java:printStatistics(492)) - Number of transactions: 8 Total time for transactions(ms): 0Number of transactions batched in Syncs: 0 Number of syncs: 7 SyncTimes(ms): 56 48 
    [junit] 2011-08-16 04:27:18,844 INFO  ipc.Server (Server.java:stop(1715)) - Stopping server on 38584
    [junit] 2011-08-16 04:27:18,844 INFO  ipc.Server (Server.java:run(1539)) - IPC Server handler 0 on 38584: exiting
    [junit] 2011-08-16 04:27:18,844 INFO  ipc.Server (Server.java:run(505)) - Stopping IPC Server listener on 38584
    [junit] 2011-08-16 04:27:18,845 INFO  ipc.Server (Server.java:run(647)) - Stopping IPC Server Responder
    [junit] 2011-08-16 04:27:18,845 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(199)) - Stopping DataNode metrics system...
    [junit] 2011-08-16 04:27:18,845 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics
    [junit] 2011-08-16 04:27:18,845 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source NameNodeActivity
    [junit] 2011-08-16 04:27:18,845 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort38584
    [junit] 2011-08-16 04:27:18,846 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort38584
    [junit] 2011-08-16 04:27:18,846 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source FSNamesystem
    [junit] 2011-08-16 04:27:18,846 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort41691
    [junit] 2011-08-16 04:27:18,846 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort41691
    [junit] 2011-08-16 04:27:18,846 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-1
    [junit] 2011-08-16 04:27:18,847 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-43076
    [junit] 2011-08-16 04:27:18,847 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort59622
    [junit] 2011-08-16 04:27:18,847 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort59622
    [junit] 2011-08-16 04:27:18,847 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-2
    [junit] 2011-08-16 04:27:18,848 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-35471
    [junit] 2011-08-16 04:27:18,848 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort34995
    [junit] 2011-08-16 04:27:18,848 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort34995
    [junit] 2011-08-16 04:27:18,848 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-3
    [junit] 2011-08-16 04:27:18,848 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-33376
    [junit] 2011-08-16 04:27:18,849 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort49871
    [junit] 2011-08-16 04:27:18,849 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort49871
    [junit] 2011-08-16 04:27:18,849 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-4
    [junit] 2011-08-16 04:27:18,849 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-47339
    [junit] 2011-08-16 04:27:18,849 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(205)) - DataNode metrics system stopped.
    [junit] 2011-08-16 04:27:18,850 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:shutdown(553)) - DataNode metrics system shutdown complete.
    [junit] Tests run: 16, Failures: 0, Errors: 0, Time elapsed: 102.733 sec

checkfailure:

-run-test-hdfs-fault-inject-withtestcaseonly:

run-test-hdfs-fault-inject:

BUILD FAILED
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build.xml:777: Tests failed!

Total time: 78 minutes 11 seconds
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
7 tests failed.
REGRESSION:  org.apache.hadoop.hdfs.TestDFSRollback.testRollback

Error Message:
File contents differed:   /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/data2/current/VERSION=d0342cd292d1c7d919b5f4b000f940de   /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/data1/current/VERSION=5baa37e94882a5fbc28d9d74bfc04db0

Stack Trace:
junit.framework.AssertionFailedError: File contents differed:
  /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/data2/current/VERSION=d0342cd292d1c7d919b5f4b000f940de
  /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/data1/current/VERSION=5baa37e94882a5fbc28d9d74bfc04db0
  /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/data2/current/VERSION=d0342cd292d1c7d919b5f4b000f940de
  /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build/test/data/dfs/data1/current/VERSION=5baa37e94882a5fbc28d9d74bfc04db0
	at org.apache.hadoop.hdfs.server.namenode.FSImageTestUtil.assertFileContentsSame(FSImageTestUtil.java:251)
	at org.apache.hadoop.hdfs.server.namenode.FSImageTestUtil.assertParallelFilesAreIdentical(FSImageTestUtil.java:237)
	at org.apache.hadoop.hdfs.TestDFSRollback.checkResult(TestDFSRollback.java:86)
	at org.apache.hadoop.hdfs.TestDFSRollback.__CLR2_4_37oj5yb1gcn(TestDFSRollback.java:171)
	at org.apache.hadoop.hdfs.TestDFSRollback.testRollback(TestDFSRollback.java:134)


REGRESSION:  org.apache.hadoop.hdfs.TestParallelRead.testParallelRead

Error Message:
Timeout occurred. Please note the time in the report does not reflect the time until the timeout.

Stack Trace:
junit.framework.AssertionFailedError: Timeout occurred. Please note the time in the report does not reflect the time until the timeout.


FAILED:  org.apache.hadoop.hdfs.TestHDFSServerPorts.testSecondaryNodePorts

Error Message:
Directory /test/dfs/namesecondary is in an inconsistent state: checkpoint directory does not exist or is not accessible.

Stack Trace:
org.apache.hadoop.hdfs.server.common.InconsistentFSStateException: Directory /test/dfs/namesecondary is in an inconsistent state: checkpoint directory does not exist or is not accessible.
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.recoverCreate(SecondaryNameNode.java:801)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:222)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:175)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:168)
	at org.apache.hadoop.hdfs.TestHDFSServerPorts.canStartSecondaryNode(TestHDFSServerPorts.java:224)
	at org.apache.hadoop.hdfs.TestHDFSServerPorts.__CLR2_4_3vpy47p1541(TestHDFSServerPorts.java:350)
	at org.apache.hadoop.hdfs.TestHDFSServerPorts.testSecondaryNodePorts(TestHDFSServerPorts.java:339)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testSeparateEditsDirLocking

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:626)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:541)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:257)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:85)
	at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:243)
	at org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.__CLR2_4_3harbaz1hcb(TestCheckpoint.java:560)
	at org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testSeparateEditsDirLocking(TestCheckpoint.java:553)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark.testNNThroughput

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark.__CLR2_4_3b2i9ur1f75(TestNNThroughputBenchmark.java:39)
	at org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark.testNNThroughput(TestNNThroughputBenchmark.java:35)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatMatchingRPCandHttpPortsThrowException

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.__CLR2_4_3b49o261bi8(TestValidateConfigurationSettings.java:49)
	at org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatMatchingRPCandHttpPortsThrowException(TestValidateConfigurationSettings.java:43)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatDifferentRPCandHttpPortsAreOK

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.__CLR2_4_3ihms9r1bii(TestValidateConfigurationSettings.java:71)
	at org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatDifferentRPCandHttpPortsAreOK(TestValidateConfigurationSettings.java:66)




Hadoop-Hdfs-trunk - Build # 752 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/752/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1466 lines...]
    [javac]   public FSDataInputStream open(Path f, int bufferSize) 
    [javac]          ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:331: cannot find symbol
    [javac] symbol  : class UnresolvedLinkException
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]       throws IOException, UnresolvedLinkException {
    [javac]                           ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:337: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst) 
    [javac]                              ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:337: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst) 
    [javac]                                        ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:338: cannot find symbol
    [javac] symbol  : class UnresolvedLinkException
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]     throws IOException, UnresolvedLinkException {
    [javac]                         ^
    [javac] /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/src/java/org/apache/hadoop/fs/Hdfs.java:343: cannot find symbol
    [javac] symbol  : class Path
    [javac] location: class org.apache.hadoop.fs.Hdfs
    [javac]   public void renameInternal(Path src, Path dst, boolean overwrite)
    [javac]                              ^
    [javac] Note: Some input files use or override a deprecated API.
    [javac] Note: Recompile with -Xlint:deprecation for details.
    [javac] 100 errors

BUILD FAILED
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build.xml:370: Compile failed; see the compiler error output for details.

Total time: 16 seconds


======================================================================
======================================================================
STORE: saving artifacts
======================================================================
======================================================================


mv: cannot stat `build/*.tar.gz': No such file or directory
mv: cannot stat `build/*.jar': No such file or directory
mv: cannot stat `build/test/findbugs': No such file or directory
mv: cannot stat `build/docs/api': No such file or directory
Build Failed
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Updating HDFS-2233
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
No tests ran.

Hadoop-Hdfs-trunk - Build # 751 - Still Failing

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/751/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 1499156 lines...]
    [junit] 2011-08-15 22:17:08,057 INFO  datanode.FSDatasetAsyncDiskService (FSDatasetAsyncDiskService.java:shutdown(142)) - All async disk service threads have been shut down.
    [junit] 2011-08-15 22:17:08,057 INFO  mortbay.log (Slf4jLog.java:info(67)) - Stopped SelectChannelConnector@localhost:0
    [junit] 2011-08-15 22:17:08,158 WARN  blockmanagement.DecommissionManager (DecommissionManager.java:run(75)) - Monitor interrupted: java.lang.InterruptedException: sleep interrupted
    [junit] 2011-08-15 22:17:08,158 WARN  blockmanagement.BlockManager (BlockManager.java:run(2614)) - ReplicationMonitor thread received InterruptedException.
    [junit] java.lang.InterruptedException: sleep interrupted
    [junit] 	at java.lang.Thread.sleep(Native Method)
    [junit] 	at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor.run(BlockManager.java:2612)
    [junit] 	at java.lang.Thread.run(Thread.java:662)
    [junit] 2011-08-15 22:17:08,159 INFO  namenode.FSEditLog (FSEditLog.java:endCurrentLogSegment(859)) - Ending log segment 1
    [junit] 2011-08-15 22:17:08,170 INFO  namenode.FSEditLog (FSEditLog.java:printStatistics(492)) - Number of transactions: 8 Total time for transactions(ms): 0Number of transactions batched in Syncs: 0 Number of syncs: 7 SyncTimes(ms): 94 83 
    [junit] 2011-08-15 22:17:08,171 INFO  ipc.Server (Server.java:stop(1715)) - Stopping server on 42042
    [junit] 2011-08-15 22:17:08,172 INFO  ipc.Server (Server.java:run(1539)) - IPC Server handler 0 on 42042: exiting
    [junit] 2011-08-15 22:17:08,172 INFO  ipc.Server (Server.java:run(505)) - Stopping IPC Server listener on 42042
    [junit] 2011-08-15 22:17:08,172 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(199)) - Stopping DataNode metrics system...
    [junit] 2011-08-15 22:17:08,172 INFO  ipc.Server (Server.java:run(647)) - Stopping IPC Server Responder
    [junit] 2011-08-15 22:17:08,172 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics
    [junit] 2011-08-15 22:17:08,173 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source NameNodeActivity
    [junit] 2011-08-15 22:17:08,173 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort42042
    [junit] 2011-08-15 22:17:08,173 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort42042
    [junit] 2011-08-15 22:17:08,173 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source FSNamesystem
    [junit] 2011-08-15 22:17:08,173 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort47877
    [junit] 2011-08-15 22:17:08,174 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort47877
    [junit] 2011-08-15 22:17:08,174 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-1
    [junit] 2011-08-15 22:17:08,174 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-34652
    [junit] 2011-08-15 22:17:08,174 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort55190
    [junit] 2011-08-15 22:17:08,174 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort55190
    [junit] 2011-08-15 22:17:08,175 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-2
    [junit] 2011-08-15 22:17:08,175 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-57909
    [junit] 2011-08-15 22:17:08,175 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort51148
    [junit] 2011-08-15 22:17:08,175 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort51148
    [junit] 2011-08-15 22:17:08,176 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-3
    [junit] 2011-08-15 22:17:08,176 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-46009
    [junit] 2011-08-15 22:17:08,176 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcActivityForPort60950
    [junit] 2011-08-15 22:17:08,176 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source RpcDetailedActivityForPort60950
    [junit] 2011-08-15 22:17:08,176 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source JvmMetrics-4
    [junit] 2011-08-15 22:17:08,177 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stopSources(408)) - Stopping metrics source DataNodeActivity-asf011.sp2.ygridcore.net-35625
    [junit] 2011-08-15 22:17:08,177 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(205)) - DataNode metrics system stopped.
    [junit] 2011-08-15 22:17:08,177 INFO  impl.MetricsSystemImpl (MetricsSystemImpl.java:shutdown(553)) - DataNode metrics system shutdown complete.
    [junit] Tests run: 16, Failures: 0, Errors: 0, Time elapsed: 102.287 sec

checkfailure:

-run-test-hdfs-fault-inject-withtestcaseonly:

run-test-hdfs-fault-inject:

BUILD FAILED
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/build.xml:777: Tests failed!

Total time: 61 minutes 22 seconds
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording fingerprints
Recording test results
Publishing Javadoc
Publishing Clover coverage report...
No Clover report will be published due to a Build Failure
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) ##############################
15 tests failed.
REGRESSION:  org.apache.hadoop.hdfs.server.datanode.TestDataDirs.testGetDataDirsFromURIs

Error Message:
org/apache/hadoop/fs/permission/FsPermission

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/fs/permission/FsPermission
	at org.apache.hadoop.hdfs.server.datanode.TestDataDirs.testGetDataDirsFromURIs(TestDataDirs.java:42)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.fs.permission.FsPermission
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestGetImageServlet.initializationError

Error Message:
org/apache/hadoop/conf/Configuration

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/conf/Configuration
	at java.lang.Class.getDeclaredMethods0(Native Method)
	at java.lang.Class.privateGetDeclaredMethods(Class.java:2427)
	at java.lang.Class.getDeclaredMethods(Class.java:1791)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.conf.Configuration
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testReplication

Error Message:
org/apache/hadoop/io/Writable

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/io/Writable
	at java.lang.ClassLoader.defineClass1(Native Method)
	at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
	at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
	at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
	at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
	at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testReplication(TestINodeFile.java:47)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.io.Writable
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testReplicationBelowLowerBound

Error Message:
Unexpected exception, expected<java.lang.IllegalArgumentException> but was<java.lang.NoClassDefFoundError>

Stack Trace:
java.lang.Exception: Unexpected exception, expected<java.lang.IllegalArgumentException> but was<java.lang.NoClassDefFoundError>
Caused by: java.lang.NoClassDefFoundError: org/apache/hadoop/io/Writable
	at java.lang.ClassLoader.defineClass1(Native Method)
	at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
	at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
	at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
	at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
	at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testReplicationBelowLowerBound(TestINodeFile.java:64)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.io.Writable
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSize

Error Message:
org/apache/hadoop/io/Writable

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/io/Writable
	at java.lang.ClassLoader.defineClass1(Native Method)
	at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
	at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
	at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
	at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
	at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSize(TestINodeFile.java:77)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.io.Writable
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSizeUpperBound

Error Message:
org/apache/hadoop/io/Writable

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/io/Writable
	at java.lang.ClassLoader.defineClass1(Native Method)
	at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
	at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
	at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
	at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
	at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSizeUpperBound(TestINodeFile.java:88)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.io.Writable
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSizeBelowLowerBound

Error Message:
Unexpected exception, expected<java.lang.IllegalArgumentException> but was<java.lang.NoClassDefFoundError>

Stack Trace:
java.lang.Exception: Unexpected exception, expected<java.lang.IllegalArgumentException> but was<java.lang.NoClassDefFoundError>
Caused by: java.lang.NoClassDefFoundError: org/apache/hadoop/io/Writable
	at java.lang.ClassLoader.defineClass1(Native Method)
	at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
	at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
	at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
	at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
	at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSizeBelowLowerBound(TestINodeFile.java:105)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.io.Writable
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSizeAboveUpperBound

Error Message:
Unexpected exception, expected<java.lang.IllegalArgumentException> but was<java.lang.NoClassDefFoundError>

Stack Trace:
java.lang.Exception: Unexpected exception, expected<java.lang.IllegalArgumentException> but was<java.lang.NoClassDefFoundError>
Caused by: java.lang.NoClassDefFoundError: org/apache/hadoop/io/Writable
	at java.lang.ClassLoader.defineClass1(Native Method)
	at java.lang.ClassLoader.defineClassCond(ClassLoader.java:631)
	at java.lang.ClassLoader.defineClass(ClassLoader.java:615)
	at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
	at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
	at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
	at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testPreferredBlockSizeAboveUpperBound(TestINodeFile.java:120)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.io.Writable
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testGetFullPathName

Error Message:
org/apache/hadoop/fs/permission/PermissionStatus

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/fs/permission/PermissionStatus
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testGetFullPathName(TestINodeFile.java:127)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.fs.permission.PermissionStatus
	at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
	at java.security.AccessController.doPrivileged(Native Method)
	at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
	at java.lang.ClassLoader.loadClass(ClassLoader.java:247)


REGRESSION:  org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testAppendBlocks

Error Message:
org/apache/hadoop/fs/permission/PermissionStatus

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/fs/permission/PermissionStatus
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.createINodeFiles(TestINodeFile.java:195)
	at org.apache.hadoop.hdfs.server.namenode.TestINodeFile.testAppendBlocks(TestINodeFile.java:157)


FAILED:  org.apache.hadoop.hdfs.TestHDFSServerPorts.testSecondaryNodePorts

Error Message:
Directory /test/dfs/namesecondary is in an inconsistent state: checkpoint directory does not exist or is not accessible.

Stack Trace:
org.apache.hadoop.hdfs.server.common.InconsistentFSStateException: Directory /test/dfs/namesecondary is in an inconsistent state: checkpoint directory does not exist or is not accessible.
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.recoverCreate(SecondaryNameNode.java:801)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:222)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:175)
	at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:168)
	at org.apache.hadoop.hdfs.TestHDFSServerPorts.canStartSecondaryNode(TestHDFSServerPorts.java:224)
	at org.apache.hadoop.hdfs.TestHDFSServerPorts.testSecondaryNodePorts(TestHDFSServerPorts.java:350)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testSeparateEditsDirLocking

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:626)
	at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:541)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:257)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:85)
	at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:243)
	at org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testSeparateEditsDirLocking(TestCheckpoint.java:560)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark.testNNThroughput

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark.testNNThroughput(TestNNThroughputBenchmark.java:39)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatMatchingRPCandHttpPortsThrowException

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatMatchingRPCandHttpPortsThrowException(TestValidateConfigurationSettings.java:49)


FAILED:  org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatDifferentRPCandHttpPortsAreOK

Error Message:
Cannot create directory /test/dfs/name/current

Stack Trace:
java.io.IOException: Cannot create directory /test/dfs/name/current
	at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:276)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:492)
	at org.apache.hadoop.hdfs.server.namenode.NNStorage.format(NNStorage.java:512)
	at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:169)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1367)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:242)
	at org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:113)
	at org.apache.hadoop.hdfs.server.namenode.TestValidateConfigurationSettings.testThatDifferentRPCandHttpPortsAreOK(TestValidateConfigurationSettings.java:71)