You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2011/12/11 15:54:47 UTC
Jenkins build is unstable: Hadoop-Hdfs-trunk #891
See <https://builds.apache.org/job/Hadoop-Hdfs-trunk/891/changes>
Hadoop-Hdfs-trunk - Build # 892 - Failure
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/892/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 11873 lines...]
[java] org.apache.hadoop.classification.InterfaceAudience
[java] org.apache.hadoop.classification.InterfaceStability$Evolving
[java] org.apache.hadoop.classification.InterfaceStability
[java] org.apache.hadoop.classification.InterfaceAudience$Public
[java] org.apache.hadoop.classification.InterfaceAudience$LimitedPrivate
[java] Warnings generated: 4
[java] Missing classes: 7
[INFO] xmlOutput is false
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Building Apache Hadoop HDFS Project 0.24.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-hdfs-project ---
[INFO] Deleting /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/target
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (create-testdirs) @ hadoop-hdfs-project ---
[INFO] Executing tasks
main:
[mkdir] Created dir: /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/target/test-dir
[INFO] Executed tasks
[INFO]
[INFO] --- maven-javadoc-plugin:2.7:jar (module-javadocs) @ hadoop-hdfs-project ---
[INFO] Not executing Javadoc as the project is not a Java classpath-capable package
[INFO]
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ hadoop-hdfs-project ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is false
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop HDFS ................................ SUCCESS [2:03:42.798s]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [31.199s]
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [0.032s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:04:14.439s
[INFO] Finished at: Mon Dec 12 13:38:09 UTC 2011
[INFO] Final Memory: 48M/808M
[INFO] ------------------------------------------------------------------------
+ /home/jenkins/tools/maven/latest/bin/mvn test -Dmaven.test.failure.ignore=true -Pclover -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results
Publishing Javadoc
Recording fingerprints
Updating HDFS-2647
Updating HDFS-2651
Sending e-mails to: hdfs-dev@hadoop.apache.org
Email was triggered for: Failure
Sending email for trigger: Failure
###################################################################################
############################## FAILED TESTS (if any) ##############################
9 tests failed.
REGRESSION: org.apache.hadoop.hdfs.TestDfsOverAvroRpc.testWorkingDirectory
Error Message:
java.lang.NoSuchFieldException: SCHEMA$
Stack Trace:
org.apache.avro.AvroRuntimeException: java.lang.NoSuchFieldException: SCHEMA$
at org.apache.avro.specific.SpecificData.createSchema(SpecificData.java:170)
at org.apache.avro.reflect.ReflectData.createSchema(ReflectData.java:265)
at org.apache.avro.reflect.ReflectData.createFieldSchema(ReflectData.java:372)
at org.apache.avro.reflect.ReflectData.createSchema(ReflectData.java:308)
at org.apache.avro.reflect.ReflectData.createFieldSchema(ReflectData.java:372)
at org.apache.avro.reflect.ReflectData.createSchema(ReflectData.java:308)
at org.apache.avro.reflect.ReflectData.createFieldSchema(ReflectData.java:372)
at org.apache.avro.reflect.ReflectData.createSchema(ReflectData.java:308)
at org.apache.avro.reflect.ReflectData.getSchema(ReflectData.java:449)
at org.apache.avro.reflect.ReflectData.getMessage(ReflectData.java:416)
at org.apache.avro.reflect.ReflectData.getProtocol(ReflectData.java:395)
at org.apache.avro.ipc.reflect.ReflectResponder.<init>(ReflectResponder.java:36)
at org.apache.hadoop.ipc.AvroRpcEngine.createResponder(AvroRpcEngine.java:187)
at org.apache.hadoop.ipc.AvroRpcEngine$TunnelResponder.addProtocol(AvroRpcEngine.java:196)
at org.apache.hadoop.ipc.AvroRpcEngine$Server.addProtocol(AvroRpcEngine.java:251)
at org.apache.hadoop.ipc.AvroRpcEngine$Server.addProtocol(AvroRpcEngine.java:229)
at org.apache.hadoop.hdfs.DFSUtil.addPBProtocol(DFSUtil.java:755)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.<init>(NameNodeRpcServer.java:194)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcServer(NameNode.java:357)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:335)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.TestLocalDFS.__CLR3_0_2hl5jzp163q(TestLocalDFS.java:66)
at org.apache.hadoop.hdfs.TestLocalDFS.testWorkingDirectory(TestLocalDFS.java:64)
at org.apache.hadoop.hdfs.TestDfsOverAvroRpc.__CLR3_0_2hl5jzp139t(TestDfsOverAvroRpc.java:33)
at org.apache.hadoop.hdfs.TestDfsOverAvroRpc.testWorkingDirectory(TestDfsOverAvroRpc.java:30)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.internal.runners.statements.FailOnTimeout$1.run(FailOnTimeout.java:28)
Caused by: java.lang.NoSuchFieldException: SCHEMA$
at java.lang.Class.getDeclaredField(Class.java:1882)
at org.apache.avro.specific.SpecificData.createSchema(SpecificData.java:168)
... 40 more
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testSingleFailedEditsDirOnFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testAllEditsDirsFailOnFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testAllEditsDirFailOnWrite
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testSingleFailedEditsDirOnSetReadyToFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testSingleRequiredFailedEditsDirOnSetReadyToFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testMultipleRedundantFailedEditsDirOnSetReadyToFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestStreamFile.testDoGetShouldWriteTheFileContentIntoServletOutputStream
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestStreamFile.__CLR3_0_2gdh3i1idv(TestStreamFile.java:206)
at org.apache.hadoop.hdfs.server.namenode.TestStreamFile.testDoGetShouldWriteTheFileContentIntoServletOutputStream(TestStreamFile.java:204)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestStreamFile.testDoGetShouldCloseTheDFSInputStreamIfResponseGetOutPutStreamThrowsAnyException
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:645)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:549)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestStreamFile.__CLR3_0_2pyj1zi1ie8(TestStreamFile.java:235)
at org.apache.hadoop.hdfs.server.namenode.TestStreamFile.testDoGetShouldCloseTheDFSInputStreamIfResponseGetOutPutStreamThrowsAnyException(TestStreamFile.java:233)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
Hadoop-Hdfs-trunk - Build # 893 - Still Failing
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/893/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 12165 lines...]
[INFO]
[INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-hdfs-project ---
[INFO] Deleting /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/target
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (create-testdirs) @ hadoop-hdfs-project ---
[INFO] Executing tasks
main:
[mkdir] Created dir: /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/target/test-dir
[INFO] Executed tasks
[INFO]
[INFO] --- maven-javadoc-plugin:2.7:jar (module-javadocs) @ hadoop-hdfs-project ---
[INFO] Not executing Javadoc as the project is not a Java classpath-capable package
[INFO]
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ hadoop-hdfs-project ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is false
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop HDFS ................................ SUCCESS [2:07:07.684s]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [33.699s]
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [0.036s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:07:41.883s
[INFO] Finished at: Tue Dec 13 13:41:48 UTC 2011
[INFO] Final Memory: 49M/749M
[INFO] ------------------------------------------------------------------------
+ /home/jenkins/tools/maven/latest/bin/mvn test -Dmaven.test.failure.ignore=true -Pclover -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results
Publishing Javadoc
Recording fingerprints
Updating MAPREDUCE-3518
Updating MAPREDUCE-3527
Updating MAPREDUCE-3537
Updating HADOOP-7878
Updating HDFS-1765
Updating HADOOP-7914
Updating HADOOP-7913
Updating HDFS-2663
Updating HADOOP-7912
Updating HADOOP-7777
Updating MAPREDUCE-3510
Updating HDFS-2653
Updating HDFS-2654
Updating MAPREDUCE-3328
Sending e-mails to: hdfs-dev@hadoop.apache.org
Email was triggered for: Failure
Sending email for trigger: Failure
###################################################################################
############################## FAILED TESTS (if any) ##############################
8 tests failed.
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestFileLimit.testFileLimit
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestFileLimit.__CLR3_0_2os0bak1g3p(TestFileLimit.java:89)
at org.apache.hadoop.hdfs.server.namenode.TestFileLimit.testFileLimit(TestFileLimit.java:78)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:168)
at junit.framework.TestCase.runBare(TestCase.java:134)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:232)
at junit.framework.TestSuite.run(TestSuite.java:227)
at org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.java:83)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestFileLimit.testFileLimitSimulated
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name2. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestFileLimit.__CLR3_0_2os0bak1g3p(TestFileLimit.java:89)
at org.apache.hadoop.hdfs.server.namenode.TestFileLimit.testFileLimit(TestFileLimit.java:78)
at org.apache.hadoop.hdfs.server.namenode.TestFileLimit.__CLR3_0_2a70u0q1g5b(TestFileLimit.java:171)
at org.apache.hadoop.hdfs.server.namenode.TestFileLimit.testFileLimitSimulated(TestFileLimit.java:169)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:168)
at junit.framework.TestCase.runBare(TestCase.java:134)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:124)
at junit.framework.TestSuite.runTest(TestSuite.java:232)
at junit.framework.TestSuite.run(TestSuite.java:227)
at org.junit.internal.runners.JUnit38ClassRunner.run(JUnit38ClassRunner.java:83)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testSingleFailedEditsDirOnFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testAllEditsDirsFailOnFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testAllEditsDirFailOnWrite
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testSingleFailedEditsDirOnSetReadyToFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testSingleRequiredFailedEditsDirOnSetReadyToFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.testMultipleRedundantFailedEditsDirOnSetReadyToFlush
Error Message:
Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
Stack Trace:
java.io.IOException: Cannot lock storage /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1. The directory is already locked.
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.lock(Storage.java:586)
at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.analyzeStorage(Storage.java:435)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverStorageDirs(FSImage.java:253)
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:169)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:385)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:328)
at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:299)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:333)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:459)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:451)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:752)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:651)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:555)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:263)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:87)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:249)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:65)
at org.apache.hadoop.hdfs.server.namenode.TestEditLogJournalFailures.setUpMiniCluster(TestEditLogJournalFailures.java:60)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:27)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
Re: Hadoop-Hdfs-trunk - Build # 894 - Still unstable
Posted by Todd Lipcon <to...@cloudera.com>.
I committed this. (r1220940)
-Todd
On Mon, Dec 19, 2011 at 11:49 AM, Todd Lipcon <to...@cloudera.com> wrote:
> I'm going to commit the following fix if no one screams otherwise:
>
> Index: hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml
> ===================================================================
> --- hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml
> Â (revision 1220312)
> +++ hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml
> Â (working copy)
> @@ -16,9 +16,9 @@
> Â <modelVersion>4.0.0</modelVersion>
> Â <parent>
> Â Â <groupId>org.apache.hadoop</groupId>
> - Â Â <artifactId>hadoop-project-dist</artifactId>
> + Â Â <artifactId>hadoop-project</artifactId>
> Â Â <version>0.24.0-SNAPSHOT</version>
> - Â Â <relativePath>../../../../hadoop-project-dist</relativePath>
> + Â Â <relativePath>../../../../../hadoop-project</relativePath>
> Â </parent>
>
> Â <groupId>org.apache.hadoop.contrib</groupId>
>
>
>
> On Mon, Dec 19, 2011 at 10:13 AM, Todd Lipcon <to...@cloudera.com> wrote:
>> HDFS Builds have been failing since HDFS-234 (bk support) went in 6
>> days ago. Can someone please fix this? Otherwise we should revert the
>> patch.
>>
>> -Todd
>>
>>
>> ---------- Forwarded message ----------
>> From: Apache Jenkins Server <je...@builds.apache.org>
>> Date: Wed, Dec 14, 2011 at 7:54 AM
>> Subject: Hadoop-Hdfs-trunk - Build # 894 - Still unstable
>> To: hdfs-dev@hadoop.apache.org
>>
>>
>> See https://builds.apache.org/job/Hadoop-Hdfs-trunk/894/
>>
>> ###################################################################################
>> ########################## LAST 60 LINES OF THE CONSOLE
>> ###########################
>> [...truncated 12545 lines...]
>> [INFO] Compiling 2 source files to
>> /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/test-classes
>> [INFO]
>> [INFO] --- maven-surefire-plugin:2.10:test (default-test) @
>> hadoop-hdfs-bkjournal ---
>> [INFO] Tests are skipped.
>> [INFO]
>> [INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default) @
>> hadoop-hdfs-bkjournal ---
>> [INFO] ****** FindBugsMojo execute *******
>> [INFO] canGenerate is true
>> [INFO] ****** FindBugsMojo executeFindbugs *******
>> [INFO] Temp File is
>> /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/findbugsTemp.xml
>> [INFO] Fork Value is true
>> [INFO] ------------------------------------------------------------------------
>> [INFO] Reactor Summary:
>> [INFO]
>> [INFO] Apache Hadoop HDFS ................................ SUCCESS
>> [2:07:40.168s]
>> [INFO] Apache Hadoop HttpFS .............................. SUCCESS [35.385s]
>> [INFO] Apache Hadoop HDFS BookKeeper Journal ............. FAILURE [1.440s]
>> [INFO] Apache Hadoop HDFS Project ........................ SKIPPED
>> [INFO] ------------------------------------------------------------------------
>> [INFO] BUILD FAILURE
>> [INFO] ------------------------------------------------------------------------
>> [INFO] Total time: 2:08:17.452s
>> [INFO] Finished at: Wed Dec 14 13:43:03 UTC 2011
>> [INFO] Final Memory: 87M/731M
>> [INFO] ------------------------------------------------------------------------
>> [ERROR] Could not find resource
>> '/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/dev-support/findbugsExcludeFile.xml'.
>> -> [Help 1]
>> [ERROR]
>> [ERROR] To see the full stack trace of the errors, re-run Maven with
>> the -e switch.
>> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
>> [ERROR]
>> [ERROR] For more information about the errors and possible solutions,
>> please read the following articles:
>> [ERROR] [Help 1]
>> http://cwiki.apache.org/confluence/display/MAVEN/ResourceNotFoundException
>> + /home/jenkins/tools/maven/latest/bin/mvn test
>> -Dmaven.test.failure.ignore=true -Pclover
>> -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
>> Archiving artifacts
>> Recording test results
>> Build step 'Publish JUnit test result report' changed build result to UNSTABLE
>> Publishing Javadoc
>> Recording fingerprints
>> Updating HDFS-2545
>> Updating MAPREDUCE-3426
>> Updating HADOOP-7892
>> Updating MAPREDUCE-2863
>> Updating MAPREDUCE-3542
>> Updating HDFS-2676
>> Updating HDFS-2663
>> Updating HADOOP-7920
>> Updating HDFS-2675
>> Updating HDFS-2661
>> Updating MAPREDUCE-3545
>> Updating MAPREDUCE-3544
>> Updating HDFS-234
>> Updating HDFS-2649
>> Updating HADOOP-7810
>> Updating HDFS-2650
>> Updating HDFS-2669
>> Updating HDFS-2666
>> Sending e-mails to: hdfs-dev@hadoop.apache.org
>> Email was triggered for: Unstable
>> Sending email for trigger: Unstable
>>
>>
>>
>> ###################################################################################
>> ############################## FAILED TESTS (if any)
>> ##############################
>> 8 tests failed.
>> FAILED: Â org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation
>>
>> Error Message:
>> expected:<401> but was:<200>
>>
>> Stack Trace:
>> junit.framework.AssertionFailedError: expected:<401> but was:<200>
>> Â Â Â Â at junit.framework.Assert.fail(Assert.java:47)
>> Â Â Â Â at junit.framework.Assert.failNotEquals(Assert.java:283)
>> Â Â Â Â at junit.framework.Assert.assertEquals(Assert.java:64)
>> Â Â Â Â at junit.framework.Assert.assertEquals(Assert.java:195)
>> Â Â Â Â at junit.framework.Assert.assertEquals(Assert.java:201)
>> Â Â Â Â at org.apache.hadoop.fs.http.server.TestHttpFSServer.__CLR3_0_2bnqkmt1wc(TestHttpFSServer.java:97)
>> Â Â Â Â at org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation(TestHttpFSServer.java:86)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.apache.hadoop.test.TestHdfsHelper$HdfsStatement.evaluate(TestHdfsHelper.java:73)
>> Â Â Â Â at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
>> Â Â Â Â at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
>> Â Â Â Â at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
>> Â Â Â Â at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>> FAILED: Â org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle
>>
>> Error Message:
>> null
>>
>> Stack Trace:
>> java.lang.NullPointerException
>> Â Â Â Â at java.util.Properties$LineReader.readLine(Properties.java:418)
>> Â Â Â Â at java.util.Properties.load0(Properties.java:337)
>> Â Â Â Â at java.util.Properties.load(Properties.java:325)
>> Â Â Â Â at org.apache.hadoop.lib.server.Server.init(Server.java:348)
>> Â Â Â Â at org.apache.hadoop.lib.servlet.ServerWebApp.contextInitialized(ServerWebApp.java:142)
>> Â Â Â Â at org.apache.hadoop.lib.servlet.TestServerWebApp.__CLR3_0_2sd9si72uk(TestServerWebApp.java:56)
>> Â Â Â Â at org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle(TestServerWebApp.java:46)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
>> Â Â Â Â at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
>> Â Â Â Â at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>> REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInDFS.testDirectory
>>
>> Error Message:
>> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
>> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>>
>> Stack Trace:
>> java.lang.ClassCastException:
>> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
>> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>> Â Â Â Â at org.apache.hadoop.hdfs.DistributedFileSystem$1.hasNext(DistributedFileSystem.java:452)
>> Â Â Â Â at org.apache.hadoop.fs.FileSystem$5.hasNext(FileSystem.java:1551)
>> Â Â Â Â at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1581)
>> Â Â Â Â at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1541)
>> Â Â Â Â at org.apache.hadoop.fs.TestListFiles.testDirectory(TestListFiles.java:146)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
>> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>> REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory
>>
>> Error Message:
>> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
>> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>>
>> Stack Trace:
>> java.lang.ClassCastException:
>> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
>> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
>> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
>> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
>> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
>> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
>> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2kt5srm165m(TestListFilesInFileContext.java:153)
>> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory(TestListFilesInFileContext.java:119)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
>> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>> REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks
>>
>> Error Message:
>> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
>> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>>
>> Stack Trace:
>> java.lang.ClassCastException:
>> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
>> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
>> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
>> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
>> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
>> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
>> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2gy0n1g166p(TestListFilesInFileContext.java:186)
>> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks(TestListFilesInFileContext.java:173)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
>> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>> REGRESSION: Â org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN
>>
>> Error Message:
>> Failed on local exception: java.io.EOFException; Host Details : local
>> host is: "asf002.sp2.ygridcore.net/67.195.138.30"; destination host
>> is: ""localhost":9929;
>>
>> Stack Trace:
>> java.io.IOException: Failed on local exception: java.io.EOFException;
>> Host Details : local host is:
>> "asf002.sp2.ygridcore.net/67.195.138.30"; destination host is:
>> ""localhost":9929;
>> Â Â Â Â at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:655)
>> Â Â Â Â at org.apache.hadoop.ipc.Client.call(Client.java:1140)
>> Â Â Â Â at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:169)
>> Â Â Â Â at $Proxy14.getDatanodeReport(Unknown Source)
>> Â Â Â Â at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:127)
>> Â Â Â Â at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:81)
>> Â Â Â Â at $Proxy14.getDatanodeReport(Unknown Source)
>> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getDatanodeReport(ClientNamenodeProtocolTranslatorPB.java:555)
>> Â Â Â Â at org.apache.hadoop.hdfs.DFSClient.datanodeReport(DFSClient.java:1443)
>> Â Â Â Â at org.apache.hadoop.hdfs.MiniDFSCluster.waitActive(MiniDFSCluster.java:1486)
>> Â Â Â Â at org.apache.hadoop.hdfs.MiniDFSCluster.addNameNode(MiniDFSCluster.java:1904)
>> Â Â Â Â at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.__CLR3_0_2yh11m91bq7(TestDataNodeMultipleRegistrations.java:237)
>> Â Â Â Â at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN(TestDataNodeMultipleRegistrations.java:215)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>> Caused by: java.io.EOFException
>> Â Â Â Â at java.io.DataInputStream.readInt(DataInputStream.java:375)
>> Â Â Â Â at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:825)
>> Â Â Â Â at org.apache.hadoop.ipc.Client$Connection.run(Client.java:761)
>>
>>
>> REGRESSION: Â org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS
>>
>> Error Message:
>> [Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
>>
>> Stack Trace:
>> java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to
>> [Ljava.lang.String;
>> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
>> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
>> Â Â Â Â at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
>> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
>> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
>> Â Â Â Â at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
>> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2ul3ta11h1o(TestListCorruptFileBlocks.java:387)
>> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS(TestListCorruptFileBlocks.java:371)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>> REGRESSION: Â org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles
>>
>> Error Message:
>> [Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
>>
>> Stack Trace:
>> java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to
>> [Ljava.lang.String;
>> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
>> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
>> Â Â Â Â at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
>> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
>> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
>> Â Â Â Â at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
>> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2mzyzlv1h3a(TestListCorruptFileBlocks.java:502)
>> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles(TestListCorruptFileBlocks.java:442)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
>> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
>> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
>> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
>> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
>> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
>> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
>> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
>> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
>> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
>> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
>> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
>> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
>> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
>> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>>
>>
>>
>>
>>
>> --
>> Todd Lipcon
>> Software Engineer, Cloudera
>
>
>
> --
> Todd Lipcon
> Software Engineer, Cloudera
--
Todd Lipcon
Software Engineer, Cloudera
Re: Hadoop-Hdfs-trunk - Build # 894 - Still unstable
Posted by Todd Lipcon <to...@cloudera.com>.
I'm going to commit the following fix if no one screams otherwise:
Index: hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml
===================================================================
--- hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml
(revision 1220312)
+++ hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/pom.xml
(working copy)
@@ -16,9 +16,9 @@
<modelVersion>4.0.0</modelVersion>
<parent>
<groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-project-dist</artifactId>
+ <artifactId>hadoop-project</artifactId>
<version>0.24.0-SNAPSHOT</version>
- <relativePath>../../../../hadoop-project-dist</relativePath>
+ <relativePath>../../../../../hadoop-project</relativePath>
</parent>
<groupId>org.apache.hadoop.contrib</groupId>
On Mon, Dec 19, 2011 at 10:13 AM, Todd Lipcon <to...@cloudera.com> wrote:
> HDFS Builds have been failing since HDFS-234 (bk support) went in 6
> days ago. Can someone please fix this? Otherwise we should revert the
> patch.
>
> -Todd
>
>
> ---------- Forwarded message ----------
> From: Apache Jenkins Server <je...@builds.apache.org>
> Date: Wed, Dec 14, 2011 at 7:54 AM
> Subject: Hadoop-Hdfs-trunk - Build # 894 - Still unstable
> To: hdfs-dev@hadoop.apache.org
>
>
> See https://builds.apache.org/job/Hadoop-Hdfs-trunk/894/
>
> ###################################################################################
> ########################## LAST 60 LINES OF THE CONSOLE
> ###########################
> [...truncated 12545 lines...]
> [INFO] Compiling 2 source files to
> /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/test-classes
> [INFO]
> [INFO] --- maven-surefire-plugin:2.10:test (default-test) @
> hadoop-hdfs-bkjournal ---
> [INFO] Tests are skipped.
> [INFO]
> [INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default) @
> hadoop-hdfs-bkjournal ---
> [INFO] ****** FindBugsMojo execute *******
> [INFO] canGenerate is true
> [INFO] ****** FindBugsMojo executeFindbugs *******
> [INFO] Temp File is
> /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/findbugsTemp.xml
> [INFO] Fork Value is true
> [INFO] ------------------------------------------------------------------------
> [INFO] Reactor Summary:
> [INFO]
> [INFO] Apache Hadoop HDFS ................................ SUCCESS
> [2:07:40.168s]
> [INFO] Apache Hadoop HttpFS .............................. SUCCESS [35.385s]
> [INFO] Apache Hadoop HDFS BookKeeper Journal ............. FAILURE [1.440s]
> [INFO] Apache Hadoop HDFS Project ........................ SKIPPED
> [INFO] ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO] ------------------------------------------------------------------------
> [INFO] Total time: 2:08:17.452s
> [INFO] Finished at: Wed Dec 14 13:43:03 UTC 2011
> [INFO] Final Memory: 87M/731M
> [INFO] ------------------------------------------------------------------------
> [ERROR] Could not find resource
> '/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/dev-support/findbugsExcludeFile.xml'.
> -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with
> the -e switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions,
> please read the following articles:
> [ERROR] [Help 1]
> http://cwiki.apache.org/confluence/display/MAVEN/ResourceNotFoundException
> + /home/jenkins/tools/maven/latest/bin/mvn test
> -Dmaven.test.failure.ignore=true -Pclover
> -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
> Archiving artifacts
> Recording test results
> Build step 'Publish JUnit test result report' changed build result to UNSTABLE
> Publishing Javadoc
> Recording fingerprints
> Updating HDFS-2545
> Updating MAPREDUCE-3426
> Updating HADOOP-7892
> Updating MAPREDUCE-2863
> Updating MAPREDUCE-3542
> Updating HDFS-2676
> Updating HDFS-2663
> Updating HADOOP-7920
> Updating HDFS-2675
> Updating HDFS-2661
> Updating MAPREDUCE-3545
> Updating MAPREDUCE-3544
> Updating HDFS-234
> Updating HDFS-2649
> Updating HADOOP-7810
> Updating HDFS-2650
> Updating HDFS-2669
> Updating HDFS-2666
> Sending e-mails to: hdfs-dev@hadoop.apache.org
> Email was triggered for: Unstable
> Sending email for trigger: Unstable
>
>
>
> ###################################################################################
> ############################## FAILED TESTS (if any)
> ##############################
> 8 tests failed.
> FAILED: Â org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation
>
> Error Message:
> expected:<401> but was:<200>
>
> Stack Trace:
> junit.framework.AssertionFailedError: expected:<401> but was:<200>
> Â Â Â Â at junit.framework.Assert.fail(Assert.java:47)
> Â Â Â Â at junit.framework.Assert.failNotEquals(Assert.java:283)
> Â Â Â Â at junit.framework.Assert.assertEquals(Assert.java:64)
> Â Â Â Â at junit.framework.Assert.assertEquals(Assert.java:195)
> Â Â Â Â at junit.framework.Assert.assertEquals(Assert.java:201)
> Â Â Â Â at org.apache.hadoop.fs.http.server.TestHttpFSServer.__CLR3_0_2bnqkmt1wc(TestHttpFSServer.java:97)
> Â Â Â Â at org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation(TestHttpFSServer.java:86)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.apache.hadoop.test.TestHdfsHelper$HdfsStatement.evaluate(TestHdfsHelper.java:73)
> Â Â Â Â at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
> Â Â Â Â at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
> Â Â Â Â at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
> Â Â Â Â at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
> FAILED: Â org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle
>
> Error Message:
> null
>
> Stack Trace:
> java.lang.NullPointerException
> Â Â Â Â at java.util.Properties$LineReader.readLine(Properties.java:418)
> Â Â Â Â at java.util.Properties.load0(Properties.java:337)
> Â Â Â Â at java.util.Properties.load(Properties.java:325)
> Â Â Â Â at org.apache.hadoop.lib.server.Server.init(Server.java:348)
> Â Â Â Â at org.apache.hadoop.lib.servlet.ServerWebApp.contextInitialized(ServerWebApp.java:142)
> Â Â Â Â at org.apache.hadoop.lib.servlet.TestServerWebApp.__CLR3_0_2sd9si72uk(TestServerWebApp.java:56)
> Â Â Â Â at org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle(TestServerWebApp.java:46)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
> Â Â Â Â at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
> Â Â Â Â at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
> REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInDFS.testDirectory
>
> Error Message:
> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>
> Stack Trace:
> java.lang.ClassCastException:
> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
> Â Â Â Â at org.apache.hadoop.hdfs.DistributedFileSystem$1.hasNext(DistributedFileSystem.java:452)
> Â Â Â Â at org.apache.hadoop.fs.FileSystem$5.hasNext(FileSystem.java:1551)
> Â Â Â Â at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1581)
> Â Â Â Â at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1541)
> Â Â Â Â at org.apache.hadoop.fs.TestListFiles.testDirectory(TestListFiles.java:146)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
> REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory
>
> Error Message:
> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>
> Stack Trace:
> java.lang.ClassCastException:
> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2kt5srm165m(TestListFilesInFileContext.java:153)
> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory(TestListFilesInFileContext.java:119)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
> REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks
>
> Error Message:
> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
>
> Stack Trace:
> java.lang.ClassCastException:
> org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
> org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
> Â Â Â Â at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
> Â Â Â Â at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2gy0n1g166p(TestListFilesInFileContext.java:186)
> Â Â Â Â at org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks(TestListFilesInFileContext.java:173)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
> Â Â Â Â at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
> REGRESSION: Â org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN
>
> Error Message:
> Failed on local exception: java.io.EOFException; Host Details : local
> host is: "asf002.sp2.ygridcore.net/67.195.138.30"; destination host
> is: ""localhost":9929;
>
> Stack Trace:
> java.io.IOException: Failed on local exception: java.io.EOFException;
> Host Details : local host is:
> "asf002.sp2.ygridcore.net/67.195.138.30"; destination host is:
> ""localhost":9929;
> Â Â Â Â at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:655)
> Â Â Â Â at org.apache.hadoop.ipc.Client.call(Client.java:1140)
> Â Â Â Â at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:169)
> Â Â Â Â at $Proxy14.getDatanodeReport(Unknown Source)
> Â Â Â Â at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:127)
> Â Â Â Â at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:81)
> Â Â Â Â at $Proxy14.getDatanodeReport(Unknown Source)
> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getDatanodeReport(ClientNamenodeProtocolTranslatorPB.java:555)
> Â Â Â Â at org.apache.hadoop.hdfs.DFSClient.datanodeReport(DFSClient.java:1443)
> Â Â Â Â at org.apache.hadoop.hdfs.MiniDFSCluster.waitActive(MiniDFSCluster.java:1486)
> Â Â Â Â at org.apache.hadoop.hdfs.MiniDFSCluster.addNameNode(MiniDFSCluster.java:1904)
> Â Â Â Â at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.__CLR3_0_2yh11m91bq7(TestDataNodeMultipleRegistrations.java:237)
> Â Â Â Â at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN(TestDataNodeMultipleRegistrations.java:215)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
> Caused by: java.io.EOFException
> Â Â Â Â at java.io.DataInputStream.readInt(DataInputStream.java:375)
> Â Â Â Â at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:825)
> Â Â Â Â at org.apache.hadoop.ipc.Client$Connection.run(Client.java:761)
>
>
> REGRESSION: Â org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS
>
> Error Message:
> [Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
>
> Stack Trace:
> java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to
> [Ljava.lang.String;
> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
> Â Â Â Â at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
> Â Â Â Â at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2ul3ta11h1o(TestListCorruptFileBlocks.java:387)
> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS(TestListCorruptFileBlocks.java:371)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
> REGRESSION: Â org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles
>
> Error Message:
> [Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
>
> Stack Trace:
> java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to
> [Ljava.lang.String;
> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
> Â Â Â Â at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
> Â Â Â Â at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
> Â Â Â Â at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
> Â Â Â Â at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2mzyzlv1h3a(TestListCorruptFileBlocks.java:502)
> Â Â Â Â at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles(TestListCorruptFileBlocks.java:442)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
> Â Â Â Â at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
> Â Â Â Â at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
> Â Â Â Â at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
> Â Â Â Â at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
> Â Â Â Â at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
> Â Â Â Â at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
> Â Â Â Â at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
> Â Â Â Â at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
> Â Â Â Â at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
> Â Â Â Â at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
> Â Â Â Â at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> Â Â Â Â at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> Â Â Â Â at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> Â Â Â Â at java.lang.reflect.Method.invoke(Method.java:597)
> Â Â Â Â at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
> Â Â Â Â at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
> Â Â Â Â at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
> Â Â Â Â at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
>
>
>
>
>
> --
> Todd Lipcon
> Software Engineer, Cloudera
--
Todd Lipcon
Software Engineer, Cloudera
Fwd: Hadoop-Hdfs-trunk - Build # 894 - Still unstable
Posted by Todd Lipcon <to...@cloudera.com>.
HDFS Builds have been failing since HDFS-234 (bk support) went in 6
days ago. Can someone please fix this? Otherwise we should revert the
patch.
-Todd
---------- Forwarded message ----------
From: Apache Jenkins Server <je...@builds.apache.org>
Date: Wed, Dec 14, 2011 at 7:54 AM
Subject: Hadoop-Hdfs-trunk - Build # 894 - Still unstable
To: hdfs-dev@hadoop.apache.org
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/894/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE
###########################
[...truncated 12545 lines...]
[INFO] Compiling 2 source files to
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/test-classes
[INFO]
[INFO] --- maven-surefire-plugin:2.10:test (default-test) @
hadoop-hdfs-bkjournal ---
[INFO] Tests are skipped.
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default) @
hadoop-hdfs-bkjournal ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is true
[INFO] ****** FindBugsMojo executeFindbugs *******
[INFO] Temp File is
/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/findbugsTemp.xml
[INFO] Fork Value is true
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop HDFS ................................ SUCCESS
[2:07:40.168s]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [35.385s]
[INFO] Apache Hadoop HDFS BookKeeper Journal ............. FAILURE [1.440s]
[INFO] Apache Hadoop HDFS Project ........................ SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:08:17.452s
[INFO] Finished at: Wed Dec 14 13:43:03 UTC 2011
[INFO] Final Memory: 87M/731M
[INFO] ------------------------------------------------------------------------
[ERROR] Could not find resource
'/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/dev-support/findbugsExcludeFile.xml'.
-> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with
the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions,
please read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/ResourceNotFoundException
+ /home/jenkins/tools/maven/latest/bin/mvn test
-Dmaven.test.failure.ignore=true -Pclover
-DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
Archiving artifacts
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Publishing Javadoc
Recording fingerprints
Updating HDFS-2545
Updating MAPREDUCE-3426
Updating HADOOP-7892
Updating MAPREDUCE-2863
Updating MAPREDUCE-3542
Updating HDFS-2676
Updating HDFS-2663
Updating HADOOP-7920
Updating HDFS-2675
Updating HDFS-2661
Updating MAPREDUCE-3545
Updating MAPREDUCE-3544
Updating HDFS-234
Updating HDFS-2649
Updating HADOOP-7810
Updating HDFS-2650
Updating HDFS-2669
Updating HDFS-2666
Sending e-mails to: hdfs-dev@hadoop.apache.org
Email was triggered for: Unstable
Sending email for trigger: Unstable
###################################################################################
############################## FAILED TESTS (if any)
##############################
8 tests failed.
FAILED: Â org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation
Error Message:
expected:<401> but was:<200>
Stack Trace:
junit.framework.AssertionFailedError: expected:<401> but was:<200>
    at junit.framework.Assert.fail(Assert.java:47)
    at junit.framework.Assert.failNotEquals(Assert.java:283)
    at junit.framework.Assert.assertEquals(Assert.java:64)
    at junit.framework.Assert.assertEquals(Assert.java:195)
    at junit.framework.Assert.assertEquals(Assert.java:201)
    at org.apache.hadoop.fs.http.server.TestHttpFSServer.__CLR3_0_2bnqkmt1wc(TestHttpFSServer.java:97)
    at org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation(TestHttpFSServer.java:86)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.apache.hadoop.test.TestHdfsHelper$HdfsStatement.evaluate(TestHdfsHelper.java:73)
    at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
    at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
    at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
    at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: Â org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle
Error Message:
null
Stack Trace:
java.lang.NullPointerException
    at java.util.Properties$LineReader.readLine(Properties.java:418)
    at java.util.Properties.load0(Properties.java:337)
    at java.util.Properties.load(Properties.java:325)
    at org.apache.hadoop.lib.server.Server.init(Server.java:348)
    at org.apache.hadoop.lib.servlet.ServerWebApp.contextInitialized(ServerWebApp.java:142)
    at org.apache.hadoop.lib.servlet.TestServerWebApp.__CLR3_0_2sd9si72uk(TestServerWebApp.java:56)
    at org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle(TestServerWebApp.java:46)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
    at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
    at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInDFS.testDirectory
Error Message:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
Stack Trace:
java.lang.ClassCastException:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
    at org.apache.hadoop.hdfs.DistributedFileSystem$1.hasNext(DistributedFileSystem.java:452)
    at org.apache.hadoop.fs.FileSystem$5.hasNext(FileSystem.java:1551)
    at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1581)
    at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1541)
    at org.apache.hadoop.fs.TestListFiles.testDirectory(TestListFiles.java:146)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
    at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory
Error Message:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
Stack Trace:
java.lang.ClassCastException:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
    at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
    at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
    at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
    at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
    at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
    at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2kt5srm165m(TestListFilesInFileContext.java:153)
    at org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory(TestListFilesInFileContext.java:119)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
    at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: Â org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks
Error Message:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
Stack Trace:
java.lang.ClassCastException:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to
org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
    at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
    at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
    at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
    at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
    at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
    at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2gy0n1g166p(TestListFilesInFileContext.java:186)
    at org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks(TestListFilesInFileContext.java:173)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
    at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: Â org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN
Error Message:
Failed on local exception: java.io.EOFException; Host Details : local
host is: "asf002.sp2.ygridcore.net/67.195.138.30"; destination host
is: ""localhost":9929;
Stack Trace:
java.io.IOException: Failed on local exception: java.io.EOFException;
Host Details : local host is:
"asf002.sp2.ygridcore.net/67.195.138.30"; destination host is:
""localhost":9929;
    at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:655)
    at org.apache.hadoop.ipc.Client.call(Client.java:1140)
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:169)
    at $Proxy14.getDatanodeReport(Unknown Source)
    at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:127)
    at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:81)
    at $Proxy14.getDatanodeReport(Unknown Source)
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getDatanodeReport(ClientNamenodeProtocolTranslatorPB.java:555)
    at org.apache.hadoop.hdfs.DFSClient.datanodeReport(DFSClient.java:1443)
    at org.apache.hadoop.hdfs.MiniDFSCluster.waitActive(MiniDFSCluster.java:1486)
    at org.apache.hadoop.hdfs.MiniDFSCluster.addNameNode(MiniDFSCluster.java:1904)
    at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.__CLR3_0_2yh11m91bq7(TestDataNodeMultipleRegistrations.java:237)
    at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN(TestDataNodeMultipleRegistrations.java:215)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
Caused by: java.io.EOFException
    at java.io.DataInputStream.readInt(DataInputStream.java:375)
    at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:825)
    at org.apache.hadoop.ipc.Client$Connection.run(Client.java:761)
REGRESSION: Â org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS
Error Message:
[Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
Stack Trace:
java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to
[Ljava.lang.String;
    at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
    at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
    at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
    at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
    at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
    at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2ul3ta11h1o(TestListCorruptFileBlocks.java:387)
    at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS(TestListCorruptFileBlocks.java:371)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: Â org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles
Error Message:
[Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
Stack Trace:
java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to
[Ljava.lang.String;
    at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
    at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
    at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
    at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
    at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
    at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2mzyzlv1h3a(TestListCorruptFileBlocks.java:502)
    at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles(TestListCorruptFileBlocks.java:442)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
    at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
    at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
    at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
    at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
    at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
    at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
    at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
--
Todd Lipcon
Software Engineer, Cloudera
Hadoop-Hdfs-trunk - Build # 894 - Still unstable
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/894/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 12545 lines...]
[INFO] Compiling 2 source files to /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/test-classes
[INFO]
[INFO] --- maven-surefire-plugin:2.10:test (default-test) @ hadoop-hdfs-bkjournal ---
[INFO] Tests are skipped.
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default) @ hadoop-hdfs-bkjournal ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is true
[INFO] ****** FindBugsMojo executeFindbugs *******
[INFO] Temp File is /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/findbugsTemp.xml
[INFO] Fork Value is true
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop HDFS ................................ SUCCESS [2:07:40.168s]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [35.385s]
[INFO] Apache Hadoop HDFS BookKeeper Journal ............. FAILURE [1.440s]
[INFO] Apache Hadoop HDFS Project ........................ SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:08:17.452s
[INFO] Finished at: Wed Dec 14 13:43:03 UTC 2011
[INFO] Final Memory: 87M/731M
[INFO] ------------------------------------------------------------------------
[ERROR] Could not find resource '/home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk/trunk/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/dev-support/findbugsExcludeFile.xml'. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/ResourceNotFoundException
+ /home/jenkins/tools/maven/latest/bin/mvn test -Dmaven.test.failure.ignore=true -Pclover -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
Archiving artifacts
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Publishing Javadoc
Recording fingerprints
Updating HDFS-2545
Updating MAPREDUCE-3426
Updating HADOOP-7892
Updating MAPREDUCE-2863
Updating MAPREDUCE-3542
Updating HDFS-2676
Updating HDFS-2663
Updating HADOOP-7920
Updating HDFS-2675
Updating HDFS-2661
Updating MAPREDUCE-3545
Updating MAPREDUCE-3544
Updating HDFS-234
Updating HDFS-2649
Updating HADOOP-7810
Updating HDFS-2650
Updating HDFS-2669
Updating HDFS-2666
Sending e-mails to: hdfs-dev@hadoop.apache.org
Email was triggered for: Unstable
Sending email for trigger: Unstable
###################################################################################
############################## FAILED TESTS (if any) ##############################
8 tests failed.
FAILED: org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation
Error Message:
expected:<401> but was:<200>
Stack Trace:
junit.framework.AssertionFailedError: expected:<401> but was:<200>
at junit.framework.Assert.fail(Assert.java:47)
at junit.framework.Assert.failNotEquals(Assert.java:283)
at junit.framework.Assert.assertEquals(Assert.java:64)
at junit.framework.Assert.assertEquals(Assert.java:195)
at junit.framework.Assert.assertEquals(Assert.java:201)
at org.apache.hadoop.fs.http.server.TestHttpFSServer.__CLR3_0_2bnqkmt1wc(TestHttpFSServer.java:97)
at org.apache.hadoop.fs.http.server.TestHttpFSServer.instrumentation(TestHttpFSServer.java:86)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.apache.hadoop.test.TestHdfsHelper$HdfsStatement.evaluate(TestHdfsHelper.java:73)
at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
FAILED: org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle
Error Message:
null
Stack Trace:
java.lang.NullPointerException
at java.util.Properties$LineReader.readLine(Properties.java:418)
at java.util.Properties.load0(Properties.java:337)
at java.util.Properties.load(Properties.java:325)
at org.apache.hadoop.lib.server.Server.init(Server.java:348)
at org.apache.hadoop.lib.servlet.ServerWebApp.contextInitialized(ServerWebApp.java:142)
at org.apache.hadoop.lib.servlet.TestServerWebApp.__CLR3_0_2sd9si72uk(TestServerWebApp.java:56)
at org.apache.hadoop.lib.servlet.TestServerWebApp.lifecycle(TestServerWebApp.java:46)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.apache.hadoop.test.TestDirHelper$1.evaluate(TestDirHelper.java:108)
at org.apache.hadoop.test.TestJettyHelper$1.evaluate(TestJettyHelper.java:51)
at org.apache.hadoop.test.TestExceptionHelper$1.evaluate(TestExceptionHelper.java:41)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.TestListFilesInDFS.testDirectory
Error Message:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
Stack Trace:
java.lang.ClassCastException: org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
at org.apache.hadoop.hdfs.DistributedFileSystem$1.hasNext(DistributedFileSystem.java:452)
at org.apache.hadoop.fs.FileSystem$5.hasNext(FileSystem.java:1551)
at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1581)
at org.apache.hadoop.fs.FileSystem$5.next(FileSystem.java:1541)
at org.apache.hadoop.fs.TestListFiles.testDirectory(TestListFiles.java:146)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory
Error Message:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
Stack Trace:
java.lang.ClassCastException: org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2kt5srm165m(TestListFilesInFileContext.java:153)
at org.apache.hadoop.hdfs.TestListFilesInFileContext.testDirectory(TestListFilesInFileContext.java:119)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks
Error Message:
org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
Stack Trace:
java.lang.ClassCastException: org.apache.hadoop.hdfs.protocol.HdfsFileStatus cannot be cast to org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus
at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:183)
at org.apache.hadoop.fs.Hdfs$1.next(Hdfs.java:179)
at org.apache.hadoop.fs.FileContext$Util$2.hasNext(FileContext.java:1763)
at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1819)
at org.apache.hadoop.fs.FileContext$Util$2.next(FileContext.java:1739)
at org.apache.hadoop.hdfs.TestListFilesInFileContext.__CLR3_0_2gy0n1g166p(TestListFilesInFileContext.java:186)
at org.apache.hadoop.hdfs.TestListFilesInFileContext.testSymbolicLinks(TestListFilesInFileContext.java:173)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:31)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN
Error Message:
Failed on local exception: java.io.EOFException; Host Details : local host is: "asf002.sp2.ygridcore.net/67.195.138.30"; destination host is: ""localhost":9929;
Stack Trace:
java.io.IOException: Failed on local exception: java.io.EOFException; Host Details : local host is: "asf002.sp2.ygridcore.net/67.195.138.30"; destination host is: ""localhost":9929;
at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:655)
at org.apache.hadoop.ipc.Client.call(Client.java:1140)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:169)
at $Proxy14.getDatanodeReport(Unknown Source)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:127)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:81)
at $Proxy14.getDatanodeReport(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getDatanodeReport(ClientNamenodeProtocolTranslatorPB.java:555)
at org.apache.hadoop.hdfs.DFSClient.datanodeReport(DFSClient.java:1443)
at org.apache.hadoop.hdfs.MiniDFSCluster.waitActive(MiniDFSCluster.java:1486)
at org.apache.hadoop.hdfs.MiniDFSCluster.addNameNode(MiniDFSCluster.java:1904)
at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.__CLR3_0_2yh11m91bq7(TestDataNodeMultipleRegistrations.java:237)
at org.apache.hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations.testMiniDFSClusterWithMultipleNN(TestDataNodeMultipleRegistrations.java:215)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:28)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
Caused by: java.io.EOFException
at java.io.DataInputStream.readInt(DataInputStream.java:375)
at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:825)
at org.apache.hadoop.ipc.Client$Connection.run(Client.java:761)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS
Error Message:
[Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
Stack Trace:
java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2ul3ta11h1o(TestListCorruptFileBlocks.java:387)
at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testlistCorruptFileBlocksDFS(TestListCorruptFileBlocks.java:371)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
REGRESSION: org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles
Error Message:
[Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
Stack Trace:
java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to [Ljava.lang.String;
at org.apache.hadoop.hdfs.protocolPB.PBHelper.convert(PBHelper.java:1217)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.listCorruptFileBlocks(ClientNamenodeProtocolTranslatorPB.java:654)
at org.apache.hadoop.hdfs.DFSClient.listCorruptFileBlocks(DFSClient.java:1438)
at org.apache.hadoop.hdfs.CorruptFileBlockIterator.loadNext(CorruptFileBlockIterator.java:67)
at org.apache.hadoop.hdfs.CorruptFileBlockIterator.<init>(CorruptFileBlockIterator.java:46)
at org.apache.hadoop.hdfs.DistributedFileSystem.listCorruptFileBlocks(DistributedFileSystem.java:618)
at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.__CLR3_0_2mzyzlv1h3a(TestListCorruptFileBlocks.java:502)
at org.apache.hadoop.hdfs.server.namenode.TestListCorruptFileBlocks.testMaxCorruptFiles(TestListCorruptFileBlocks.java:442)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:44)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:15)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:41)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:20)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:76)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:193)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:52)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:191)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:42)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:184)
at org.junit.runners.ParentRunner.run(ParentRunner.java:236)
at org.apache.maven.surefire.junit4.JUnit4TestSet.execute(JUnit4TestSet.java:53)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:123)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:104)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:164)
at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:110)
at org.apache.maven.surefire.booter.SurefireStarter.invokeProvider(SurefireStarter.java:175)
at org.apache.maven.surefire.booter.SurefireStarter.runSuitesInProcessWhenForked(SurefireStarter.java:81)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:68)
Jenkins build is unstable: Hadoop-Hdfs-trunk #894
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Hadoop-Hdfs-trunk/894/changes>
Build failed in Jenkins: Hadoop-Hdfs-trunk #893
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Hadoop-Hdfs-trunk/893/changes>
Changes:
[szetszwo] Add .classpath, .externalToolBuilders, .launches, .project and .settings to svn:ignore of hadoop-mapreduce-project/hadoop-mapreduce-examples.
[sradia] HADOOP-7913 Fix bug in ProtoBufRpcEngine (sanjay)
[szetszwo] HADOOP-7914. Remove the duplicated declaration of hadoop-hdfs test-jar in hadoop-project/pom.xml.
[eli] Add missing file to previous commit.
[eli] HDFS-2654. Make BlockReaderLocal not extend RemoteBlockReader2. Contributed by Eli Collins
[eli] HDFS-2653. DFSClient should cache whether addrs are non-local when short-circuiting is enabled. Contributed by Eli Collins
[acmurthy] MAPREDUCE-3537. Fix race condition in DefaultContainerExecutor which led to container localization occuring in wrong directories.
[eli] Put HDFS-1765 in the right place in the changelog.
[eli] HDFS-1765. Block Replication should respect under-replication block priority. Contributed by Uma Maheswara Rao G
[suresh] HDFS-2663. Handle protobuf optional parameters correctly. Contributed by Suresh Srinivas.
[mahadev] MAPREDUCE-3510. Capacity Scheduler inherited ACLs not displayed by mapred queue -showacls (Jonathan Eagles via mahadev)
[mahadev] MAPREDUCE-3328. mapred queue -list output inconsistent and missing child queues. (Ravi Prakash via mahadev)
[mahadev] MAPREDUCE-3518. mapred queue -info <queue> -showJobs throws NPE. (Jonathan Eagles via mahadev)
[tomwhite] HADOOP-7912. test-patch should run eclipse:eclipse to verify that it does not break again. Contributed by Robert Joseph Evans
[tomwhite] MAPREDUCE-3527. Fix minor API incompatibilities between 1.0 and 0.23.
[stevel] HADOOP-7878 Regression: HADOOP-7777 switch changes break HDFS tests when the isSingleSwitch() predicate is used
------------------------------------------
[...truncated 11972 lines...]
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSDelete.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSFileChecksum.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSFileStatus.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSHomeDir.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSListStatus.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSMkdirs.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSOpen.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSRename.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSSetOwner.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSSetPermission.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSSetReplication.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/FSOperations.FSSetTimes.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSServerWebApp.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSExceptionProvider.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSServer.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/lang//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/server//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/hadoop//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/instrumentation//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/scheduler//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/security//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/servlet//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/util//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/wsrs//package-use.html...>
Building index for all the packages and classes...
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/overview-tree.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/index-all.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/deprecated-list.html...>
Building index for all classes...
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/allclasses-frame.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/allclasses-noframe.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/index.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/overview-summary.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/help-doc.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/stylesheet.css...>
90 warnings
[WARNING] Javadoc Warnings
[WARNING] org/apache/hadoop/fs/FileSystem.class(org/apache/hadoop/fs:FileSystem.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate': class file for org.apache.hadoop.classification.InterfaceAudience not found
[WARNING] org/apache/hadoop/fs/FileSystem.class(org/apache/hadoop/fs:FileSystem.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/UserGroupInformation.class(org/apache/hadoop/security:UserGroupInformation.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/UserGroupInformation.class(org/apache/hadoop/security:UserGroupInformation.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/UserGroupInformation.class(org/apache/hadoop/security:UserGroupInformation.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/Groups.class(org/apache/hadoop/security:Groups.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSDataInputStream.class(org/apache/hadoop/fs:FSDataInputStream.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSDataOutputStream.class(org/apache/hadoop/fs:FSDataOutputStream.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/Path.class(org/apache/hadoop/fs:Path.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/RunnableCallable.java>:76: warning - @return tag cannot be used in method with void return type.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java>:708: warning - @return tag has no arguments.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java>:262: warning - Tag @link: can't find serverStatusChange(Status, Status) in org.apache.hadoop.lib.server.Service
[WARNING] org/apache/hadoop/fs/UnresolvedLinkException.class(org/apache/hadoop/fs:UnresolvedLinkException.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.class(org/apache/hadoop/fs:MD5MD5CRC32FileChecksum.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/LocalDirAllocator.class(org/apache/hadoop/fs:LocalDirAllocator.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FileContext.class(org/apache/hadoop/fs:FileContext.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FileContext.class(org/apache/hadoop/fs:FileContext.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSOutputSummer.class(org/apache/hadoop/fs:FSOutputSummer.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSInputStream.class(org/apache/hadoop/fs:FSInputStream.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSInputChecker.class(org/apache/hadoop/fs:FSInputChecker.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/DU.class(org/apache/hadoop/fs:DU.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/DF.class(org/apache/hadoop/fs:DF.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/AbstractFileSystem.class(org/apache/hadoop/fs:AbstractFileSystem.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParams.java>:234: warning - @returns is an unknown tag.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:111: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:111: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#LISTSTATUS
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#INSTRUMENTATION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#LISTSTATUS
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - @param argument "override," is not a parameter name.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETPERMISSION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETREPLICATION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:379: warning - Tag @link: reference not found: HttpFSFileSystem.DeleteOpValues#DELETE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues#APPEND
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#RENAME
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETTIMES
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETOWNER
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETOWNER
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - @param argument "override," is not a parameter name.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETPERMISSION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETREPLICATION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETTIMES
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETTIMES
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues#APPEND
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues#APPEND
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[INFO] Building jar: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT-javadoc.jar>
[INFO]
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-httpfs ---
[INFO] Building jar: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT-sources.jar>
[INFO]
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ hadoop-hdfs-httpfs ---
[INFO]
[INFO] --- maven-assembly-plugin:2.2-beta-3:single (dist) @ hadoop-hdfs-httpfs ---
[INFO] Copying files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT>
[WARNING] Assembly file: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT> is not a regular file (it may be a directory). It cannot be attached to the project build for installation or deployment.
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (dist) @ hadoop-hdfs-httpfs ---
[INFO] Executing tasks
main:
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/downloads>
[get] Getting: http://archive.apache.org/dist/tomcat/tomcat-6/v6.0.32/bin/apache-tomcat-6.0.32.tar.gz
[get] To: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/downloads/tomcat.tar.gz>
.........................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/tomcat.exp>
[delete] Deleting directory <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/tomcat.exp>
[delete] Deleting directory <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps>
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps>
[delete] Deleting: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/conf/server.xml>
[copy] Copying 1 file to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/conf>
[copy] Copying 1 file to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/conf>
[copy] Copying 2 files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps/ROOT>
[copy] Copying 166 files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps/webhdfs>
[copy] Copied 23 empty directories to 1 empty directory under <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps/webhdfs>
[INFO] Executed tasks
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (tar) @ hadoop-hdfs-httpfs ---
[INFO] Executing tasks
main:
[INFO] Executed tasks
[INFO]
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ hadoop-hdfs-httpfs ---
[INFO]
[INFO] There are 1655 checkstyle errors.
[WARNING] Unable to locate Source XRef to link to - DISABLED
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ hadoop-hdfs-httpfs ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is true
[INFO] ****** FindBugsMojo executeFindbugs *******
[INFO] Temp File is <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/findbugsTemp.xml>
[INFO] Fork Value is true
[java] The following classes needed for analysis were missing:
[java] org.apache.hadoop.classification.InterfaceAudience$Private
[java] org.apache.hadoop.classification.InterfaceStability$Stable
[java] org.apache.hadoop.classification.InterfaceAudience
[java] org.apache.hadoop.classification.InterfaceStability$Evolving
[java] org.apache.hadoop.classification.InterfaceStability
[java] org.apache.hadoop.classification.InterfaceAudience$Public
[java] org.apache.hadoop.classification.InterfaceAudience$LimitedPrivate
[java] Warnings generated: 4
[java] Missing classes: 7
[INFO] xmlOutput is false
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Building Apache Hadoop HDFS Project 0.24.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-hdfs-project ---
[INFO] Deleting <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/target>
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (create-testdirs) @ hadoop-hdfs-project ---
[INFO] Executing tasks
main:
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/target/test-dir>
[INFO] Executed tasks
[INFO]
[INFO] --- maven-javadoc-plugin:2.7:jar (module-javadocs) @ hadoop-hdfs-project ---
[INFO] Not executing Javadoc as the project is not a Java classpath-capable package
[INFO]
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ hadoop-hdfs-project ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is false
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop HDFS ................................ SUCCESS [2:07:07.684s]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [33.699s]
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [0.036s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:07:41.883s
[INFO] Finished at: Tue Dec 13 13:41:48 UTC 2011
[INFO] Final Memory: 49M/749M
[INFO] ------------------------------------------------------------------------
+ /home/jenkins/tools/maven/latest/bin/mvn test -Dmaven.test.failure.ignore=true -Pclover -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results
Publishing Javadoc
Recording fingerprints
Updating MAPREDUCE-3518
Updating MAPREDUCE-3527
Updating MAPREDUCE-3537
Updating HADOOP-7878
Updating HDFS-1765
Updating HADOOP-7914
Updating HADOOP-7913
Updating HDFS-2663
Updating HADOOP-7912
Updating HADOOP-7777
Updating MAPREDUCE-3510
Updating HDFS-2653
Updating HDFS-2654
Updating MAPREDUCE-3328
Build failed in Jenkins: Hadoop-Hdfs-trunk #892
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Hadoop-Hdfs-trunk/892/changes>
Changes:
[shv] JDiff output for release 0.22.0
[sradia] HDFS-2651 ClientNameNodeProtocol Translators for Protocol Buffers (sanjay)
[shv] Preparing for release 0.22.0
[suresh] HDFS-2647. Used protobuf based RPC for InterDatanodeProtocol, ClientDatanodeProtocol, JournalProtocol, NamenodeProtocol. Contributed by Suresh Srinivas.
------------------------------------------
[...truncated 11680 lines...]
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.FilterParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.FsPathParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.GetOpParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.GroupParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.LenParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.ModifiedTimeParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.OffsetParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.OverwriteParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.OwnerParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.PermissionParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.PostOpParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.PutOpParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.ReplicationParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSParams.ToPathParam.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSServerWebApp.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSExceptionProvider.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/AuthFilter.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSReleaseFilter.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//class-use/HttpFSServer.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpPseudoAuthenticator.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpKerberosAuthenticator.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpFSFileSystem.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpFSFileSystem.FILE_TYPE.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpFSFileSystem.GetOpValues.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpFSFileSystem.PostOpValues.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpFSFileSystem.PutOpValues.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//class-use/HttpFSFileSystem.DeleteOpValues.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/client//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/fs/http/server//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/lang//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/server//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/hadoop//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/instrumentation//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/scheduler//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/service/security//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/servlet//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/util//package-use.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/org/apache/hadoop/lib/wsrs//package-use.html...>
Building index for all the packages and classes...
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/overview-tree.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/index-all.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/deprecated-list.html...>
Building index for all classes...
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/allclasses-frame.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/allclasses-noframe.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/index.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/overview-summary.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/help-doc.html...>
Generating <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/apidocs/stylesheet.css...>
90 warnings
[WARNING] Javadoc Warnings
[WARNING] org/apache/hadoop/fs/FileSystem.class(org/apache/hadoop/fs:FileSystem.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate': class file for org.apache.hadoop.classification.InterfaceAudience not found
[WARNING] org/apache/hadoop/fs/FileSystem.class(org/apache/hadoop/fs:FileSystem.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/Groups.class(org/apache/hadoop/security:Groups.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/UserGroupInformation.class(org/apache/hadoop/security:UserGroupInformation.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/UserGroupInformation.class(org/apache/hadoop/security:UserGroupInformation.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/security/UserGroupInformation.class(org/apache/hadoop/security:UserGroupInformation.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSDataInputStream.class(org/apache/hadoop/fs:FSDataInputStream.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSDataOutputStream.class(org/apache/hadoop/fs:FSDataOutputStream.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/Path.class(org/apache/hadoop/fs:Path.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java>:708: warning - @return tag has no arguments.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/server/Server.java>:262: warning - Tag @link: can't find serverStatusChange(Status, Status) in org.apache.hadoop.lib.server.Service
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/lib/lang/RunnableCallable.java>:76: warning - @return tag cannot be used in method with void return type.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSParams.java>:234: warning - @returns is an unknown tag.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:111: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:111: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#LISTSTATUS
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#INSTRUMENTATION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#OPEN
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues#LISTSTATUS
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - @param argument "override," is not a parameter name.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETPERMISSION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETREPLICATION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:379: warning - Tag @link: reference not found: HttpFSFileSystem.DeleteOpValues#DELETE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues#APPEND
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#RENAME
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETTIMES
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETOWNER
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETOWNER
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - @param argument "override," is not a parameter name.
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#CREATE
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETPERMISSION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETREPLICATION
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETTIMES
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues#SETTIMES
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues#APPEND
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues#APPEND
[WARNING] org/apache/hadoop/fs/UnresolvedLinkException.class(org/apache/hadoop/fs:UnresolvedLinkException.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.class(org/apache/hadoop/fs:MD5MD5CRC32FileChecksum.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/LocalDirAllocator.class(org/apache/hadoop/fs:LocalDirAllocator.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FileContext.class(org/apache/hadoop/fs:FileContext.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FileContext.class(org/apache/hadoop/fs:FileContext.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSOutputSummer.class(org/apache/hadoop/fs:FSOutputSummer.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSInputStream.class(org/apache/hadoop/fs:FSInputStream.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/FSInputChecker.class(org/apache/hadoop/fs:FSInputChecker.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/DU.class(org/apache/hadoop/fs:DU.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/DF.class(org/apache/hadoop/fs:DF.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] org/apache/hadoop/fs/AbstractFileSystem.class(org/apache/hadoop/fs:AbstractFileSystem.class): warning: Cannot find annotation method 'value()' in type 'org.apache.hadoop.classification.InterfaceAudience.LimitedPrivate'
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:241: warning - Tag @link: reference not found: HttpFSFileSystem.GetOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:573: warning - Tag @link: reference not found: HttpFSFileSystem.PostOpValues
[WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/java/org/apache/hadoop/fs/http/server/HttpFSServer.java>:450: warning - Tag @link: reference not found: HttpFSFileSystem.PutOpValues
[INFO] Building jar: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT-javadoc.jar>
[INFO]
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-httpfs ---
[INFO] Building jar: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT-sources.jar>
[INFO]
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ hadoop-hdfs-httpfs ---
[INFO]
[INFO] --- maven-assembly-plugin:2.2-beta-3:single (dist) @ hadoop-hdfs-httpfs ---
[INFO] Copying files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT>
[WARNING] Assembly file: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT> is not a regular file (it may be a directory). It cannot be attached to the project build for installation or deployment.
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (dist) @ hadoop-hdfs-httpfs ---
[INFO] Executing tasks
main:
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/downloads>
[get] Getting: http://archive.apache.org/dist/tomcat/tomcat-6/v6.0.32/bin/apache-tomcat-6.0.32.tar.gz
[get] To: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/downloads/tomcat.tar.gz>
..............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/tomcat.exp>
[delete] Deleting directory <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/tomcat.exp>
[delete] Deleting directory <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps>
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps>
[delete] Deleting: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/conf/server.xml>
[copy] Copying 1 file to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/conf>
[copy] Copying 1 file to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/conf>
[copy] Copying 2 files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps/ROOT>
[copy] Copying 166 files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps/webhdfs>
[copy] Copied 23 empty directories to 1 empty directory under <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-0.24.0-SNAPSHOT/share/hadoop/httpfs/tomcat/webapps/webhdfs>
[INFO] Executed tasks
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (tar) @ hadoop-hdfs-httpfs ---
[INFO] Executing tasks
main:
[INFO] Executed tasks
[INFO]
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ hadoop-hdfs-httpfs ---
[INFO]
[INFO] There are 1655 checkstyle errors.
[WARNING] Unable to locate Source XRef to link to - DISABLED
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ hadoop-hdfs-httpfs ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is true
[INFO] ****** FindBugsMojo executeFindbugs *******
[INFO] Temp File is <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/findbugsTemp.xml>
[INFO] Fork Value is true
[java] The following classes needed for analysis were missing:
[java] org.apache.hadoop.classification.InterfaceAudience$Private
[java] org.apache.hadoop.classification.InterfaceStability$Stable
[java] org.apache.hadoop.classification.InterfaceAudience
[java] org.apache.hadoop.classification.InterfaceStability$Evolving
[java] org.apache.hadoop.classification.InterfaceStability
[java] org.apache.hadoop.classification.InterfaceAudience$Public
[java] org.apache.hadoop.classification.InterfaceAudience$LimitedPrivate
[java] Warnings generated: 4
[java] Missing classes: 7
[INFO] xmlOutput is false
[INFO]
[INFO] ------------------------------------------------------------------------
[INFO] Building Apache Hadoop HDFS Project 0.24.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-hdfs-project ---
[INFO] Deleting <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/target>
[INFO]
[INFO] --- maven-antrun-plugin:1.6:run (create-testdirs) @ hadoop-hdfs-project ---
[INFO] Executing tasks
main:
[mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk/ws/trunk/hadoop-hdfs-project/target/test-dir>
[INFO] Executed tasks
[INFO]
[INFO] --- maven-javadoc-plugin:2.7:jar (module-javadocs) @ hadoop-hdfs-project ---
[INFO] Not executing Javadoc as the project is not a Java classpath-capable package
[INFO]
[INFO] --- maven-source-plugin:2.1.2:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-site-plugin:3.0:attach-descriptor (attach-descriptor) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- maven-checkstyle-plugin:2.6:checkstyle (default-cli) @ hadoop-hdfs-project ---
[INFO]
[INFO] --- findbugs-maven-plugin:2.3.2:findbugs (default-cli) @ hadoop-hdfs-project ---
[INFO] ****** FindBugsMojo execute *******
[INFO] canGenerate is false
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop HDFS ................................ SUCCESS [2:03:42.798s]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [31.199s]
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [0.032s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:04:14.439s
[INFO] Finished at: Mon Dec 12 13:38:09 UTC 2011
[INFO] Final Memory: 48M/808M
[INFO] ------------------------------------------------------------------------
+ /home/jenkins/tools/maven/latest/bin/mvn test -Dmaven.test.failure.ignore=true -Pclover -DcloverLicenseLocation=/home/jenkins/tools/clover/latest/lib/clover.license
Build step 'Execute shell' marked build as failure
Archiving artifacts
Recording test results
Publishing Javadoc
Recording fingerprints
Updating HDFS-2647
Updating HDFS-2651