You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-dev@hadoop.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2016/03/15 07:41:51 UTC
Hadoop-Mapreduce-trunk-Java8 - Build # 1155 - Still Failing
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/1155/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 8391 lines...]
Running org.apache.hadoop.mapred.TestJobInfo
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.555 sec - in org.apache.hadoop.mapred.TestJobInfo
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestFileOutputCommitter
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.724 sec - in org.apache.hadoop.mapred.TestFileOutputCommitter
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestJobAclsManager
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.924 sec - in org.apache.hadoop.mapred.TestJobAclsManager
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestJobEndNotifier
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.078 sec - in org.apache.hadoop.mapred.TestJobEndNotifier
Results :
Failed tests:
TestCLI.testGetJob:175 null
Tests run: 225, Failures: 1, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.305 s]
[INFO] Apache Hadoop MapReduce Core ...................... FAILURE [01:36 min]
[INFO] Apache Hadoop MapReduce Common .................... SKIPPED
[INFO] Apache Hadoop MapReduce Shuffle ................... SKIPPED
[INFO] Apache Hadoop MapReduce App ....................... SKIPPED
[INFO] Apache Hadoop MapReduce HistoryServer ............. SKIPPED
[INFO] Apache Hadoop MapReduce JobClient ................. SKIPPED
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:39 min
[INFO] Finished at: 2016-03-15T06:41:42+00:00
[INFO] Final Memory: 32M/200M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-core: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk-Java8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-mapreduce-client-core
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
###################################################################################
############################## FAILED TESTS (if any) ##############################
1 tests failed.
FAILED: org.apache.hadoop.mapreduce.tools.TestCLI.testGetJob
Error Message:
null
Stack Trace:
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at org.apache.hadoop.mapreduce.tools.TestCLI.testGetJob(TestCLI.java:175)
Hadoop-Mapreduce-trunk-Java8 - Build # 1160 - Still Failing
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/1160/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 8394 lines...]
Running org.apache.hadoop.mapred.TestJobInfo
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.537 sec - in org.apache.hadoop.mapred.TestJobInfo
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestJobAclsManager
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.913 sec - in org.apache.hadoop.mapred.TestJobAclsManager
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestSkipBadRecords
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.657 sec - in org.apache.hadoop.mapred.TestSkipBadRecords
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestJobQueueClient
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.086 sec - in org.apache.hadoop.mapred.TestJobQueueClient
Results :
Failed tests:
TestCLI.testGetJob:175 null
Tests run: 225, Failures: 1, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.246 s]
[INFO] Apache Hadoop MapReduce Core ...................... FAILURE [01:36 min]
[INFO] Apache Hadoop MapReduce Common .................... SKIPPED
[INFO] Apache Hadoop MapReduce Shuffle ................... SKIPPED
[INFO] Apache Hadoop MapReduce App ....................... SKIPPED
[INFO] Apache Hadoop MapReduce HistoryServer ............. SKIPPED
[INFO] Apache Hadoop MapReduce JobClient ................. SKIPPED
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:39 min
[INFO] Finished at: 2016-03-16T06:53:19+00:00
[INFO] Final Memory: 32M/185M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-core: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk-Java8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-mapreduce-client-core
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
###################################################################################
############################## FAILED TESTS (if any) ##############################
1 tests failed.
FAILED: org.apache.hadoop.mapreduce.tools.TestCLI.testGetJob
Error Message:
null
Stack Trace:
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at org.apache.hadoop.mapreduce.tools.TestCLI.testGetJob(TestCLI.java:175)
Hadoop-Mapreduce-trunk-Java8 - Build # 1159 - Still Failing
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/1159/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 8739 lines...]
Running org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobConf
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.893 sec - in org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobConf
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestTaskAttemptListenerImpl
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.996 sec - in org.apache.hadoop.mapred.TestTaskAttemptListenerImpl
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestTaskAttemptFinishingMonitor
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.393 sec - in org.apache.hadoop.mapred.TestTaskAttemptFinishingMonitor
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestLocalContainerLauncher
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.659 sec - in org.apache.hadoop.mapred.TestLocalContainerLauncher
Results :
Failed tests:
TestTaskAttempt.testMRAppHistoryForTAFailedInAssigned:177->testTaskAttemptAssignedKilledHistory:388 No Ta Started JH Event
Tests run: 340, Failures: 1, Errors: 0, Skipped: 0
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.123 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:38 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 24.115 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 4.376 s]
[INFO] Apache Hadoop MapReduce App ....................... FAILURE [11:04 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SKIPPED
[INFO] Apache Hadoop MapReduce JobClient ................. SKIPPED
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 13:14 min
[INFO] Finished at: 2016-03-16T03:46:56+00:00
[INFO] Final Memory: 35M/197M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-app: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk-Java8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-mapreduce-client-app
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
###################################################################################
############################## FAILED TESTS (if any) ##############################
1 tests failed.
FAILED: org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt.testMRAppHistoryForTAFailedInAssigned
Error Message:
No Ta Started JH Event
Stack Trace:
java.lang.AssertionError: No Ta Started JH Event
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt.testTaskAttemptAssignedKilledHistory(TestTaskAttempt.java:388)
at org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt.testMRAppHistoryForTAFailedInAssigned(TestTaskAttempt.java:177)
Hadoop-Mapreduce-trunk-Java8 - Build # 1158 - Still Failing
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/1158/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 9452 lines...]
Running org.apache.hadoop.mapred.pipes.TestPipeApplication
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.33 sec - in org.apache.hadoop.mapred.pipes.TestPipeApplication
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestJavaSerialization
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.979 sec - in org.apache.hadoop.mapred.TestJavaSerialization
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.131 sec - in org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.ipc.TestMRCJCSocketFactory
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.168 sec - in org.apache.hadoop.ipc.TestMRCJCSocketFactory
Results :
Failed tests:
TestNetworkedJob.testNetworkedJob:174 expected:<[[Tue Mar 15 23:47:46 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
Tests run: 527, Failures: 1, Errors: 0, Skipped: 11
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.314 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:36 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 24.474 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 4.452 s]
[INFO] Apache Hadoop MapReduce App ....................... SUCCESS [11:08 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [06:08 min]
[INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [ 01:40 h]
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 02:00 h
[INFO] Finished at: 2016-03-16T00:25:53+00:00
[INFO] Final Memory: 34M/158M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-jobclient: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk-Java8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
###################################################################################
############################## FAILED TESTS (if any) ##############################
1 tests failed.
FAILED: org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob
Error Message:
expected:<[[Tue Mar 15 23:47:46 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
Stack Trace:
org.junit.ComparisonFailure: expected:<[[Tue Mar 15 23:47:46 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
at org.junit.Assert.assertEquals(Assert.java:115)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob(TestNetworkedJob.java:174)
Hadoop-Mapreduce-trunk-Java8 - Build # 1157 - Still Failing
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/1157/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 21100 lines...]
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.621 sec - in org.apache.hadoop.util.TestMRCJCReflectionUtils
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.util.TestMRCJCRunJar
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.253 sec - in org.apache.hadoop.util.TestMRCJCRunJar
Results :
Failed tests:
TestNetworkedJob.testNetworkedJob:174 expected:<[[Tue Mar 15 18:03:50 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
TestMRJobs.testJobWithChangePriority:276 expected:<DEFAULT> but was:<HIGH>
Tests in error:
TestMRJobs.testContainerRollingLog:715 » test timed out after 120000 millisec...
TestMRJobs.testConfVerificationWithJobClientLocal:331->testConfVerification:355 » NoClassDefFound
TestMRJobs.testConfVerificationWithClassloader:310->testConfVerification:414 »
TestMRJobs.tearDown:177 » NoClassDefFound org/apache/hadoop/service/ServiceOpe...
Tests run: 528, Failures: 2, Errors: 4, Skipped: 11
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.553 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:38 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 24.311 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 5.563 s]
[INFO] Apache Hadoop MapReduce App ....................... SUCCESS [11:09 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [06:14 min]
[INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [ 01:57 h]
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 02:17 h
[INFO] Finished at: 2016-03-15T19:53:30+00:00
[INFO] Final Memory: 34M/151M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-jobclient: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk-Java8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
###################################################################################
############################## FAILED TESTS (if any) ##############################
6 tests failed.
FAILED: org.apache.hadoop.mapreduce.v2.TestMRJobs.testContainerRollingLog
Error Message:
test timed out after 120000 milliseconds
Stack Trace:
java.lang.Exception: test timed out after 120000 milliseconds
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1404)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1362)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.testContainerRollingLog(TestMRJobs.java:715)
FAILED: org.apache.hadoop.mapreduce.v2.TestMRJobs.testConfVerificationWithJobClientLocal
Error Message:
org/apache/hadoop/metrics/Updater
Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/metrics/Updater
at java.net.URLClassLoader$1.run(URLClassLoader.java:372)
at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:360)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClass(ClassLoader.java:760)
at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:455)
at java.net.URLClassLoader.access$100(URLClassLoader.java:73)
at java.net.URLClassLoader$1.run(URLClassLoader.java:367)
at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:360)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at org.apache.hadoop.mapred.LocalJobRunner.<init>(LocalJobRunner.java:771)
at org.apache.hadoop.mapred.LocalJobRunner.<init>(LocalJobRunner.java:764)
at org.apache.hadoop.mapred.LocalClientProtocolProvider.create(LocalClientProtocolProvider.java:42)
at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:111)
at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:98)
at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:91)
at org.apache.hadoop.mapred.JobClient.init(JobClient.java:475)
at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:465)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.testConfVerification(TestMRJobs.java:355)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.testConfVerificationWithJobClientLocal(TestMRJobs.java:331)
FAILED: org.apache.hadoop.mapreduce.v2.TestMRJobs.testJobWithChangePriority
Error Message:
expected:<DEFAULT> but was:<HIGH>
Stack Trace:
java.lang.AssertionError: expected:<DEFAULT> but was:<HIGH>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.testJobWithChangePriority(TestMRJobs.java:276)
FAILED: org.apache.hadoop.mapreduce.v2.TestMRJobs.testConfVerificationWithClassloader
Error Message:
test timed out after 300000 milliseconds
Stack Trace:
java.lang.Exception: test timed out after 300000 milliseconds
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1404)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1362)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.testConfVerification(TestMRJobs.java:414)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.testConfVerificationWithClassloader(TestMRJobs.java:310)
FAILED: org.apache.hadoop.mapreduce.v2.TestMRJobs.org.apache.hadoop.mapreduce.v2.TestMRJobs
Error Message:
org/apache/hadoop/service/ServiceOperations
Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/service/ServiceOperations
at java.net.URLClassLoader$1.run(URLClassLoader.java:372)
at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:360)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
at org.apache.hadoop.service.CompositeService.stop(CompositeService.java:157)
at org.apache.hadoop.service.CompositeService.serviceStop(CompositeService.java:131)
at org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
at org.apache.hadoop.mapreduce.v2.TestMRJobs.tearDown(TestMRJobs.java:177)
FAILED: org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob
Error Message:
expected:<[[Tue Mar 15 18:03:50 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
Stack Trace:
org.junit.ComparisonFailure: expected:<[[Tue Mar 15 18:03:50 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
at org.junit.Assert.assertEquals(Assert.java:115)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob(TestNetworkedJob.java:174)
Hadoop-Mapreduce-trunk-Java8 - Build # 1156 - Still Failing
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/1156/
###################################################################################
########################## LAST 60 LINES OF THE CONSOLE ###########################
[...truncated 9544 lines...]
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.826 sec - in org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0
Running org.apache.hadoop.mapred.TestReporter
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.111 sec - in org.apache.hadoop.mapred.TestReporter
Results :
Failed tests:
TestNetworkedJob.testNetworkedJob:174 expected:<[[Tue Mar 15 14:45:31 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
Tests in error:
TestMRCJCFileInputFormat.testLocality:55->newDFSCluster:50 » IO Unable to init...
TestMRCJCFileInputFormat.testMultiLevelInput:164 » IO Unable to initialize Web...
TestMRCJCFileInputFormat.testNumInputs:114->newDFSCluster:50 » FileNotFound we...
Tests run: 527, Failures: 1, Errors: 3, Skipped: 11
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 3.111 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:55 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 27.798 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 5.015 s]
[INFO] Apache Hadoop MapReduce App ....................... SUCCESS [11:26 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [06:34 min]
[INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [ 01:44 h]
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 02:05 h
[INFO] Finished at: 2016-03-15T15:23:55+00:00
[INFO] Final Memory: 34M/173M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-jobclient: There are test failures.
[ERROR]
[ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk-Java8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any
###################################################################################
############################## FAILED TESTS (if any) ##############################
4 tests failed.
FAILED: org.apache.hadoop.mapred.TestMRCJCFileInputFormat.testLocality
Error Message:
Unable to initialize WebAppContext
Stack Trace:
java.io.IOException: Unable to initialize WebAppContext
at org.mortbay.jetty.servlet.Holder.doStart(Holder.java:79)
at org.mortbay.jetty.servlet.FilterHolder.doStart(FilterHolder.java:81)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.jetty.servlet.ServletHandler.initialize(ServletHandler.java:713)
at org.mortbay.jetty.servlet.Context.startContext(Context.java:140)
at org.mortbay.jetty.webapp.WebAppContext.startContext(WebAppContext.java:1282)
at org.mortbay.jetty.handler.ContextHandler.doStart(ContextHandler.java:518)
at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:499)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
at org.mortbay.jetty.Server.doStart(Server.java:224)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:863)
at org.apache.hadoop.hdfs.server.namenode.NameNodeHttpServer.start(NameNodeHttpServer.java:158)
at org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:835)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:693)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:897)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:876)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1593)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:1247)
at org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1016)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:891)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:823)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:482)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441)
at org.apache.hadoop.mapred.TestMRCJCFileInputFormat.newDFSCluster(TestMRCJCFileInputFormat.java:50)
at org.apache.hadoop.mapred.TestMRCJCFileInputFormat.testLocality(TestMRCJCFileInputFormat.java:55)
FAILED: org.apache.hadoop.mapred.TestMRCJCFileInputFormat.testMultiLevelInput
Error Message:
Unable to initialize WebAppContext
Stack Trace:
java.io.IOException: Unable to initialize WebAppContext
at java.util.zip.ZipFile.getZipEntry(ZipFile.java:567)
at java.util.zip.ZipFile.access$900(ZipFile.java:61)
at java.util.zip.ZipFile$ZipEntryIterator.next(ZipFile.java:525)
at java.util.zip.ZipFile$ZipEntryIterator.nextElement(ZipFile.java:500)
at java.util.zip.ZipFile$ZipEntryIterator.nextElement(ZipFile.java:481)
at java.util.jar.JarFile$JarEntryIterator.next(JarFile.java:257)
at java.util.jar.JarFile$JarEntryIterator.nextElement(JarFile.java:266)
at java.util.jar.JarFile$JarEntryIterator.nextElement(JarFile.java:247)
at org.mortbay.resource.JarFileResource.exists(JarFileResource.java:159)
at org.mortbay.jetty.webapp.WebAppContext.resolveWebApp(WebAppContext.java:937)
at org.mortbay.jetty.webapp.WebAppContext.getWebInf(WebAppContext.java:832)
at org.mortbay.jetty.webapp.WebInfConfiguration.configureClassLoader(WebInfConfiguration.java:62)
at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:489)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
at org.mortbay.jetty.Server.doStart(Server.java:224)
at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:863)
at org.apache.hadoop.hdfs.server.namenode.NameNodeHttpServer.start(NameNodeHttpServer.java:158)
at org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:835)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:693)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:897)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:876)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1593)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:1247)
at org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1016)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:891)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:823)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:482)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441)
at org.apache.hadoop.mapred.TestMRCJCFileInputFormat.testMultiLevelInput(TestMRCJCFileInputFormat.java:164)
FAILED: org.apache.hadoop.mapred.TestMRCJCFileInputFormat.testNumInputs
Error Message:
webapps/hdfs not found in CLASSPATH
Stack Trace:
java.io.FileNotFoundException: webapps/hdfs not found in CLASSPATH
at org.apache.hadoop.http.HttpServer2.getWebAppsPath(HttpServer2.java:794)
at org.apache.hadoop.http.HttpServer2.<init>(HttpServer2.java:347)
at org.apache.hadoop.http.HttpServer2.<init>(HttpServer2.java:115)
at org.apache.hadoop.http.HttpServer2$Builder.build(HttpServer2.java:297)
at org.apache.hadoop.hdfs.server.namenode.NameNodeHttpServer.start(NameNodeHttpServer.java:142)
at org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:835)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:693)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:897)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:876)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1593)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:1247)
at org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1016)
at org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:891)
at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:823)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:482)
at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441)
at org.apache.hadoop.mapred.TestMRCJCFileInputFormat.newDFSCluster(TestMRCJCFileInputFormat.java:50)
at org.apache.hadoop.mapred.TestMRCJCFileInputFormat.testNumInputs(TestMRCJCFileInputFormat.java:114)
FAILED: org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob
Error Message:
expected:<[[Tue Mar 15 14:45:31 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
Stack Trace:
org.junit.ComparisonFailure: expected:<[[Tue Mar 15 14:45:31 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]>
at org.junit.Assert.assertEquals(Assert.java:115)
at org.junit.Assert.assertEquals(Assert.java:144)
at org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob(TestNetworkedJob.java:174)