You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@hbase.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2015/12/02 20:07:03 UTC
Build failed in Jenkins: HBase-1.3-IT » IntegrationTestImportTsv,latest1.8,Hadoop #351
See <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=latest1.8,label=Hadoop/351/changes>
Changes:
[stack] HBASE-14819 hbase-it tests failing with OOME; permgen
[stack] HBASE-14908 TestRowCounter flakey especially on branch-1
------------------------------------------
[...truncated 2338 lines...]
bool UseLargePagesIndividualAllocation = false {pd product}
bool UseLinuxPosixThreadCPUClocks = true {product}
bool UseLockedTracing = false {product}
bool UseLoopCounter = true {product}
bool UseLoopInvariantCodeMotion = true {C1 product}
bool UseLoopPredicate = true {C2 product}
bool UseMathExactIntrinsics = true {C2 product}
bool UseMaximumCompactionOnSystemGC = true {product}
bool UseMembar = false {pd product}
bool UseMultiplyToLenIntrinsic = true {C2 product}
bool UseNUMA = false {product}
bool UseNUMAInterleaving = false {product}
bool UseNewLongLShift = false {ARCH product}
bool UseOSErrorReporting = false {pd product}
bool UseOldInlining = true {C2 product}
bool UseOnStackReplacement = true {pd product}
bool UseOnlyInlinedBimorphic = true {C2 product}
bool UseOprofile = false {product}
bool UseOptoBiasInlining = true {C2 product}
bool UsePSAdaptiveSurvivorSizePolicy = true {product}
bool UseParNewGC = false {product}
bool UseParallelGC := true {product}
bool UseParallelOldGC = true {product}
bool UsePerfData = true {product}
bool UsePopCountInstruction = true {product}
bool UseRDPCForConstantTableBase = false {C2 product}
bool UseRTMDeopt = false {ARCH product}
bool UseRTMLocking = false {ARCH product}
bool UseSHA = false {product}
bool UseSHA1Intrinsics = false {product}
bool UseSHA256Intrinsics = false {product}
bool UseSHA512Intrinsics = false {product}
bool UseSHM = false {product}
intx UseSSE = 4 {product}
bool UseSSE42Intrinsics = true {product}
bool UseSerialGC = false {product}
bool UseSharedSpaces = false {product}
bool UseSignalChaining = true {product}
bool UseStoreImmI16 = false {ARCH product}
bool UseStringDeduplication = false {product}
bool UseSuperWord = true {C2 product}
bool UseTLAB = true {pd product}
bool UseThreadPriorities = true {pd product}
bool UseTransparentHugePages = false {product}
bool UseTypeProfile = true {product}
bool UseTypeSpeculation = true {C2 product}
bool UseUnalignedLoadStores = true {ARCH product}
bool UseVMInterruptibleIO = false {product}
bool UseXMMForArrayCopy = true {product}
bool UseXmmI2D = false {ARCH product}
bool UseXmmI2F = false {ARCH product}
bool UseXmmLoadAndClearUpper = true {ARCH product}
bool UseXmmRegToRegMoveAll = true {ARCH product}
bool VMThreadHintNoPreempt = false {product}
intx VMThreadPriority = -1 {product}
intx VMThreadStackSize = 1024 {pd product}
intx ValueMapInitialSize = 11 {C1 product}
intx ValueMapMaxLoopSize = 8 {C1 product}
intx ValueSearchLimit = 1000 {C2 product}
bool VerifyMergedCPBytecodes = true {product}
bool VerifySharedSpaces = false {product}
intx WorkAroundNPTLTimedWaitHang = 1 {product}
uintx YoungGenerationSizeIncrement = 20 {product}
uintx YoungGenerationSizeSupplement = 80 {product}
uintx YoungGenerationSizeSupplementDecay = 8 {product}
uintx YoungPLABSize = 4096 {product}
bool ZeroTLAB = false {product}
intx hashCode = 5 {product}
Running org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv
SUREFIRE-859: [GC (Allocation Failure) 193536K->67782K(741888K), 0.1388339 secs]
SUREFIRE-859: [GC (Metadata GC Threshold) 102209K->43070K(935424K), 0.0161945 secs]
SUREFIRE-859: [Full GC (Metadata GC Threshold) 43070K->7675K(707072K), 0.0479951 secs]
SUREFIRE-859: [GC (Metadata GC Threshold) 295135K->88276K(707072K), 0.1598161 secs]
SUREFIRE-859: [Full GC (Metadata GC Threshold) 88276K->85369K(936448K), 0.2708152 secs]
SUREFIRE-859: [GC (Allocation Failure) 472441K->123626K(1093120K), 0.0580246 secs]
SUREFIRE-859: [GC (Metadata GC Threshold) 202956K->109853K(1251840K), 0.0417314 secs]
SUREFIRE-859: [Full GC (Metadata GC Threshold) 109853K->91778K(1384960K), 0.2045113 secs]
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 1,262.298 sec <<< FAILURE! - in org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv
testGenerateAndLoad(org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv) Time elapsed: 1,214.896 sec <<< ERROR!
java.net.ConnectException: Call From asf903.gq1.ygridcore.net/67.195.81.147 to asf903.gq1.ygridcore.net:0 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:529)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:493)
at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:606)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:700)
at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:367)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1463)
at org.apache.hadoop.ipc.Client.call(Client.java:1382)
at org.apache.hadoop.ipc.Client.call(Client.java:1364)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
at com.sun.proxy.$Proxy51.getNewApplication(Unknown Source)
at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getNewApplication(ApplicationClientProtocolPBClientImpl.java:193)
at sun.reflect.GeneratedMethodAccessor35.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
at com.sun.proxy.$Proxy52.getNewApplication(Unknown Source)
at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getNewApplication(YarnClientImpl.java:191)
at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.createApplication(YarnClientImpl.java:199)
at org.apache.hadoop.mapred.ResourceMgrDelegate.getNewJobID(ResourceMgrDelegate.java:179)
at org.apache.hadoop.mapred.YARNRunner.getNewJobID(YARNRunner.java:230)
at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:357)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1303)
at org.apache.hadoop.hbase.mapreduce.ImportTsv.run(ImportTsv.java:775)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.hbase.mapreduce.TestImportTsv.doMROnTableTest(TestImportTsv.java:428)
at org.apache.hadoop.hbase.mapreduce.TestImportTsv.doMROnTableTest(TestImportTsv.java:381)
at org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv.testGenerateAndLoad(IntegrationTestImportTsv.java:203)
Results :
Tests in error:
IntegrationTestImportTsv.testGenerateAndLoad:203 » Connect Call From asf903.gq...
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0
[WARNING] File encoding has not been set, using platform encoding ISO-8859-1, i.e. build is platform dependent! The file encoding for reports output files should be provided by the POM property ${project.reporting.outputEncoding}.
[INFO]
[INFO] --- maven-failsafe-plugin:2.18.1:verify (verify) @ hbase-it ---
[INFO] Failsafe report directory: <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=latest1.8,label=Hadoop/ws/hbase-it/target/failsafe-reports>
[WARNING] File encoding has not been set, using platform encoding ISO-8859-1, i.e. build is platform dependent! The file encoding for reports output files should be provided by the POM property ${project.reporting.outputEncoding}.
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache HBase ...................................... SUCCESS [5.654s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [0.988s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [0.347s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [1.597s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [16.604s]
[INFO] Apache HBase - Common ............................. SUCCESS [11.552s]
[INFO] Apache HBase - Procedure .......................... SUCCESS [1.803s]
[INFO] Apache HBase - Client ............................. SUCCESS [7.503s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [0.495s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [1.447s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [1.507s]
[INFO] Apache HBase - Server ............................. SUCCESS [43.563s]
[INFO] Apache HBase - Testing Util ....................... SUCCESS [4.934s]
[INFO] Apache HBase - Thrift ............................. SUCCESS [9.631s]
[INFO] Apache HBase - Rest ............................... SUCCESS [6.167s]
[INFO] Apache HBase - Shell .............................. SUCCESS [2.053s]
[INFO] Apache HBase - Integration Tests .................. FAILURE [21:07.383s]
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 23:05.389s
[INFO] Finished at: Wed Dec 02 19:07:03 UTC 2015
[INFO] Final Memory: 183M/2883M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.18.1:verify (verify) on project hbase-it: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=latest1.8,label=Hadoop/ws/hbase-it/target/failsafe-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hbase-it
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script : # Post-build task script. TODO: Check this in and have all builds reference check-in.
pwd && ls
# NOTE!!!! The below code has been copied and pasted from ./dev-tools/run-test.sh
# Do not change here without syncing there and vice-versa.
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo "Suspicious java process found - waiting 30s to see if there are just slow to stop"
sleep 30
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo " {color:red}There appear to be $ZOMBIE_TESTS_COUNT zombie tests{color}, they should have been killed by surefire but survived"
jps -v | grep surefirebooter | grep -e '-Dhbase.test'
jps -v | grep surefirebooter | grep -e '-Dhbase.test' | cut -d ' ' -f 1 | xargs -n 1 jstack
# Exit with error
exit 1
else
echo "We're ok: there is no zombie test, but some tests took some time to stop"
fi
else
echo "We're ok: there is no zombie test"
fi
[Hadoop] $ /bin/bash -xe /tmp/hudson4550326020892243906.sh
+ pwd
<https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=latest1.8,label=Hadoop/ws/>
+ ls
bin
CHANGES.txt
conf
dev-support
hbase-annotations
hbase-assembly
hbase-checkstyle
hbase-client
hbase-common
hbase-examples
hbase-external-blockcache
hbase-hadoop2-compat
hbase-hadoop-compat
hbase-it
hbase-native-client
hbase-prefix-tree
hbase-procedure
hbase-protocol
hbase-resource-bundle
hbase-rest
hbase-server
hbase-shaded
hbase-shell
hbase-spark
hbase-testing-util
hbase-thrift
LICENSE.txt
NOTICE.txt
pom.xml
README.txt
src
target
++ jps -v
++ grep surefirebooter
++ grep -e -Dhbase.test
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Jenkins build is back to normal : HBase-1.3-IT » IntegrationTestImportTsv,latest1.8,Hadoop #352
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=latest1.8,label=Hadoop/352/changes>