You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@hbase.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2017/08/17 11:06:57 UTC
Build failed in Jenkins: HBase-1.3-IT » IntegrationTestImportTsv,JDK 1.8 (latest),Hadoop #172
See <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=JDK%201.8%20(latest),label=Hadoop/172/display/redirect?page=changes>
Changes:
[chia7712] HBASE-17617 Backport HBASE-16731 (Inconsistent results from the Get/Scan
------------------------------------------
[...truncated 294.26 KB...]
bool UseMaximumCompactionOnSystemGC = true {product}
bool UseMembar = false {pd product}
bool UseMontgomeryMultiplyIntrinsic = false {C2 product}
bool UseMontgomerySquareIntrinsic = false {C2 product}
bool UseMulAddIntrinsic = false {C2 product}
bool UseMultiplyToLenIntrinsic = true {C2 product}
bool UseNUMA = false {product}
bool UseNUMAInterleaving = false {product}
bool UseNewLongLShift = false {ARCH product}
bool UseOSErrorReporting = false {pd product}
bool UseOldInlining = true {C2 product}
bool UseOnStackReplacement = true {pd product}
bool UseOnlyInlinedBimorphic = true {C2 product}
bool UseOprofile = false {product}
bool UseOptoBiasInlining = true {C2 product}
bool UsePSAdaptiveSurvivorSizePolicy = true {product}
bool UseParNewGC = false {product}
bool UseParallelGC := true {product}
bool UseParallelOldGC = true {product}
bool UsePerfData = true {product}
bool UsePopCountInstruction = true {product}
bool UseRDPCForConstantTableBase = false {C2 product}
bool UseRTMDeopt = false {ARCH product}
bool UseRTMLocking = false {ARCH product}
bool UseSHA = false {product}
bool UseSHA1Intrinsics = false {product}
bool UseSHA256Intrinsics = false {product}
bool UseSHA512Intrinsics = false {product}
bool UseSHM = false {product}
intx UseSSE = 4 {product}
bool UseSSE42Intrinsics = true {product}
bool UseSerialGC = false {product}
bool UseSharedSpaces = false {product}
bool UseSignalChaining = true {product}
bool UseSquareToLenIntrinsic = false {C2 product}
bool UseStoreImmI16 = false {ARCH product}
bool UseStringDeduplication = false {product}
bool UseSuperWord = true {C2 product}
bool UseTLAB = true {pd product}
bool UseThreadPriorities = true {pd product}
bool UseTransparentHugePages = false {product}
bool UseTypeProfile = true {product}
bool UseTypeSpeculation = true {C2 product}
bool UseUnalignedLoadStores = true {ARCH product}
bool UseVMInterruptibleIO = false {product}
bool UseXMMForArrayCopy = true {product}
bool UseXmmI2D = false {ARCH product}
bool UseXmmI2F = false {ARCH product}
bool UseXmmLoadAndClearUpper = true {ARCH product}
bool UseXmmRegToRegMoveAll = true {ARCH product}
bool VMThreadHintNoPreempt = false {product}
intx VMThreadPriority = -1 {product}
intx VMThreadStackSize = 1024 {pd product}
intx ValueMapInitialSize = 11 {C1 product}
intx ValueMapMaxLoopSize = 8 {C1 product}
intx ValueSearchLimit = 1000 {C2 product}
bool VerifyMergedCPBytecodes = true {product}
bool VerifySharedSpaces = false {product}
intx WorkAroundNPTLTimedWaitHang = 1 {product}
uintx YoungGenerationSizeIncrement = 20 {product}
uintx YoungGenerationSizeSupplement = 80 {product}
uintx YoungGenerationSizeSupplementDecay = 8 {product}
uintx YoungPLABSize = 4096 {product}
bool ZeroTLAB = false {product}
intx hashCode = 5 {product}
Running org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv
SUREFIRE-859: [GC (Allocation Failure) 193536K->68348K(741888K), 0.1579545 secs]
SUREFIRE-859: [GC (Metadata GC Threshold) 101907K->44724K(935424K), 0.0490487 secs]
SUREFIRE-859: [Full GC (Metadata GC Threshold) 44724K->8665K(706048K), 0.1271550 secs]
SUREFIRE-859: [GC (Metadata GC Threshold) 299273K->89009K(706048K), 0.1614518 secs]
SUREFIRE-859: [Full GC (Metadata GC Threshold) 89009K->85945K(891392K), 0.2000270 secs]
SUREFIRE-859: [GC (Metadata GC Threshold) 454787K->125448K(1116672K), 0.1703650 secs]
SUREFIRE-859: [Full GC (Metadata GC Threshold) 125448K->112257K(1222656K), 0.3267988 secs]
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 1,263.046 sec <<< FAILURE! - in org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv
testGenerateAndLoad(org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv) Time elapsed: 1,213.19 sec <<< ERROR!
java.net.ConnectException: Call From asf911.gq1.ygridcore.net/67.195.81.155 to asf911.gq1.ygridcore.net:0 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:529)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:493)
at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:606)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:700)
at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:367)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1463)
at org.apache.hadoop.ipc.Client.call(Client.java:1382)
at org.apache.hadoop.ipc.Client.call(Client.java:1364)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
at com.sun.proxy.$Proxy51.getNewApplication(Unknown Source)
at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getNewApplication(ApplicationClientProtocolPBClientImpl.java:193)
at sun.reflect.GeneratedMethodAccessor39.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
at com.sun.proxy.$Proxy52.getNewApplication(Unknown Source)
at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getNewApplication(YarnClientImpl.java:191)
at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.createApplication(YarnClientImpl.java:199)
at org.apache.hadoop.mapred.ResourceMgrDelegate.getNewJobID(ResourceMgrDelegate.java:179)
at org.apache.hadoop.mapred.YARNRunner.getNewJobID(YARNRunner.java:230)
at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:357)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1303)
at org.apache.hadoop.hbase.mapreduce.ImportTsv.run(ImportTsv.java:782)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.hbase.mapreduce.TestImportTsv.doMROnTableTest(TestImportTsv.java:440)
at org.apache.hadoop.hbase.mapreduce.TestImportTsv.doMROnTableTest(TestImportTsv.java:393)
at org.apache.hadoop.hbase.mapreduce.IntegrationTestImportTsv.testGenerateAndLoad(IntegrationTestImportTsv.java:203)
Results :
Tests in error:
IntegrationTestImportTsv.testGenerateAndLoad:203 » Connect Call From asf911.gq...
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0
[WARNING] File encoding has not been set, using platform encoding ISO-8859-1, i.e. build is platform dependent! The file encoding for reports output files should be provided by the POM property ${project.reporting.outputEncoding}.
[INFO]
[INFO] --- maven-failsafe-plugin:2.18.1:verify (verify) @ hbase-it ---
[INFO] Failsafe report directory: <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=JDK%201.8%20(latest),label=Hadoop/ws/hbase-it/target/failsafe-reports>
[WARNING] File encoding has not been set, using platform encoding ISO-8859-1, i.e. build is platform dependent! The file encoding for reports output files should be provided by the POM property ${project.reporting.outputEncoding}.
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache HBase ...................................... SUCCESS [5.926s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [0.798s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [0.380s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [1.148s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [14.188s]
[INFO] Apache HBase - Common ............................. SUCCESS [9.306s]
[INFO] Apache HBase - Procedure .......................... SUCCESS [1.655s]
[INFO] Apache HBase - Client ............................. SUCCESS [6.832s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [0.496s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [1.696s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [1.554s]
[INFO] Apache HBase - Server ............................. SUCCESS [46.071s]
[INFO] Apache HBase - Testing Util ....................... SUCCESS [4.015s]
[INFO] Apache HBase - Thrift ............................. SUCCESS [9.174s]
[INFO] Apache HBase - Rest ............................... SUCCESS [3.387s]
[INFO] Apache HBase - Shell .............................. SUCCESS [1.169s]
[INFO] Apache HBase - Integration Tests .................. FAILURE [21:10.662s]
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] Apache HBase - Archetypes ......................... SKIPPED
[INFO] Apache HBase - Exemplar for hbase-client archetype SKIPPED
[INFO] Apache HBase - Exemplar for hbase-shaded-client archetype SKIPPED
[INFO] Apache HBase - Archetype builder .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 23:00.975s
[INFO] Finished at: Thu Aug 17 11:06:51 UTC 2017
[INFO] Final Memory: 188M/3480M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.18.1:verify (verify) on project hbase-it: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=JDK%201.8%20(latest),label=Hadoop/ws/hbase-it/target/failsafe-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hbase-it
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script : # Post-build task script. TODO: Check this in and have all builds reference check-in.
pwd && ls
# NOTE!!!! The below code has been copied and pasted from ./dev-tools/run-test.sh
# Do not change here without syncing there and vice-versa.
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo "Suspicious java process found - waiting 30s to see if there are just slow to stop"
sleep 30
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo " {color:red}There appear to be $ZOMBIE_TESTS_COUNT zombie tests{color}, they should have been killed by surefire but survived"
jps -v | grep surefirebooter | grep -e '-Dhbase.test'
jps -v | grep surefirebooter | grep -e '-Dhbase.test' | cut -d ' ' -f 1 | xargs -n 1 jstack
# Exit with error
exit 1
else
echo "We're ok: there is no zombie test, but some tests took some time to stop"
fi
else
echo "We're ok: there is no zombie test"
fi
[25f87dfd] $ /bin/bash -xe /tmp/jenkins9133349362795617559.sh
+ pwd
<https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=JDK%201.8%20(latest),label=Hadoop/ws/>
+ ls
bin
CHANGES.txt
conf
dev-support
hbase-annotations
hbase-archetypes
hbase-assembly
hbase-checkstyle
hbase-client
hbase-common
hbase-examples
hbase-external-blockcache
hbase-hadoop2-compat
hbase-hadoop-compat
hbase-it
hbase-native-client
hbase-prefix-tree
hbase-procedure
hbase-protocol
hbase-resource-bundle
hbase-rest
hbase-server
hbase-shaded
hbase-shell
hbase-testing-util
hbase-thrift
LICENSE.txt
NOTICE.txt
pom.xml
README.txt
src
target
++ jps -v
++ grep surefirebooter
++ grep -e -Dhbase.test
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Jenkins build is back to normal : HBase-1.3-IT » IntegrationTestImportTsv,JDK 1.8 (latest),Hadoop #173
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-1.3-IT/it.test=IntegrationTestImportTsv,jdk=JDK%201.8%20(latest),label=Hadoop/173/display/redirect>