You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@hbase.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2014/02/26 06:48:13 UTC

Build failed in Jenkins: HBase-TRUNK #4955

See <https://builds.apache.org/job/HBase-TRUNK/4955/changes>

Changes:

[mbertozzi] HBASE-10608 Acquire the FS Delegation Token for Secure ExportSnapshot

[jmhsieh] HBASE-10436 restore regionserver lists removed from hbase 0.96.0 jmx

[stack] HBASE-10604 Fix parseArgs javadoc (Jean-Marc Spaggiari)

------------------------------------------
[...truncated 4660 lines...]
Running org.apache.hadoop.hbase.client.TestHTableMultiplexer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.785 sec
Running org.apache.hadoop.hbase.client.TestMetaScanner
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 69.254 sec
Running org.apache.hadoop.hbase.client.TestHTableUtil
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.401 sec
Running org.apache.hadoop.hbase.client.TestHCM
Tests run: 19, Failures: 0, Errors: 0, Skipped: 3, Time elapsed: 755.536 sec
Running org.apache.hadoop.hbase.client.TestTableSnapshotScanner
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 134.312 sec
Running org.apache.hadoop.hbase.client.TestSnapshotFromClient
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 71.155 sec
Running org.apache.hadoop.hbase.client.TestHTablePool$TestHTableReusablePool
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.165 sec
Running org.apache.hadoop.hbase.client.TestHTablePool$TestHTableThreadLocalPool
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 51.668 sec
Running org.apache.hadoop.hbase.client.TestSnapshotMetadata
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 75.017 sec
Running org.apache.hadoop.hbase.client.TestMultipleTimestamps
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 47.944 sec
Running org.apache.hadoop.hbase.client.TestRestoreSnapshotFromClient
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 585.106 sec
Running org.apache.hadoop.hbase.client.TestClientScannerRPCTimeout
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.929 sec
Running org.apache.hadoop.hbase.client.TestSnapshotCloneIndependence
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 137.34 sec
Running org.apache.hadoop.hbase.client.replication.TestReplicationAdmin
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.589 sec
Running org.apache.hadoop.hbase.client.TestClientOperationInterrupt
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 76.189 sec
Running org.apache.hadoop.hbase.client.TestMultiParallel
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 101.337 sec
Running org.apache.hadoop.hbase.client.TestCloneSnapshotFromClient
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 384.2 sec
Running org.apache.hadoop.hbase.client.TestFromClientSideNoCodec
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.958 sec
Running org.apache.hadoop.hbase.constraint.TestConstraint
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.874 sec
Running org.apache.hadoop.hbase.client.TestAdmin
Tests run: 44, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 534.389 sec
Running org.apache.hadoop.hbase.rest.TestGzipFilter
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.509 sec
Running org.apache.hadoop.hbase.rest.TestGetAndPutResource
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 106.778 sec
Running org.apache.hadoop.hbase.rest.TestTableScan
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.464 sec
Running org.apache.hadoop.hbase.rest.TestMultiRowResource
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.555 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithFilters
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.449 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithLabels
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.989 sec
Running org.apache.hadoop.hbase.rest.TestResourceFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.718 sec
Running org.apache.hadoop.hbase.rest.TestDeleteRow
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 46.993 sec
Running org.apache.hadoop.hbase.rest.TestVersionResource
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.603 sec
Running org.apache.hadoop.hbase.rest.client.TestRemoteTable
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 67.892 sec
Running org.apache.hadoop.hbase.rest.TestScannerResource
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.873 sec
Running org.apache.hadoop.hbase.rest.TestStatusResource
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.204 sec
Running org.apache.hadoop.hbase.rest.TestTableResource
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.977 sec
Running org.apache.hadoop.hbase.TestIOFencing
Tests run: 2, Failures: 0, Errors: 0, Skipped: 2, Time elapsed: 0.015 sec
Running org.apache.hadoop.hbase.rest.TestSchemaResource
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.307 sec
Running org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.452 sec
Running org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 104.936 sec
Running org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.538 sec
Running org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 40, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 183.97 sec
Running org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.798 sec
Running org.apache.hadoop.hbase.io.hfile.slab.TestSlabCache
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.701 sec
Running org.apache.hadoop.hbase.io.hfile.slab.TestSingleSizeCache
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.018 sec
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.719 sec
Running org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 62.272 sec
Running org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.61 sec
Running org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 236.233 sec
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Tests run: 28, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 101.546 sec
Running org.apache.hadoop.hbase.io.TestFileLink
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.197 sec
Running org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Tests run: 36, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 110.383 sec
Running org.apache.hadoop.hbase.migration.TestUpgradeTo96
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.471 sec
Running org.apache.hadoop.hbase.TestClusterBootOrder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 40.831 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRSCompressed
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 70.913 sec
Running org.apache.hadoop.hbase.migration.TestNamespaceUpgrade
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 83.547 sec
Running org.apache.hadoop.hbase.replication.TestReplicationDisableInactivePeer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 63.88 sec
Running org.apache.hadoop.hbase.replication.TestPerTableCFReplication
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 72.96 sec
Running org.apache.hadoop.hbase.replication.TestReplicationChangingPeerRegionservers
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.665 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 90.108 sec
Running org.apache.hadoop.hbase.replication.TestMultiSlaveReplication
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 69.988 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSyncUpTool
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 137.187 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSourceManager
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.797 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSink
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.292 sec
Running org.apache.hadoop.hbase.replication.TestReplicationTrackerZKImpl
Tests run: 4, Failures: 0, Errors: 0, Skipped: 2, Time elapsed: 3.343 sec
Running org.apache.hadoop.hbase.replication.TestReplicationStateZKImpl
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.006 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationHLogReaderManager
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 218.707 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSource
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.281 sec
Running org.apache.hadoop.hbase.replication.TestMasterReplication
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 155.704 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSmallTests
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 102.933 sec
Running org.apache.hadoop.hbase.replication.TestReplicationWithTags
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.974 sec
Running org.apache.hadoop.hbase.ipc.TestDelayedRpc
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.726 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillSlaveRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 68.296 sec
Running org.apache.hadoop.hbase.ipc.TestHBaseClient
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.363 sec
Running org.apache.hadoop.hbase.ipc.TestProtoBufRpc
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.656 sec
Running org.apache.hadoop.hbase.TestHBaseTestingUtility
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 84.481 sec
Running org.apache.hadoop.hbase.TestZooKeeper
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 133.354 sec

Results :

Tests in error: 
  org.apache.hadoop.hbase.master.TestMasterMetrics: Shutting down

Tests run: 1836, Failures: 0, Errors: 1, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [12.280s]
[INFO] HBase - Common .................................... SUCCESS [27.250s]
[INFO] HBase - Protocol .................................. SUCCESS [9.284s]
[INFO] HBase - Client .................................... SUCCESS [38.752s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.423s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [2.939s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.903s]
[INFO] HBase - Server .................................... FAILURE [3:02:53.288s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 3:04:33.914s
[INFO] Finished at: Wed Feb 26 05:45:38 UTC 2014
[INFO] Final Memory: 111M/823M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: Failure or timeout -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson4583443606918032540.sh
++ wc -l
++ grep surefirebooter
++ jps
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10436
Updating HBASE-10608
Updating HBASE-10604

Jenkins build is back to normal : HBase-TRUNK #4975

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4975/changes>


Build failed in Jenkins: HBase-TRUNK #4974

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4974/changes>

Changes:

[enis] HBASE-10635 thrift#TestThriftServer fails due to TTL validity check (Addendum patch to TestThriftServerCmdLine)

[tedyu] HBASE-10609 Remove filterKeyValue(Cell ignored) from FilterBase

------------------------------------------
[...truncated 4654 lines...]
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRSCompressed
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 67.71 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillSlaveRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 66.872 sec
Running org.apache.hadoop.hbase.replication.TestMasterReplication
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 91.731 sec
Running org.apache.hadoop.hbase.replication.TestReplicationStateZKImpl
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.135 sec
Running org.apache.hadoop.hbase.replication.TestReplicationChangingPeerRegionservers
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.706 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSyncUpTool
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 89.432 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSourceManager
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.825 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSink
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.861 sec
Running org.apache.hadoop.hbase.replication.TestReplicationWithTags
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.234 sec
Running org.apache.hadoop.hbase.replication.TestMultiSlaveReplication
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 41.719 sec
Running org.apache.hadoop.hbase.replication.TestReplicationDisableInactivePeer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.463 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSource
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.952 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 68.805 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationHLogReaderManager
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 274.665 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSmallTests
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 84.447 sec
Running org.apache.hadoop.hbase.replication.TestPerTableCFReplication
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 59.134 sec
Running org.apache.hadoop.hbase.TestLocalHBaseCluster
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.713 sec
Running org.apache.hadoop.hbase.fs.TestBlockReorder
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.916 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.985 sec
Running org.apache.hadoop.hbase.TestAcidGuarantees
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 111.505 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterObserver
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 32.361 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverInterface
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.613 sec
Running org.apache.hadoop.hbase.coprocessor.TestCoprocessorEndpoint
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.6 sec
Running org.apache.hadoop.hbase.coprocessor.TestWALObserver
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.657 sec
Running org.apache.hadoop.hbase.coprocessor.TestHTableWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.268 sec
Running org.apache.hadoop.hbase.coprocessor.TestBigDecimalColumnInterpreter
Tests run: 38, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.884 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverBypass
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.731 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithAbort
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.054 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerObserver
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.09 sec
Running org.apache.hadoop.hbase.coprocessor.TestOpenTableInCoprocessor
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.111 sec
Running org.apache.hadoop.hbase.coprocessor.TestAggregateProtocol
Tests run: 44, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.46 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.526 sec
Running org.apache.hadoop.hbase.coprocessor.TestClassLoading
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.649 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.211 sec
Running org.apache.hadoop.hbase.coprocessor.TestRowProcessorEndpoint
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.062 sec
Running org.apache.hadoop.hbase.constraint.TestConstraint
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.043 sec
Running org.apache.hadoop.hbase.procedure.TestZKProcedureControllers
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.12 sec
Running org.apache.hadoop.hbase.procedure.TestZKProcedure
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.972 sec
Running org.apache.hadoop.hbase.snapshot.TestRestoreFlushSnapshotFromClient
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 106.86 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 543.177 sec
Running org.apache.hadoop.hbase.snapshot.TestSecureExportSnapshot
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 288.362 sec
Running org.apache.hadoop.hbase.snapshot.TestSnapshotDescriptionUtils
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.92 sec
Running org.apache.hadoop.hbase.snapshot.TestFlushSnapshotFromClient
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.443 sec
Running org.apache.hadoop.hbase.mapred.TestTableMapReduce
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 149.966 sec
Running org.apache.hadoop.hbase.mapred.TestTableMapReduceUtil
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 131.127 sec
Running org.apache.hadoop.hbase.snapshot.TestExportSnapshot
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 319.593 sec
Running org.apache.hadoop.hbase.mapred.TestTableInputFormat
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.602 sec
Running org.apache.hadoop.hbase.master.TestOpenedRegionHandler
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.419 sec
Running org.apache.hadoop.hbase.master.TestActiveMasterManager
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.993 sec
Running org.apache.hadoop.hbase.master.TestAssignmentManager
Tests run: 19, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.178 sec
Running org.apache.hadoop.hbase.master.TestMasterRestartAfterDisablingTable
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.034 sec
Running org.apache.hadoop.hbase.master.TestMasterFailoverBalancerPersistence
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.553 sec
Running org.apache.hadoop.hbase.master.TestMasterFailover
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 70.316 sec
Running org.apache.hadoop.hbase.master.TestRollingRestart
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 41.95 sec
Running org.apache.hadoop.hbase.master.TestMasterTransitions
Tests run: 3, Failures: 0, Errors: 0, Skipped: 3, Time elapsed: 26.411 sec
Running org.apache.hadoop.hbase.master.TestRestartCluster
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.262 sec
Running org.apache.hadoop.hbase.master.TestMasterMetrics
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.966 sec
Running org.apache.hadoop.hbase.master.TestMasterStatusServlet
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.673 sec
Running org.apache.hadoop.hbase.master.TestHMasterRPCException
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.942 sec
Running org.apache.hadoop.hbase.master.TestTableLockManager
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 117.54 sec
Running org.apache.hadoop.hbase.master.TestRegionPlacement
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.533 sec
Running org.apache.hadoop.hbase.master.TestAssignmentManagerOnCluster
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.027 sec
Running org.apache.hadoop.hbase.master.balancer.TestBaseLoadBalancer
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.079 sec
Running org.apache.hadoop.hbase.master.balancer.TestStochasticLoadBalancer
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 103.076 sec
Running org.apache.hadoop.hbase.master.balancer.TestDefaultLoadBalancer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.459 sec
Running org.apache.hadoop.hbase.master.TestMaster
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.29 sec
Running org.apache.hadoop.hbase.master.cleaner.TestHFileCleaner
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.05 sec
Running org.apache.hadoop.hbase.master.cleaner.TestLogsCleaner
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.469 sec
Running org.apache.hadoop.hbase.master.cleaner.TestSnapshotFromMaster
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 40.07 sec
Running org.apache.hadoop.hbase.master.TestMasterShutdown
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.244 sec
Running org.apache.hadoop.hbase.master.TestMasterFileSystem
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.112 sec
Running org.apache.hadoop.hbase.master.handler.TestTableDeleteFamilyHandler
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.912 sec
Running org.apache.hadoop.hbase.master.TestDistributedLogSplitting
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 301.821 sec
Running org.apache.hadoop.hbase.master.handler.TestCreateTableHandler
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.536 sec
Running org.apache.hadoop.hbase.master.handler.TestTableDescriptorModification
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.188 sec
Running org.apache.hadoop.hbase.master.TestClusterStatusPublisher
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.584 sec
Running org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.751 sec
Running org.apache.hadoop.hbase.master.TestMasterMetricsWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.98 sec
Running org.apache.hadoop.hbase.master.TestMasterNoCluster
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.429 sec
Running org.apache.hadoop.hbase.master.TestDeadServer
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.587 sec
Running org.apache.hadoop.hbase.master.TestZKBasedOpenCloseRegion
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.156 sec
Running org.apache.hadoop.hbase.master.TestSplitLogManager
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 87.025 sec

Results :

Failed tests:   testFavoredNodes(org.apache.hadoop.hbase.regionserver.TestRegionFavoredNodes): Block location 127.0.0.1:36893 not a favored node

Tests run: 1828, Failures: 1, Errors: 0, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [6.812s]
[INFO] HBase - Common .................................... SUCCESS [34.780s]
[INFO] HBase - Protocol .................................. SUCCESS [11.644s]
[INFO] HBase - Client .................................... SUCCESS [40.191s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.842s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [2.763s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [4.490s]
[INFO] HBase - Server .................................... FAILURE [2:32:14.436s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:34:02.003s
[INFO] Finished at: Tue Mar 04 03:36:55 UTC 2014
[INFO] Final Memory: 37M/845M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson2450927639234625773.sh
++ grep surefirebooter
++ wc -l
++ jps
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10609
Updating HBASE-10635

Build failed in Jenkins: HBase-TRUNK #4973

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4973/changes>

Changes:

[enis] HBASE-10632 Region lost in limbo after ArrayIndexOutOfBoundsException during assignment

[jmhsieh] HBASE-10639 Unload script displays wrong counts (off by one) when unloading regions (Srikanth Srungarapu)

------------------------------------------
[...truncated 4826 lines...]
[INFO] --- maven-source-plugin:2.2.1:jar-no-fork (attach-sources) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.411 sec
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.046 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.343 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 170.028 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 170.028 sec <<< FAILURE!

Results :

Tests in error: 
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 32, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [8.471s]
[INFO] HBase - Common .................................... SUCCESS [39.507s]
[INFO] HBase - Protocol .................................. SUCCESS [12.938s]
[INFO] HBase - Client .................................... SUCCESS [42.210s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [6.178s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [3.287s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [5.819s]
[INFO] HBase - Server .................................... SUCCESS [2:37:43.705s]
[INFO] HBase - Testing Util .............................. SUCCESS [7.578s]
[INFO] HBase - Thrift .................................... FAILURE [2:57.311s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:42:48.187s
[INFO] Finished at: Mon Mar 03 23:25:37 UTC 2014
[INFO] Final Memory: 53M/832M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson372105742896416834.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10639
Updating HBASE-10632

Build failed in Jenkins: HBase-TRUNK #4972

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4972/changes>

Changes:

[enis] HBASE-10635 thrift#TestThriftServer fails due to TTL validity check

------------------------------------------
[...truncated 5194 lines...]
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xdf9496c8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-9-thread-1" prio=10 tid=0xa7f6e800 nid=0x269a waiting on condition [0xa6b18000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xde6fa910> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-8-thread-1" prio=10 tid=0xa7f6c400 nid=0x2695 waiting on condition [0xa6c5c000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xde6daf70> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-7-thread-1" prio=10 tid=0xa7f69400 nid=0x2683 waiting on condition [0xa815c000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xde654198> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"RMI RenewClean-[140.211.11.27:48196]" daemon prio=10 tid=0xa7c0a000 nid=0x2679 in Object.wait() [0xadd7d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xde36ed40> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0xde36ed40> (a java.lang.ref.ReferenceQueue$Lock)
	at sun.rmi.transport.DGCClient$EndpointEntry$RenewCleanThread.run(DGCClient.java:535)
	at java.lang.Thread.run(Thread.java:724)

"RMI Scheduler(0)" daemon prio=10 tid=0xa781ac00 nid=0x2678 waiting on condition [0xad57d000]
   java.lang.Thread.State: TIMED_WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb37ee460> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
	at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1090)
	at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:807)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"GC Daemon" daemon prio=10 tid=0xa7804000 nid=0x2676 in Object.wait() [0xae57d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xb3ae4ef0> (a sun.misc.GC$LatencyLock)
	at sun.misc.GC$Daemon.run(GC.java:117)
	- locked <0xb3ae4ef0> (a sun.misc.GC$LatencyLock)

"RMI TCP Accept-0" daemon prio=10 tid=0xa7802c00 nid=0x2674 runnable [0xa967d000]
   java.lang.Thread.State: RUNNABLE
	at java.net.PlainSocketImpl.socketAccept(Native Method)
	at java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:398)
	at java.net.ServerSocket.implAccept(ServerSocket.java:530)
	at java.net.ServerSocket.accept(ServerSocket.java:498)
	at sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:387)
	at sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:359)
	at java.lang.Thread.run(Thread.java:724)

"ActiveMQ Connection Executor: tcp://localhost/127.0.0.1:13103@38391" prio=10 tid=0xa7600800 nid=0x2665 waiting on condition [0xae37d000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb3ae55a8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"ActiveMQ Connection Executor: tcp://localhost/127.0.0.1:13103@38390" prio=10 tid=0xa7900800 nid=0x2662 waiting on condition [0xad97d000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb3b19c30> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"ActiveMQ Connection Executor: tcp://localhost/127.0.0.1:13101@55652" prio=10 tid=0xa8000c00 nid=0x2450 waiting on condition [0xadb7d000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb3b401b8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0xa9726c00 nid=0x23cc runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0xa9725000 nid=0x23cb waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0xa9723000 nid=0x23ca waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0xa9721400 nid=0x23c9 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0xa9ce6c00 nid=0x23c7 in Object.wait() [0xaed7d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xb3aeee50> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0xb3aeee50> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0xa9ce5000 nid=0x23c6 in Object.wait() [0xaf07d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xb3ae4ee8> (a java.lang.ref.Reference$Lock)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0xb3ae4ee8> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6806800 nid=0x23ae waiting on condition [0xf695b000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xe04b6388> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:374)
	at org.apache.activemq.transport.FutureResponse.getResult(FutureResponse.java:40)
	at org.apache.activemq.transport.ResponseCorrelator.request(ResponseCorrelator.java:87)
	at org.apache.activemq.ActiveMQConnection.syncSendPacket(ActiveMQConnection.java:1394)
	at org.apache.activemq.ActiveMQConnection.ensureConnectionInfoSent(ActiveMQConnection.java:1510)
	- locked <0xe04b5420> (a java.lang.Object)
	at org.apache.activemq.ActiveMQConnection.start(ActiveMQConnection.java:524)
	at org.apache.cxf.transport.jms.util.TestReceiver.drainQueue(TestReceiver.java:71)
	at org.apache.cxf.transport.jms.util.TestReceiver.runAsync(TestReceiver.java:120)
	at org.apache.cxf.transport.jms.uri.URIConfiguredConduitTest.sendAndReceive(URIConfiguredConduitTest.java:84)
	at org.apache.cxf.transport.jms.uri.URIConfiguredConduitTest.testSendReceive(URIConfiguredConduitTest.java:63)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
	at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
	at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
	at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)

"VM Thread" prio=10 tid=0xa9ff8400 nid=0x23c5 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6810800 nid=0x23b0 runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6812000 nid=0x23b1 runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6813800 nid=0x23b2 runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6814c00 nid=0x23b3 runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6816400 nid=0x23b4 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6817c00 nid=0x23b5 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6819000 nid=0x23b6 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf681a800 nid=0x23b8 runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf681c000 nid=0x23b9 runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf681d400 nid=0x23ba runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf681ec00 nid=0x23bd runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6820400 nid=0x23be runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6821800 nid=0x23bf runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6823000 nid=0x23c0 runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6824800 nid=0x23c1 runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6825c00 nid=0x23c2 runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6827400 nid=0x23c3 runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6828c00 nid=0x23c4 runnable 

"VM Periodic Task Thread" prio=10 tid=0xa9730c00 nid=0x23cd waiting on condition 

JNI global references: 207

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s): 	at org.apache.cxf.transport.jms.uri.URIConfiguredConduitTest.testSendReceive(URIConfiguredConduitTest.java:63)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10635

Build failed in Jenkins: HBase-TRUNK #4971

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4971/>

------------------------------------------
Started by an SCM change
Building remotely on ubuntu2 in workspace <https://builds.apache.org/job/HBase-TRUNK/ws/>
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@72cf4805:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@72cf4805:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@72cf4805:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson5442153936504093802.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
1906: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4970

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4970/changes>

Changes:

[anoopsamjohn] HBASE-10451 Enable back Tag compression on HFiles.(Anoop)

------------------------------------------
[...truncated 4673 lines...]
Running org.apache.hadoop.hbase.rest.client.TestRemoteTable
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.929 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithFilters
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.46 sec
Running org.apache.hadoop.hbase.rest.TestTableResource
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.567 sec
Running org.apache.hadoop.hbase.rest.TestGetAndPutResource
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.916 sec
Running org.apache.hadoop.hbase.rest.TestMultiRowResource
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.941 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithLabels
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.726 sec
Running org.apache.hadoop.hbase.rest.TestDeleteRow
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.195 sec
Running org.apache.hadoop.hbase.coprocessor.TestWALObserver
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.367 sec
Running org.apache.hadoop.hbase.coprocessor.TestRowProcessorEndpoint
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.331 sec
Running org.apache.hadoop.hbase.coprocessor.TestOpenTableInCoprocessor
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.739 sec
Running org.apache.hadoop.hbase.coprocessor.TestClassLoading
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.416 sec
Running org.apache.hadoop.hbase.coprocessor.TestAggregateProtocol
Tests run: 44, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.638 sec
Running org.apache.hadoop.hbase.coprocessor.TestCoprocessorEndpoint
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.884 sec
Running org.apache.hadoop.hbase.coprocessor.TestHTableWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.603 sec
Running org.apache.hadoop.hbase.coprocessor.TestBigDecimalColumnInterpreter
Tests run: 38, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.948 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverInterface
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 60.519 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverBypass
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.178 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerObserver
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.549 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterObserver
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.161 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.008 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.901 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.841 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithAbort
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.365 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.515 sec
Running org.apache.hadoop.hbase.TestGlobalMemStoreSize
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.137 sec
Running org.apache.hadoop.hbase.catalog.TestMetaReaderEditorNoCluster
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.28 sec
Running org.apache.hadoop.hbase.catalog.TestMetaMigrationConvertingToPB
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.38 sec
Running org.apache.hadoop.hbase.catalog.TestCatalogTracker
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.943 sec
Running org.apache.hadoop.hbase.catalog.TestMetaReaderEditor
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.756 sec
Running org.apache.hadoop.hbase.regionserver.TestReversibleScanners
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.529 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionMergeTransactionOnCluster
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.654 sec
Running org.apache.hadoop.hbase.regionserver.TestCompaction
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 92.916 sec
Running org.apache.hadoop.hbase.regionserver.TestStoreScanner
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.697 sec
Running org.apache.hadoop.hbase.regionserver.TestMemStore
Tests run: 23, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.258 sec
Running org.apache.hadoop.hbase.regionserver.TestPriorityRpc
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.86 sec
Running org.apache.hadoop.hbase.regionserver.TestParallelPut
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.342 sec
Running org.apache.hadoop.hbase.regionserver.TestClusterId
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.99 sec
Running org.apache.hadoop.hbase.regionserver.TestStore
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.844 sec
Running org.apache.hadoop.hbase.regionserver.TestCacheOnWriteInSchema
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.549 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionBusyWait
Tests run: 66, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 69.534 sec
Running org.apache.hadoop.hbase.regionserver.TestServerCustomProtocol
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.575 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLog
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.528 sec
Running org.apache.hadoop.hbase.regionserver.TestSplitTransactionOnCluster
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.907 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogFiltering
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.934 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRollPeriod
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 73.445 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogSplitCompressed
Tests run: 33, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 132.406 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestWALReplay
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 72.335 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestSecureWALReplay
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.715 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogSplit
Tests run: 33, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 119.409 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestReadOldRootAndMetaEdits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.341 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestSecureHLog
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.759 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestDurability
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.986 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRollAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.171 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestWALReplayCompressed
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 60.425 sec
Running org.apache.hadoop.hbase.regionserver.TestJoinedScanners
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.079 sec
Running org.apache.hadoop.hbase.regionserver.TestGetClosestAtOrBefore
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.831 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionOnCluster
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.723 sec
Running org.apache.hadoop.hbase.regionserver.handler.TestCloseRegionHandler
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.678 sec
Running org.apache.hadoop.hbase.regionserver.handler.TestOpenRegionHandler
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.767 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 45.605 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionFavoredNodes
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.714 sec
Running org.apache.hadoop.hbase.regionserver.TestCompoundBloomFilter
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.733 sec
Running org.apache.hadoop.hbase.regionserver.TestCompactionState
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 56.44 sec
Running org.apache.hadoop.hbase.regionserver.TestStoreFileBlockCacheSummary
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.429 sec
Running org.apache.hadoop.hbase.regionserver.TestBlocksRead
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.112 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRolling
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 477.864 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionServerMetrics
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.595 sec
Running org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 44.538 sec
Running org.apache.hadoop.hbase.regionserver.TestEncryptionRandomKeying
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.139 sec
Running org.apache.hadoop.hbase.regionserver.TestFSErrorsExposed
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.861 sec
Running org.apache.hadoop.hbase.regionserver.TestEncryptionKeyRotation
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.686 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegion
Tests run: 65, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 68.217 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionServerNoMaster
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.753 sec
Running org.apache.hadoop.hbase.regionserver.TestSplitLogWorker
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.851 sec
Running org.apache.hadoop.hbase.regionserver.TestCompactionWithCoprocessor
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 96.341 sec
Running org.apache.hadoop.hbase.regionserver.TestAtomicOperation
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.127 sec
Running org.apache.hadoop.hbase.regionserver.TestSeekOptimizations
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.836 sec
Running org.apache.hadoop.hbase.regionserver.TestTags
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.336 sec
Running org.apache.hadoop.hbase.regionserver.TestEndToEndSplitTransaction
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.749 sec
Running org.apache.hadoop.hbase.regionserver.TestMasterAddressManager
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.527 sec
Running org.apache.hadoop.hbase.regionserver.TestRSKilledWhenInitializing
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.883 sec
Running org.apache.hadoop.hbase.TestClusterBootOrder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.927 sec

Results :

Tests in error: 
  org.apache.hadoop.hbase.master.TestMasterMetrics: Shutting down

Tests run: 1820, Failures: 0, Errors: 1, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [4.229s]
[INFO] HBase - Common .................................... SUCCESS [39.552s]
[INFO] HBase - Protocol .................................. SUCCESS [12.994s]
[INFO] HBase - Client .................................... SUCCESS [38.519s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.773s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [3.276s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.529s]
[INFO] HBase - Server .................................... FAILURE [2:34:07.777s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:35:56.454s
[INFO] Finished at: Sat Mar 01 17:13:00 UTC 2014
[INFO] Final Memory: 37M/860M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson3773190383010929138.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10451

Build failed in Jenkins: HBase-TRUNK #4969

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4969/>

------------------------------------------
[...truncated 3285 lines...]
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)

"main.logSyncer" daemon prio=10 tid=0x57362800 nid=0x41f5 in Object.wait() [0x54894000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x6861ff30> (a java.util.concurrent.atomic.AtomicBoolean)
	at org.apache.hadoop.hbase.regionserver.wal.HLog$LogSyncer.run(HLog.java:1256)
	- locked <0x6861ff30> (a java.util.concurrent.atomic.AtomicBoolean)
	at java.lang.Thread.run(Thread.java:724)

"testReplayEditsWrittenViaHRegion-server-1234-index-writer-pool-31-thread-1-EventThread" daemon prio=10 tid=0x59dd9800 nid=0x41c0 waiting on condition [0x546ff000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x6861ffd8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)

"testReplayEditsWrittenViaHRegion-server-1234-index-writer-pool-31-thread-1-SendThread(localhost:53471)" daemon prio=10 tid=0x59dd8800 nid=0x41bf waiting on condition [0x54750000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86)
	at org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)

"LeaseChecker" daemon prio=10 tid=0x57b18400 nid=0x40c2 waiting on condition [0x57969000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.ipc.Client$Connection.handleConnectionFailure(Client.java:663)
	at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:456)
	- locked <0x6d63d958> (a org.apache.hadoop.ipc.Client$Connection)
	at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560)
	- locked <0x6d63d958> (a org.apache.hadoop.ipc.Client$Connection)
	at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184)
	at org.apache.hadoop.ipc.Client.getConnection(Client.java:1206)
	at org.apache.hadoop.ipc.Client.call(Client.java:1050)
	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
	at com.sun.proxy.$Proxy16.renewLease(Unknown Source)
	at sun.reflect.GeneratedMethodAccessor44.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
	at com.sun.proxy.$Proxy16.renewLease(Unknown Source)
	at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.renew(DFSClient.java:1359)
	at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.run(DFSClient.java:1371)
	at java.lang.Thread.run(Thread.java:724)

"pool-4-thread-1" prio=10 tid=0x59b61c00 nid=0x40b1 runnable [0x57ef8000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x68575af8> (a sun.nio.ch.Util$2)
	- locked <0x68575ae8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x685758d8> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x685758c0> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-3-thread-1" prio=10 tid=0xf6758800 nid=0x4095 runnable [0x583b7000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x685bde78> (a sun.nio.ch.Util$2)
	- locked <0x685bde68> (a java.util.Collections$UnmodifiableSet)
	- locked <0x685bdc58> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x685bdc40> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-2-thread-1" prio=10 tid=0x5a72c400 nid=0x4079 runnable [0x58876000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x66fd4870> (a sun.nio.ch.Util$2)
	- locked <0x66fd4860> (a java.util.Collections$UnmodifiableSet)
	- locked <0x66fd4650> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x66fd4638> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x58e42800 nid=0x4053 runnable [0x59e5c000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x66f47400> (a sun.nio.ch.Util$2)
	- locked <0x66f47410> (a java.util.Collections$UnmodifiableSet)
	- locked <0x66f473c0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x66f47420> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x5b1d8400 nid=0x404a runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x5b1d6400 nid=0x4049 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x5b1d4400 nid=0x4048 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x5b1d2c00 nid=0x4047 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x5b198000 nid=0x4046 in Object.wait() [0x6207d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x66c93238> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x66c93238> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x5b196400 nid=0x4045 in Object.wait() [0x6227d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x66c96690> (a java.lang.ref.Reference$Lock)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x66c96690> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6706800 nid=0x403b runnable [0xf6937000]
   java.lang.Thread.State: RUNNABLE
	at java.io.FileOutputStream.writeBytes(Native Method)
	at java.io.FileOutputStream.write(FileOutputStream.java:318)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	- locked <0x66caab70> (a java.io.BufferedOutputStream)
	at java.io.PrintStream.write(PrintStream.java:482)
	- locked <0x66caab50> (a java.io.PrintStream)
	at org.apache.maven.surefire.booter.ForkingRunListener.writeTestOutput(ForkingRunListener.java:178)
	- locked <0x66caab50> (a java.io.PrintStream)
	at org.apache.maven.surefire.report.ConsoleOutputCapture$ForwardingPrintStream.write(ConsoleOutputCapture.java:58)
	at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221)
	at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291)
	at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295)
	at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141)
	- locked <0x66ca85e0> (a java.io.OutputStreamWriter)
	at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229)
	at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59)
	at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324)
	at org.apache.log4j.WriterAppender.append(WriterAppender.java:162)
	at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251)
	- locked <0x66ca7ed0> (a org.apache.log4j.ConsoleAppender)
	at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66)
	at org.apache.log4j.Category.callAppenders(Category.java:206)
	- locked <0x66ca4290> (a org.apache.log4j.spi.RootLogger)
	at org.apache.log4j.Category.forcedLog(Category.java:391)
	at org.apache.log4j.Category.log(Category.java:856)
	at org.slf4j.impl.Log4jLoggerAdapter.debug(Log4jLoggerAdapter.java:207)
	at org.mortbay.log.Slf4jLog.debug(Slf4jLog.java:40)
	at org.mortbay.log.Log.debug(Log.java:105)
	at org.mortbay.resource.JarResource.extract(JarResource.java:184)
	at org.mortbay.jetty.webapp.WebAppContext.resolveWebApp(WebAppContext.java:974)
	at org.mortbay.jetty.webapp.WebAppContext.getWebInf(WebAppContext.java:832)
	at org.mortbay.jetty.webapp.WebInfConfiguration.configureClassLoader(WebInfConfiguration.java:62)
	at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:489)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	- locked <0xc87593d0> (a java.lang.Object)
	at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
	at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	- locked <0xc8751860> (a java.lang.Object)
	at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
	at org.mortbay.jetty.Server.doStart(Server.java:224)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	- locked <0xc873eb98> (a java.lang.Object)
	at org.apache.hadoop.http.HttpServer.start(HttpServer.java:617)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:481)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:299)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1496)
	at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:417)
	- locked <0x6d971530> (a org.apache.hadoop.hdfs.MiniDFSCluster)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
	at org.apache.phoenix.hbase.index.covered.TestEndToEndCoveredColumnsIndexBuilder.setupCluster(TestEndToEndCoveredColumnsIndexBuilder.java:106)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
	at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
	at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
	at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)

"VM Thread" prio=10 tid=0x5b190c00 nid=0x4044 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6710800 nid=0x403c runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6712000 nid=0x403d runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6713400 nid=0x403e runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6714c00 nid=0x403f runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6716400 nid=0x4040 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6717800 nid=0x4041 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6719000 nid=0x4042 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf671a800 nid=0x4043 runnable 

"VM Periodic Task Thread" prio=10 tid=0x5b1e2400 nid=0x404b waiting on condition 

JNI global references: 398

POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4968

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4968/changes>

Changes:

[tedyu] HBASE-10644 TestSecureExportSnapshot#testExportFileSystemState fails on hadoop-1

------------------------------------------
[...truncated 9365 lines...]
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1408)

"IPC Server listener on 37626" daemon prio=10 tid=0x6ba42c00 nid=0x1b45 runnable [0x6db7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7efcabd0> (a sun.nio.ch.Util$2)
	- locked <0x7efcab50> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7efca900> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener.run(Server.java:479)

"IPC Server Responder" daemon prio=10 tid=0x6ba3b400 nid=0x1b44 runnable [0x6dd7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7efd4f50> (a sun.nio.ch.Util$2)
	- locked <0x7efd4f40> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7efd4d40> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.apache.hadoop.ipc.Server$Responder.run(Server.java:645)

"Timer-0" daemon prio=10 tid=0x6ba2d800 nid=0x1b42 in Object.wait() [0x73d7d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x7f81f350> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"14915048@qtp-24350088-1 - Acceptor0 SelectChannelConnector@localhost:49842" prio=10 tid=0x6ba12400 nid=0x1b34 runnable [0x73f7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7f82a6e8> (a sun.nio.ch.Util$2)
	- locked <0x7f82a6d8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7f82a298> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
	at org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
	at org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
	at org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)

"32715295@qtp-24350088-0" prio=10 tid=0x6ba13800 nid=0x1b33 in Object.wait() [0x7417d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626)
	- locked <0x7f829de0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)

"pool-2-thread-1" prio=10 tid=0x6c381800 nid=0x1b25 runnable [0x7437d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7efcc4d8> (a sun.nio.ch.Util$2)
	- locked <0x7efcc4c8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7efcc2c8> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:373)
	- locked <0x7efcc6c0> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor@148d148" daemon prio=10 tid=0x6c356800 nid=0x1b24 waiting on condition [0x7457d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor@533643" daemon prio=10 tid=0x6c355800 nid=0x1b23 waiting on condition [0x7477d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor.run(FSNamesystem.java:3167)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@14cc6" daemon prio=10 tid=0x6c354c00 nid=0x1b22 waiting on condition [0x7497d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:368)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor@4ed34" daemon prio=10 tid=0x6c353400 nid=0x1b21 waiting on condition [0x74b7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor.run(FSNamesystem.java:3146)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor@112c7fb" daemon prio=10 tid=0x6c351400 nid=0x1b1f waiting on condition [0x74d7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:197)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x6d34c800 nid=0x1a4a in Object.wait() [0x7517d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.Thread.join(Thread.java:1268)
	- locked <0xceb8df00> (a org.junit.internal.runners.statements.FailOnTimeout$StatementThread)
	at org.junit.internal.runners.statements.FailOnTimeout.evaluateStatement(FailOnTimeout.java:26)
	at org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:17)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runners.Suite.runChild(Suite.java:127)
	at org.junit.runners.Suite.runChild(Suite.java:26)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x6de27000 nid=0x1a45 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x6de25400 nid=0x1a44 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x6de23400 nid=0x1a43 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x6de21800 nid=0x1a42 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x6e2e7000 nid=0x1a41 in Object.wait() [0x75f5a000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x7ca75db0> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x6e2e5400 nid=0x1a40 in Object.wait() [0x6e76d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x7ca597b0> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6706c00 nid=0x1a2c waiting on condition [0xf6940000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x7ca75f58> (a java.util.concurrent.FutureTask$Sync)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:834)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:994)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1303)
	at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:248)
	at java.util.concurrent.FutureTask.get(FutureTask.java:111)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.waitForCompletion(AsynchronousRunner.java:73)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.finished(AsynchronousRunner.java:58)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:242)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:138)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:61)
	at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:137)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:188)
	at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:166)
	at org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:86)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:101)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:74)

"VM Thread" prio=10 tid=0x6e5f8800 nid=0x1a3f runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6710c00 nid=0x1a2d runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6712000 nid=0x1a2e runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6713800 nid=0x1a2f runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6715000 nid=0x1a30 runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6716400 nid=0x1a31 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6717c00 nid=0x1a32 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6719400 nid=0x1a33 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf671ac00 nid=0x1a34 runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf671c000 nid=0x1a35 runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf671d800 nid=0x1a36 runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf671f000 nid=0x1a37 runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6720400 nid=0x1a38 runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6721c00 nid=0x1a39 runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6723400 nid=0x1a3a runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6724800 nid=0x1a3b runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6726000 nid=0x1a3c runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6727800 nid=0x1a3d runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6728c00 nid=0x1a3e runnable 

"VM Periodic Task Thread" prio=10 tid=0x6de29000 nid=0x1a47 waiting on condition 

JNI global references: 359

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 2 zombie test(s): 	at org.apache.hadoop.hbase.regionserver.wal.TestWALReplay.testReplayEditsWrittenIntoWAL(TestWALReplay.java:698)
	at org.apache.hadoop.hbase.regionserver.wal.TestHLogSplit.testEmptyOpenLogFiles(TestHLogSplit.java:435)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10644
Updating HADOOP-1

Build failed in Jenkins: HBase-TRUNK #4967

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4967/>

------------------------------------------
Started by an SCM change
Building remotely on ubuntu2 in workspace <https://builds.apache.org/job/HBase-TRUNK/ws/>
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@5733e030:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@5733e030:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@5733e030:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson2782120265627618185.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
1906: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4966

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4966/changes>

Changes:

[tedyu] HBASE-10627 A logic mistake in HRegionServer isHealthy (Shaohui)

[tedyu] HBASE-10638 Improve error message when there is no region server available for move

------------------------------------------
[...truncated 4824 lines...]
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.17 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.483 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 116.189 sec <<< FAILURE!
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 116.189 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 116.189 sec <<< FAILURE!

Results :

Tests in error: 
  testAll(org.apache.hadoop.hbase.thrift.TestThriftServer): org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 33, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [10.702s]
[INFO] HBase - Common .................................... SUCCESS [37.944s]
[INFO] HBase - Protocol .................................. SUCCESS [12.320s]
[INFO] HBase - Client .................................... SUCCESS [40.905s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [6.404s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [3.033s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [5.278s]
[INFO] HBase - Server .................................... SUCCESS [2:27:52.980s]
[INFO] HBase - Testing Util .............................. SUCCESS [3.258s]
[INFO] HBase - Thrift .................................... FAILURE [2:03.245s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:31:57.559s
[INFO] Finished at: Fri Feb 28 20:39:20 UTC 2014
[INFO] Final Memory: 53M/768M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson7831537771808749339.sh
++ jps
++ wc -l
++ grep surefirebooter
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10627
Updating HBASE-10638

Build failed in Jenkins: HBase-TRUNK #4965

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4965/changes>

Changes:

[mbertozzi] HBASE-10631 Avoid extra seek on FileLink open

------------------------------------------
[...truncated 4823 lines...]
[INFO] --- maven-source-plugin:2.2.1:jar-no-fork (attach-sources) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.266 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.173 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 116.222 sec <<< FAILURE!
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 116.222 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 116.222 sec <<< FAILURE!

Results :

Tests in error: 
  testAll(org.apache.hadoop.hbase.thrift.TestThriftServer): org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 33, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [2.870s]
[INFO] HBase - Common .................................... SUCCESS [24.681s]
[INFO] HBase - Protocol .................................. SUCCESS [10.118s]
[INFO] HBase - Client .................................... SUCCESS [38.465s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.385s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [1.954s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.441s]
[INFO] HBase - Server .................................... SUCCESS [2:25:26.609s]
[INFO] HBase - Testing Util .............................. SUCCESS [2.406s]
[INFO] HBase - Thrift .................................... FAILURE [2:00.860s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:28:57.557s
[INFO] Finished at: Fri Feb 28 12:36:03 UTC 2014
[INFO] Final Memory: 55M/898M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson490619549740657010.sh
++ grep surefirebooter
++ wc -l
++ jps
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10631

Build failed in Jenkins: HBase-TRUNK #4964

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4964/changes>

Changes:

[nkeywal] HBASE-10080 Unnecessary call to locateRegion when creating an HTable instance

[nkeywal] HBASE-9990 HTable uses the conf for each call creation

------------------------------------------
[...truncated 4824 lines...]
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.061 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.474 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 113.925 sec <<< FAILURE!
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 113.925 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 113.925 sec <<< FAILURE!

Results :

Tests in error: 
  testAll(org.apache.hadoop.hbase.thrift.TestThriftServer): org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 33, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [7.536s]
[INFO] HBase - Common .................................... SUCCESS [38.618s]
[INFO] HBase - Protocol .................................. SUCCESS [13.025s]
[INFO] HBase - Client .................................... SUCCESS [41.196s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [6.116s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [4.102s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [5.662s]
[INFO] HBase - Server .................................... SUCCESS [2:25:40.767s]
[INFO] HBase - Testing Util .............................. SUCCESS [5.121s]
[INFO] HBase - Thrift .................................... FAILURE [2:01.352s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:29:44.801s
[INFO] Finished at: Fri Feb 28 01:00:28 UTC 2014
[INFO] Final Memory: 52M/719M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson7334869491645478187.sh
++ jps
++ wc -l
++ grep surefirebooter
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10080
Updating HBASE-9990

Build failed in Jenkins: HBase-TRUNK #4963

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4963/>

------------------------------------------
Started by an SCM change
Building remotely on ubuntu2 in workspace <https://builds.apache.org/job/HBase-TRUNK/ws/>
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@420bd0c5:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@420bd0c5:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@420bd0c5:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson8535546868435223615.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
1906: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4962

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4962/changes>

Changes:

[larsh] HBASE-10594 Revert due to test instability

[tedyu] HBASE-10621 Unable to grant user permission to namespace

------------------------------------------
[...truncated 18187 lines...]

"IPC Server Responder" daemon prio=10 tid=0x6f62e400 nid=0x1040 runnable [0x70d68000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7eb95e88> (a sun.nio.ch.Util$2)
	- locked <0x7eb95e78> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7eb95c78> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.apache.hadoop.ipc.Server$Responder.run(Server.java:605)

"Timer-0" daemon prio=10 tid=0x6f647400 nid=0x1035 in Object.wait() [0x7617d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x80dd0cd0> (a java.util.TaskQueue)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x80dd0cd0> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"25740701@qtp-23498697-1" prio=10 tid=0x6f64d800 nid=0xfeb in Object.wait() [0x7637d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x80f2e360> (a org.mortbay.thread.QueuedThreadPool$PoolThread)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626)
	- locked <0x80f2e360> (a org.mortbay.thread.QueuedThreadPool$PoolThread)

"14766724@qtp-23498697-0 - Acceptor0 SelectChannelConnector@localhost:35490" prio=10 tid=0x6f653400 nid=0xfea runnable [0x7657d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x80f0e0c0> (a sun.nio.ch.Util$2)
	- locked <0x80f0e0d0> (a java.util.Collections$UnmodifiableSet)
	- locked <0x80f0e080> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
	at org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
	at org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
	at org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)

"pool-2-thread-1" prio=10 tid=0x6f6b1c00 nid=0xfc0 runnable [0x7677d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7ec19910> (a sun.nio.ch.Util$2)
	- locked <0x7ec19920> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7ec198d0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x7ec19930> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor@76bfbd" daemon prio=10 tid=0x6f672c00 nid=0xfbe waiting on condition [0x7697d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor@7790ec" daemon prio=10 tid=0x6f6a1c00 nid=0xfbd waiting on condition [0x76b7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor.run(FSNamesystem.java:2716)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@8bf924" daemon prio=10 tid=0x6f69f400 nid=0xfbc waiting on condition [0x76d7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:368)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor@15dfb9d" daemon prio=10 tid=0x6f69e800 nid=0xfbb waiting on condition [0x76f7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor.run(FSNamesystem.java:2693)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor@f3424a" daemon prio=10 tid=0x6f69ac00 nid=0xfb9 waiting on condition [0x7717d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:186)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x6f884000 nid=0xd55 waiting on condition [0x7757d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.mapred.JobClient.monitorAndPrintJob(JobClient.java:1295)
	at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:533)
	at org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat.testScan(TestMultiTableInputFormat.java:250)
	at org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat.testScanEmptyToEmpty(TestMultiTableInputFormat.java:177)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runners.Suite.runChild(Suite.java:127)
	at org.junit.runners.Suite.runChild(Suite.java:26)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x70326c00 nid=0xd31 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x70325000 nid=0xd30 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x70323000 nid=0xd2f waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x70321400 nid=0xd2e runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x708e6c00 nid=0xd2c in Object.wait() [0x77f7d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x7ec57f20> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x7ec57f20> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x708e5000 nid=0xd2b in Object.wait() [0x7817d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x7ec57a40> (a java.lang.ref.Reference$Lock)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x7ec57a40> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6706800 nid=0xd16 waiting on condition [0xf68fd000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x7ea550b8> (a java.util.concurrent.FutureTask$Sync)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:834)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:994)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1303)
	at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:248)
	at java.util.concurrent.FutureTask.get(FutureTask.java:111)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.waitForCompletion(AsynchronousRunner.java:73)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.finished(AsynchronousRunner.java:58)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:242)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:138)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:61)
	at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:137)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:188)
	at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:166)
	at org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:86)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:101)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:74)

"VM Thread" prio=10 tid=0x70bf8400 nid=0xd2a runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6710800 nid=0xd17 runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6712000 nid=0xd18 runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6713800 nid=0xd19 runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6714c00 nid=0xd1a runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6716400 nid=0xd1b runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6717c00 nid=0xd1c runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6719000 nid=0xd1d runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf671a800 nid=0xd1e runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf671c000 nid=0xd1f runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf671d400 nid=0xd20 runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf671ec00 nid=0xd21 runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6720400 nid=0xd22 runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6721800 nid=0xd23 runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6723000 nid=0xd24 runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6724800 nid=0xd25 runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6726000 nid=0xd26 runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6727400 nid=0xd27 runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6728c00 nid=0xd28 runnable 

"VM Periodic Task Thread" prio=10 tid=0x70328800 nid=0xd32 waiting on condition 

JNI global references: 169

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 4 zombie test(s): 	at org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScanBase.testScan(TestTableInputFormatScanBase.java:234)
	at org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1.testScanEmptyToOPP(TestTableInputFormatScan1.java:96)
	at org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed.testTimeRangeMapRed(TestTimeRangeMapRed.java:152)
	at org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat.testScan(TestMultiTableInputFormat.java:250)
	at org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat.testScanEmptyToEmpty(TestMultiTableInputFormat.java:177)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10594
Updating HBASE-10621

Build failed in Jenkins: HBase-TRUNK #4961

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4961/changes>

Changes:

[tedyu] HBASE-10612 Remove unnecessary dependency on org.eclipse.jdt:core

[apurtell] HBASE-10618 User should not be allowed to disable/drop visibility labels table (Anoop Sam John)

[stack] HBASE-10598 Written data can not be read out because MemStore#timeRangeTracker might be updated concurrently (cuijianwei)

[stack] HBASE-8803 region_mover.rb should move multiple regions at a time

------------------------------------------
[...truncated 8867 lines...]
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x68438a68> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)

"Master:0;hemera.apache.org,47289,1393479972468-SendThread(localhost:51801)" daemon prio=10 tid=0x56faac00 nid=0x3611 waiting on condition [0x523b4000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:984)

"LeaseChecker" daemon prio=10 tid=0x55202000 nid=0x3551 waiting on condition [0x5510b000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.ipc.Client$Connection.handleConnectionFailure(Client.java:663)
	at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:456)
	- locked <0xc65d6ba8> (a org.apache.hadoop.ipc.Client$Connection)
	at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560)
	- locked <0xc65d6ba8> (a org.apache.hadoop.ipc.Client$Connection)
	at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184)
	at org.apache.hadoop.ipc.Client.getConnection(Client.java:1206)
	at org.apache.hadoop.ipc.Client.call(Client.java:1050)
	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
	at com.sun.proxy.$Proxy16.renewLease(Unknown Source)
	at sun.reflect.GeneratedMethodAccessor45.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
	at com.sun.proxy.$Proxy16.renewLease(Unknown Source)
	at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.renew(DFSClient.java:1359)
	at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.run(DFSClient.java:1371)
	at java.lang.Thread.run(Thread.java:724)

"pool-4-thread-1" prio=10 tid=0xf6765000 nid=0x353f runnable [0x55690000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x684bf9e8> (a sun.nio.ch.Util$2)
	- locked <0x684bf9f8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x684bf9a8> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x684bfa08> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-3-thread-1" prio=10 tid=0x58aa6000 nid=0x3486 runnable [0x55c9d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x68529778> (a sun.nio.ch.Util$2)
	- locked <0x68529768> (a java.util.Collections$UnmodifiableSet)
	- locked <0x68529558> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x68529540> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-2-thread-1" prio=10 tid=0x56ca8000 nid=0x3406 runnable [0x5615c000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x684d8d10> (a sun.nio.ch.Util$2)
	- locked <0x684d8d00> (a java.util.Collections$UnmodifiableSet)
	- locked <0x684d8af0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x684d8ad8> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x58011000 nid=0x3336 runnable [0x6087d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x66f1bd40> (a sun.nio.ch.Util$2)
	- locked <0x66f1bd30> (a java.util.Collections$UnmodifiableSet)
	- locked <0x66f1bb28> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x66ec9d40> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x59626c00 nid=0x31b8 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x59625000 nid=0x31b7 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x59623000 nid=0x31b6 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x59621400 nid=0x31b5 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x59be6c00 nid=0x319b in Object.wait() [0x6207d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x66edf1f8> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x59be5000 nid=0x319a in Object.wait() [0x6227d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x66ec9d38> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6706800 nid=0x317e runnable [0xf6947000]
   java.lang.Thread.State: RUNNABLE
	at com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl$NSContentDriver.scanRootElementHook(XMLNSDocumentScannerImpl.java:601)
	at com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl$FragmentContentDriver.next(XMLDocumentFragmentScannerImpl.java:3065)
	at com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl$PrologDriver.next(XMLDocumentScannerImpl.java:881)
	at com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl.next(XMLDocumentScannerImpl.java:607)
	at com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.next(XMLNSDocumentScannerImpl.java:116)
	at com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:489)
	at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:835)
	at com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:764)
	at com.sun.org.apache.xerces.internal.parsers.XMLParser.parse(XMLParser.java:123)
	at com.sun.org.apache.xerces.internal.parsers.DOMParser.parse(DOMParser.java:237)
	at com.sun.org.apache.xerces.internal.jaxp.DocumentBuilderImpl.parse(DocumentBuilderImpl.java:300)
	at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:177)
	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:1168)
	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:1119)
	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:1063)
	- locked <0xe23a52e8> (a org.apache.hadoop.conf.Configuration)
	at org.apache.hadoop.conf.Configuration.get(Configuration.java:416)
	at org.apache.hadoop.conf.Configuration.getBoolean(Configuration.java:604)
	at org.apache.hadoop.hbase.HBaseConfiguration.checkDefaultsVersion(HBaseConfiguration.java:64)
	at org.apache.hadoop.hbase.HBaseConfiguration.addHbaseResources(HBaseConfiguration.java:100)
	at org.apache.hadoop.hbase.HBaseConfiguration.create(HBaseConfiguration.java:111)
	at org.apache.phoenix.query.ConfigurationFactory$ConfigurationFactoryImpl.getConfiguration(ConfigurationFactory.java:41)
	at org.apache.phoenix.end2end.BaseConnectedQueryTest.getUrl(BaseConnectedQueryTest.java:99)
	at org.apache.phoenix.end2end.QueryTest.initTable(QueryTest.java:137)
	at sun.reflect.GeneratedMethodAccessor54.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runners.Suite.runChild(Suite.java:127)
	at org.junit.runners.Suite.runChild(Suite.java:26)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
	at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
	at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
	at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)

"VM Thread" prio=10 tid=0x59ef8400 nid=0x3193 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6710800 nid=0x317f runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6712000 nid=0x3180 runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6713800 nid=0x3181 runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6714c00 nid=0x3182 runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6716400 nid=0x3183 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6717c00 nid=0x3184 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6719000 nid=0x3185 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf671a800 nid=0x3186 runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf671c000 nid=0x3187 runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf671d400 nid=0x3188 runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf671ec00 nid=0x3189 runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6720400 nid=0x318a runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6721800 nid=0x318b runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6723000 nid=0x318c runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6724800 nid=0x318d runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6726000 nid=0x318e runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6727400 nid=0x318f runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6728c00 nid=0x3190 runnable 

"VM Periodic Task Thread" prio=10 tid=0x59628c00 nid=0x31b9 waiting on condition 

JNI global references: 259

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 2 zombie test(s): 	at org.apache.phoenix.end2end.QueryTest.testDistinctGroupedAggregation(QueryTest.java:1177)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10618
Updating HBASE-10598
Updating HBASE-8803
Updating HBASE-10612

Build failed in Jenkins: HBase-TRUNK #4960

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4960/changes>

Changes:

[enis] HBASE-10591 Sanity check table configuration in createTable

[ndimiduk] HBASE-10592 Refactor PerformanceEvaluation tool

[mbertozzi] HBASE-10608 Acquire the FS Delegation Token for Secure ExportSnapshot (addendum)

------------------------------------------
[...truncated 4825 lines...]
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.257 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.28 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 133.271 sec <<< FAILURE!
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 133.271 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 133.271 sec <<< FAILURE!

Results :

Tests in error: 
  testAll(org.apache.hadoop.hbase.thrift.TestThriftServer): org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 33, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [2.634s]
[INFO] HBase - Common .................................... SUCCESS [28.317s]
[INFO] HBase - Protocol .................................. SUCCESS [8.962s]
[INFO] HBase - Client .................................... SUCCESS [38.305s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.336s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [2.035s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.833s]
[INFO] HBase - Server .................................... SUCCESS [3:00:31.991s]
[INFO] HBase - Testing Util .............................. SUCCESS [6.170s]
[INFO] HBase - Thrift .................................... FAILURE [2:17.720s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 3:04:26.699s
[INFO] Finished at: Thu Feb 27 02:02:50 UTC 2014
[INFO] Final Memory: 51M/779M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson814437241229081016.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10608
Updating HBASE-10592
Updating HBASE-10591

Build failed in Jenkins: HBase-TRUNK #4959

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4959/changes>

Changes:

[stack] Add link to nice blog post by our Enis on region split/merge into the region split section

------------------------------------------
[...truncated 4648 lines...]
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.402 sec
Running org.apache.hadoop.hbase.rest.TestVersionResource
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.948 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithFilters
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.639 sec
Running org.apache.hadoop.hbase.rest.TestTableResource
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.119 sec
Running org.apache.hadoop.hbase.rest.TestMultiRowResource
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.499 sec
Running org.apache.hadoop.hbase.rest.TestGetAndPutResource
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.193 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithLabels
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.014 sec
Running org.apache.hadoop.hbase.rest.TestDeleteRow
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 25.65 sec
Running org.apache.hadoop.hbase.coprocessor.TestWALObserver
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.028 sec
Running org.apache.hadoop.hbase.coprocessor.TestRowProcessorEndpoint
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.909 sec
Running org.apache.hadoop.hbase.coprocessor.TestOpenTableInCoprocessor
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.729 sec
Running org.apache.hadoop.hbase.coprocessor.TestAggregateProtocol
Tests run: 44, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.944 sec
Running org.apache.hadoop.hbase.coprocessor.TestClassLoading
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.67 sec
Running org.apache.hadoop.hbase.coprocessor.TestCoprocessorEndpoint
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.637 sec
Running org.apache.hadoop.hbase.coprocessor.TestHTableWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.744 sec
Running org.apache.hadoop.hbase.coprocessor.TestBigDecimalColumnInterpreter
Tests run: 38, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.566 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverInterface
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.647 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverBypass
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.741 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerObserver
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.531 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterObserver
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 42.159 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.628 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.967 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.229 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithAbort
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.989 sec
Running org.apache.hadoop.hbase.TestGlobalMemStoreSize
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.439 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.307 sec
Running org.apache.hadoop.hbase.catalog.TestMetaReaderEditorNoCluster
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.141 sec
Running org.apache.hadoop.hbase.catalog.TestMetaMigrationConvertingToPB
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.804 sec
Running org.apache.hadoop.hbase.catalog.TestCatalogTracker
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.094 sec
Running org.apache.hadoop.hbase.catalog.TestMetaReaderEditor
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.036 sec
Running org.apache.hadoop.hbase.regionserver.TestReversibleScanners
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.544 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionMergeTransactionOnCluster
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.045 sec
Running org.apache.hadoop.hbase.regionserver.TestMemStore
Tests run: 23, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.696 sec
Running org.apache.hadoop.hbase.regionserver.TestCompaction
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 92.508 sec
Running org.apache.hadoop.hbase.regionserver.TestStoreScanner
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.671 sec
Running org.apache.hadoop.hbase.regionserver.TestPriorityRpc
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.772 sec
Running org.apache.hadoop.hbase.regionserver.TestParallelPut
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.793 sec
Running org.apache.hadoop.hbase.regionserver.TestClusterId
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.12 sec
Running org.apache.hadoop.hbase.regionserver.TestStore
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.114 sec
Running org.apache.hadoop.hbase.regionserver.TestCacheOnWriteInSchema
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.555 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionBusyWait
Tests run: 66, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 83.451 sec
Running org.apache.hadoop.hbase.regionserver.TestServerCustomProtocol
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 44.226 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLog
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 46.609 sec
Running org.apache.hadoop.hbase.regionserver.TestSplitTransactionOnCluster
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 88.736 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogFiltering
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.922 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogSplitCompressed
Tests run: 33, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 120.222 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRollPeriod
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 66.679 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestWALReplay
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 64.883 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestSecureWALReplay
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.94 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogSplit
Tests run: 33, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 104.96 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestReadOldRootAndMetaEdits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.347 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestSecureHLog
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.733 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestDurability
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.732 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRollAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.358 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestWALReplayCompressed
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 57.415 sec
Running org.apache.hadoop.hbase.regionserver.TestJoinedScanners
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.121 sec
Running org.apache.hadoop.hbase.regionserver.TestGetClosestAtOrBefore
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.155 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionOnCluster
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.706 sec
Running org.apache.hadoop.hbase.regionserver.handler.TestCloseRegionHandler
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.358 sec
Running org.apache.hadoop.hbase.regionserver.handler.TestOpenRegionHandler
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.779 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 44.433 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionFavoredNodes
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.403 sec
Running org.apache.hadoop.hbase.regionserver.TestCompoundBloomFilter
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.882 sec
Running org.apache.hadoop.hbase.regionserver.TestCompactionState
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 57.731 sec
Running org.apache.hadoop.hbase.regionserver.TestStoreFileBlockCacheSummary
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.248 sec
Running org.apache.hadoop.hbase.regionserver.TestBlocksRead
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.112 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRolling
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 476.386 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionServerMetrics
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.696 sec
Running org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 46.19 sec
Running org.apache.hadoop.hbase.regionserver.TestEncryptionRandomKeying
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.443 sec
Running org.apache.hadoop.hbase.regionserver.TestFSErrorsExposed
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 40.702 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegion
Tests run: 65, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 68.17 sec
Running org.apache.hadoop.hbase.regionserver.TestEncryptionKeyRotation
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.113 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionServerNoMaster
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.444 sec
Running org.apache.hadoop.hbase.regionserver.TestSplitLogWorker
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.494 sec
Running org.apache.hadoop.hbase.regionserver.TestCompactionWithCoprocessor
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 96.835 sec
Running org.apache.hadoop.hbase.regionserver.TestAtomicOperation
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 44.89 sec
Running org.apache.hadoop.hbase.regionserver.TestSeekOptimizations
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.55 sec
Running org.apache.hadoop.hbase.regionserver.TestEndToEndSplitTransaction
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.85 sec
Running org.apache.hadoop.hbase.regionserver.TestMasterAddressManager
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.505 sec
Running org.apache.hadoop.hbase.regionserver.TestTags
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.16 sec
Running org.apache.hadoop.hbase.regionserver.TestRSKilledWhenInitializing
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.136 sec
Running org.apache.hadoop.hbase.TestClusterBootOrder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.31 sec

Results :

Tests in error: 
  org.apache.hadoop.hbase.master.TestMasterMetrics: Shutting down

Tests run: 1841, Failures: 0, Errors: 1, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [6.532s]
[INFO] HBase - Common .................................... SUCCESS [30.409s]
[INFO] HBase - Protocol .................................. SUCCESS [9.172s]
[INFO] HBase - Client .................................... SUCCESS [38.223s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.466s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [2.522s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.419s]
[INFO] HBase - Server .................................... FAILURE [2:38:16.155s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:39:52.739s
[INFO] Finished at: Wed Feb 26 22:30:34 UTC 2014
[INFO] Final Memory: 38M/506M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson3441151644633016059.sh
++ grep surefirebooter
++ wc -l
++ jps
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4958

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4958/changes>

Changes:

[nkeywal] HBASE-10606 Bad timeout in RpcRetryingCaller#callWithRetries w/o parameters

------------------------------------------
[...truncated 18380 lines...]
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7ef0a340> (a sun.nio.ch.Util$2)
	- locked <0x7ef0a2c0> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7ef0a070> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener.run(Server.java:439)

"IPC Server Responder" daemon prio=10 tid=0x6ea7f800 nid=0x467f runnable [0x70a7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7ef0e3f0> (a sun.nio.ch.Util$2)
	- locked <0x7ef0e3e0> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7ef0e1e0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.apache.hadoop.ipc.Server$Responder.run(Server.java:605)

"Timer-0" daemon prio=10 tid=0x6f02c000 nid=0x467e in Object.wait() [0x70e68000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x7ef87e98> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"10473802@qtp-31849863-1 - Acceptor0 SelectChannelConnector@localhost:55130" prio=10 tid=0x6f785000 nid=0x4660 runnable [0x7627d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7ef93730> (a sun.nio.ch.Util$2)
	- locked <0x7ef93720> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7ef883f8> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
	at org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
	at org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
	at org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)

"14554797@qtp-31849863-0" prio=10 tid=0x6f784800 nid=0x465f in Object.wait() [0x7647d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626)
	- locked <0x7ef938e0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)

"pool-2-thread-1" prio=10 tid=0x6f78ec00 nid=0x4648 runnable [0x7667d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7ef0ac60> (a sun.nio.ch.Util$2)
	- locked <0x7ef0ac50> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7ef0aa50> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x7ef0ae48> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor@13dd116" daemon prio=10 tid=0x6f3fb000 nid=0x4647 waiting on condition [0x7687d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor@1b8f748" daemon prio=10 tid=0x6f3fa000 nid=0x4646 waiting on condition [0x76a7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor.run(FSNamesystem.java:2716)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@d062fa" daemon prio=10 tid=0x6f3f8800 nid=0x4645 waiting on condition [0x76c7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:368)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor@7e23d1" daemon prio=10 tid=0x6f3f7400 nid=0x4644 waiting on condition [0x76e7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor.run(FSNamesystem.java:2693)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor@11a2ff1" daemon prio=10 tid=0x6f3f6000 nid=0x4643 waiting on condition [0x7707d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:186)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x6f999c00 nid=0x447a waiting on condition [0x77470000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hbase.client.TestFromClientSide.testUpdatesWithMajorCompaction(TestFromClientSide.java:3459)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
	at org.junit.rules.RunRules.evaluate(RunRules.java:20)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runners.Suite.runChild(Suite.java:127)
	at org.junit.runners.Suite.runChild(Suite.java:26)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x70426c00 nid=0x445c runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x70425000 nid=0x445b waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x70423000 nid=0x445a waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x70421400 nid=0x4459 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x709e6c00 nid=0x4458 in Object.wait() [0x77e7d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x7cadce08> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x709e5000 nid=0x4457 in Object.wait() [0x7817d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x7cad5708> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6806800 nid=0x4443 waiting on condition [0xf696a000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x7cadcfd0> (a java.util.concurrent.FutureTask$Sync)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:834)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:994)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1303)
	at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:248)
	at java.util.concurrent.FutureTask.get(FutureTask.java:111)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.waitForCompletion(AsynchronousRunner.java:73)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.finished(AsynchronousRunner.java:58)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:242)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:138)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:61)
	at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:137)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:188)
	at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:166)
	at org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:86)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:101)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:74)

"VM Thread" prio=10 tid=0x70cf8400 nid=0x4456 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6810800 nid=0x4444 runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6812000 nid=0x4445 runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6813800 nid=0x4446 runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6814c00 nid=0x4447 runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6816400 nid=0x4448 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6817c00 nid=0x4449 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6819000 nid=0x444a runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf681a800 nid=0x444b runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf681c000 nid=0x444c runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf681d400 nid=0x444d runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf681ec00 nid=0x444e runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6820400 nid=0x444f runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6821800 nid=0x4450 runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6823000 nid=0x4451 runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6824800 nid=0x4452 runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6826000 nid=0x4453 runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6827400 nid=0x4454 runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6828c00 nid=0x4455 runnable 

"VM Periodic Task Thread" prio=10 tid=0x70428800 nid=0x445d waiting on condition 

JNI global references: 294

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 4 zombie test(s): 	at org.apache.hadoop.hbase.client.TestFromClientSide.testJiraTest861(TestFromClientSide.java:2183)
	at org.apache.hadoop.hbase.client.TestMetaScanner.testConcurrentMetaScannerAndCatalogJanitor(TestMetaScanner.java:235)
	at org.apache.hadoop.hbase.client.TestFromClientSide.testUpdatesWithMajorCompaction(TestFromClientSide.java:3459)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10606

Build failed in Jenkins: HBase-TRUNK #4957

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4957/changes>

Changes:

[nkeywal] HBASE-10566 cleanup rpcTimeout in the client - addendum

[stack] HBASE-10614 Master could not be stopped (Jingcheng Du)

[tedyu] HBASE-10600 HTable#batch() should perform validation on empty Put

------------------------------------------
[...truncated 4650 lines...]
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.134 sec
Running org.apache.hadoop.hbase.security.access.TestNamespaceCommands
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.928 sec
Running org.apache.hadoop.hbase.TestFullLogReconstruction
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.344 sec
Running org.apache.hadoop.hbase.ipc.TestDelayedRpc
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.752 sec
Running org.apache.hadoop.hbase.ipc.TestHBaseClient
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.358 sec
Running org.apache.hadoop.hbase.ipc.TestProtoBufRpc
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.522 sec
Running org.apache.hadoop.hbase.TestGlobalMemStoreSize
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.63 sec
Running org.apache.hadoop.hbase.constraint.TestConstraint
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.529 sec
Running org.apache.hadoop.hbase.rest.TestSchemaResource
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.771 sec
Running org.apache.hadoop.hbase.security.access.TestAccessController
Tests run: 43, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 162.191 sec
Running org.apache.hadoop.hbase.rest.TestTableResource
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.087 sec
Running org.apache.hadoop.hbase.rest.client.TestRemoteTable
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.119 sec
Running org.apache.hadoop.hbase.rest.TestMultiRowResource
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.338 sec
Running org.apache.hadoop.hbase.rest.TestGzipFilter
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.375 sec
Running org.apache.hadoop.hbase.rest.TestResourceFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.036 sec
Running org.apache.hadoop.hbase.rest.TestGetAndPutResource
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.383 sec
Running org.apache.hadoop.hbase.rest.TestDeleteRow
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.029 sec
Running org.apache.hadoop.hbase.rest.TestTableScan
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.17 sec
Running org.apache.hadoop.hbase.rest.TestScannerResource
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.186 sec
Running org.apache.hadoop.hbase.rest.TestVersionResource
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.4 sec
Running org.apache.hadoop.hbase.rest.TestStatusResource
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.305 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithLabels
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.113 sec
Running org.apache.hadoop.hbase.trace.TestHTraceHooks
Running org.apache.hadoop.hbase.rest.TestScannersWithFilters
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.572 sec
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.494 sec
Running org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.801 sec
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.614 sec
Running org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.358 sec
Running org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Tests run: 36, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 80.622 sec
Running org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 62.636 sec
Running org.apache.hadoop.hbase.io.hfile.slab.TestSlabCache
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.023 sec
Running org.apache.hadoop.hbase.io.hfile.slab.TestSingleSizeCache
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.176 sec
Running org.apache.hadoop.hbase.io.TestFileLink
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.891 sec
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Tests run: 28, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 100.11 sec
Running org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 90.158 sec
Running org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 40, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 183.536 sec
Running org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.136 sec
Running org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 25.056 sec
Running org.apache.hadoop.hbase.TestRegionRebalancing
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.888 sec
Running org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 191.16 sec
Running org.apache.hadoop.hbase.mapred.TestTableInputFormat
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.673 sec
Running org.apache.hadoop.hbase.mapred.TestTableMapReduceUtil
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 103.302 sec
Running org.apache.hadoop.hbase.replication.TestReplicationTrackerZKImpl
Tests run: 4, Failures: 0, Errors: 0, Skipped: 2, Time elapsed: 1.808 sec
Running org.apache.hadoop.hbase.replication.TestReplicationDisableInactivePeer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 48.465 sec
Running org.apache.hadoop.hbase.mapred.TestTableMapReduce
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 124.407 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRSCompressed
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.079 sec
Running org.apache.hadoop.hbase.replication.TestPerTableCFReplication
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.508 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSink
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.166 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSourceManager
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.739 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSource
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.222 sec
Running org.apache.hadoop.hbase.replication.TestReplicationChangingPeerRegionservers
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.367 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSyncUpTool
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.152 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationHLogReaderManager
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 223.358 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 48.634 sec
Running org.apache.hadoop.hbase.replication.TestReplicationWithTags
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.44 sec
Running org.apache.hadoop.hbase.replication.TestMultiSlaveReplication
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.456 sec
Running org.apache.hadoop.hbase.replication.TestReplicationStateZKImpl
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.535 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillSlaveRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 60.329 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSmallTests
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 69.343 sec
Running org.apache.hadoop.hbase.util.TestMergeTool
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.317 sec
Running org.apache.hadoop.hbase.util.TestFSTableDescriptors
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.342 sec
Running org.apache.hadoop.hbase.util.TestHBaseFsckEncryption
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.695 sec
Running org.apache.hadoop.hbase.replication.TestMasterReplication
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 80.163 sec
Running org.apache.hadoop.hbase.util.TestIdLock
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.528 sec
Running org.apache.hadoop.hbase.util.TestMergeTable
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.487 sec
Running org.apache.hadoop.hbase.util.TestMiniClusterLoadSequential
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 59.855 sec
Running org.apache.hadoop.hbase.util.TestMiniClusterLoadParallel
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 53.656 sec
Running org.apache.hadoop.hbase.util.TestFSUtils
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.238 sec
Running org.apache.hadoop.hbase.util.TestRegionSplitter
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.606 sec
Running org.apache.hadoop.hbase.util.TestProcessBasedCluster
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.649 sec
Running org.apache.hadoop.hbase.util.TestMiniClusterLoadEncoded
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 61.777 sec
Running org.apache.hadoop.hbase.util.TestCoprocessorScanPolicy
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.988 sec
Running org.apache.hadoop.hbase.util.TestFSVisitor
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.13 sec
Running org.apache.hadoop.hbase.util.hbck.TestOfflineMetaRebuildBase
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.41 sec
Running org.apache.hadoop.hbase.util.hbck.TestOfflineMetaRebuildOverlap
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 25.483 sec
Running org.apache.hadoop.hbase.util.hbck.TestOfflineMetaRebuildHole
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.696 sec
Running org.apache.hadoop.hbase.util.TestDefaultEnvironmentEdge
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.377 sec
Running org.apache.hadoop.hbase.util.TestFSHDFSUtils
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.279 sec
Running org.apache.hadoop.hbase.migration.TestNamespaceUpgrade
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 47.128 sec
Running org.apache.hadoop.hbase.migration.TestUpgradeTo96
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.959 sec
Running org.apache.hadoop.hbase.TestNamespace
Tests run: 9, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 30.119 sec
Running org.apache.hadoop.hbase.TestInfoServers
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.325 sec
Running org.apache.hadoop.hbase.util.TestHBaseFsck
Tests run: 40, Failures: 1, Errors: 0, Skipped: 1, Time elapsed: 373.779 sec <<< FAILURE!

Results :

Failed tests:   testSplitDaughtersNotInMeta(org.apache.hadoop.hbase.util.TestHBaseFsck)

Tests run: 1842, Failures: 1, Errors: 0, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [12.438s]
[INFO] HBase - Common .................................... SUCCESS [35.258s]
[INFO] HBase - Protocol .................................. SUCCESS [8.767s]
[INFO] HBase - Client .................................... SUCCESS [38.460s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.559s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [2.201s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.674s]
[INFO] HBase - Server .................................... FAILURE [2:16:20.945s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:18:08.072s
[INFO] Finished at: Wed Feb 26 15:24:49 UTC 2014
[INFO] Final Memory: 38M/534M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to /x1/jenkins/jenkins-slave/workspace/HBase-TRUNK/trunk/hbase-server/target/surefire-reports for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson2886889436664242477.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10566
Updating HBASE-10600
Updating HBASE-10614

Build failed in Jenkins: HBase-TRUNK #4956

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4956/>

------------------------------------------
Started by an SCM change
Started by an SCM change
Building remotely on ubuntu2 in workspace <https://builds.apache.org/job/HBase-TRUNK/ws/>
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@1bb9147f:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.GeneratedMethodAccessor350.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@1bb9147f:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.GeneratedMethodAccessor350.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@1bb9147f:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.GeneratedMethodAccessor350.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson5100696741431418595.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
1906: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results