You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@hbase.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2014/02/28 13:36:54 UTC

Build failed in Jenkins: HBase-TRUNK #4965

See <https://builds.apache.org/job/HBase-TRUNK/4965/changes>

Changes:

[mbertozzi] HBASE-10631 Avoid extra seek on FileLink open

------------------------------------------
[...truncated 4823 lines...]
[INFO] --- maven-source-plugin:2.2.1:jar-no-fork (attach-sources) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.266 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.173 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 116.222 sec <<< FAILURE!
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 116.222 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 116.222 sec <<< FAILURE!

Results :

Tests in error: 
  testAll(org.apache.hadoop.hbase.thrift.TestThriftServer): org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 33, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [2.870s]
[INFO] HBase - Common .................................... SUCCESS [24.681s]
[INFO] HBase - Protocol .................................. SUCCESS [10.118s]
[INFO] HBase - Client .................................... SUCCESS [38.465s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.385s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [1.954s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.441s]
[INFO] HBase - Server .................................... SUCCESS [2:25:26.609s]
[INFO] HBase - Testing Util .............................. SUCCESS [2.406s]
[INFO] HBase - Thrift .................................... FAILURE [2:00.860s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:28:57.557s
[INFO] Finished at: Fri Feb 28 12:36:03 UTC 2014
[INFO] Final Memory: 55M/898M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson490619549740657010.sh
++ grep surefirebooter
++ wc -l
++ jps
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10631

Jenkins build is back to normal : HBase-TRUNK #4975

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4975/changes>


Build failed in Jenkins: HBase-TRUNK #4974

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4974/changes>

Changes:

[enis] HBASE-10635 thrift#TestThriftServer fails due to TTL validity check (Addendum patch to TestThriftServerCmdLine)

[tedyu] HBASE-10609 Remove filterKeyValue(Cell ignored) from FilterBase

------------------------------------------
[...truncated 4654 lines...]
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRSCompressed
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 67.71 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillSlaveRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 66.872 sec
Running org.apache.hadoop.hbase.replication.TestMasterReplication
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 91.731 sec
Running org.apache.hadoop.hbase.replication.TestReplicationStateZKImpl
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.135 sec
Running org.apache.hadoop.hbase.replication.TestReplicationChangingPeerRegionservers
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.706 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSyncUpTool
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 89.432 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSourceManager
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.825 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationSink
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.861 sec
Running org.apache.hadoop.hbase.replication.TestReplicationWithTags
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.234 sec
Running org.apache.hadoop.hbase.replication.TestMultiSlaveReplication
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 41.719 sec
Running org.apache.hadoop.hbase.replication.TestReplicationDisableInactivePeer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.463 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSource
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.952 sec
Running org.apache.hadoop.hbase.replication.TestReplicationKillMasterRS
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 68.805 sec
Running org.apache.hadoop.hbase.replication.regionserver.TestReplicationHLogReaderManager
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 274.665 sec
Running org.apache.hadoop.hbase.replication.TestReplicationSmallTests
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 84.447 sec
Running org.apache.hadoop.hbase.replication.TestPerTableCFReplication
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 59.134 sec
Running org.apache.hadoop.hbase.TestLocalHBaseCluster
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.713 sec
Running org.apache.hadoop.hbase.fs.TestBlockReorder
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.916 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.985 sec
Running org.apache.hadoop.hbase.TestAcidGuarantees
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 111.505 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterObserver
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 32.361 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverInterface
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.613 sec
Running org.apache.hadoop.hbase.coprocessor.TestCoprocessorEndpoint
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.6 sec
Running org.apache.hadoop.hbase.coprocessor.TestWALObserver
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.657 sec
Running org.apache.hadoop.hbase.coprocessor.TestHTableWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.268 sec
Running org.apache.hadoop.hbase.coprocessor.TestBigDecimalColumnInterpreter
Tests run: 38, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.884 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverBypass
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.731 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithAbort
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.054 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerObserver
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.09 sec
Running org.apache.hadoop.hbase.coprocessor.TestOpenTableInCoprocessor
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.111 sec
Running org.apache.hadoop.hbase.coprocessor.TestAggregateProtocol
Tests run: 44, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.46 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.526 sec
Running org.apache.hadoop.hbase.coprocessor.TestClassLoading
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.649 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.211 sec
Running org.apache.hadoop.hbase.coprocessor.TestRowProcessorEndpoint
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.062 sec
Running org.apache.hadoop.hbase.constraint.TestConstraint
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.043 sec
Running org.apache.hadoop.hbase.procedure.TestZKProcedureControllers
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.12 sec
Running org.apache.hadoop.hbase.procedure.TestZKProcedure
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.972 sec
Running org.apache.hadoop.hbase.snapshot.TestRestoreFlushSnapshotFromClient
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 106.86 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 543.177 sec
Running org.apache.hadoop.hbase.snapshot.TestSecureExportSnapshot
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 288.362 sec
Running org.apache.hadoop.hbase.snapshot.TestSnapshotDescriptionUtils
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.92 sec
Running org.apache.hadoop.hbase.snapshot.TestFlushSnapshotFromClient
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.443 sec
Running org.apache.hadoop.hbase.mapred.TestTableMapReduce
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 149.966 sec
Running org.apache.hadoop.hbase.mapred.TestTableMapReduceUtil
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 131.127 sec
Running org.apache.hadoop.hbase.snapshot.TestExportSnapshot
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 319.593 sec
Running org.apache.hadoop.hbase.mapred.TestTableInputFormat
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.602 sec
Running org.apache.hadoop.hbase.master.TestOpenedRegionHandler
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.419 sec
Running org.apache.hadoop.hbase.master.TestActiveMasterManager
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.993 sec
Running org.apache.hadoop.hbase.master.TestAssignmentManager
Tests run: 19, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.178 sec
Running org.apache.hadoop.hbase.master.TestMasterRestartAfterDisablingTable
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.034 sec
Running org.apache.hadoop.hbase.master.TestMasterFailoverBalancerPersistence
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.553 sec
Running org.apache.hadoop.hbase.master.TestMasterFailover
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 70.316 sec
Running org.apache.hadoop.hbase.master.TestRollingRestart
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 41.95 sec
Running org.apache.hadoop.hbase.master.TestMasterTransitions
Tests run: 3, Failures: 0, Errors: 0, Skipped: 3, Time elapsed: 26.411 sec
Running org.apache.hadoop.hbase.master.TestRestartCluster
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.262 sec
Running org.apache.hadoop.hbase.master.TestMasterMetrics
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.966 sec
Running org.apache.hadoop.hbase.master.TestMasterStatusServlet
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.673 sec
Running org.apache.hadoop.hbase.master.TestHMasterRPCException
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.942 sec
Running org.apache.hadoop.hbase.master.TestTableLockManager
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 117.54 sec
Running org.apache.hadoop.hbase.master.TestRegionPlacement
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.533 sec
Running org.apache.hadoop.hbase.master.TestAssignmentManagerOnCluster
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.027 sec
Running org.apache.hadoop.hbase.master.balancer.TestBaseLoadBalancer
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.079 sec
Running org.apache.hadoop.hbase.master.balancer.TestStochasticLoadBalancer
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 103.076 sec
Running org.apache.hadoop.hbase.master.balancer.TestDefaultLoadBalancer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.459 sec
Running org.apache.hadoop.hbase.master.TestMaster
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.29 sec
Running org.apache.hadoop.hbase.master.cleaner.TestHFileCleaner
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.05 sec
Running org.apache.hadoop.hbase.master.cleaner.TestLogsCleaner
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.469 sec
Running org.apache.hadoop.hbase.master.cleaner.TestSnapshotFromMaster
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 40.07 sec
Running org.apache.hadoop.hbase.master.TestMasterShutdown
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.244 sec
Running org.apache.hadoop.hbase.master.TestMasterFileSystem
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.112 sec
Running org.apache.hadoop.hbase.master.handler.TestTableDeleteFamilyHandler
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.912 sec
Running org.apache.hadoop.hbase.master.TestDistributedLogSplitting
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 301.821 sec
Running org.apache.hadoop.hbase.master.handler.TestCreateTableHandler
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.536 sec
Running org.apache.hadoop.hbase.master.handler.TestTableDescriptorModification
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.188 sec
Running org.apache.hadoop.hbase.master.TestClusterStatusPublisher
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.584 sec
Running org.apache.hadoop.hbase.master.snapshot.TestSnapshotFileCache
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.751 sec
Running org.apache.hadoop.hbase.master.TestMasterMetricsWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.98 sec
Running org.apache.hadoop.hbase.master.TestMasterNoCluster
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.429 sec
Running org.apache.hadoop.hbase.master.TestDeadServer
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.587 sec
Running org.apache.hadoop.hbase.master.TestZKBasedOpenCloseRegion
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.156 sec
Running org.apache.hadoop.hbase.master.TestSplitLogManager
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 87.025 sec

Results :

Failed tests:   testFavoredNodes(org.apache.hadoop.hbase.regionserver.TestRegionFavoredNodes): Block location 127.0.0.1:36893 not a favored node

Tests run: 1828, Failures: 1, Errors: 0, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [6.812s]
[INFO] HBase - Common .................................... SUCCESS [34.780s]
[INFO] HBase - Protocol .................................. SUCCESS [11.644s]
[INFO] HBase - Client .................................... SUCCESS [40.191s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.842s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [2.763s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [4.490s]
[INFO] HBase - Server .................................... FAILURE [2:32:14.436s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:34:02.003s
[INFO] Finished at: Tue Mar 04 03:36:55 UTC 2014
[INFO] Final Memory: 37M/845M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson2450927639234625773.sh
++ grep surefirebooter
++ wc -l
++ jps
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10609
Updating HBASE-10635

Build failed in Jenkins: HBase-TRUNK #4973

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4973/changes>

Changes:

[enis] HBASE-10632 Region lost in limbo after ArrayIndexOutOfBoundsException during assignment

[jmhsieh] HBASE-10639 Unload script displays wrong counts (off by one) when unloading regions (Srikanth Srungarapu)

------------------------------------------
[...truncated 4826 lines...]
[INFO] --- maven-source-plugin:2.2.1:jar-no-fork (attach-sources) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.411 sec
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.046 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.343 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 170.028 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 170.028 sec <<< FAILURE!

Results :

Tests in error: 
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 32, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [8.471s]
[INFO] HBase - Common .................................... SUCCESS [39.507s]
[INFO] HBase - Protocol .................................. SUCCESS [12.938s]
[INFO] HBase - Client .................................... SUCCESS [42.210s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [6.178s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [3.287s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [5.819s]
[INFO] HBase - Server .................................... SUCCESS [2:37:43.705s]
[INFO] HBase - Testing Util .............................. SUCCESS [7.578s]
[INFO] HBase - Thrift .................................... FAILURE [2:57.311s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:42:48.187s
[INFO] Finished at: Mon Mar 03 23:25:37 UTC 2014
[INFO] Final Memory: 53M/832M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson372105742896416834.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10639
Updating HBASE-10632

Build failed in Jenkins: HBase-TRUNK #4972

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4972/changes>

Changes:

[enis] HBASE-10635 thrift#TestThriftServer fails due to TTL validity check

------------------------------------------
[...truncated 5194 lines...]
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xdf9496c8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-9-thread-1" prio=10 tid=0xa7f6e800 nid=0x269a waiting on condition [0xa6b18000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xde6fa910> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-8-thread-1" prio=10 tid=0xa7f6c400 nid=0x2695 waiting on condition [0xa6c5c000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xde6daf70> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-7-thread-1" prio=10 tid=0xa7f69400 nid=0x2683 waiting on condition [0xa815c000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xde654198> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"RMI RenewClean-[140.211.11.27:48196]" daemon prio=10 tid=0xa7c0a000 nid=0x2679 in Object.wait() [0xadd7d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xde36ed40> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0xde36ed40> (a java.lang.ref.ReferenceQueue$Lock)
	at sun.rmi.transport.DGCClient$EndpointEntry$RenewCleanThread.run(DGCClient.java:535)
	at java.lang.Thread.run(Thread.java:724)

"RMI Scheduler(0)" daemon prio=10 tid=0xa781ac00 nid=0x2678 waiting on condition [0xad57d000]
   java.lang.Thread.State: TIMED_WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb37ee460> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
	at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1090)
	at java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:807)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"GC Daemon" daemon prio=10 tid=0xa7804000 nid=0x2676 in Object.wait() [0xae57d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xb3ae4ef0> (a sun.misc.GC$LatencyLock)
	at sun.misc.GC$Daemon.run(GC.java:117)
	- locked <0xb3ae4ef0> (a sun.misc.GC$LatencyLock)

"RMI TCP Accept-0" daemon prio=10 tid=0xa7802c00 nid=0x2674 runnable [0xa967d000]
   java.lang.Thread.State: RUNNABLE
	at java.net.PlainSocketImpl.socketAccept(Native Method)
	at java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:398)
	at java.net.ServerSocket.implAccept(ServerSocket.java:530)
	at java.net.ServerSocket.accept(ServerSocket.java:498)
	at sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:387)
	at sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:359)
	at java.lang.Thread.run(Thread.java:724)

"ActiveMQ Connection Executor: tcp://localhost/127.0.0.1:13103@38391" prio=10 tid=0xa7600800 nid=0x2665 waiting on condition [0xae37d000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb3ae55a8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"ActiveMQ Connection Executor: tcp://localhost/127.0.0.1:13103@38390" prio=10 tid=0xa7900800 nid=0x2662 waiting on condition [0xad97d000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb3b19c30> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"ActiveMQ Connection Executor: tcp://localhost/127.0.0.1:13101@55652" prio=10 tid=0xa8000c00 nid=0x2450 waiting on condition [0xadb7d000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xb3b401b8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0xa9726c00 nid=0x23cc runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0xa9725000 nid=0x23cb waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0xa9723000 nid=0x23ca waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0xa9721400 nid=0x23c9 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0xa9ce6c00 nid=0x23c7 in Object.wait() [0xaed7d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xb3aeee50> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0xb3aeee50> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0xa9ce5000 nid=0x23c6 in Object.wait() [0xaf07d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0xb3ae4ee8> (a java.lang.ref.Reference$Lock)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0xb3ae4ee8> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6806800 nid=0x23ae waiting on condition [0xf695b000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0xe04b6388> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:374)
	at org.apache.activemq.transport.FutureResponse.getResult(FutureResponse.java:40)
	at org.apache.activemq.transport.ResponseCorrelator.request(ResponseCorrelator.java:87)
	at org.apache.activemq.ActiveMQConnection.syncSendPacket(ActiveMQConnection.java:1394)
	at org.apache.activemq.ActiveMQConnection.ensureConnectionInfoSent(ActiveMQConnection.java:1510)
	- locked <0xe04b5420> (a java.lang.Object)
	at org.apache.activemq.ActiveMQConnection.start(ActiveMQConnection.java:524)
	at org.apache.cxf.transport.jms.util.TestReceiver.drainQueue(TestReceiver.java:71)
	at org.apache.cxf.transport.jms.util.TestReceiver.runAsync(TestReceiver.java:120)
	at org.apache.cxf.transport.jms.uri.URIConfiguredConduitTest.sendAndReceive(URIConfiguredConduitTest.java:84)
	at org.apache.cxf.transport.jms.uri.URIConfiguredConduitTest.testSendReceive(URIConfiguredConduitTest.java:63)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
	at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
	at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
	at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)

"VM Thread" prio=10 tid=0xa9ff8400 nid=0x23c5 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6810800 nid=0x23b0 runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6812000 nid=0x23b1 runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6813800 nid=0x23b2 runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6814c00 nid=0x23b3 runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6816400 nid=0x23b4 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6817c00 nid=0x23b5 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6819000 nid=0x23b6 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf681a800 nid=0x23b8 runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf681c000 nid=0x23b9 runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf681d400 nid=0x23ba runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf681ec00 nid=0x23bd runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6820400 nid=0x23be runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6821800 nid=0x23bf runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6823000 nid=0x23c0 runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6824800 nid=0x23c1 runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6825c00 nid=0x23c2 runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6827400 nid=0x23c3 runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6828c00 nid=0x23c4 runnable 

"VM Periodic Task Thread" prio=10 tid=0xa9730c00 nid=0x23cd waiting on condition 

JNI global references: 207

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 1 zombie test(s): 	at org.apache.cxf.transport.jms.uri.URIConfiguredConduitTest.testSendReceive(URIConfiguredConduitTest.java:63)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10635

Build failed in Jenkins: HBase-TRUNK #4971

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4971/>

------------------------------------------
Started by an SCM change
Building remotely on ubuntu2 in workspace <https://builds.apache.org/job/HBase-TRUNK/ws/>
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@72cf4805:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@72cf4805:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@72cf4805:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson5442153936504093802.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
1906: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4970

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4970/changes>

Changes:

[anoopsamjohn] HBASE-10451 Enable back Tag compression on HFiles.(Anoop)

------------------------------------------
[...truncated 4673 lines...]
Running org.apache.hadoop.hbase.rest.client.TestRemoteTable
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.929 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithFilters
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.46 sec
Running org.apache.hadoop.hbase.rest.TestTableResource
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.567 sec
Running org.apache.hadoop.hbase.rest.TestGetAndPutResource
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.916 sec
Running org.apache.hadoop.hbase.rest.TestMultiRowResource
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.941 sec
Running org.apache.hadoop.hbase.rest.TestScannersWithLabels
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.726 sec
Running org.apache.hadoop.hbase.rest.TestDeleteRow
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.195 sec
Running org.apache.hadoop.hbase.coprocessor.TestWALObserver
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.367 sec
Running org.apache.hadoop.hbase.coprocessor.TestRowProcessorEndpoint
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.331 sec
Running org.apache.hadoop.hbase.coprocessor.TestOpenTableInCoprocessor
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.739 sec
Running org.apache.hadoop.hbase.coprocessor.TestClassLoading
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.416 sec
Running org.apache.hadoop.hbase.coprocessor.TestAggregateProtocol
Tests run: 44, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.638 sec
Running org.apache.hadoop.hbase.coprocessor.TestCoprocessorEndpoint
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.884 sec
Running org.apache.hadoop.hbase.coprocessor.TestHTableWrapper
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.603 sec
Running org.apache.hadoop.hbase.coprocessor.TestBigDecimalColumnInterpreter
Tests run: 38, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.948 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverInterface
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 60.519 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverBypass
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.178 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerObserver
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.549 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterObserver
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.161 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.008 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.901 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithRemove
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.841 sec
Running org.apache.hadoop.hbase.coprocessor.TestRegionServerCoprocessorExceptionWithAbort
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.365 sec
Running org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.515 sec
Running org.apache.hadoop.hbase.TestGlobalMemStoreSize
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.137 sec
Running org.apache.hadoop.hbase.catalog.TestMetaReaderEditorNoCluster
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.28 sec
Running org.apache.hadoop.hbase.catalog.TestMetaMigrationConvertingToPB
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.38 sec
Running org.apache.hadoop.hbase.catalog.TestCatalogTracker
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.943 sec
Running org.apache.hadoop.hbase.catalog.TestMetaReaderEditor
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.756 sec
Running org.apache.hadoop.hbase.regionserver.TestReversibleScanners
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.529 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionMergeTransactionOnCluster
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.654 sec
Running org.apache.hadoop.hbase.regionserver.TestCompaction
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 92.916 sec
Running org.apache.hadoop.hbase.regionserver.TestStoreScanner
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.697 sec
Running org.apache.hadoop.hbase.regionserver.TestMemStore
Tests run: 23, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.258 sec
Running org.apache.hadoop.hbase.regionserver.TestPriorityRpc
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.86 sec
Running org.apache.hadoop.hbase.regionserver.TestParallelPut
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.342 sec
Running org.apache.hadoop.hbase.regionserver.TestClusterId
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.99 sec
Running org.apache.hadoop.hbase.regionserver.TestStore
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.844 sec
Running org.apache.hadoop.hbase.regionserver.TestCacheOnWriteInSchema
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.549 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionBusyWait
Tests run: 66, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 69.534 sec
Running org.apache.hadoop.hbase.regionserver.TestServerCustomProtocol
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.575 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLog
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.528 sec
Running org.apache.hadoop.hbase.regionserver.TestSplitTransactionOnCluster
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.907 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogFiltering
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.934 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRollPeriod
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 73.445 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogSplitCompressed
Tests run: 33, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 132.406 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestWALReplay
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 72.335 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestSecureWALReplay
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.715 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestHLogSplit
Tests run: 33, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 119.409 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestReadOldRootAndMetaEdits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.341 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestSecureHLog
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.759 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestDurability
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.986 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRollAbort
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.171 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestWALReplayCompressed
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 60.425 sec
Running org.apache.hadoop.hbase.regionserver.TestJoinedScanners
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.079 sec
Running org.apache.hadoop.hbase.regionserver.TestGetClosestAtOrBefore
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.831 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionOnCluster
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.723 sec
Running org.apache.hadoop.hbase.regionserver.handler.TestCloseRegionHandler
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.678 sec
Running org.apache.hadoop.hbase.regionserver.handler.TestOpenRegionHandler
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.767 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegionServerBulkLoad
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 45.605 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionFavoredNodes
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.714 sec
Running org.apache.hadoop.hbase.regionserver.TestCompoundBloomFilter
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.733 sec
Running org.apache.hadoop.hbase.regionserver.TestCompactionState
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 56.44 sec
Running org.apache.hadoop.hbase.regionserver.TestStoreFileBlockCacheSummary
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.429 sec
Running org.apache.hadoop.hbase.regionserver.TestBlocksRead
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.112 sec
Running org.apache.hadoop.hbase.regionserver.wal.TestLogRolling
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 477.864 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionServerMetrics
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 17.595 sec
Running org.apache.hadoop.hbase.regionserver.TestMultiColumnScanner
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 44.538 sec
Running org.apache.hadoop.hbase.regionserver.TestEncryptionRandomKeying
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.139 sec
Running org.apache.hadoop.hbase.regionserver.TestFSErrorsExposed
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.861 sec
Running org.apache.hadoop.hbase.regionserver.TestEncryptionKeyRotation
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.686 sec
Running org.apache.hadoop.hbase.regionserver.TestHRegion
Tests run: 65, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 68.217 sec
Running org.apache.hadoop.hbase.regionserver.TestRegionServerNoMaster
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.753 sec
Running org.apache.hadoop.hbase.regionserver.TestSplitLogWorker
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.851 sec
Running org.apache.hadoop.hbase.regionserver.TestCompactionWithCoprocessor
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 96.341 sec
Running org.apache.hadoop.hbase.regionserver.TestAtomicOperation
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.127 sec
Running org.apache.hadoop.hbase.regionserver.TestSeekOptimizations
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.836 sec
Running org.apache.hadoop.hbase.regionserver.TestTags
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.336 sec
Running org.apache.hadoop.hbase.regionserver.TestEndToEndSplitTransaction
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 37.749 sec
Running org.apache.hadoop.hbase.regionserver.TestMasterAddressManager
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.527 sec
Running org.apache.hadoop.hbase.regionserver.TestRSKilledWhenInitializing
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.883 sec
Running org.apache.hadoop.hbase.TestClusterBootOrder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 31.927 sec

Results :

Tests in error: 
  org.apache.hadoop.hbase.master.TestMasterMetrics: Shutting down

Tests run: 1820, Failures: 0, Errors: 1, Skipped: 24

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [4.229s]
[INFO] HBase - Common .................................... SUCCESS [39.552s]
[INFO] HBase - Protocol .................................. SUCCESS [12.994s]
[INFO] HBase - Client .................................... SUCCESS [38.519s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [5.773s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [3.276s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [3.529s]
[INFO] HBase - Server .................................... FAILURE [2:34:07.777s]
[INFO] HBase - Testing Util .............................. SKIPPED
[INFO] HBase - Thrift .................................... SKIPPED
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:35:56.454s
[INFO] Finished at: Sat Mar 01 17:13:00 UTC 2014
[INFO] Final Memory: 37M/860M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson3773190383010929138.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10451

Build failed in Jenkins: HBase-TRUNK #4969

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4969/>

------------------------------------------
[...truncated 3285 lines...]
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)

"main.logSyncer" daemon prio=10 tid=0x57362800 nid=0x41f5 in Object.wait() [0x54894000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x6861ff30> (a java.util.concurrent.atomic.AtomicBoolean)
	at org.apache.hadoop.hbase.regionserver.wal.HLog$LogSyncer.run(HLog.java:1256)
	- locked <0x6861ff30> (a java.util.concurrent.atomic.AtomicBoolean)
	at java.lang.Thread.run(Thread.java:724)

"testReplayEditsWrittenViaHRegion-server-1234-index-writer-pool-31-thread-1-EventThread" daemon prio=10 tid=0x59dd9800 nid=0x41c0 waiting on condition [0x546ff000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x6861ffd8> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043)
	at java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:491)

"testReplayEditsWrittenViaHRegion-server-1234-index-writer-pool-31-thread-1-SendThread(localhost:53471)" daemon prio=10 tid=0x59dd8800 nid=0x41bf waiting on condition [0x54750000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.zookeeper.client.StaticHostProvider.next(StaticHostProvider.java:86)
	at org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:937)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)

"LeaseChecker" daemon prio=10 tid=0x57b18400 nid=0x40c2 waiting on condition [0x57969000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.ipc.Client$Connection.handleConnectionFailure(Client.java:663)
	at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:456)
	- locked <0x6d63d958> (a org.apache.hadoop.ipc.Client$Connection)
	at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:560)
	- locked <0x6d63d958> (a org.apache.hadoop.ipc.Client$Connection)
	at org.apache.hadoop.ipc.Client$Connection.access$2000(Client.java:184)
	at org.apache.hadoop.ipc.Client.getConnection(Client.java:1206)
	at org.apache.hadoop.ipc.Client.call(Client.java:1050)
	at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
	at com.sun.proxy.$Proxy16.renewLease(Unknown Source)
	at sun.reflect.GeneratedMethodAccessor44.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
	at com.sun.proxy.$Proxy16.renewLease(Unknown Source)
	at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.renew(DFSClient.java:1359)
	at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.run(DFSClient.java:1371)
	at java.lang.Thread.run(Thread.java:724)

"pool-4-thread-1" prio=10 tid=0x59b61c00 nid=0x40b1 runnable [0x57ef8000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x68575af8> (a sun.nio.ch.Util$2)
	- locked <0x68575ae8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x685758d8> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x685758c0> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-3-thread-1" prio=10 tid=0xf6758800 nid=0x4095 runnable [0x583b7000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x685bde78> (a sun.nio.ch.Util$2)
	- locked <0x685bde68> (a java.util.Collections$UnmodifiableSet)
	- locked <0x685bdc58> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x685bdc40> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-2-thread-1" prio=10 tid=0x5a72c400 nid=0x4079 runnable [0x58876000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x66fd4870> (a sun.nio.ch.Util$2)
	- locked <0x66fd4860> (a java.util.Collections$UnmodifiableSet)
	- locked <0x66fd4650> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x66fd4638> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x58e42800 nid=0x4053 runnable [0x59e5c000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x66f47400> (a sun.nio.ch.Util$2)
	- locked <0x66f47410> (a java.util.Collections$UnmodifiableSet)
	- locked <0x66f473c0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:333)
	- locked <0x66f47420> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x5b1d8400 nid=0x404a runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x5b1d6400 nid=0x4049 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x5b1d4400 nid=0x4048 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x5b1d2c00 nid=0x4047 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x5b198000 nid=0x4046 in Object.wait() [0x6207d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x66c93238> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x66c93238> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x5b196400 nid=0x4045 in Object.wait() [0x6227d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x66c96690> (a java.lang.ref.Reference$Lock)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x66c96690> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6706800 nid=0x403b runnable [0xf6937000]
   java.lang.Thread.State: RUNNABLE
	at java.io.FileOutputStream.writeBytes(Native Method)
	at java.io.FileOutputStream.write(FileOutputStream.java:318)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	- locked <0x66caab70> (a java.io.BufferedOutputStream)
	at java.io.PrintStream.write(PrintStream.java:482)
	- locked <0x66caab50> (a java.io.PrintStream)
	at org.apache.maven.surefire.booter.ForkingRunListener.writeTestOutput(ForkingRunListener.java:178)
	- locked <0x66caab50> (a java.io.PrintStream)
	at org.apache.maven.surefire.report.ConsoleOutputCapture$ForwardingPrintStream.write(ConsoleOutputCapture.java:58)
	at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221)
	at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291)
	at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295)
	at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141)
	- locked <0x66ca85e0> (a java.io.OutputStreamWriter)
	at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229)
	at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59)
	at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324)
	at org.apache.log4j.WriterAppender.append(WriterAppender.java:162)
	at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251)
	- locked <0x66ca7ed0> (a org.apache.log4j.ConsoleAppender)
	at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66)
	at org.apache.log4j.Category.callAppenders(Category.java:206)
	- locked <0x66ca4290> (a org.apache.log4j.spi.RootLogger)
	at org.apache.log4j.Category.forcedLog(Category.java:391)
	at org.apache.log4j.Category.log(Category.java:856)
	at org.slf4j.impl.Log4jLoggerAdapter.debug(Log4jLoggerAdapter.java:207)
	at org.mortbay.log.Slf4jLog.debug(Slf4jLog.java:40)
	at org.mortbay.log.Log.debug(Log.java:105)
	at org.mortbay.resource.JarResource.extract(JarResource.java:184)
	at org.mortbay.jetty.webapp.WebAppContext.resolveWebApp(WebAppContext.java:974)
	at org.mortbay.jetty.webapp.WebAppContext.getWebInf(WebAppContext.java:832)
	at org.mortbay.jetty.webapp.WebInfConfiguration.configureClassLoader(WebInfConfiguration.java:62)
	at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:489)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	- locked <0xc87593d0> (a java.lang.Object)
	at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152)
	at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	- locked <0xc8751860> (a java.lang.Object)
	at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130)
	at org.mortbay.jetty.Server.doStart(Server.java:224)
	at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
	- locked <0xc873eb98> (a java.lang.Object)
	at org.apache.hadoop.http.HttpServer.start(HttpServer.java:617)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:481)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:299)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
	at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1496)
	at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:417)
	- locked <0x6d971530> (a org.apache.hadoop.hdfs.MiniDFSCluster)
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
	at org.apache.phoenix.hbase.index.covered.TestEndToEndCoveredColumnsIndexBuilder.setupCluster(TestEndToEndCoveredColumnsIndexBuilder.java:106)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
	at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
	at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
	at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)

"VM Thread" prio=10 tid=0x5b190c00 nid=0x4044 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6710800 nid=0x403c runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6712000 nid=0x403d runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6713400 nid=0x403e runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6714c00 nid=0x403f runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6716400 nid=0x4040 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6717800 nid=0x4041 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6719000 nid=0x4042 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf671a800 nid=0x4043 runnable 

"VM Periodic Task Thread" prio=10 tid=0x5b1e2400 nid=0x404b waiting on condition 

JNI global references: 398

POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4968

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4968/changes>

Changes:

[tedyu] HBASE-10644 TestSecureExportSnapshot#testExportFileSystemState fails on hadoop-1

------------------------------------------
[...truncated 9365 lines...]
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1408)

"IPC Server listener on 37626" daemon prio=10 tid=0x6ba42c00 nid=0x1b45 runnable [0x6db7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7efcabd0> (a sun.nio.ch.Util$2)
	- locked <0x7efcab50> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7efca900> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener.run(Server.java:479)

"IPC Server Responder" daemon prio=10 tid=0x6ba3b400 nid=0x1b44 runnable [0x6dd7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7efd4f50> (a sun.nio.ch.Util$2)
	- locked <0x7efd4f40> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7efd4d40> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.apache.hadoop.ipc.Server$Responder.run(Server.java:645)

"Timer-0" daemon prio=10 tid=0x6ba2d800 nid=0x1b42 in Object.wait() [0x73d7d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x7f81f350> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"14915048@qtp-24350088-1 - Acceptor0 SelectChannelConnector@localhost:49842" prio=10 tid=0x6ba12400 nid=0x1b34 runnable [0x73f7d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7f82a6e8> (a sun.nio.ch.Util$2)
	- locked <0x7f82a6d8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7f82a298> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
	at org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
	at org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
	at org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)

"32715295@qtp-24350088-0" prio=10 tid=0x6ba13800 nid=0x1b33 in Object.wait() [0x7417d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626)
	- locked <0x7f829de0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)

"pool-2-thread-1" prio=10 tid=0x6c381800 nid=0x1b25 runnable [0x7437d000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:228)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:81)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x7efcc4d8> (a sun.nio.ch.Util$2)
	- locked <0x7efcc4c8> (a java.util.Collections$UnmodifiableSet)
	- locked <0x7efcc2c8> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:373)
	- locked <0x7efcc6c0> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor@148d148" daemon prio=10 tid=0x6c356800 nid=0x1b24 waiting on condition [0x7457d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor@533643" daemon prio=10 tid=0x6c355800 nid=0x1b23 waiting on condition [0x7477d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$ReplicationMonitor.run(FSNamesystem.java:3167)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@14cc6" daemon prio=10 tid=0x6c354c00 nid=0x1b22 waiting on condition [0x7497d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:368)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor@4ed34" daemon prio=10 tid=0x6c353400 nid=0x1b21 waiting on condition [0x74b7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem$HeartbeatMonitor.run(FSNamesystem.java:3146)
	at java.lang.Thread.run(Thread.java:724)

"org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor@112c7fb" daemon prio=10 tid=0x6c351400 nid=0x1b1f waiting on condition [0x74d7d000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.namenode.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:197)
	at java.lang.Thread.run(Thread.java:724)

"pool-1-thread-1" prio=10 tid=0x6d34c800 nid=0x1a4a in Object.wait() [0x7517d000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.Thread.join(Thread.java:1268)
	- locked <0xceb8df00> (a org.junit.internal.runners.statements.FailOnTimeout$StatementThread)
	at org.junit.internal.runners.statements.FailOnTimeout.evaluateStatement(FailOnTimeout.java:26)
	at org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:17)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runners.Suite.runChild(Suite.java:127)
	at org.junit.runners.Suite.runChild(Suite.java:26)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
	at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
	at java.util.concurrent.FutureTask.run(FutureTask.java:166)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:724)

"Service Thread" daemon prio=10 tid=0x6de27000 nid=0x1a45 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x6de25400 nid=0x1a44 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x6de23400 nid=0x1a43 waiting on condition [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x6de21800 nid=0x1a42 runnable [0x00000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x6e2e7000 nid=0x1a41 in Object.wait() [0x75f5a000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x7ca75db0> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189)

"Reference Handler" daemon prio=10 tid=0x6e2e5400 nid=0x1a40 in Object.wait() [0x6e76d000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x7ca597b0> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0xf6706c00 nid=0x1a2c waiting on condition [0xf6940000]
   java.lang.Thread.State: WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x7ca75f58> (a java.util.concurrent.FutureTask$Sync)
	at java.util.concurrent.locks.LockSupport.park(LockSupport.java:186)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:834)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:994)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1303)
	at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:248)
	at java.util.concurrent.FutureTask.get(FutureTask.java:111)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.waitForCompletion(AsynchronousRunner.java:73)
	at org.apache.maven.surefire.junitcore.AsynchronousRunner.finished(AsynchronousRunner.java:58)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:242)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:138)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:61)
	at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:137)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.maven.surefire.util.ReflectionUtils.invokeMethodWithArray(ReflectionUtils.java:188)
	at org.apache.maven.surefire.booter.ProviderFactory$ProviderProxy.invoke(ProviderFactory.java:166)
	at org.apache.maven.surefire.booter.ProviderFactory.invokeProvider(ProviderFactory.java:86)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:101)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:74)

"VM Thread" prio=10 tid=0x6e5f8800 nid=0x1a3f runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0xf6710c00 nid=0x1a2d runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0xf6712000 nid=0x1a2e runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0xf6713800 nid=0x1a2f runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0xf6715000 nid=0x1a30 runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0xf6716400 nid=0x1a31 runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0xf6717c00 nid=0x1a32 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0xf6719400 nid=0x1a33 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0xf671ac00 nid=0x1a34 runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0xf671c000 nid=0x1a35 runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0xf671d800 nid=0x1a36 runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0xf671f000 nid=0x1a37 runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0xf6720400 nid=0x1a38 runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0xf6721c00 nid=0x1a39 runnable 

"GC task thread#13 (ParallelGC)" prio=10 tid=0xf6723400 nid=0x1a3a runnable 

"GC task thread#14 (ParallelGC)" prio=10 tid=0xf6724800 nid=0x1a3b runnable 

"GC task thread#15 (ParallelGC)" prio=10 tid=0xf6726000 nid=0x1a3c runnable 

"GC task thread#16 (ParallelGC)" prio=10 tid=0xf6727800 nid=0x1a3d runnable 

"GC task thread#17 (ParallelGC)" prio=10 tid=0xf6728c00 nid=0x1a3e runnable 

"VM Periodic Task Thread" prio=10 tid=0x6de29000 nid=0x1a47 waiting on condition 

JNI global references: 359

+ echo '************ END  zombies jstack extract'
************ END  zombies jstack extract
+ JIRA_COMMENT='

     {color:red}-1 core zombie tests{color}.  There are 2 zombie test(s): 	at org.apache.hadoop.hbase.regionserver.wal.TestWALReplay.testReplayEditsWrittenIntoWAL(TestWALReplay.java:698)
	at org.apache.hadoop.hbase.regionserver.wal.TestHLogSplit.testEmptyOpenLogFiles(TestHLogSplit.java:435)'
+ BAD=1
+ jps
+ grep surefirebooter
+ cut -d ' ' -f 1
+ xargs kill -9
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10644
Updating HADOOP-1

Build failed in Jenkins: HBase-TRUNK #4967

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4967/>

------------------------------------------
Started by an SCM change
Building remotely on ubuntu2 in workspace <https://builds.apache.org/job/HBase-TRUNK/ws/>
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@5733e030:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@5733e030:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Retrying after 10 seconds
Cleaning local Directory trunk
java.io.IOException: remote file operation failed: <https://builds.apache.org/job/HBase-TRUNK/ws/> at hudson.remoting.Channel@5733e030:ubuntu2
	at hudson.FilePath.act(FilePath.java:910)
	at hudson.FilePath.act(FilePath.java:887)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:848)
	at hudson.scm.SubversionSCM.checkout(SubversionSCM.java:786)
	at hudson.model.AbstractProject.checkout(AbstractProject.java:1411)
	at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:651)
	at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:560)
	at hudson.model.Run.execute(Run.java:1670)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:46)
	at hudson.model.ResourceController.execute(ResourceController.java:88)
	at hudson.model.Executor.run(Executor.java:231)
Caused by: java.nio.file.DirectoryNotEmptyException: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/.svn/tmp>
	at sun.nio.fs.UnixFileSystemProvider.implDelete(UnixFileSystemProvider.java:242)
	at sun.nio.fs.AbstractFileSystemProvider.delete(AbstractFileSystemProvider.java:103)
	at java.nio.file.Files.delete(Files.java:1077)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at hudson.Util.deleteFile(Util.java:238)
	at hudson.Util.deleteRecursive(Util.java:301)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.Util.deleteRecursive(Util.java:292)
	at hudson.Util.deleteContentsRecursive(Util.java:203)
	at hudson.scm.subversion.CheckoutUpdater$1.perform(CheckoutUpdater.java:75)
	at hudson.scm.subversion.WorkspaceUpdater$UpdateTask.delegateTo(WorkspaceUpdater.java:161)
	at hudson.scm.SubversionSCM$CheckOutTask.perform(SubversionSCM.java:908)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:889)
	at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:872)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2461)
	at hudson.remoting.UserRequest.perform(UserRequest.java:118)
	at hudson.remoting.UserRequest.perform(UserRequest.java:48)
	at hudson.remoting.Request$2.run(Request.java:328)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:744)
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson2782120265627618185.sh
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'Suspicious java process found - waiting 10s to see if there are just slow to stop'
Suspicious java process found - waiting 10s to see if there are just slow to stop
+ sleep 10
++ jps
++ grep surefirebooter
++ wc -l
+ ZOMBIE_TESTS_COUNT=2
+ [[ 2 != 0 ]]
+ echo 'There are 2 zombie tests, they should have been killed by surefire but survived'
There are 2 zombie tests, they should have been killed by surefire but survived
+ echo '************ BEGIN zombies jstack extract'
************ BEGIN zombies jstack extract
++ jps
++ grep surefirebooter
++ cut -d ' ' -f 1
++ xargs -n 1 jstack
++ grep .test
++ grep '\.java'
1906: Unable to open socket file: target process not responding or HotSpot VM not loaded
The -F option can be used when the target process is not responding
+ ZB_STACK=
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Build failed in Jenkins: HBase-TRUNK #4966

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-TRUNK/4966/changes>

Changes:

[tedyu] HBASE-10627 A logic mistake in HRegionServer isHealthy (Shaohui)

[tedyu] HBASE-10638 Improve error message when there is no region server available for move

------------------------------------------
[...truncated 4824 lines...]
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-server ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-server ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-tests.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-server/target/hbase-server-0.99.0-SNAPSHOT-sources.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-server/0.99.0-SNAPSHOT/hbase-server-0.99.0-SNAPSHOT-sources.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Testing Util 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-testing-util ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/main/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-testing-util ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-testing-util ---
[INFO] No sources to compile
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (secondPartTestsExecution) @ hbase-testing-util ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ hbase-testing-util ---
[INFO] Building jar: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar>
[INFO] 
[INFO] --- maven-install-plugin:2.3.1:install (default-install) @ hbase-testing-util ---
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.jar
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/pom.xml> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT.pom
[INFO] Installing <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-testing-util/target/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar> to /home/jenkins/.m2/repository/org/apache/hbase/hbase-testing-util/0.99.0-SNAPSHOT/hbase-testing-util-0.99.0-SNAPSHOT-tests.jar
[INFO]                                                                         
[INFO] ------------------------------------------------------------------------
[INFO] Building HBase - Thrift 0.99.0-SNAPSHOT
[INFO] ------------------------------------------------------------------------
[INFO] 
[INFO] --- maven-antrun-plugin:1.6:run (generate) @ hbase-thrift ---
[INFO] Executing tasks

main:
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
     [copy] Copying 17 files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps>
log4j:WARN No appenders could be found for logger (org.mortbay.log).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
    [mkdir] Created dir: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/hbase-webapps/thrift/WEB-INF>
[INFO] Executed tasks
[INFO] 
[INFO] --- build-helper-maven-plugin:1.5:add-source (jspcSource-packageInfo-source) @ hbase-thrift ---
[INFO] Source directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/generated-sources/java> added.
[INFO] 
[INFO] --- maven-remote-resources-plugin:1.4:process (default) @ hbase-thrift ---
[INFO] 
[INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] Copying 18 resources
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hbase-thrift ---
[INFO] Compiling 47 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/classes>
[INFO] 
[INFO] --- maven-dependency-plugin:2.1:build-classpath (create-mrapp-generated-classpath) @ hbase-thrift ---
[INFO] Wrote classpath file '<https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes/mrapp-generated-classpath'.>
[INFO] 
[INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hbase-thrift ---
[INFO] Using 'UTF-8' encoding to copy filtered resources.
[INFO] skip non existing resourceDirectory <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/src/test/resources>
[INFO] Copying 3 resources
[INFO] 
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) @ hbase-thrift ---
[INFO] Compiling 5 source files to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/test-classes>
[INFO] 
[INFO] --- maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) @ hbase-thrift ---
[INFO] Surefire report directory: <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports>
[INFO] Using configured provider org.apache.maven.surefire.junitcore.JUnitCoreProvider

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.thrift.TestCallQueue
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.17 sec
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandlerWithLabels
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.483 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServer
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 116.189 sec <<< FAILURE!
Running org.apache.hadoop.hbase.thrift2.TestThriftHBaseServiceHandler
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 116.189 sec
Running org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine
Tests run: 32, Failures: 0, Errors: 32, Skipped: 0, Time elapsed: 116.189 sec <<< FAILURE!

Results :

Tests in error: 
  testAll(org.apache.hadoop.hbase.thrift.TestThriftServer): org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[0](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[1](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[2](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[3](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[4](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[5](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[6](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[7](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[8](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[9](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[10](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[11](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[12](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[13](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[14](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[15](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[16](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[17](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[18](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[19](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[20](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[21](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[22](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[23](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[24](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[25](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[26](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[27](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[28](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[29](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[30](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)
  testRunThriftServer[31](org.apache.hadoop.hbase.thrift.TestThriftServerCmdLine): IOError(message:org.apache.hadoop.hbase.DoNotRetryIOException: TTL for column family columnA  must be positive. Set hbase.table.sanity.checks to false at conf or table descriptor if you want to bypass sanity checks(..)

Tests run: 77, Failures: 0, Errors: 33, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] HBase ............................................. SUCCESS [10.702s]
[INFO] HBase - Common .................................... SUCCESS [37.944s]
[INFO] HBase - Protocol .................................. SUCCESS [12.320s]
[INFO] HBase - Client .................................... SUCCESS [40.905s]
[INFO] HBase - Hadoop Compatibility ...................... SUCCESS [6.404s]
[INFO] HBase - Hadoop Two Compatibility .................. SUCCESS [3.033s]
[INFO] HBase - Prefix Tree ............................... SUCCESS [5.278s]
[INFO] HBase - Server .................................... SUCCESS [2:27:52.980s]
[INFO] HBase - Testing Util .............................. SUCCESS [3.258s]
[INFO] HBase - Thrift .................................... FAILURE [2:03.245s]
[INFO] HBase - Shell ..................................... SKIPPED
[INFO] HBase - Integration Tests ......................... SKIPPED
[INFO] HBase - Examples .................................. SKIPPED
[INFO] HBase - Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:31:57.559s
[INFO] Finished at: Fri Feb 28 20:39:20 UTC 2014
[INFO] Final Memory: 53M/768M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.12-TRUNK-HBASE-2:test (default-test) on project hbase-thrift: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/HBase-TRUNK/ws/trunk/hbase-thrift/target/surefire-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-thrift
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  :   ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
  if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
    #It seems sometimes the tests are not dying immediately. Let's give them 10s
    echo "Suspicious java process found - waiting 10s to see if there are just slow to stop"
    sleep 10   
    ZOMBIE_TESTS_COUNT=`jps | grep surefirebooter | wc -l`
    if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
      echo "There are $ZOMBIE_TESTS_COUNT zombie tests, they should have been killed by surefire but survived"
      echo "************ BEGIN zombies jstack extract"
      ZB_STACK=`jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack | grep ".test" | grep "\.java"`
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs -n 1 jstack
      echo "************ END  zombies jstack extract"
      JIRA_COMMENT="$JIRA_COMMENT

     {color:red}-1 core zombie tests{color}.  There are ${ZOMBIE_TESTS_COUNT} zombie test(s): ${ZB_STACK}"
      BAD=1
      jps | grep surefirebooter | cut -d ' ' -f 1 | xargs kill -9
    else
      echo "We're ok: there is no zombie test, but some tests took some time to stop"
    fi
  else
    echo "We're ok: there is no zombie test"
  fi
[HBase-TRUNK] $ /bin/bash -xe /tmp/hudson7831537771808749339.sh
++ jps
++ wc -l
++ grep surefirebooter
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Updating HBASE-10627
Updating HBASE-10638