You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@hbase.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2015/12/16 19:57:25 UTC

Build failed in Jenkins: HBase-1.3 » latest1.8,Hadoop #440

See <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/440/changes>

Changes:

[busbey] HBASE-14838 Add clarification docs to SimpleRegionNormalizer.

------------------------------------------
[...truncated 58405 lines...]
	- parking to wait for  <0x00000007f8be1530> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
	at java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467)
	at org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1985)

"IPC Server handler 0 on 47491" daemon prio=10 tid=0x00007f9fdd218000 nid=0xe23 waiting on condition [0x00007f9f637f6000]
   java.lang.Thread.State: TIMED_WAITING (parking)
	at sun.misc.Unsafe.park(Native Method)
	- parking to wait for  <0x00000007f8be1530> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
	at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
	at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
	at java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467)
	at org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1985)

"IPC Server listener on 47491" daemon prio=10 tid=0x00007f9fdd216800 nid=0xe1c runnable [0x00007f9f638f7000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x0000000756421468> (a sun.nio.ch.Util$2)
	- locked <0x0000000756421478> (a java.util.Collections$UnmodifiableSet)
	- locked <0x0000000756421420> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener.run(Server.java:668)

"IPC Server Responder" daemon prio=10 tid=0x00007f9fdd215000 nid=0xe19 runnable [0x00007f9f63dfc000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x000000075644c208> (a sun.nio.ch.Util$2)
	- locked <0x000000075644c218> (a java.util.Collections$UnmodifiableSet)
	- locked <0x000000075644c1c0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:835)
	at org.apache.hadoop.ipc.Server$Responder.run(Server.java:818)

"org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@139e0d30" daemon prio=10 tid=0x00007f9fdd209800 nid=0xe13 waiting on condition [0x00007f9f639f8000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor.run(BlockManager.java:3397)
	at java.lang.Thread.run(Thread.java:745)

"org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@6d73d45f" daemon prio=10 tid=0x00007f9fdd207800 nid=0xe12 waiting on condition [0x00007f9f63af9000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:319)
	at java.lang.Thread.run(Thread.java:745)

"org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@3f779cb" daemon prio=10 tid=0x00007f9fdd205800 nid=0xe11 waiting on condition [0x00007f9f63bfa000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor.run(DecommissionManager.java:76)
	at java.lang.Thread.run(Thread.java:745)

"org.apache.hadoop.hdfs.server.blockmanagement.PendingReplicationBlocks$PendingReplicationMonitor@748ede42" daemon prio=10 tid=0x00007f9fdd203800 nid=0xe10 waiting on condition [0x00007f9f63cfb000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.hdfs.server.blockmanagement.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:221)
	at java.lang.Thread.run(Thread.java:745)

"org.apache.hadoop.util.JvmPauseMonitor$Monitor@138301d0" daemon prio=10 tid=0x00007f9fdcb52000 nid=0xe06 waiting on condition [0x00007f9f63efd000]
   java.lang.Thread.State: TIMED_WAITING (sleeping)
	at java.lang.Thread.sleep(Native Method)
	at org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:180)
	at java.lang.Thread.run(Thread.java:745)

"IPC Server idle connection scanner for port 47491" daemon prio=10 tid=0x00007f9fdce1b000 nid=0xdfe in Object.wait() [0x00007f9f63ffe000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x000000075646a048> (a java.util.TaskQueue)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x000000075646a048> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"Socket Reader #1 for port 47491" prio=10 tid=0x00007f9fdce19800 nid=0xdfd runnable [0x00007f9f9815e000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x0000000756422040> (a sun.nio.ch.Util$2)
	- locked <0x0000000756422030> (a java.util.Collections$UnmodifiableSet)
	- locked <0x00000007564216c0> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
	at org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:616)
	- locked <0x000000075644ac00> (a org.apache.hadoop.ipc.Server$Listener$Reader)
	at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:595)

"Timer-0" daemon prio=10 tid=0x00007f9fdd164000 nid=0xd5e in Object.wait() [0x00007f9f9825f000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x000000075646ae28> (a java.util.TaskQueue)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x000000075646ae28> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"570337240@qtp-928158108-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:51338" daemon prio=10 tid=0x00007f9fdd13f000 nid=0xd36 runnable [0x00007f9f98360000]
   java.lang.Thread.State: RUNNABLE
	at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
	at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
	at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
	at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
	- locked <0x0000000758059c48> (a sun.nio.ch.Util$2)
	- locked <0x0000000758059c38> (a java.util.Collections$UnmodifiableSet)
	- locked <0x0000000758057a70> (a sun.nio.ch.EPollSelectorImpl)
	at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
	at org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
	at org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
	at org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
	at org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)

"1654926621@qtp-928158108-0" daemon prio=10 tid=0x00007f9fdd144800 nid=0xd35 in Object.wait() [0x00007f9f98461000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x000000075646f8e0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)
	at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626)
	- locked <0x000000075646f8e0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)

"Timer for 'NameNode' metrics system" daemon prio=10 tid=0x00007f9fdd0c3800 nid=0xd18 in Object.wait() [0x00007f9f98bd4000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x000000075805a080> (a java.util.TaskQueue)
	at java.util.TimerThread.mainLoop(Timer.java:552)
	- locked <0x000000075805a080> (a java.util.TaskQueue)
	at java.util.TimerThread.run(Timer.java:505)

"Service Thread" daemon prio=10 tid=0x00007f9fdc293800 nid=0xa2e runnable [0x0000000000000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread1" daemon prio=10 tid=0x00007f9fdc291800 nid=0xa2d waiting on condition [0x0000000000000000]
   java.lang.Thread.State: RUNNABLE

"C2 CompilerThread0" daemon prio=10 tid=0x00007f9fdc28e800 nid=0xa2c waiting on condition [0x0000000000000000]
   java.lang.Thread.State: RUNNABLE

"Signal Dispatcher" daemon prio=10 tid=0x00007f9fdc28c000 nid=0xa2b runnable [0x0000000000000000]
   java.lang.Thread.State: RUNNABLE

"Finalizer" daemon prio=10 tid=0x00007f9fdc26b800 nid=0xa1d in Object.wait() [0x00007f9f9a2e1000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x00000007f712edb0> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
	- locked <0x00000007f712edb0> (a java.lang.ref.ReferenceQueue$Lock)
	at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
	at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:209)

"Reference Handler" daemon prio=10 tid=0x00007f9fdc269800 nid=0xa1a in Object.wait() [0x00007f9f9a3e2000]
   java.lang.Thread.State: WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x0000000757fae488> (a java.lang.ref.Reference$Lock)
	at java.lang.Object.wait(Object.java:503)
	at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
	- locked <0x0000000757fae488> (a java.lang.ref.Reference$Lock)

"main" prio=10 tid=0x00007f9fdc00a800 nid=0xa0a in Object.wait() [0x00007f9fe2d98000]
   java.lang.Thread.State: TIMED_WAITING (on object monitor)
	at java.lang.Object.wait(Native Method)
	- waiting on <0x00000007d6d74328> (a org.apache.hadoop.hbase.util.JVMClusterUtil$MasterThread)
	at java.lang.Thread.join(Thread.java:1289)
	- locked <0x00000007d6d74328> (a org.apache.hadoop.hbase.util.JVMClusterUtil$MasterThread)
	at org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:131)
	at org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:434)
	at org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:758)
	at org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1161)
	at org.apache.hadoop.hbase.TestZooKeeper.after(TestZooKeeper.java:123)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
	at org.junit.runners.Suite.runChild(Suite.java:128)
	at org.junit.runners.Suite.runChild(Suite.java:27)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:108)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:78)
	at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:54)
	at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:144)
	at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:203)
	at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:155)
	at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)

"VM Thread" prio=10 tid=0x00007f9fdc265000 nid=0xa19 runnable 

"GC task thread#0 (ParallelGC)" prio=10 tid=0x00007f9fdc020800 nid=0xa0b runnable 

"GC task thread#1 (ParallelGC)" prio=10 tid=0x00007f9fdc022800 nid=0xa0c runnable 

"GC task thread#2 (ParallelGC)" prio=10 tid=0x00007f9fdc024000 nid=0xa0d runnable 

"GC task thread#3 (ParallelGC)" prio=10 tid=0x00007f9fdc026000 nid=0xa0e runnable 

"GC task thread#4 (ParallelGC)" prio=10 tid=0x00007f9fdc028000 nid=0xa0f runnable 

"GC task thread#5 (ParallelGC)" prio=10 tid=0x00007f9fdc029800 nid=0xa10 runnable 

"GC task thread#6 (ParallelGC)" prio=10 tid=0x00007f9fdc02b800 nid=0xa11 runnable 

"GC task thread#7 (ParallelGC)" prio=10 tid=0x00007f9fdc02d800 nid=0xa12 runnable 

"GC task thread#8 (ParallelGC)" prio=10 tid=0x00007f9fdc02f800 nid=0xa13 runnable 

"GC task thread#9 (ParallelGC)" prio=10 tid=0x00007f9fdc031000 nid=0xa14 runnable 

"GC task thread#10 (ParallelGC)" prio=10 tid=0x00007f9fdc033000 nid=0xa15 runnable 

"GC task thread#11 (ParallelGC)" prio=10 tid=0x00007f9fdc035000 nid=0xa16 runnable 

"GC task thread#12 (ParallelGC)" prio=10 tid=0x00007f9fdc036800 nid=0xa17 runnable 

"VM Periodic Task Thread" prio=10 tid=0x00007f9fdc29e800 nid=0xa30 waiting on condition 

JNI global references: 169

+ exit 1
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results

Jenkins build is back to normal : HBase-1.3 » latest1.8,Hadoop #442

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/442/changes>


Build failed in Jenkins: HBase-1.3 » latest1.8,Hadoop #441

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/441/changes>

Changes:

[busbey] HBASE-14952 ensure correct components in source assembly.

[busbey] HBASE-14952. ADDENDUM hbase-spark not present in branch-1 yet.

[eclark] HBASE-14984 Allow memcached block cache to set optimze to false

------------------------------------------
[...truncated 49032 lines...]
	at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:642)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:613)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1007)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:888)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:870)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:857)
	at org.apache.hadoop.hbase.filter.TestFilterWrapper.setUp(TestFilterWrapper.java:190)

Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=256m; support was removed in 8.0
Running org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.068 sec <<< FAILURE! - in org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter  Time elapsed: 0.063 sec  <<< ERROR!
java.io.IOException: Shutting down
	at org.apache.hadoop.hbase.util.JVMClusterUtil.startup(JVMClusterUtil.java:225)
	at org.apache.hadoop.hbase.LocalHBaseCluster.startup(LocalHBaseCluster.java:447)
	at org.apache.hadoop.hbase.MiniHBaseCluster.init(MiniHBaseCluster.java:225)
	at org.apache.hadoop.hbase.MiniHBaseCluster.<init>(MiniHBaseCluster.java:94)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniHBaseCluster(HBaseTestingUtility.java:1056)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1016)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:888)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:882)
	at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:826)
	at org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter.setUpBeforeClass(TestMultiRowRangeFilter.java:61)


Results :

Failed tests: 
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testLoadTooMayHFiles(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testLoadTooMayHFiles:537 null
  Run 2: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testLoadTooMayHFiles:537 null

Tests in error: 
org.apache.hadoop.hbase.TestGlobalMemStoreSize.testGlobalMemStore(org.apache.hadoop.hbase.TestGlobalMemStoreSize)
  Run 1: TestGlobalMemStoreSize.testGlobalMemStore » Remote unable to create new native...
  Run 2: TestGlobalMemStoreSize.testGlobalMemStore:69 » OutOfMemory unable to create ne...
  Run 3: TestGlobalMemStoreSize.testGlobalMemStore:69 » OutOfMemory unable to create ne...

  TestNamespace.setUp:72 » Exit java.lang.OutOfMemoryError: unable to create new...
  TestConstraint.setUpBeforeClass:64 » OutOfMemory unable to create new native t...
  TestFilterWrapper.setUp:190 » OutOfMemory unable to create new native thread
  TestFuzzyRowAndColumnRangeFilter.setUpBeforeClass:64 » OutOfMemory unable to c...
  TestMultiRowRangeFilter.setUpBeforeClass:61 » IO Shutting down
  TestCellCounter.testCellCounteEndTimeRange:177 » FileNotFound target/test-data...
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplit(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:173->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:217->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » TestTimedOut
  Run 2: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:173->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:217->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:270 » IO

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom:182->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:217->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » RetriesExhausted
  Run 2: TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom » Remote ...

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowColBloom(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowColBloom » Remote
  Run 2: TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowColBloom » IO U...

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testSimpleHFileSplit(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testSimpleHFileSplit:155->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » RetriesExhausted
  Run 2: TestSecureLoadIncrementalHFiles.testSimpleHFileSplit » Remote unable to create...

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testSplitALot(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testSplitALot:196->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » RetriesExhausted
  Run 2: TestSecureLoadIncrementalHFiles.testSplitALot » Remote unable to create new na...

  TestTableInputFormat.beforeClass:88 » OutOfMemory unable to create new native ...
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testRestoreSnapshotDoesNotCreateBackRefLinks(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
  Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testRestoreSnapshotDoesNotCreateBackRefLinks:117->TableSnapshotInputFormatTestBase.setupCluster:63 » IO
  Run 2: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testRestoreSnapshotDoesNotCreateBackRefLinks:117->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
  Run 3: TestTableSnapshotInputFormat.testRestoreSnapshotDoesNotCreateBackRefLinks » OutOfMemory

org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMapReduceAndOfflineHBaseMultiRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
  Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceAndOfflineHBaseMultiRegion:111->TableSnapshotInputFormatTestBase.testWithMapReduce:161 » YarnRuntime
  Run 2: TestTableSnapshotInputFormat.testWithMapReduceAndOfflineHBaseMultiRegion » OutOfMemory
  Run 3: TestTableSnapshotInputFormat.testWithMapReduceAndOfflineHBaseMultiRegion » OutOfMemory

org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMapReduceMultiRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
  Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceMultiRegion:105->TableSnapshotInputFormatTestBase.testWithMapReduce:165->testWithMapReduceImpl:256->doTestWithMapReduce:289 null
  Run 2: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceMultiRegion:105->TableSnapshotInputFormatTestBase.testWithMapReduce:160->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
  Run 3: TestTableSnapshotInputFormat.testWithMapReduceMultiRegion » OutOfMemory unable...

org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMapReduceSingleRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
  Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceSingleRegion:100->TableSnapshotInputFormatTestBase.testWithMapReduce:160->TableSnapshotInputFormatTestBase.setupCluster:63 » IO
  Run 2: TestTableSnapshotInputFormat.testWithMapReduceSingleRegion » OutOfMemory unabl...
  Run 3: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceSingleRegion:100->TableSnapshotInputFormatTestBase.testWithMapReduce:160->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory

org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMockedMapReduceMultiRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
  Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMockedMapReduceMultiRegion:95->testWithMockedMapReduce:192->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
  Run 2: TestTableSnapshotInputFormat.testWithMockedMapReduceMultiRegion » OutOfMemory ...
  Run 3: TestTableSnapshotInputFormat.testWithMockedMapReduceMultiRegion » OutOfMemory ...

org.apache.hadoop.hbase.security.access.TestNamespaceCommands.org.apache.hadoop.hbase.security.access.TestNamespaceCommands
  Run 1: TestNamespaceCommands.beforeClass:149 » IO Shutting down
  Run 2: TestNamespaceCommands.afterClass:185 » RetriesExhausted Failed after attempts=...

  TestScanEarlyTermination.setupBeforeClass:91 » Runtime java.lang.OutOfMemoryEr...
org.apache.hadoop.hbase.security.access.TestScanEarlyTermination.testEarlyScanTermination(org.apache.hadoop.hbase.security.access.TestScanEarlyTermination)
  Run 1: TestScanEarlyTermination.testEarlyScanTermination » Remote unable to create ne...
  Run 2: TestScanEarlyTermination.tearDown:144 » IO java.lang.reflect.InvocationTargetE...

  TestTablePermissions.beforeClass:101 » Runtime java.lang.RuntimeException: jav...
org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelReplicationWithExpAsString.testVisibilityReplication(org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelReplicationWithExpAsString)
  Run 1: TestVisibilityLabelReplicationWithExpAsString.setup:139 » IO Shutting down
  Run 2: TestVisibilityLabelReplicationWithExpAsString.setup:114 » OutOfMemory unable t...
  Run 3: TestVisibilityLabelReplicationWithExpAsString.setup:116 » OutOfMemory unable t...

  TestFlushSnapshotFromClient.setupCluster:86 » IO Shutting down
  TestRestoreFlushSnapshotFromClient.setUpBeforeClass:77 » OutOfMemory unable to...
Flaked tests: 
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testNonHfileFolderWithUnmatchedFamilyName(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testNonHfileFolderWithUnmatchedFamilyName:345->TestLoadIncrementalHFiles.testNonHfileFolder:377 » RetriesExhausted
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testNonexistentColumnFamilyLoad(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testNonexistentColumnFamilyLoad:331->TestLoadIncrementalHFiles.runTest:276 » TestTimedOut
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingLoad(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingLoad:118->TestLoadIncrementalHFiles.runTest:231->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:280 » RetriesExhausted
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testTableWithCFNameStartWithUnderScore(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testTableWithCFNameStartWithUnderScore:563 » RetriesExhausted
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testWithoutAnExistingTableAndCreateTableSetToNo(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
  Run 1: TestSecureLoadIncrementalHFiles.testWithoutAnExistingTableAndCreateTableSetToNo » 
  Run 2: PASS

org.apache.hadoop.hbase.regionserver.TestHRegion.testBatchPut_whileNoRowLocksHeld(org.apache.hadoop.hbase.regionserver.TestHRegion)
  Run 1: TestHRegion.testBatchPut_whileNoRowLocksHeld:1487 Metrics Counters should be equal expected:<1056> but was:<1055>
  Run 2: PASS

org.apache.hadoop.hbase.security.access.TestAccessController.testNamespaceUserGrant(org.apache.hadoop.hbase.security.access.TestAccessController)
  Run 1: TestAccessController.testNamespaceUserGrant » OutOfMemory unable to create new...
  Run 2: PASS

org.apache.hadoop.hbase.security.access.TestAccessController.testPostGrantRevoke(org.apache.hadoop.hbase.security.access.TestAccessController)
  Run 1: TestAccessController.testPostGrantRevoke » Remote java.lang.OutOfMemoryError: ...
  Run 2: PASS

org.apache.hadoop.hbase.security.access.TestAccessController.testTableDeletion(org.apache.hadoop.hbase.security.access.TestAccessController)
  Run 1: TestAccessController.testTableDeletion:2225->createTestTable:2257->SecureTestUtil.createTable:669->SecureTestUtil.createTable:691 » IO
  Run 2: PASS

org.apache.hadoop.hbase.snapshot.TestSnapshotClientRetries.testSnapshotAlreadyExist(org.apache.hadoop.hbase.snapshot.TestSnapshotClientRetries)
  Run 1: TestSnapshotClientRetries.setUp:55 » IO Shutting down
  Run 2: PASS


Tests run: 1755, Failures: 1, Errors: 25, Skipped: 37, Flakes: 10

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache HBase ...................................... SUCCESS [1:11.895s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [6.525s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [0.617s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [3.843s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [46.959s]
[INFO] Apache HBase - Common ............................. SUCCESS [4:32.502s]
[INFO] Apache HBase - Procedure .......................... SUCCESS [5:35.464s]
[INFO] Apache HBase - Client ............................. SUCCESS [2:09.039s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [13.502s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [21.779s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [25.554s]
[INFO] Apache HBase - Server ............................. FAILURE [2:51:42.220s]
[INFO] Apache HBase - Testing Util ....................... SKIPPED
[INFO] Apache HBase - Thrift ............................. SKIPPED
[INFO] Apache HBase - Rest ............................... SKIPPED
[INFO] Apache HBase - Shell .............................. SKIPPED
[INFO] Apache HBase - Integration Tests .................. SKIPPED
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 3:07:34.353s
[INFO] Finished at: Wed Dec 16 23:40:37 UTC 2015
[INFO] Final Memory: 377M/3212M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test (secondPartTestsExecution) on project hbase-server: ExecutionException: java.lang.RuntimeException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server> && /home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.8/jre/bin/java -enableassertions -Dhbase.test -Xmx2800m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom -Djava.net.preferIPv4Stack=true -Djava.awt.headless=true -jar <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server/target/surefire/surefirebooter5933983239854825811.jar> <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server/target/surefire/surefire5820196020613052737tmp> <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server/target/surefire/surefire_10924691193655747684866tmp>
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  : # Post-build task script. TODO: Check this in and have all builds reference check-in.
pwd && ls
# NOTE!!!! The below code has been copied and pasted from ./dev-tools/run-test.sh
# Do not change here without syncing there and vice-versa.
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
 echo "Suspicious java process found - waiting 30s to see if there are just slow to stop"
 sleep 30
 ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
 if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
   echo " {color:red}There appear to be $ZOMBIE_TESTS_COUNT zombie tests{color}, they should have been killed by surefire but survived"
   jps -v | grep surefirebooter | grep -e '-Dhbase.test'
   jps -v | grep surefirebooter | grep -e '-Dhbase.test' | cut -d ' ' -f 1 | xargs -n 1 jstack
   # Exit with error
   exit 1
 else
   echo "We're ok: there is no zombie test, but some tests took some time to stop"
 fi
else
  echo "We're ok: there is no zombie test"
fi
[Hadoop] $ /bin/bash -xe /tmp/hudson5055114759237812355.sh
+ pwd
<https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/>
+ ls
hbase-server
++ jps -v
++ grep surefirebooter
++ grep -e -Dhbase.test
++ wc -l
+ ZOMBIE_TESTS_COUNT=1
+ [[ 1 != 0 ]]
+ echo 'Suspicious java process found - waiting 30s to see if there are just slow to stop'
Suspicious java process found - waiting 30s to see if there are just slow to stop
+ sleep 30
++ wc -l
++ grep -e -Dhbase.test
++ grep surefirebooter
++ jps -v
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test, but some tests took some time to stop'
We're ok: there is no zombie test, but some tests took some time to stop
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
ERROR: Publisher 'Publish JUnit test result report' failed: No test report files were found. Configuration error?