You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@hbase.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2015/12/16 19:57:25 UTC
Build failed in Jenkins: HBase-1.3 » latest1.8,Hadoop #440
See <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/440/changes>
Changes:
[busbey] HBASE-14838 Add clarification docs to SimpleRegionNormalizer.
------------------------------------------
[...truncated 58405 lines...]
- parking to wait for <0x00000007f8be1530> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
at java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467)
at org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1985)
"IPC Server handler 0 on 47491" daemon prio=10 tid=0x00007f9fdd218000 nid=0xe23 waiting on condition [0x00007f9f637f6000]
java.lang.Thread.State: TIMED_WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for <0x00000007f8be1530> (a java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
at java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467)
at org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1985)
"IPC Server listener on 47491" daemon prio=10 tid=0x00007f9fdd216800 nid=0xe1c runnable [0x00007f9f638f7000]
java.lang.Thread.State: RUNNABLE
at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
- locked <0x0000000756421468> (a sun.nio.ch.Util$2)
- locked <0x0000000756421478> (a java.util.Collections$UnmodifiableSet)
- locked <0x0000000756421420> (a sun.nio.ch.EPollSelectorImpl)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
at org.apache.hadoop.ipc.Server$Listener.run(Server.java:668)
"IPC Server Responder" daemon prio=10 tid=0x00007f9fdd215000 nid=0xe19 runnable [0x00007f9f63dfc000]
java.lang.Thread.State: RUNNABLE
at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
- locked <0x000000075644c208> (a sun.nio.ch.Util$2)
- locked <0x000000075644c218> (a java.util.Collections$UnmodifiableSet)
- locked <0x000000075644c1c0> (a sun.nio.ch.EPollSelectorImpl)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
at org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:835)
at org.apache.hadoop.ipc.Server$Responder.run(Server.java:818)
"org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@139e0d30" daemon prio=10 tid=0x00007f9fdd209800 nid=0xe13 waiting on condition [0x00007f9f639f8000]
java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor.run(BlockManager.java:3397)
at java.lang.Thread.run(Thread.java:745)
"org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@6d73d45f" daemon prio=10 tid=0x00007f9fdd207800 nid=0xe12 waiting on condition [0x00007f9f63af9000]
java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:319)
at java.lang.Thread.run(Thread.java:745)
"org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@3f779cb" daemon prio=10 tid=0x00007f9fdd205800 nid=0xe11 waiting on condition [0x00007f9f63bfa000]
java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor.run(DecommissionManager.java:76)
at java.lang.Thread.run(Thread.java:745)
"org.apache.hadoop.hdfs.server.blockmanagement.PendingReplicationBlocks$PendingReplicationMonitor@748ede42" daemon prio=10 tid=0x00007f9fdd203800 nid=0xe10 waiting on condition [0x00007f9f63cfb000]
java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.hdfs.server.blockmanagement.PendingReplicationBlocks$PendingReplicationMonitor.run(PendingReplicationBlocks.java:221)
at java.lang.Thread.run(Thread.java:745)
"org.apache.hadoop.util.JvmPauseMonitor$Monitor@138301d0" daemon prio=10 tid=0x00007f9fdcb52000 nid=0xe06 waiting on condition [0x00007f9f63efd000]
java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:180)
at java.lang.Thread.run(Thread.java:745)
"IPC Server idle connection scanner for port 47491" daemon prio=10 tid=0x00007f9fdce1b000 nid=0xdfe in Object.wait() [0x00007f9f63ffe000]
java.lang.Thread.State: TIMED_WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x000000075646a048> (a java.util.TaskQueue)
at java.util.TimerThread.mainLoop(Timer.java:552)
- locked <0x000000075646a048> (a java.util.TaskQueue)
at java.util.TimerThread.run(Timer.java:505)
"Socket Reader #1 for port 47491" prio=10 tid=0x00007f9fdce19800 nid=0xdfd runnable [0x00007f9f9815e000]
java.lang.Thread.State: RUNNABLE
at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
- locked <0x0000000756422040> (a sun.nio.ch.Util$2)
- locked <0x0000000756422030> (a java.util.Collections$UnmodifiableSet)
- locked <0x00000007564216c0> (a sun.nio.ch.EPollSelectorImpl)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:102)
at org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:616)
- locked <0x000000075644ac00> (a org.apache.hadoop.ipc.Server$Listener$Reader)
at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:595)
"Timer-0" daemon prio=10 tid=0x00007f9fdd164000 nid=0xd5e in Object.wait() [0x00007f9f9825f000]
java.lang.Thread.State: TIMED_WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x000000075646ae28> (a java.util.TaskQueue)
at java.util.TimerThread.mainLoop(Timer.java:552)
- locked <0x000000075646ae28> (a java.util.TaskQueue)
at java.util.TimerThread.run(Timer.java:505)
"570337240@qtp-928158108-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:51338" daemon prio=10 tid=0x00007f9fdd13f000 nid=0xd36 runnable [0x00007f9f98360000]
java.lang.Thread.State: RUNNABLE
at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:79)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:87)
- locked <0x0000000758059c48> (a sun.nio.ch.Util$2)
- locked <0x0000000758059c38> (a java.util.Collections$UnmodifiableSet)
- locked <0x0000000758057a70> (a sun.nio.ch.EPollSelectorImpl)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:98)
at org.mortbay.io.nio.SelectorManager$SelectSet.doSelect(SelectorManager.java:498)
at org.mortbay.io.nio.SelectorManager.doSelect(SelectorManager.java:192)
at org.mortbay.jetty.nio.SelectChannelConnector.accept(SelectChannelConnector.java:124)
at org.mortbay.jetty.AbstractConnector$Acceptor.run(AbstractConnector.java:708)
at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
"1654926621@qtp-928158108-0" daemon prio=10 tid=0x00007f9fdd144800 nid=0xd35 in Object.wait() [0x00007f9f98461000]
java.lang.Thread.State: TIMED_WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x000000075646f8e0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)
at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:626)
- locked <0x000000075646f8e0> (a org.mortbay.thread.QueuedThreadPool$PoolThread)
"Timer for 'NameNode' metrics system" daemon prio=10 tid=0x00007f9fdd0c3800 nid=0xd18 in Object.wait() [0x00007f9f98bd4000]
java.lang.Thread.State: TIMED_WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x000000075805a080> (a java.util.TaskQueue)
at java.util.TimerThread.mainLoop(Timer.java:552)
- locked <0x000000075805a080> (a java.util.TaskQueue)
at java.util.TimerThread.run(Timer.java:505)
"Service Thread" daemon prio=10 tid=0x00007f9fdc293800 nid=0xa2e runnable [0x0000000000000000]
java.lang.Thread.State: RUNNABLE
"C2 CompilerThread1" daemon prio=10 tid=0x00007f9fdc291800 nid=0xa2d waiting on condition [0x0000000000000000]
java.lang.Thread.State: RUNNABLE
"C2 CompilerThread0" daemon prio=10 tid=0x00007f9fdc28e800 nid=0xa2c waiting on condition [0x0000000000000000]
java.lang.Thread.State: RUNNABLE
"Signal Dispatcher" daemon prio=10 tid=0x00007f9fdc28c000 nid=0xa2b runnable [0x0000000000000000]
java.lang.Thread.State: RUNNABLE
"Finalizer" daemon prio=10 tid=0x00007f9fdc26b800 nid=0xa1d in Object.wait() [0x00007f9f9a2e1000]
java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x00000007f712edb0> (a java.lang.ref.ReferenceQueue$Lock)
at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:135)
- locked <0x00000007f712edb0> (a java.lang.ref.ReferenceQueue$Lock)
at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:151)
at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:209)
"Reference Handler" daemon prio=10 tid=0x00007f9fdc269800 nid=0xa1a in Object.wait() [0x00007f9f9a3e2000]
java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x0000000757fae488> (a java.lang.ref.Reference$Lock)
at java.lang.Object.wait(Object.java:503)
at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133)
- locked <0x0000000757fae488> (a java.lang.ref.Reference$Lock)
"main" prio=10 tid=0x00007f9fdc00a800 nid=0xa0a in Object.wait() [0x00007f9fe2d98000]
java.lang.Thread.State: TIMED_WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x00000007d6d74328> (a org.apache.hadoop.hbase.util.JVMClusterUtil$MasterThread)
at java.lang.Thread.join(Thread.java:1289)
- locked <0x00000007d6d74328> (a org.apache.hadoop.hbase.util.JVMClusterUtil$MasterThread)
at org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:131)
at org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:434)
at org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:758)
at org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1161)
at org.apache.hadoop.hbase.TestZooKeeper.after(TestZooKeeper.java:123)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:108)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:78)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:54)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:144)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:203)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:155)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
"VM Thread" prio=10 tid=0x00007f9fdc265000 nid=0xa19 runnable
"GC task thread#0 (ParallelGC)" prio=10 tid=0x00007f9fdc020800 nid=0xa0b runnable
"GC task thread#1 (ParallelGC)" prio=10 tid=0x00007f9fdc022800 nid=0xa0c runnable
"GC task thread#2 (ParallelGC)" prio=10 tid=0x00007f9fdc024000 nid=0xa0d runnable
"GC task thread#3 (ParallelGC)" prio=10 tid=0x00007f9fdc026000 nid=0xa0e runnable
"GC task thread#4 (ParallelGC)" prio=10 tid=0x00007f9fdc028000 nid=0xa0f runnable
"GC task thread#5 (ParallelGC)" prio=10 tid=0x00007f9fdc029800 nid=0xa10 runnable
"GC task thread#6 (ParallelGC)" prio=10 tid=0x00007f9fdc02b800 nid=0xa11 runnable
"GC task thread#7 (ParallelGC)" prio=10 tid=0x00007f9fdc02d800 nid=0xa12 runnable
"GC task thread#8 (ParallelGC)" prio=10 tid=0x00007f9fdc02f800 nid=0xa13 runnable
"GC task thread#9 (ParallelGC)" prio=10 tid=0x00007f9fdc031000 nid=0xa14 runnable
"GC task thread#10 (ParallelGC)" prio=10 tid=0x00007f9fdc033000 nid=0xa15 runnable
"GC task thread#11 (ParallelGC)" prio=10 tid=0x00007f9fdc035000 nid=0xa16 runnable
"GC task thread#12 (ParallelGC)" prio=10 tid=0x00007f9fdc036800 nid=0xa17 runnable
"VM Periodic Task Thread" prio=10 tid=0x00007f9fdc29e800 nid=0xa30 waiting on condition
JNI global references: 169
+ exit 1
POST BUILD TASK : FAILURE
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
Jenkins build is back to normal : HBase-1.3 » latest1.8,Hadoop #442
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/442/changes>
Build failed in Jenkins: HBase-1.3 » latest1.8,Hadoop #441
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/441/changes>
Changes:
[busbey] HBASE-14952 ensure correct components in source assembly.
[busbey] HBASE-14952. ADDENDUM hbase-spark not present in branch-1 yet.
[eclark] HBASE-14984 Allow memcached block cache to set optimze to false
------------------------------------------
[...truncated 49032 lines...]
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:642)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:613)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1007)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:888)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:870)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:857)
at org.apache.hadoop.hbase.filter.TestFilterWrapper.setUp(TestFilterWrapper.java:190)
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=256m; support was removed in 8.0
Running org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.068 sec <<< FAILURE! - in org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter Time elapsed: 0.063 sec <<< ERROR!
java.io.IOException: Shutting down
at org.apache.hadoop.hbase.util.JVMClusterUtil.startup(JVMClusterUtil.java:225)
at org.apache.hadoop.hbase.LocalHBaseCluster.startup(LocalHBaseCluster.java:447)
at org.apache.hadoop.hbase.MiniHBaseCluster.init(MiniHBaseCluster.java:225)
at org.apache.hadoop.hbase.MiniHBaseCluster.<init>(MiniHBaseCluster.java:94)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniHBaseCluster(HBaseTestingUtility.java:1056)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1016)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:888)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:882)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:826)
at org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter.setUpBeforeClass(TestMultiRowRangeFilter.java:61)
Results :
Failed tests:
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testLoadTooMayHFiles(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testLoadTooMayHFiles:537 null
Run 2: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testLoadTooMayHFiles:537 null
Tests in error:
org.apache.hadoop.hbase.TestGlobalMemStoreSize.testGlobalMemStore(org.apache.hadoop.hbase.TestGlobalMemStoreSize)
Run 1: TestGlobalMemStoreSize.testGlobalMemStore » Remote unable to create new native...
Run 2: TestGlobalMemStoreSize.testGlobalMemStore:69 » OutOfMemory unable to create ne...
Run 3: TestGlobalMemStoreSize.testGlobalMemStore:69 » OutOfMemory unable to create ne...
TestNamespace.setUp:72 » Exit java.lang.OutOfMemoryError: unable to create new...
TestConstraint.setUpBeforeClass:64 » OutOfMemory unable to create new native t...
TestFilterWrapper.setUp:190 » OutOfMemory unable to create new native thread
TestFuzzyRowAndColumnRangeFilter.setUpBeforeClass:64 » OutOfMemory unable to c...
TestMultiRowRangeFilter.setUpBeforeClass:61 » IO Shutting down
TestCellCounter.testCellCounteEndTimeRange:177 » FileNotFound target/test-data...
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplit(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:173->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:217->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » TestTimedOut
Run 2: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:173->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:217->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:270 » IO
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom:182->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:217->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » RetriesExhausted
Run 2: TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom » Remote ...
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowColBloom(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowColBloom » Remote
Run 2: TestSecureLoadIncrementalHFiles.testRegionCrossingHFileSplitRowColBloom » IO U...
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testSimpleHFileSplit(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testSimpleHFileSplit:155->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » RetriesExhausted
Run 2: TestSecureLoadIncrementalHFiles.testSimpleHFileSplit » Remote unable to create...
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testSplitALot(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testSplitALot:196->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:276 » RetriesExhausted
Run 2: TestSecureLoadIncrementalHFiles.testSplitALot » Remote unable to create new na...
TestTableInputFormat.beforeClass:88 » OutOfMemory unable to create new native ...
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testRestoreSnapshotDoesNotCreateBackRefLinks(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testRestoreSnapshotDoesNotCreateBackRefLinks:117->TableSnapshotInputFormatTestBase.setupCluster:63 » IO
Run 2: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testRestoreSnapshotDoesNotCreateBackRefLinks:117->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
Run 3: TestTableSnapshotInputFormat.testRestoreSnapshotDoesNotCreateBackRefLinks » OutOfMemory
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMapReduceAndOfflineHBaseMultiRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceAndOfflineHBaseMultiRegion:111->TableSnapshotInputFormatTestBase.testWithMapReduce:161 » YarnRuntime
Run 2: TestTableSnapshotInputFormat.testWithMapReduceAndOfflineHBaseMultiRegion » OutOfMemory
Run 3: TestTableSnapshotInputFormat.testWithMapReduceAndOfflineHBaseMultiRegion » OutOfMemory
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMapReduceMultiRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceMultiRegion:105->TableSnapshotInputFormatTestBase.testWithMapReduce:165->testWithMapReduceImpl:256->doTestWithMapReduce:289 null
Run 2: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceMultiRegion:105->TableSnapshotInputFormatTestBase.testWithMapReduce:160->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
Run 3: TestTableSnapshotInputFormat.testWithMapReduceMultiRegion » OutOfMemory unable...
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMapReduceSingleRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceSingleRegion:100->TableSnapshotInputFormatTestBase.testWithMapReduce:160->TableSnapshotInputFormatTestBase.setupCluster:63 » IO
Run 2: TestTableSnapshotInputFormat.testWithMapReduceSingleRegion » OutOfMemory unabl...
Run 3: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMapReduceSingleRegion:100->TableSnapshotInputFormatTestBase.testWithMapReduce:160->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat.testWithMockedMapReduceMultiRegion(org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat)
Run 1: TestTableSnapshotInputFormat>TableSnapshotInputFormatTestBase.testWithMockedMapReduceMultiRegion:95->testWithMockedMapReduce:192->TableSnapshotInputFormatTestBase.setupCluster:63 » OutOfMemory
Run 2: TestTableSnapshotInputFormat.testWithMockedMapReduceMultiRegion » OutOfMemory ...
Run 3: TestTableSnapshotInputFormat.testWithMockedMapReduceMultiRegion » OutOfMemory ...
org.apache.hadoop.hbase.security.access.TestNamespaceCommands.org.apache.hadoop.hbase.security.access.TestNamespaceCommands
Run 1: TestNamespaceCommands.beforeClass:149 » IO Shutting down
Run 2: TestNamespaceCommands.afterClass:185 » RetriesExhausted Failed after attempts=...
TestScanEarlyTermination.setupBeforeClass:91 » Runtime java.lang.OutOfMemoryEr...
org.apache.hadoop.hbase.security.access.TestScanEarlyTermination.testEarlyScanTermination(org.apache.hadoop.hbase.security.access.TestScanEarlyTermination)
Run 1: TestScanEarlyTermination.testEarlyScanTermination » Remote unable to create ne...
Run 2: TestScanEarlyTermination.tearDown:144 » IO java.lang.reflect.InvocationTargetE...
TestTablePermissions.beforeClass:101 » Runtime java.lang.RuntimeException: jav...
org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelReplicationWithExpAsString.testVisibilityReplication(org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelReplicationWithExpAsString)
Run 1: TestVisibilityLabelReplicationWithExpAsString.setup:139 » IO Shutting down
Run 2: TestVisibilityLabelReplicationWithExpAsString.setup:114 » OutOfMemory unable t...
Run 3: TestVisibilityLabelReplicationWithExpAsString.setup:116 » OutOfMemory unable t...
TestFlushSnapshotFromClient.setupCluster:86 » IO Shutting down
TestRestoreFlushSnapshotFromClient.setUpBeforeClass:77 » OutOfMemory unable to...
Flaked tests:
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testNonHfileFolderWithUnmatchedFamilyName(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testNonHfileFolderWithUnmatchedFamilyName:345->TestLoadIncrementalHFiles.testNonHfileFolder:377 » RetriesExhausted
Run 2: PASS
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testNonexistentColumnFamilyLoad(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testNonexistentColumnFamilyLoad:331->TestLoadIncrementalHFiles.runTest:276 » TestTimedOut
Run 2: PASS
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testRegionCrossingLoad(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testRegionCrossingLoad:118->TestLoadIncrementalHFiles.runTest:231->TestLoadIncrementalHFiles.runTest:241->TestLoadIncrementalHFiles.runTest:254->TestLoadIncrementalHFiles.runTest:280 » RetriesExhausted
Run 2: PASS
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testTableWithCFNameStartWithUnderScore(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles>TestLoadIncrementalHFiles.testTableWithCFNameStartWithUnderScore:563 » RetriesExhausted
Run 2: PASS
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles.testWithoutAnExistingTableAndCreateTableSetToNo(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFiles)
Run 1: TestSecureLoadIncrementalHFiles.testWithoutAnExistingTableAndCreateTableSetToNo »
Run 2: PASS
org.apache.hadoop.hbase.regionserver.TestHRegion.testBatchPut_whileNoRowLocksHeld(org.apache.hadoop.hbase.regionserver.TestHRegion)
Run 1: TestHRegion.testBatchPut_whileNoRowLocksHeld:1487 Metrics Counters should be equal expected:<1056> but was:<1055>
Run 2: PASS
org.apache.hadoop.hbase.security.access.TestAccessController.testNamespaceUserGrant(org.apache.hadoop.hbase.security.access.TestAccessController)
Run 1: TestAccessController.testNamespaceUserGrant » OutOfMemory unable to create new...
Run 2: PASS
org.apache.hadoop.hbase.security.access.TestAccessController.testPostGrantRevoke(org.apache.hadoop.hbase.security.access.TestAccessController)
Run 1: TestAccessController.testPostGrantRevoke » Remote java.lang.OutOfMemoryError: ...
Run 2: PASS
org.apache.hadoop.hbase.security.access.TestAccessController.testTableDeletion(org.apache.hadoop.hbase.security.access.TestAccessController)
Run 1: TestAccessController.testTableDeletion:2225->createTestTable:2257->SecureTestUtil.createTable:669->SecureTestUtil.createTable:691 » IO
Run 2: PASS
org.apache.hadoop.hbase.snapshot.TestSnapshotClientRetries.testSnapshotAlreadyExist(org.apache.hadoop.hbase.snapshot.TestSnapshotClientRetries)
Run 1: TestSnapshotClientRetries.setUp:55 » IO Shutting down
Run 2: PASS
Tests run: 1755, Failures: 1, Errors: 25, Skipped: 37, Flakes: 10
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache HBase ...................................... SUCCESS [1:11.895s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [6.525s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [0.617s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [3.843s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [46.959s]
[INFO] Apache HBase - Common ............................. SUCCESS [4:32.502s]
[INFO] Apache HBase - Procedure .......................... SUCCESS [5:35.464s]
[INFO] Apache HBase - Client ............................. SUCCESS [2:09.039s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [13.502s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [21.779s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [25.554s]
[INFO] Apache HBase - Server ............................. FAILURE [2:51:42.220s]
[INFO] Apache HBase - Testing Util ....................... SKIPPED
[INFO] Apache HBase - Thrift ............................. SKIPPED
[INFO] Apache HBase - Rest ............................... SKIPPED
[INFO] Apache HBase - Shell .............................. SKIPPED
[INFO] Apache HBase - Integration Tests .................. SKIPPED
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 3:07:34.353s
[INFO] Finished at: Wed Dec 16 23:40:37 UTC 2015
[INFO] Final Memory: 377M/3212M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test (secondPartTestsExecution) on project hbase-server: ExecutionException: java.lang.RuntimeException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server> && /home/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.8/jre/bin/java -enableassertions -Dhbase.test -Xmx2800m -XX:MaxPermSize=256m -Djava.security.egd=file:/dev/./urandom -Djava.net.preferIPv4Stack=true -Djava.awt.headless=true -jar <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server/target/surefire/surefirebooter5933983239854825811.jar> <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server/target/surefire/surefire5820196020613052737tmp> <https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/hbase-server/target/surefire/surefire_10924691193655747684866tmp>
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script : # Post-build task script. TODO: Check this in and have all builds reference check-in.
pwd && ls
# NOTE!!!! The below code has been copied and pasted from ./dev-tools/run-test.sh
# Do not change here without syncing there and vice-versa.
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo "Suspicious java process found - waiting 30s to see if there are just slow to stop"
sleep 30
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc -l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo " {color:red}There appear to be $ZOMBIE_TESTS_COUNT zombie tests{color}, they should have been killed by surefire but survived"
jps -v | grep surefirebooter | grep -e '-Dhbase.test'
jps -v | grep surefirebooter | grep -e '-Dhbase.test' | cut -d ' ' -f 1 | xargs -n 1 jstack
# Exit with error
exit 1
else
echo "We're ok: there is no zombie test, but some tests took some time to stop"
fi
else
echo "We're ok: there is no zombie test"
fi
[Hadoop] $ /bin/bash -xe /tmp/hudson5055114759237812355.sh
+ pwd
<https://builds.apache.org/job/HBase-1.3/jdk=latest1.8,label=Hadoop/ws/>
+ ls
hbase-server
++ jps -v
++ grep surefirebooter
++ grep -e -Dhbase.test
++ wc -l
+ ZOMBIE_TESTS_COUNT=1
+ [[ 1 != 0 ]]
+ echo 'Suspicious java process found - waiting 30s to see if there are just slow to stop'
Suspicious java process found - waiting 30s to see if there are just slow to stop
+ sleep 30
++ wc -l
++ grep -e -Dhbase.test
++ grep surefirebooter
++ jps -v
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test, but some tests took some time to stop'
We're ok: there is no zombie test, but some tests took some time to stop
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
ERROR: Publisher 'Publish JUnit test result report' failed: No test report files were found. Configuration error?