You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@phoenix.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/07/09 04:16:45 UTC

Build failed in Jenkins: Phoenix-4.x-HBase-1.3 #464

See <https://builds.apache.org/job/Phoenix-4.x-HBase-1.3/464/display/redirect?page=changes>

Changes:

[chinmayskulkarni] PHOENIX-5103: Can't create/drop table using 4.14 client against 4.15

------------------------------------------
[...truncated 579.76 KB...]
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA2.N000012: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2467)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17053)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8350)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2170)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2152)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:35076)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2376)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:123)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:188)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:168)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:220)
	at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:314)
	at org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:289)
	at org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:164)
	at org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:159)
	at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:796)
	at org.apache.hadoop.hbase.client.HTableWrapper.getScanner(HTableWrapper.java:215)
	at org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:93)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropChildViews(MetaDataEndpointImpl.java:2505)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:2106)
	... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:717)
	at java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:957)
	at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1367)
	at org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:146)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:287)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:170)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:60)
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
	... 18 more


[INFO] Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 244.513 s - in org.apache.phoenix.end2end.ViewMetadataIT
[ERROR] Tests run: 10, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 269.671 s <<< FAILURE! - in org.apache.phoenix.end2end.TenantSpecificViewIndexIT
[ERROR] testMultiCFViewLocalIndex(org.apache.phoenix.end2end.TenantSpecificViewIndexIT)  Time elapsed: 22.901 s  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA2.N000002: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:656)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17038)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8350)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2170)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2152)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:35076)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2376)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:123)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:188)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:168)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:220)
	at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:314)
	at org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:289)
	at org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:164)
	at org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:159)
	at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:796)
	at org.apache.hadoop.hbase.client.HTableWrapper.getScanner(HTableWrapper.java:215)
	at org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:93)
	at org.apache.phoenix.coprocessor.ViewFinder.findAllRelatives(ViewFinder.java:65)
	at org.apache.phoenix.coprocessor.ViewFinder.findAllRelatives(ViewFinder.java:59)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.findAncestorViews(MetaDataEndpointImpl.java:2564)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.findAncestorViewsOfIndex(MetaDataEndpointImpl.java:2552)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.addDerivedColumnsFromAncestors(MetaDataEndpointImpl.java:744)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.combineColumns(MetaDataEndpointImpl.java:680)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTableFromCache(MetaDataEndpointImpl.java:1929)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:3714)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.combineColumns(MetaDataEndpointImpl.java:695)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTableFromCache(MetaDataEndpointImpl.java:1929)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:3714)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:611)
	... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:717)
	at java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:957)
	at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1367)
	at org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:146)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:287)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:170)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:60)
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
	... 28 more

	at org.apache.phoenix.end2end.TenantSpecificViewIndexIT.createViewAndIndexesWithTenantId(TenantSpecificViewIndexIT.java:191)
	at org.apache.phoenix.end2end.TenantSpecificViewIndexIT.testMultiCFViewIndex(TenantSpecificViewIndexIT.java:127)
	at org.apache.phoenix.end2end.TenantSpecificViewIndexIT.testMultiCFViewLocalIndex(TenantSpecificViewIndexIT.java:92)
Caused by: org.apache.hadoop.hbase.DoNotRetryIOException: 
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA2.N000002: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:656)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17038)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8350)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2170)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2152)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:35076)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2376)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:123)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:188)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:168)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:220)
	at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:314)
	at org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:289)
	at org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:164)
	at org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:159)
	at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:796)
	at org.apache.hadoop.hbase.client.HTableWrapper.getScanner(HTableWrapper.java:215)
	at org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:93)
	at org.apache.phoenix.coprocessor.ViewFinder.findAllRelatives(ViewFinder.java:65)
	at org.apache.phoenix.coprocessor.ViewFinder.findAllRelatives(ViewFinder.java:59)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.findAncestorViews(MetaDataEndpointImpl.java:2564)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.findAncestorViewsOfIndex(MetaDataEndpointImpl.java:2552)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.addDerivedColumnsFromAncestors(MetaDataEndpointImpl.java:744)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.combineColumns(MetaDataEndpointImpl.java:680)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTableFromCache(MetaDataEndpointImpl.java:1929)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:3714)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.combineColumns(MetaDataEndpointImpl.java:695)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTableFromCache(MetaDataEndpointImpl.java:1929)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:3714)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:611)
	... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:717)
	at java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:957)
	at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1367)
	at org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:146)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:287)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:170)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:60)
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
	... 28 more

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
org.apache.hadoop.hbase.DoNotRetryIOException: SCHEMA2.N000002: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:120)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:656)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:17038)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:8350)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execServiceOnRegion(RSRpcServices.java:2170)
	at org.apache.hadoop.hbase.regionserver.RSRpcServices.execService(RSRpcServices.java:2152)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:35076)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2376)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:123)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:188)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:168)
Caused by: java.lang.RuntimeException: java.lang.OutOfMemoryError: unable to create new native thread
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:220)
	at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:314)
	at org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:289)
	at org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:164)
	at org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:159)
	at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:796)
	at org.apache.hadoop.hbase.client.HTableWrapper.getScanner(HTableWrapper.java:215)
	at org.apache.phoenix.coprocessor.ViewFinder.findRelatedViews(ViewFinder.java:93)
	at org.apache.phoenix.coprocessor.ViewFinder.findAllRelatives(ViewFinder.java:65)
	at org.apache.phoenix.coprocessor.ViewFinder.findAllRelatives(ViewFinder.java:59)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.findAncestorViews(MetaDataEndpointImpl.java:2564)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.findAncestorViewsOfIndex(MetaDataEndpointImpl.java:2552)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.addDerivedColumnsFromAncestors(MetaDataEndpointImpl.java:744)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.combineColumns(MetaDataEndpointImpl.java:680)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTableFromCache(MetaDataEndpointImpl.java:1929)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:3714)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.combineColumns(MetaDataEndpointImpl.java:695)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTableFromCache(MetaDataEndpointImpl.java:1929)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.doGetTable(MetaDataEndpointImpl.java:3714)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:611)
	... 9 more
Caused by: java.lang.OutOfMemoryError: unable to create new native thread
	at java.lang.Thread.start0(Native Method)
	at java.lang.Thread.start(Thread.java:717)
	at java.util.concurrent.ThreadPoolExecutor.addWorker(ThreadPoolExecutor.java:957)
	at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1367)
	at org.apache.hadoop.hbase.client.ResultBoundedCompletionService.submit(ResultBoundedCompletionService.java:146)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.addCallsForCurrentReplica(ScannerCallableWithReplicas.java:287)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:170)
	at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:60)
	at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
	... 28 more


[INFO] Running org.apache.phoenix.end2end.index.ViewIndexIT
[WARNING] Tests run: 16, Failures: 0, Errors: 0, Skipped: 2, Time elapsed: 223.85 s - in org.apache.phoenix.end2end.index.ViewIndexIT
[ERROR] Tests run: 56, Failures: 1, Errors: 1, Skipped: 0, Time elapsed: 703.49 s <<< FAILURE! - in org.apache.phoenix.end2end.AlterTableWithViewsIT
[ERROR] testDivergedViewsStayDiverged[AlterTableWithViewsIT_columnEncoded=false, multiTenant=false, salted=false](org.apache.phoenix.end2end.AlterTableWithViewsIT)  Time elapsed: 6.305 s  <<< FAILURE!
java.lang.AssertionError: expected:<504> but was:<101>
	at org.apache.phoenix.end2end.AlterTableWithViewsIT.testDivergedViewsStayDiverged(AlterTableWithViewsIT.java:913)

[ERROR] testAddNewColumnsToBaseTableWithViews[AlterTableWithViewsIT_columnEncoded=false, multiTenant=false, salted=false](org.apache.phoenix.end2end.AlterTableWithViewsIT)  Time elapsed: 4.388 s  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: unable to create new native thread
	at org.apache.phoenix.end2end.AlterTableWithViewsIT.testAddNewColumnsToBaseTableWithViews(AlterTableWithViewsIT.java:146)
Caused by: org.apache.hadoop.ipc.RemoteException: unable to create new native thread

[INFO] 
[INFO] Results:
[INFO] 
[ERROR] Failures: 
[ERROR]   AlterTableWithViewsIT.testDivergedViewsStayDiverged:913 expected:<504> but was:<101>
[ERROR]   ViewIT.doSetup:140->SplitSystemCatalogIT.splitSystemCatalog:77->BaseTest.splitSystemCatalog:1924->BaseTest.splitTable:1895 Multiple regions on asf933.gq1.ygridcore.net,42826,1562644931028
[ERROR] Errors: 
[ERROR]   AlterMultiTenantTableWithViewsIT.testChangingPKOfBaseTableChangesPKForAllViews:197 » PhoenixIO
[ERROR]   AlterTableWithViewsIT.testAddNewColumnsToBaseTableWithViews:146 » PhoenixIO un...
[ERROR]   DropTableWithViewsIT.testDropTableWithChildViews:127 » PhoenixIO org.apache.ha...
[ERROR]   DropTableWithViewsIT.testDropTableWithChildViews:117 » PhoenixIO unable to cre...
[ERROR]   TenantSpecificViewIndexIT.testMultiCFViewLocalIndex:92->testMultiCFViewIndex:127->createViewAndIndexesWithTenantId:191 » PhoenixIO
[ERROR]   TenantSpecificViewIndexSaltedIT.testUpdatableSaltedViewWithLocalIndex:38->BaseTenantSpecificViewIndexIT.testUpdatableView:52->BaseTenantSpecificViewIndexIT.createBaseTable:116 » PhoenixIO
[ERROR]   TenantSpecificViewIndexSaltedIT.testUpdatableViewsWithSameNameDifferentTenants:33->BaseTenantSpecificViewIndexIT.testUpdatableViewsWithSameNameDifferentTenants:78->BaseTenantSpecificViewIndexIT.testUpdatableViewsWithSameNameDifferentTenants:85->BaseTenantSpecificViewIndexIT.createBaseTable:116 » PhoenixIO
[INFO] 
[ERROR] Tests run: 120, Failures: 2, Errors: 7, Skipped: 2
[INFO] 
[INFO] 
[INFO] --- maven-failsafe-plugin:2.20:verify (ParallelStatsEnabledTest) @ phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary for Apache Phoenix 4.15.0-HBase-1.3-SNAPSHOT:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  3.082 s]
[INFO] Phoenix Core ....................................... FAILURE [  02:53 h]
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time:  02:53 h
[INFO] Finished at: 2019-07-09T04:16:37Z
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.20:verify (ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/Phoenix-4.x-HBase-1.3/ws/phoenix-core/target/failsafe-reports> for the individual test results.
[ERROR] Please refer to dump files (if any exist) [date]-jvmRun[N].dump, [date].dumpstream and [date]-jvmRun[N].dumpstream.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
[JIRA] Updating issue PHOENIX-5103
Recording test results

Jenkins build is back to normal : Phoenix-4.x-HBase-1.3 #465

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Phoenix-4.x-HBase-1.3/465/display/redirect?page=changes>