You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@phoenix.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2016/11/05 23:10:58 UTC

Build failed in Jenkins: Phoenix | 4.x-HBase-0.98 #1369

See <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1369/changes>

Changes:

[jamestaylor] PHOENIX-3449 Ignore hanging IndexExtendedIT tests until they can be

[jamestaylor] PHOENIX-3457 Adjust build settings in pom to improve consistency

------------------------------------------
[...truncated 1242 lines...]
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 2,406.296 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1202390: row '  TABLE5_IDX' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
	at org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneIndexTable(CsvBulkLoadToolIT.java:309)
	at org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneLocalIndexTable(CsvBulkLoadToolIT.java:297)
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1202390: row '  TABLE5_IDX' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.net.SocketTimeoutException: Call to penates.apache.org/67.195.81.186:47021 failed because java.net.SocketTimeoutException: 1200000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/67.195.81.186:41574 remote=penates.apache.org/67.195.81.186:47021]
Caused by: java.net.SocketTimeoutException: 1200000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/67.195.81.186:41574 remote=penates.apache.org/67.195.81.186:47021]

testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 2,406.297 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1221676: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1221676: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testInvalidArguments(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 1,200.399 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1222423: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1222423: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testImportWithIndex(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 1,206.216 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1222055: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1222055: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testImportOneIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 1,204.948 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1221818: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1221818: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478375043796.a060877d4e5fc5f1f88f29a6ed2826fc., hostname=penates.apache.org,47021,1478375036900, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)



Results :

Tests in error: 
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)
  Run 1: CsvBulkLoadToolIT.testImportOneLocalIndexTable:297->testImportOneIndexTable:309 » PhoenixIO
  Run 2: CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO

  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportWithLocalIndex(org.apache.phoenix.end2end.CsvBulkLoadToolIT)
  Run 1: CsvBulkLoadToolIT.testImportWithLocalIndex:258 » PhoenixIO callTimeout=1200000...
  Run 2: CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO

  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  MutableIndexFailureIT.testWriteFailureDisablesIndex:134->helpTestWriteFailureDisablesIndex:232 » Commit

Tests run: 171, Failures: 0, Errors: 9, Skipped: 17

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:verify (ParallelStatsEnabledTest) @ phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  2.246 s]
[INFO] Phoenix Core ....................................... FAILURE [  03:59 h]
[INFO] Phoenix - Flume .................................... SKIPPED
[INFO] Phoenix - Pig ...................................... SKIPPED
[INFO] Phoenix Query Server Client ........................ SKIPPED
[INFO] Phoenix Query Server ............................... SKIPPED
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix - Spark .................................... SKIPPED
[INFO] Phoenix - Hive ..................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:59 h
[INFO] Finished at: 2016-11-05T23:05:15+00:00
[INFO] Final Memory: 59M/640M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.19.1:verify (ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/ws/phoenix-core/target/failsafe-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Compressed 1.31 GB of artifacts by 31.7% relative to #1272
Recording test results

Build failed in Jenkins: Phoenix | 4.x-HBase-0.98 #1371

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1371/changes>

Changes:

[jamestaylor] PHOENIX-3457 Adjust build settings in pom to improve consistency

------------------------------------------
[...truncated 1071 lines...]
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 2,406.256 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1202401: row '  TABLE5_IDX' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
	at org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneIndexTable(CsvBulkLoadToolIT.java:309)
	at org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneLocalIndexTable(CsvBulkLoadToolIT.java:297)
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1202401: row '  TABLE5_IDX' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.net.SocketTimeoutException: Call to penates.apache.org/67.195.81.186:46904 failed because java.net.SocketTimeoutException: 1200000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/67.195.81.186:53442 remote=penates.apache.org/67.195.81.186:46904]
Caused by: java.net.SocketTimeoutException: 1200000 millis timeout while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/67.195.81.186:53442 remote=penates.apache.org/67.195.81.186:46904]

testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 2,406.257 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1222100: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1222100: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testInvalidArguments(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 1,199.791 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1221740: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1221740: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testImportWithIndex(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 1,205.839 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1222115: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1222115: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)


testImportOneIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time elapsed: 1,204.477 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, callDuration=1221804: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, callDuration=1221804: row '  TABLE5' on table 'SYSTEM.CATALOG' at region=SYSTEM.CATALOG,,1478411796335.9aed887e4a6afdc0014fee9b4c65b977., hostname=penates.apache.org,46904,1478411786329, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
	at org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1568)
	at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1710)
	at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
	at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
	at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
	at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
	at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
	at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
	at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
	at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
	at java.lang.Thread.run(Thread.java:745)



Results :

Failed tests: 
  MutableIndexFailureIT.testWriteFailureDisablesIndex:134->helpTestWriteFailureDisablesIndex:248 expected:<x[2]> but was:<x[]>
Tests in error: 
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)
  Run 1: CsvBulkLoadToolIT.testImportOneLocalIndexTable:297->testImportOneIndexTable:309 » PhoenixIO
  Run 2: CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO

  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportWithLocalIndex(org.apache.phoenix.end2end.CsvBulkLoadToolIT)
  Run 1: CsvBulkLoadToolIT.testImportWithLocalIndex:258 » PhoenixIO callTimeout=1200000...
  Run 2: CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO

  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO
  CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921 » PhoenixIO

Tests run: 171, Failures: 1, Errors: 8, Skipped: 17

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:verify (ParallelStatsEnabledTest) @ phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  2.368 s]
[INFO] Phoenix Core ....................................... FAILURE [  04:07 h]
[INFO] Phoenix - Flume .................................... SKIPPED
[INFO] Phoenix - Pig ...................................... SKIPPED
[INFO] Phoenix Query Server Client ........................ SKIPPED
[INFO] Phoenix Query Server ............................... SKIPPED
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix - Spark .................................... SKIPPED
[INFO] Phoenix - Hive ..................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 04:07 h
[INFO] Finished at: 2016-11-06T09:17:48+00:00
[INFO] Final Memory: 59M/681M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.19.1:verify (ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/ws/phoenix-core/target/failsafe-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Compressed 1.38 GB of artifacts by 30.0% relative to #1272
Recording test results

Build failed in Jenkins: Phoenix | 4.x-HBase-0.98 #1370

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1370/changes>

Changes:

[jamestaylor] PHOENIX-3439 Query using an RVC based on the base table PK is

[jamestaylor] PHOENIX-3457 Adjust build settings in pom to improve consistency

------------------------------------------
[...truncated 317 lines...]
Running org.apache.phoenix.schema.stats.StatisticsScannerTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.957 sec - in org.apache.phoenix.mapreduce.FormatToBytesWritableMapperTest
Running org.apache.phoenix.schema.RowKeyValueAccessorTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.359 sec - in org.apache.phoenix.filter.SkipScanBigFilterTest
Running org.apache.phoenix.schema.PCharPadTest
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.017 sec - in org.apache.phoenix.schema.PCharPadTest
Running org.apache.phoenix.schema.SortOrderTest
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.018 sec - in org.apache.phoenix.schema.SortOrderTest
Running org.apache.phoenix.schema.types.PrimitiveIntPhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.027 sec - in org.apache.phoenix.schema.types.PrimitiveIntPhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PrimitiveShortPhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.034 sec - in org.apache.phoenix.schema.types.PrimitiveShortPhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PDataTypeForArraysTest
Tests run: 68, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.32 sec - in org.apache.phoenix.schema.types.PDataTypeForArraysTest
Running org.apache.phoenix.schema.types.PVarcharArrayToStringTest
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.003 sec - in org.apache.phoenix.schema.types.PVarcharArrayToStringTest
Running org.apache.phoenix.schema.types.PrimitiveBytePhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.024 sec - in org.apache.phoenix.schema.types.PrimitiveBytePhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PDateArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.03 sec - in org.apache.phoenix.schema.types.PDateArrayToStringTest
Running org.apache.phoenix.schema.types.PrimitiveLongPhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.015 sec - in org.apache.phoenix.schema.types.PrimitiveLongPhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PrimitiveFloatPhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.003 sec - in org.apache.phoenix.schema.types.PrimitiveFloatPhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PrimitiveDoublePhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.038 sec - in org.apache.phoenix.schema.types.PrimitiveDoublePhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PrimitiveBooleanPhoenixArrayToStringTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.027 sec - in org.apache.phoenix.schema.types.PrimitiveBooleanPhoenixArrayToStringTest
Running org.apache.phoenix.schema.types.PDataTypeTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.532 sec - in org.apache.phoenix.schema.RowKeySchemaTest
Running org.apache.phoenix.schema.SchemaUtilTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.019 sec - in org.apache.phoenix.schema.SchemaUtilTest
Running org.apache.phoenix.schema.ValueBitSetTest
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.009 sec - in org.apache.phoenix.schema.ValueBitSetTest
Running org.apache.phoenix.schema.PMetaDataImplTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.02 sec - in org.apache.phoenix.schema.PMetaDataImplTest
Running org.apache.phoenix.schema.SequenceAllocationTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.018 sec - in org.apache.phoenix.schema.SequenceAllocationTest
Running org.apache.phoenix.schema.MutationTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.17 sec - in org.apache.phoenix.schema.RowKeyValueAccessorTest
Running org.apache.phoenix.schema.SaltingUtilTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.012 sec - in org.apache.phoenix.schema.SaltingUtilTest
Running org.apache.phoenix.memory.MemoryManagerTest
Tests run: 35, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.492 sec - in org.apache.phoenix.schema.types.PDataTypeTest
Running org.apache.phoenix.index.IndexMaintainerTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 3, Time elapsed: 0.247 sec - in org.apache.phoenix.memory.MemoryManagerTest
Running org.apache.phoenix.index.automated.MRJobSubmitterTest
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.82 sec - in org.apache.phoenix.schema.stats.StatisticsScannerTest
Running org.apache.phoenix.expression.LikeExpressionTest
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.12 sec - in org.apache.phoenix.mapreduce.util.PhoenixConfigurationUtilTest
Running org.apache.phoenix.expression.ArrayPrependFunctionTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.879 sec - in org.apache.phoenix.index.automated.MRJobSubmitterTest
Running org.apache.phoenix.expression.ArithmeticOperationTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.895 sec - in org.apache.phoenix.expression.LikeExpressionTest
Running org.apache.phoenix.expression.ArrayConcatFunctionTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.104 sec - in org.apache.phoenix.expression.ArithmeticOperationTest
Running org.apache.phoenix.expression.ExpFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.044 sec - in org.apache.phoenix.expression.ExpFunctionTest
Running org.apache.phoenix.expression.DeterminismTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.01 sec - in org.apache.phoenix.expression.DeterminismTest
Running org.apache.phoenix.expression.GetSetByteBitFunctionTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.486 sec - in org.apache.phoenix.schema.MutationTest
Running org.apache.phoenix.expression.util.regex.PatternPerformanceTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.019 sec - in org.apache.phoenix.expression.util.regex.PatternPerformanceTest
Running org.apache.phoenix.expression.ArrayAppendFunctionTest
Tests run: 37, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.933 sec - in org.apache.phoenix.expression.ArrayPrependFunctionTest
Running org.apache.phoenix.expression.RegexpSplitFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.848 sec - in org.apache.phoenix.expression.RegexpSplitFunctionTest
Running org.apache.phoenix.expression.ILikeExpressionTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.1 sec - in org.apache.phoenix.expression.ILikeExpressionTest
Running org.apache.phoenix.expression.SignFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.118 sec - in org.apache.phoenix.expression.SignFunctionTest
Running org.apache.phoenix.expression.NullValueTest
Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.343 sec - in org.apache.phoenix.index.IndexMaintainerTest
Running org.apache.phoenix.expression.ArrayToStringFunctionTest
Tests run: 35, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.275 sec - in org.apache.phoenix.expression.ArrayConcatFunctionTest
Running org.apache.phoenix.expression.RoundFloorCeilExpressionsTest
Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.18 sec - in org.apache.phoenix.expression.ArrayToStringFunctionTest
Running org.apache.phoenix.expression.RegexpSubstrFunctionTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.446 sec - in org.apache.phoenix.expression.NullValueTest
Running org.apache.phoenix.expression.OctetLengthFunctionTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.95 sec - in org.apache.phoenix.expression.GetSetByteBitFunctionTest
Running org.apache.phoenix.expression.CoerceExpressionTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.007 sec - in org.apache.phoenix.expression.CoerceExpressionTest
Running org.apache.phoenix.expression.SortOrderExpressionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.209 sec - in org.apache.phoenix.expression.OctetLengthFunctionTest
Running org.apache.phoenix.expression.ArrayConstructorExpressionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.004 sec - in org.apache.phoenix.expression.ArrayConstructorExpressionTest
Running org.apache.phoenix.expression.PowerFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.171 sec - in org.apache.phoenix.expression.PowerFunctionTest
Running org.apache.phoenix.expression.function.ExternalSqlTypeIdFunctionTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.009 sec - in org.apache.phoenix.expression.function.ExternalSqlTypeIdFunctionTest
Running org.apache.phoenix.expression.function.BuiltinFunctionConstructorTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.094 sec - in org.apache.phoenix.expression.RegexpSubstrFunctionTest
Running org.apache.phoenix.expression.function.InstrFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.008 sec - in org.apache.phoenix.expression.function.InstrFunctionTest
Running org.apache.phoenix.expression.ArrayFillFunctionTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.204 sec - in org.apache.phoenix.expression.function.BuiltinFunctionConstructorTest
Running org.apache.phoenix.expression.RegexpReplaceFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.026 sec - in org.apache.phoenix.expression.RegexpReplaceFunctionTest
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.11 sec - in org.apache.phoenix.expression.ArrayFillFunctionTest
Running org.apache.phoenix.expression.SqrtFunctionTest
Running org.apache.phoenix.expression.CbrtFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.047 sec - in org.apache.phoenix.expression.SqrtFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.043 sec - in org.apache.phoenix.expression.CbrtFunctionTest
Running org.apache.phoenix.expression.ColumnExpressionTest
Running org.apache.phoenix.expression.LnLogFunctionTest
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.033 sec - in org.apache.phoenix.expression.ColumnExpressionTest
Running org.apache.phoenix.expression.AbsFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.083 sec - in org.apache.phoenix.expression.AbsFunctionTest
Running org.apache.phoenix.expression.StringToArrayFunctionTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.118 sec - in org.apache.phoenix.expression.LnLogFunctionTest
Running org.apache.phoenix.query.OrderByTest
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.139 sec - in org.apache.phoenix.expression.StringToArrayFunctionTest
Running org.apache.phoenix.query.ConnectionlessTest
Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.917 sec - in org.apache.phoenix.expression.SortOrderExpressionTest
Running org.apache.phoenix.query.KeyRangeIntersectTest
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.184 sec - in org.apache.phoenix.query.KeyRangeIntersectTest
Running org.apache.phoenix.query.KeyRangeUnionTest
Tests run: 25, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6 sec - in org.apache.phoenix.expression.ArrayAppendFunctionTest
Running org.apache.phoenix.query.HBaseFactoryProviderTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.009 sec - in org.apache.phoenix.query.HBaseFactoryProviderTest
Running org.apache.phoenix.query.ScannerLeaseRenewalTest
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.3 sec - in org.apache.phoenix.query.KeyRangeUnionTest
Running org.apache.phoenix.query.ParallelIteratorsSplitTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.324 sec - in org.apache.phoenix.query.ConnectionlessTest
Running org.apache.phoenix.query.QueryPlanTest
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.777 sec - in org.apache.phoenix.query.OrderByTest
Running org.apache.phoenix.query.PhoenixStatsCacheRemovalListenerTest
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.006 sec - in org.apache.phoenix.query.PhoenixStatsCacheRemovalListenerTest
Running org.apache.phoenix.query.KeyRangeCoalesceTest
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.301 sec - in org.apache.phoenix.query.ParallelIteratorsSplitTest
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.255 sec - in org.apache.phoenix.query.KeyRangeCoalesceTest
Tests run: 23, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.827 sec - in org.apache.phoenix.expression.RoundFloorCeilExpressionsTest
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.852 sec - in org.apache.phoenix.jdbc.SecureUserConnectionsTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.555 sec - in org.apache.phoenix.query.QueryPlanTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.652 sec - in org.apache.phoenix.query.ScannerLeaseRenewalTest

Results :

Tests run: 1432, Failures: 0, Errors: 0, Skipped: 6

[INFO] 
[INFO] --- maven-source-plugin:2.2.1:jar-no-fork (attach-sources) @ phoenix-core ---
[INFO] Building jar: <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1370/artifact/phoenix-core/target/phoenix-core-4.9.0-HBase-0.98-sources.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:test-jar (default) @ phoenix-core ---
[INFO] Building jar: <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1370/artifact/phoenix-core/target/phoenix-core-4.9.0-HBase-0.98-tests.jar>
[INFO] 
[INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ phoenix-core ---
[INFO] Building jar: <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1370/artifact/phoenix-core/target/phoenix-core-4.9.0-HBase-0.98.jar>
[INFO] 
[INFO] --- maven-site-plugin:3.2:attach-descriptor (attach-descriptor) @ phoenix-core ---
[INFO] 
[INFO] --- maven-assembly-plugin:2.5.2:single (core) @ phoenix-core ---
[INFO] Reading assembly descriptor: src/build/phoenix-core.xml
[WARNING] Artifact: org.apache.phoenix:phoenix-core:jar:4.9.0-HBase-0.98 references the same file as the assembly destination file. Moving it to a temporary location for inclusion.
[INFO] Building jar: <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/1370/artifact/phoenix-core/target/phoenix-core-4.9.0-HBase-0.98.jar>
[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:integration-test (ParallelStatsEnabledTest) @ phoenix-core ---

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.phoenix.end2end.MultiCfQueryExecIT
Running org.apache.phoenix.end2end.ParallelIteratorsIT
Running org.apache.phoenix.end2end.KeyOnlyIT
Running org.apache.phoenix.coprocessor.StatisticsCollectionRunTrackerIT
Running org.apache.phoenix.end2end.StatsCollectorIT
Running org.apache.phoenix.end2end.SaltedViewIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.488 sec - in org.apache.phoenix.end2end.ParallelIteratorsIT
Running org.apache.phoenix.end2end.TenantSpecificTablesDDLIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.002 sec - in org.apache.phoenix.end2end.KeyOnlyIT
Running org.apache.phoenix.end2end.TenantSpecificTablesDMLIT
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.938 sec - in org.apache.phoenix.coprocessor.StatisticsCollectionRunTrackerIT
Running org.apache.phoenix.end2end.TransactionalViewIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.122 sec - in org.apache.phoenix.end2end.TransactionalViewIT
Running org.apache.phoenix.end2end.ViewIT
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.236 sec - in org.apache.phoenix.end2end.MultiCfQueryExecIT
Running org.apache.phoenix.end2end.index.ImmutableIndexWithStatsIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.793 sec - in org.apache.phoenix.end2end.index.ImmutableIndexWithStatsIT
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 28.028 sec - in org.apache.phoenix.end2end.SaltedViewIT
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.642 sec - in org.apache.phoenix.end2end.StatsCollectorIT
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 47.455 sec - in org.apache.phoenix.end2end.TenantSpecificTablesDDLIT
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 50.114 sec - in org.apache.phoenix.end2end.TenantSpecificTablesDMLIT
Tests run: 46, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 98.076 sec - in org.apache.phoenix.end2end.ViewIT

Results :

Tests run: 124, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:integration-test (ParallelStatsDisabledTest) @ phoenix-core ---

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:integration-test (ClientManagedTimeTests) @ phoenix-core ---

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Java HotSpot(TM) 64-Bit Server VM warning: INFO: os::commit_memory(0x00000007df800000, 180355072, 0) failed; error='Cannot allocate memory' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (malloc) failed to allocate 180355072 bytes for committing reserved memory.
# An error report file with more information is saved as:
# <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/ws/phoenix-core/hs_err_pid27309.log>
Java HotSpot(TM) 64-Bit Server VM warning: INFO: os::commit_memory(0x00000007e7600000, 331350016, 0) failed; error='Cannot allocate memory' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (malloc) failed to allocate 331350016 bytes for committing reserved memory.
# An error report file with more information is saved as:
# <https://builds.apache.org/job/Phoenix-4.x-HBase-0.98/ws/phoenix-core/hs_err_pid27297.log>
Running org.apache.phoenix.end2end.AggregateQueryIT
Running org.apache.phoenix.end2end.CastAndCoerceIT
Running org.apache.phoenix.end2end.ArrayIT

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:integration-test (HBaseManagedTimeTests) @ phoenix-core ---

-------------------------------------------------------
 T E S T S
-------------------------------------------------------

Results :

Tests run: 0, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:integration-test (NeedTheirOwnClusterTests) @ phoenix-core ---

-------------------------------------------------------
 T E S T S
-------------------------------------------------------
Running org.apache.hadoop.hbase.regionserver.wal.WALReplayWithIndexWritesAndCompressedWALIT
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Recording test results