You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@phoenix.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2014/10/14 10:38:13 UTC
Build failed in Jenkins: Phoenix | 3.0 | Hadoop1 #256
See <https://builds.apache.org/job/Phoenix-3.0-hadoop1/256/changes>
Changes:
[jtaylor] Remove unused imports and local vars
[jtaylor] PHOENIX-1338 Logic to group together parallel scans is incorrect
[jtaylor] PHOENIX-1313 Investigate why LocalIndexIT.testLocalIndexScanAfterRegionSplit() is failing(Rajeshbabu)
[jtaylor] PHOENIX-619 Support DELETE over table with immutable index when possible
[jtaylor] Backport minor fixes from 4.0
------------------------------------------
[...truncated 28319 lines...]
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:94)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:58)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:134)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 231,481.297 sec <<< FAILURE! - in org.apache.phoenix.end2end.ParallelIteratorsIT
org.apache.phoenix.end2end.ParallelIteratorsIT Time elapsed: 231,481.297 sec <<< ERROR!
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:546)
at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:480)
at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:466)
at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:520)
at org.apache.phoenix.end2end.ParallelIteratorsIT.doSetup(ParallelIteratorsIT.java:70)
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 11.678 sec <<< FAILURE! - in org.apache.phoenix.end2end.TenantSpecificTablesDDLIT
org.apache.phoenix.end2end.TenantSpecificTablesDDLIT Time elapsed: 7.693 sec <<< ERROR!
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:563)
at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:480)
at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:466)
at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:520)
at org.apache.phoenix.end2end.BaseTenantSpecificTablesIT.doSetup(BaseTenantSpecificTablesIT.java:80)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runners.Suite.runChild(Suite.java:127)
at org.junit.runners.Suite.runChild(Suite.java:26)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
at org.junit.runner.JUnitCore.run(JUnitCore.java:138)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:113)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:94)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:58)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:134)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Caused by: java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.query.BaseTest.initMiniCluster(BaseTest.java:546)
at org.apache.phoenix.query.BaseTest.setUpTestCluster(BaseTest.java:480)
at org.apache.phoenix.query.BaseTest.checkClusterInitialized(BaseTest.java:466)
at org.apache.phoenix.query.BaseTest.setUpTestDriver(BaseTest.java:520)
at org.apache.phoenix.end2end.BaseTenantSpecificTablesIT.doSetup(BaseTenantSpecificTablesIT.java:80)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runners.Suite.runChild(Suite.java:127)
at org.junit.runners.Suite.runChild(Suite.java:26)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
at org.junit.runner.JUnitCore.run(JUnitCore.java:138)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:113)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:94)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:58)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:134)
at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 231,482.109 sec <<< FAILURE! - in org.apache.phoenix.end2end.ContextClassloaderIT
org.apache.phoenix.end2end.ContextClassloaderIT Time elapsed: 231,482.109 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.end2end.ContextClassloaderIT.setUpBeforeClass(ContextClassloaderIT.java:62)
Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 231,482.278 sec <<< FAILURE! - in org.apache.phoenix.mapreduce.CsvBulkLoadToolIT
org.apache.phoenix.mapreduce.CsvBulkLoadToolIT Time elapsed: 231,482.277 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.mapreduce.CsvBulkLoadToolIT.setUp(CsvBulkLoadToolIT.java:58)
org.apache.phoenix.mapreduce.CsvBulkLoadToolIT Time elapsed: 231,482.278 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.phoenix.mapreduce.CsvBulkLoadToolIT.tearDownAfterClass(CsvBulkLoadToolIT.java:70)
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 231,485.12 sec <<< FAILURE! - in org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT
org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT Time elapsed: 231,485.12 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT.setupCluster(EndToEndCoveredColumnsIndexBuilderIT.java:105)
Running org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 15.708 sec <<< FAILURE! - in org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT
org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT Time elapsed: 11.723 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT.setupCluster(FailWithoutRetriesIT.java:94)
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 231,486.421 sec <<< FAILURE! - in org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT
org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT Time elapsed: 231,486.421 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT.setupCluster(EndtoEndIndexingWithCompressionIT.java:49)
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 231,486.545 sec <<< FAILURE! - in org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT
org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT Time elapsed: 231,486.545 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT.setupCluster(EndToEndCoveredIndexingIT.java:112)
Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 1.199 sec <<< FAILURE! - in org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT
testDoesNotStartRegionServerForUnsupportedCompressionAndVersion(org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT) Time elapsed: 1.164 sec <<< ERROR!
java.lang.NullPointerException: null
at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:422)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:280)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:452)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:620)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:576)
at org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:545)
at org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT.testDoesNotStartRegionServerForUnsupportedCompressionAndVersion(FailForUnsupportedHBaseVersionsIT.java:121)
Results :
Tests in error:
KeyOnlyIT.doSetup:55->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
StatsCollectorIT.doSetup:33->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
MultiCfQueryExecIT.doSetup:56->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
MutableIndexFailureIT.doSetup:99 » NullPointer
MutableIndexFailureIT.doSetup:99 » NullPointer
SaltedViewIT>BaseViewIT.doSetup:51->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
TenantSpecificTablesDMLIT>BaseTenantSpecificTablesIT.doSetup:80->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
ViewIT>BaseViewIT.doSetup:51->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
ParallelIteratorsIT.doSetup:70->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
TenantSpecificTablesDDLIT>BaseTenantSpecificTablesIT.doSetup:80->BaseTest.setUpTestDriver:520->BaseTest.checkClusterInitialized:466->BaseTest.setUpTestCluster:480->BaseTest.initMiniCluster:563 » Runtime
ContextClassloaderIT.setUpBeforeClass:62 » NullPointer
CsvBulkLoadToolIT.setUp:58 » NullPointer
CsvBulkLoadToolIT.tearDownAfterClass:70 NullPointer
EndToEndCoveredColumnsIndexBuilderIT.setupCluster:105 » NullPointer
FailWithoutRetriesIT.setupCluster:94 » NullPointer
EndtoEndIndexingWithCompressionIT.setupCluster:49 » NullPointer
EndToEndCoveredIndexingIT.setupCluster:112 » NullPointer
FailForUnsupportedHBaseVersionsIT.testDoesNotStartRegionServerForUnsupportedCompressionAndVersion:121 » NullPointer
Tests run: 19, Failures: 0, Errors: 18, Skipped: 0
[INFO]
[INFO] --- maven-failsafe-plugin:2.17:verify (HBaseManagedTimeTests) @ phoenix-core ---
[INFO] Failsafe report directory: <https://builds.apache.org/job/Phoenix-3.0-hadoop1/ws/phoenix-core/target/failsafe-reports>
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Phoenix .................................... SUCCESS [7.358s]
[INFO] Phoenix Hadoop Compatibility ...................... SUCCESS [1.043s]
[INFO] Phoenix Core ...................................... FAILURE [3:42.136s]
[INFO] Phoenix - Flume ................................... SKIPPED
[INFO] Phoenix - Pig ..................................... SKIPPED
[INFO] Phoenix Hadoop1 Compatibility ..................... SKIPPED
[INFO] Phoenix Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 3:52.308s
[INFO] Finished at: Tue Oct 14 08:35:31 UTC 2014
[INFO] Final Memory: 36M/501M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.17:verify (HBaseManagedTimeTests) on project phoenix-core: There are test failures.
[ERROR]
[ERROR] Please refer to <https://builds.apache.org/job/Phoenix-3.0-hadoop1/ws/phoenix-core/target/failsafe-reports> for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Sending artifact delta relative to Phoenix | 3.0 | Hadoop1 #255
Archived 430 artifacts
Archive block size is 32768
Received 8 blocks and 47369712 bytes
Compression is 0.6%
Took 37 sec
Updating PHOENIX-1313
Updating PHOENIX-619
Updating PHOENIX-1338
Recording test results
Jenkins build is back to normal : Phoenix | 3.0 | Hadoop1 #257
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Phoenix-3.0-hadoop1/257/changes>