You are viewing a plain text version of this content. The canonical link for it is here.
Posted to notifications@asterixdb.apache.org by "Wenhai (JIRA)" <ji...@apache.org> on 2016/11/01 16:48:58 UTC
[jira] [Updated] (ASTERIXDB-1724) Unexpected exception in
ExecutionTest
[ https://issues.apache.org/jira/browse/ASTERIXDB-1724?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Wenhai updated ASTERIXDB-1724:
------------------------------
Description:
Hi, all
Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following:
{noformat}
java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED!
at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3:
< "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
> "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
... 34 more
{noformat}
The excepted result of runtimets/results/api/cluster_state_1 is
{noformat}
{
"cc": {
"configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
"statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
},
"config": {
"api.port": 19002,
"cluster.partitions": {
"0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
"1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1",
"2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
"3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
},
"compiler.framesize": 32768,
"compiler.groupmemory": 163840,
"compiler.joinmemory": 163840,
"compiler.pregelix.home": "~/pregelix",
"compiler.sortmemory": 327680,
"core.dump.paths": {},
"feed.central.manager.port": 4500,
"feed.max.threshold.period": 5,
"feed.memory.available.wait.timeout": 10,
"feed.memory.global.budget": 67108864,
"feed.pending.work.threshold": 50,
"feed.port": 19003,
"instance.name": null,
"log.level": "INFO",
"max.wait.active.cluster": 60,
"metadata.callback.port": 0,
"metadata.node": "asterix_nc1",
"metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
"metadata.port": 0,
"metadata.registration.timeout.secs": 60,
"node.partitions": {
"asterix_nc1": [
"ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
"ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1"
],
"asterix_nc2": [
"ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
"ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
]
},
"node.stores": {
"asterix_nc1": [
"iodevice0",
"iodevice1"
],
"asterix_nc2": [
"iodevice0",
"iodevice1"
]
},
"plot.activate": false,
"storage.buffercache.maxopenfiles": 2147483647,
"storage.buffercache.pagesize": 32768,
"storage.buffercache.size": 33554432,
"storage.lsm.bloomfilter.falsepositiverate": 0.01,
"storage.memorycomponent.globalbudget": 536870912,
"storage.memorycomponent.numcomponents": 2,
"storage.memorycomponent.numpages": 8,
"storage.memorycomponent.pagesize": 131072,
"storage.metadata.memorycomponent.numpages": 256,
"transaction.log.dirs": {
"asterix_nc1": "target/txnLogDir/asterix_nc1",
"asterix_nc2": "target/txnLogDir/asterix_nc2"
},
"txn.commitprofiler.reportinterval": 5,
"txn.job.recovery.memorysize": 67108864,
"txn.lock.escalationthreshold": 1000,
"txn.lock.shrinktimer": 5000,
"txn.lock.timeout.sweepthreshold": 10000,
"txn.lock.timeout.waitthreshold": 60000,
"txn.log.buffer.numpages": 8,
"txn.log.buffer.pagesize": 131072,
"txn.log.checkpoint.history": 0,
"txn.log.checkpoint.lsnthreshold": 67108864,
"txn.log.checkpoint.pollfrequency": 120,
"txn.log.partitionsize": 268435456,
"web.port": 19001,
"web.queryinterface.port": 19006,
"web.secondary.port": 19005
},
"diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
"fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
"metadata_node": "asterix_nc1",
"ncs": [
{
"configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
"node_id": "asterix_nc1",
"partitions": [
{
"active": true,
"partition_id": "partition_0"
},
{
"active": true,
"partition_id": "partition_1"
}
],
"state": "ACTIVE",
"statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
},
{
"configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
"node_id": "asterix_nc2",
"partitions": [
{
"active": true,
"partition_id": "partition_2"
},
{
"active": true,
"partition_id": "partition_3"
}
],
"state": "ACTIVE",
"statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
}
],
"replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
"shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
"state": "ACTIVE",
"versionUri": "http://127.0.0.1:19002/admin/version"
{noformat}
But I got:
{noformat}
{
"cc": {
"statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
"configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
},
"replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
"fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
"state": "ACTIVE",
"metadata_node": "asterix_nc1",
"shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
"versionUri": "http://127.0.0.1:19002/admin/version",
"config": {
"feed.port": 19003,
"txn.log.checkpoint.history": 0,
"compiler.joinmemory": 163840,
"feed.central.manager.port": 4500,
"compiler.sortmemory": 327680,
"node.partitions": {
"asterix_nc1": [
{
"partitionId": 0,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
{
"partitionId": 1,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
}
],
"asterix_nc2": [
{
"partitionId": 2,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
},
{
"partitionId": 3,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
}
]
},
"txn.lock.shrinktimer": 5000,
"metadata.node": "asterix_nc1",
"metadata.port": 0,
"log.level": "INFO",
"api.port": 19002,
"cluster.partitions": {
"0": {
"partitionId": 0,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
"1": {
"partitionId": 1,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
"2": {
"partitionId": 2,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
},
"3": {
"partitionId": 3,
"IODeviceNum": 1,
"active": true,
"activeNodeId": "asterix_nc2",
"nodeId": "asterix_nc2"
}
},
"storage.buffercache.size": 33554432,
"web.port": 19001,
"txn.log.checkpoint.lsnthreshold": 67108864,
"txn.log.partitionsize": 268435456,
"compiler.groupmemory": 163840,
"max.wait.active.cluster": 60,
"metadata.registration.timeout.secs": 60,
"storage.lsm.bloomfilter.falsepositiverate": 0.01,
"plot.activate": false,
"storage.memorycomponent.globalbudget": 536870912,
"web.queryinterface.port": 19006,
"txn.log.buffer.numpages": 8,
"web.secondary.port": 19005,
"feed.memory.global.budget": 67108864,
"compiler.pregelix.home": "~/pregelix",
"compiler.framesize": 32768,
"feed.memory.available.wait.timeout": 10,
"core.dump.paths": {},
"txn.log.buffer.pagesize": 131072,
"txn.job.recovery.memorysize": 67108864,
"storage.metadata.memorycomponent.numpages": 256,
"metadata.partition": {
"partitionId": 0,
"IODeviceNum": 0,
"active": true,
"activeNodeId": "asterix_nc1",
"nodeId": "asterix_nc1"
},
"txn.log.checkpoint.pollfrequency": 120,
"feed.pending.work.threshold": 50,
"feed.max.threshold.period": 5,
"storage.buffercache.maxopenfiles": 2147483647,
"txn.lock.escalationthreshold": 1000,
"txn.lock.timeout.waitthreshold": 60000,
"storage.memorycomponent.numcomponents": 2,
"storage.buffercache.pagesize": 32768,
"storage.memorycomponent.numpages": 8,
"transaction.log.dirs": {
"asterix_nc1": "target/txnLogDir/asterix_nc1",
"asterix_nc2": "target/txnLogDir/asterix_nc2"
},
"txn.commitprofiler.reportinterval": 5,
"metadata.callback.port": 0,
"txn.lock.timeout.sweepthreshold": 10000,
"storage.memorycomponent.pagesize": 131072,
"node.stores": {
"asterix_nc1": [
"iodevice0",
"iodevice1"
],
"asterix_nc2": [
"iodevice0",
"iodevice1"
]
}
},
"diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
"ncs": [
{
"partitions": [
{
"partition_id": "partition_0",
"active": true
},
{
"partition_id": "partition_1",
"active": true
}
],
"statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
"configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
"state": "ACTIVE",
"node_id": "asterix_nc1",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
},
{
"partitions": [
{
"partition_id": "partition_2",
"active": true
},
{
"partition_id": "partition_3",
"active": true
}
],
"statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
"configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
"state": "ACTIVE",
"node_id": "asterix_nc2",
"threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
}
]
}
{noformat}
was:
Hi, all
Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following:
{noformat}
java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED!
at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3:
< "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
> "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
... 34 more
{noformat}
> Unexpected exception in ExecutionTest
> -------------------------------------
>
> Key: ASTERIXDB-1724
> URL: https://issues.apache.org/jira/browse/ASTERIXDB-1724
> Project: Apache AsterixDB
> Issue Type: Test
> Environment: General system.
> Reporter: Wenhai
>
> Hi, all
> Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following:
> {noformat}
> java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED!
> at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
> at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at org.junit.runners.Suite.runChild(Suite.java:128)
> at org.junit.runners.Suite.runChild(Suite.java:27)
> at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
> at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
> at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
> at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
> at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
> at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
> Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3:
> < "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
> > "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
> at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
> at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
> at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
> at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
> ... 34 more
> {noformat}
> The excepted result of runtimets/results/api/cluster_state_1 is
> {noformat}
> {
> "cc": {
> "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
> "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
> "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
> },
> "config": {
> "api.port": 19002,
> "cluster.partitions": {
> "0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
> "1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1",
> "2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
> "3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
> },
> "compiler.framesize": 32768,
> "compiler.groupmemory": 163840,
> "compiler.joinmemory": 163840,
> "compiler.pregelix.home": "~/pregelix",
> "compiler.sortmemory": 327680,
> "core.dump.paths": {},
> "feed.central.manager.port": 4500,
> "feed.max.threshold.period": 5,
> "feed.memory.available.wait.timeout": 10,
> "feed.memory.global.budget": 67108864,
> "feed.pending.work.threshold": 50,
> "feed.port": 19003,
> "instance.name": null,
> "log.level": "INFO",
> "max.wait.active.cluster": 60,
> "metadata.callback.port": 0,
> "metadata.node": "asterix_nc1",
> "metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
> "metadata.port": 0,
> "metadata.registration.timeout.secs": 60,
> "node.partitions": {
> "asterix_nc1": [
> "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
> "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1"
> ],
> "asterix_nc2": [
> "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
> "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
> ]
> },
> "node.stores": {
> "asterix_nc1": [
> "iodevice0",
> "iodevice1"
> ],
> "asterix_nc2": [
> "iodevice0",
> "iodevice1"
> ]
> },
> "plot.activate": false,
> "storage.buffercache.maxopenfiles": 2147483647,
> "storage.buffercache.pagesize": 32768,
> "storage.buffercache.size": 33554432,
> "storage.lsm.bloomfilter.falsepositiverate": 0.01,
> "storage.memorycomponent.globalbudget": 536870912,
> "storage.memorycomponent.numcomponents": 2,
> "storage.memorycomponent.numpages": 8,
> "storage.memorycomponent.pagesize": 131072,
> "storage.metadata.memorycomponent.numpages": 256,
> "transaction.log.dirs": {
> "asterix_nc1": "target/txnLogDir/asterix_nc1",
> "asterix_nc2": "target/txnLogDir/asterix_nc2"
> },
> "txn.commitprofiler.reportinterval": 5,
> "txn.job.recovery.memorysize": 67108864,
> "txn.lock.escalationthreshold": 1000,
> "txn.lock.shrinktimer": 5000,
> "txn.lock.timeout.sweepthreshold": 10000,
> "txn.lock.timeout.waitthreshold": 60000,
> "txn.log.buffer.numpages": 8,
> "txn.log.buffer.pagesize": 131072,
> "txn.log.checkpoint.history": 0,
> "txn.log.checkpoint.lsnthreshold": 67108864,
> "txn.log.checkpoint.pollfrequency": 120,
> "txn.log.partitionsize": 268435456,
> "web.port": 19001,
> "web.queryinterface.port": 19006,
> "web.secondary.port": 19005
> },
> "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
> "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
> "metadata_node": "asterix_nc1",
> "ncs": [
> {
> "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
> "node_id": "asterix_nc1",
> "partitions": [
> {
> "active": true,
> "partition_id": "partition_0"
> },
> {
> "active": true,
> "partition_id": "partition_1"
> }
> ],
> "state": "ACTIVE",
> "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
> "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
> },
> {
> "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
> "node_id": "asterix_nc2",
> "partitions": [
> {
> "active": true,
> "partition_id": "partition_2"
> },
> {
> "active": true,
> "partition_id": "partition_3"
> }
> ],
> "state": "ACTIVE",
> "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
> "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
> }
> ],
> "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
> "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
> "state": "ACTIVE",
> "versionUri": "http://127.0.0.1:19002/admin/version"
> {noformat}
> But I got:
> {noformat}
> {
> "cc": {
> "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
> "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
> "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
> },
> "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
> "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
> "state": "ACTIVE",
> "metadata_node": "asterix_nc1",
> "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
> "versionUri": "http://127.0.0.1:19002/admin/version",
> "config": {
> "feed.port": 19003,
> "txn.log.checkpoint.history": 0,
> "compiler.joinmemory": 163840,
> "feed.central.manager.port": 4500,
> "compiler.sortmemory": 327680,
> "node.partitions": {
> "asterix_nc1": [
> {
> "partitionId": 0,
> "IODeviceNum": 0,
> "active": true,
> "activeNodeId": "asterix_nc1",
> "nodeId": "asterix_nc1"
> },
> {
> "partitionId": 1,
> "IODeviceNum": 1,
> "active": true,
> "activeNodeId": "asterix_nc1",
> "nodeId": "asterix_nc1"
> }
> ],
> "asterix_nc2": [
> {
> "partitionId": 2,
> "IODeviceNum": 0,
> "active": true,
> "activeNodeId": "asterix_nc2",
> "nodeId": "asterix_nc2"
> },
> {
> "partitionId": 3,
> "IODeviceNum": 1,
> "active": true,
> "activeNodeId": "asterix_nc2",
> "nodeId": "asterix_nc2"
> }
> ]
> },
> "txn.lock.shrinktimer": 5000,
> "metadata.node": "asterix_nc1",
> "metadata.port": 0,
> "log.level": "INFO",
> "api.port": 19002,
> "cluster.partitions": {
> "0": {
> "partitionId": 0,
> "IODeviceNum": 0,
> "active": true,
> "activeNodeId": "asterix_nc1",
> "nodeId": "asterix_nc1"
> },
> "1": {
> "partitionId": 1,
> "IODeviceNum": 1,
> "active": true,
> "activeNodeId": "asterix_nc1",
> "nodeId": "asterix_nc1"
> },
> "2": {
> "partitionId": 2,
> "IODeviceNum": 0,
> "active": true,
> "activeNodeId": "asterix_nc2",
> "nodeId": "asterix_nc2"
> },
> "3": {
> "partitionId": 3,
> "IODeviceNum": 1,
> "active": true,
> "activeNodeId": "asterix_nc2",
> "nodeId": "asterix_nc2"
> }
> },
> "storage.buffercache.size": 33554432,
> "web.port": 19001,
> "txn.log.checkpoint.lsnthreshold": 67108864,
> "txn.log.partitionsize": 268435456,
> "compiler.groupmemory": 163840,
> "max.wait.active.cluster": 60,
> "metadata.registration.timeout.secs": 60,
> "storage.lsm.bloomfilter.falsepositiverate": 0.01,
> "plot.activate": false,
> "storage.memorycomponent.globalbudget": 536870912,
> "web.queryinterface.port": 19006,
> "txn.log.buffer.numpages": 8,
> "web.secondary.port": 19005,
> "feed.memory.global.budget": 67108864,
> "compiler.pregelix.home": "~/pregelix",
> "compiler.framesize": 32768,
> "feed.memory.available.wait.timeout": 10,
> "core.dump.paths": {},
> "txn.log.buffer.pagesize": 131072,
> "txn.job.recovery.memorysize": 67108864,
> "storage.metadata.memorycomponent.numpages": 256,
> "metadata.partition": {
> "partitionId": 0,
> "IODeviceNum": 0,
> "active": true,
> "activeNodeId": "asterix_nc1",
> "nodeId": "asterix_nc1"
> },
> "txn.log.checkpoint.pollfrequency": 120,
> "feed.pending.work.threshold": 50,
> "feed.max.threshold.period": 5,
> "storage.buffercache.maxopenfiles": 2147483647,
> "txn.lock.escalationthreshold": 1000,
> "txn.lock.timeout.waitthreshold": 60000,
> "storage.memorycomponent.numcomponents": 2,
> "storage.buffercache.pagesize": 32768,
> "storage.memorycomponent.numpages": 8,
> "transaction.log.dirs": {
> "asterix_nc1": "target/txnLogDir/asterix_nc1",
> "asterix_nc2": "target/txnLogDir/asterix_nc2"
> },
> "txn.commitprofiler.reportinterval": 5,
> "metadata.callback.port": 0,
> "txn.lock.timeout.sweepthreshold": 10000,
> "storage.memorycomponent.pagesize": 131072,
> "node.stores": {
> "asterix_nc1": [
> "iodevice0",
> "iodevice1"
> ],
> "asterix_nc2": [
> "iodevice0",
> "iodevice1"
> ]
> }
> },
> "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
> "ncs": [
> {
> "partitions": [
> {
> "partition_id": "partition_0",
> "active": true
> },
> {
> "partition_id": "partition_1",
> "active": true
> }
> ],
> "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
> "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
> "state": "ACTIVE",
> "node_id": "asterix_nc1",
> "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
> },
> {
> "partitions": [
> {
> "partition_id": "partition_2",
> "active": true
> },
> {
> "partition_id": "partition_3",
> "active": true
> }
> ],
> "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
> "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
> "state": "ACTIVE",
> "node_id": "asterix_nc2",
> "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
> }
> ]
> }
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)