You are viewing a plain text version of this content. The canonical link for it is here.
Posted to notifications@asterixdb.apache.org by "Wenhai (JIRA)" <ji...@apache.org> on 2016/11/01 16:48:58 UTC

[jira] [Updated] (ASTERIXDB-1724) Unexpected exception in ExecutionTest

     [ https://issues.apache.org/jira/browse/ASTERIXDB-1724?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Wenhai updated ASTERIXDB-1724:
------------------------------
    Description: 
Hi, all
Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following:
{noformat}
java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED!
	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
	at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:497)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
	at org.junit.runners.Suite.runChild(Suite.java:128)
	at org.junit.runners.Suite.runChild(Suite.java:27)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
	at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
	at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3:
<         "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
>         "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
	at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
	at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
	... 34 more


{noformat}

The excepted result of runtimets/results/api/cluster_state_1 is
{noformat}
{
    "cc": {
        "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
        "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
        "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
    },
    "config": {
        "api.port": 19002,
        "cluster.partitions": {
            "0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
            "1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1",
            "2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
            "3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
        },
        "compiler.framesize": 32768,
        "compiler.groupmemory": 163840,
        "compiler.joinmemory": 163840,
        "compiler.pregelix.home": "~/pregelix",
        "compiler.sortmemory": 327680,
        "core.dump.paths": {},
        "feed.central.manager.port": 4500,
        "feed.max.threshold.period": 5,
        "feed.memory.available.wait.timeout": 10,
        "feed.memory.global.budget": 67108864,
        "feed.pending.work.threshold": 50,
        "feed.port": 19003,
        "instance.name": null,
        "log.level": "INFO",
        "max.wait.active.cluster": 60,
        "metadata.callback.port": 0,
        "metadata.node": "asterix_nc1",
        "metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
        "metadata.port": 0,
        "metadata.registration.timeout.secs": 60,
        "node.partitions": {
            "asterix_nc1": [
                "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
                "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1"
            ],
            "asterix_nc2": [
                "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
                "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
            ]
        },
        "node.stores": {
            "asterix_nc1": [
                "iodevice0",
                "iodevice1"
            ],
            "asterix_nc2": [
                "iodevice0",
                "iodevice1"
            ]
        },
        "plot.activate": false,
        "storage.buffercache.maxopenfiles": 2147483647,
        "storage.buffercache.pagesize": 32768,
        "storage.buffercache.size": 33554432,
        "storage.lsm.bloomfilter.falsepositiverate": 0.01,
        "storage.memorycomponent.globalbudget": 536870912,
        "storage.memorycomponent.numcomponents": 2,
        "storage.memorycomponent.numpages": 8,
        "storage.memorycomponent.pagesize": 131072,
        "storage.metadata.memorycomponent.numpages": 256,
        "transaction.log.dirs": {
            "asterix_nc1": "target/txnLogDir/asterix_nc1",
            "asterix_nc2": "target/txnLogDir/asterix_nc2"
        },
        "txn.commitprofiler.reportinterval": 5,
        "txn.job.recovery.memorysize": 67108864,
        "txn.lock.escalationthreshold": 1000,
        "txn.lock.shrinktimer": 5000,
        "txn.lock.timeout.sweepthreshold": 10000,
        "txn.lock.timeout.waitthreshold": 60000,
        "txn.log.buffer.numpages": 8,
        "txn.log.buffer.pagesize": 131072,
        "txn.log.checkpoint.history": 0,
        "txn.log.checkpoint.lsnthreshold": 67108864,
        "txn.log.checkpoint.pollfrequency": 120,
        "txn.log.partitionsize": 268435456,
        "web.port": 19001,
        "web.queryinterface.port": 19006,
        "web.secondary.port": 19005
    },
    "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
    "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
    "metadata_node": "asterix_nc1",
    "ncs": [
        {
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
            "node_id": "asterix_nc1",
            "partitions": [
                {
                    "active": true,
                    "partition_id": "partition_0"
                },
                {
                    "active": true,
                    "partition_id": "partition_1"
                }
            ],
            "state": "ACTIVE",
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
        },
        {
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
            "node_id": "asterix_nc2",
            "partitions": [
                {
                    "active": true,
                    "partition_id": "partition_2"
                },
                {
                    "active": true,
                    "partition_id": "partition_3"
                }
            ],
            "state": "ACTIVE",
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
        }
    ],
    "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
    "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
    "state": "ACTIVE",
    "versionUri": "http://127.0.0.1:19002/admin/version"
{noformat}

But I got:
{noformat}
{
    "cc": {
        "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
        "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
        "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
    },
    "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
    "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
    "state": "ACTIVE",
    "metadata_node": "asterix_nc1",
    "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
    "versionUri": "http://127.0.0.1:19002/admin/version",
    "config": {
        "feed.port": 19003,
        "txn.log.checkpoint.history": 0,
        "compiler.joinmemory": 163840,
        "feed.central.manager.port": 4500,
        "compiler.sortmemory": 327680,
        "node.partitions": {
            "asterix_nc1": [
                {
                    "partitionId": 0,
                    "IODeviceNum": 0,
                    "active": true,
                    "activeNodeId": "asterix_nc1",
                    "nodeId": "asterix_nc1"
                },
                {
                    "partitionId": 1,
                    "IODeviceNum": 1,
                    "active": true,
                    "activeNodeId": "asterix_nc1",
                    "nodeId": "asterix_nc1"
                }
            ],
            "asterix_nc2": [
                {
                    "partitionId": 2,
                    "IODeviceNum": 0,
                    "active": true,
                    "activeNodeId": "asterix_nc2",
                    "nodeId": "asterix_nc2"
                },
                {
                    "partitionId": 3,
                    "IODeviceNum": 1,
                    "active": true,
                    "activeNodeId": "asterix_nc2",
                    "nodeId": "asterix_nc2"
                }
            ]
        },
        "txn.lock.shrinktimer": 5000,
        "metadata.node": "asterix_nc1",
        "metadata.port": 0,
        "log.level": "INFO",
        "api.port": 19002,
        "cluster.partitions": {
            "0": {
                "partitionId": 0,
                "IODeviceNum": 0,
                "active": true,
                "activeNodeId": "asterix_nc1",
                "nodeId": "asterix_nc1"
            },
            "1": {
                "partitionId": 1,
                "IODeviceNum": 1,
                "active": true,
                "activeNodeId": "asterix_nc1",
                "nodeId": "asterix_nc1"
            },
            "2": {
                "partitionId": 2,
                "IODeviceNum": 0,
                "active": true,
                "activeNodeId": "asterix_nc2",
                "nodeId": "asterix_nc2"
            },
            "3": {
                "partitionId": 3,
                "IODeviceNum": 1,
                "active": true,
                "activeNodeId": "asterix_nc2",
                "nodeId": "asterix_nc2"
            }
        },
        "storage.buffercache.size": 33554432,
        "web.port": 19001,
        "txn.log.checkpoint.lsnthreshold": 67108864,
        "txn.log.partitionsize": 268435456,
        "compiler.groupmemory": 163840,
        "max.wait.active.cluster": 60,
        "metadata.registration.timeout.secs": 60,
        "storage.lsm.bloomfilter.falsepositiverate": 0.01,
        "plot.activate": false,
        "storage.memorycomponent.globalbudget": 536870912,
        "web.queryinterface.port": 19006,
        "txn.log.buffer.numpages": 8,
        "web.secondary.port": 19005,
        "feed.memory.global.budget": 67108864,
        "compiler.pregelix.home": "~/pregelix",
        "compiler.framesize": 32768,
        "feed.memory.available.wait.timeout": 10,
        "core.dump.paths": {},
        "txn.log.buffer.pagesize": 131072,
        "txn.job.recovery.memorysize": 67108864,
        "storage.metadata.memorycomponent.numpages": 256,
        "metadata.partition": {
            "partitionId": 0,
            "IODeviceNum": 0,
            "active": true,
            "activeNodeId": "asterix_nc1",
            "nodeId": "asterix_nc1"
        },
        "txn.log.checkpoint.pollfrequency": 120,
        "feed.pending.work.threshold": 50,
        "feed.max.threshold.period": 5,
        "storage.buffercache.maxopenfiles": 2147483647,
        "txn.lock.escalationthreshold": 1000,
        "txn.lock.timeout.waitthreshold": 60000,
        "storage.memorycomponent.numcomponents": 2,
        "storage.buffercache.pagesize": 32768,
        "storage.memorycomponent.numpages": 8,
        "transaction.log.dirs": {
            "asterix_nc1": "target/txnLogDir/asterix_nc1",
            "asterix_nc2": "target/txnLogDir/asterix_nc2"
        },
        "txn.commitprofiler.reportinterval": 5,
        "metadata.callback.port": 0,
        "txn.lock.timeout.sweepthreshold": 10000,
        "storage.memorycomponent.pagesize": 131072,
        "node.stores": {
            "asterix_nc1": [
                "iodevice0",
                "iodevice1"
            ],
            "asterix_nc2": [
                "iodevice0",
                "iodevice1"
            ]
        }
    },
    "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
    "ncs": [
        {
            "partitions": [
                {
                    "partition_id": "partition_0",
                    "active": true
                },
                {
                    "partition_id": "partition_1",
                    "active": true
                }
            ],
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
            "state": "ACTIVE",
            "node_id": "asterix_nc1",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
        },
        {
            "partitions": [
                {
                    "partition_id": "partition_2",
                    "active": true
                },
                {
                    "partition_id": "partition_3",
                    "active": true
                }
            ],
            "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
            "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
            "state": "ACTIVE",
            "node_id": "asterix_nc2",
            "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
        }
    ]
}
{noformat}

  was:
Hi, all
Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following:
{noformat}
java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED!
	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
	at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:497)
	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
	at org.junit.runners.Suite.runChild(Suite.java:128)
	at org.junit.runners.Suite.runChild(Suite.java:27)
	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
	at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
	at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3:
<         "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
>         "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
	at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
	at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
	... 34 more


{noformat}


> Unexpected exception in ExecutionTest
> -------------------------------------
>
>                 Key: ASTERIXDB-1724
>                 URL: https://issues.apache.org/jira/browse/ASTERIXDB-1724
>             Project: Apache AsterixDB
>          Issue Type: Test
>         Environment: General system.
>            Reporter: Wenhai
>
> Hi, all
> Yestoday, after I applying the latest master onto the branch. I got a series of error (13 testcases) that should not be posed by my branch, like the following:
> {noformat}
> java.lang.Exception: Test "src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql" FAILED!
> 	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1028)
> 	at org.apache.asterix.test.runtime.ExecutionTest.test(ExecutionTest.java:127)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:497)
> 	at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 	at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 	at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 	at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 	at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 	at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 	at org.junit.runners.Suite.runChild(Suite.java:128)
> 	at org.junit.runners.Suite.runChild(Suite.java:27)
> 	at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 	at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 	at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 	at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 	at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 	at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 	at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 	at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 	at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:86)
> 	at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
> 	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:459)
> 	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:675)
> 	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:382)
> 	at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:192)
> Caused by: org.apache.asterix.test.base.ComparisonException: Result for src/test/resources/runtimets/queries/api/cluster_state_1/cluster_state_1.1.cstate.aql changed at line 3:
> <         "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
> >         "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
> 	at org.apache.asterix.test.aql.TestExecutor.throwLineChanged(TestExecutor.java:203)
> 	at org.apache.asterix.test.aql.TestExecutor.runScriptAndCompareWithResult(TestExecutor.java:163)
> 	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:870)
> 	at org.apache.asterix.test.aql.TestExecutor.executeTest(TestExecutor.java:1000)
> 	... 34 more
> {noformat}
> The excepted result of runtimets/results/api/cluster_state_1 is
> {noformat}
> {
>     "cc": {
>         "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
>         "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
>         "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
>     },
>     "config": {
>         "api.port": 19002,
>         "cluster.partitions": {
>             "0": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
>             "1": "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1",
>             "2": "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
>             "3": "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
>         },
>         "compiler.framesize": 32768,
>         "compiler.groupmemory": 163840,
>         "compiler.joinmemory": 163840,
>         "compiler.pregelix.home": "~/pregelix",
>         "compiler.sortmemory": 327680,
>         "core.dump.paths": {},
>         "feed.central.manager.port": 4500,
>         "feed.max.threshold.period": 5,
>         "feed.memory.available.wait.timeout": 10,
>         "feed.memory.global.budget": 67108864,
>         "feed.pending.work.threshold": 50,
>         "feed.port": 19003,
>         "instance.name": null,
>         "log.level": "INFO",
>         "max.wait.active.cluster": 60,
>         "metadata.callback.port": 0,
>         "metadata.node": "asterix_nc1",
>         "metadata.partition": "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
>         "metadata.port": 0,
>         "metadata.registration.timeout.secs": 60,
>         "node.partitions": {
>             "asterix_nc1": [
>                 "ID:0, Original Node: asterix_nc1, IODevice: 0, Active Node: asterix_nc1",
>                 "ID:1, Original Node: asterix_nc1, IODevice: 1, Active Node: asterix_nc1"
>             ],
>             "asterix_nc2": [
>                 "ID:2, Original Node: asterix_nc2, IODevice: 0, Active Node: asterix_nc2",
>                 "ID:3, Original Node: asterix_nc2, IODevice: 1, Active Node: asterix_nc2"
>             ]
>         },
>         "node.stores": {
>             "asterix_nc1": [
>                 "iodevice0",
>                 "iodevice1"
>             ],
>             "asterix_nc2": [
>                 "iodevice0",
>                 "iodevice1"
>             ]
>         },
>         "plot.activate": false,
>         "storage.buffercache.maxopenfiles": 2147483647,
>         "storage.buffercache.pagesize": 32768,
>         "storage.buffercache.size": 33554432,
>         "storage.lsm.bloomfilter.falsepositiverate": 0.01,
>         "storage.memorycomponent.globalbudget": 536870912,
>         "storage.memorycomponent.numcomponents": 2,
>         "storage.memorycomponent.numpages": 8,
>         "storage.memorycomponent.pagesize": 131072,
>         "storage.metadata.memorycomponent.numpages": 256,
>         "transaction.log.dirs": {
>             "asterix_nc1": "target/txnLogDir/asterix_nc1",
>             "asterix_nc2": "target/txnLogDir/asterix_nc2"
>         },
>         "txn.commitprofiler.reportinterval": 5,
>         "txn.job.recovery.memorysize": 67108864,
>         "txn.lock.escalationthreshold": 1000,
>         "txn.lock.shrinktimer": 5000,
>         "txn.lock.timeout.sweepthreshold": 10000,
>         "txn.lock.timeout.waitthreshold": 60000,
>         "txn.log.buffer.numpages": 8,
>         "txn.log.buffer.pagesize": 131072,
>         "txn.log.checkpoint.history": 0,
>         "txn.log.checkpoint.lsnthreshold": 67108864,
>         "txn.log.checkpoint.pollfrequency": 120,
>         "txn.log.partitionsize": 268435456,
>         "web.port": 19001,
>         "web.queryinterface.port": 19006,
>         "web.secondary.port": 19005
>     },
>     "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
>     "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
>     "metadata_node": "asterix_nc1",
>     "ncs": [
>         {
>             "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
>             "node_id": "asterix_nc1",
>             "partitions": [
>                 {
>                     "active": true,
>                     "partition_id": "partition_0"
>                 },
>                 {
>                     "active": true,
>                     "partition_id": "partition_1"
>                 }
>             ],
>             "state": "ACTIVE",
>             "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
>             "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
>         },
>         {
>             "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
>             "node_id": "asterix_nc2",
>             "partitions": [
>                 {
>                     "active": true,
>                     "partition_id": "partition_2"
>                 },
>                 {
>                     "active": true,
>                     "partition_id": "partition_3"
>                 }
>             ],
>             "state": "ACTIVE",
>             "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
>             "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
>         }
>     ],
>     "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
>     "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
>     "state": "ACTIVE",
>     "versionUri": "http://127.0.0.1:19002/admin/version"
> {noformat}
> But I got:
> {noformat}
> {
>     "cc": {
>         "statsUri": "http://127.0.0.1:19002/admin/cluster/cc/stats",
>         "configUri": "http://127.0.0.1:19002/admin/cluster/cc/config",
>         "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/cc/threaddump"
>     },
>     "replicationUri": "http://127.0.0.1:19002/admin/cluster/replication",
>     "fullShutdownUri": "http://127.0.0.1:19002/admin/shutdown?all=true",
>     "state": "ACTIVE",
>     "metadata_node": "asterix_nc1",
>     "shutdownUri": "http://127.0.0.1:19002/admin/shutdown",
>     "versionUri": "http://127.0.0.1:19002/admin/version",
>     "config": {
>         "feed.port": 19003,
>         "txn.log.checkpoint.history": 0,
>         "compiler.joinmemory": 163840,
>         "feed.central.manager.port": 4500,
>         "compiler.sortmemory": 327680,
>         "node.partitions": {
>             "asterix_nc1": [
>                 {
>                     "partitionId": 0,
>                     "IODeviceNum": 0,
>                     "active": true,
>                     "activeNodeId": "asterix_nc1",
>                     "nodeId": "asterix_nc1"
>                 },
>                 {
>                     "partitionId": 1,
>                     "IODeviceNum": 1,
>                     "active": true,
>                     "activeNodeId": "asterix_nc1",
>                     "nodeId": "asterix_nc1"
>                 }
>             ],
>             "asterix_nc2": [
>                 {
>                     "partitionId": 2,
>                     "IODeviceNum": 0,
>                     "active": true,
>                     "activeNodeId": "asterix_nc2",
>                     "nodeId": "asterix_nc2"
>                 },
>                 {
>                     "partitionId": 3,
>                     "IODeviceNum": 1,
>                     "active": true,
>                     "activeNodeId": "asterix_nc2",
>                     "nodeId": "asterix_nc2"
>                 }
>             ]
>         },
>         "txn.lock.shrinktimer": 5000,
>         "metadata.node": "asterix_nc1",
>         "metadata.port": 0,
>         "log.level": "INFO",
>         "api.port": 19002,
>         "cluster.partitions": {
>             "0": {
>                 "partitionId": 0,
>                 "IODeviceNum": 0,
>                 "active": true,
>                 "activeNodeId": "asterix_nc1",
>                 "nodeId": "asterix_nc1"
>             },
>             "1": {
>                 "partitionId": 1,
>                 "IODeviceNum": 1,
>                 "active": true,
>                 "activeNodeId": "asterix_nc1",
>                 "nodeId": "asterix_nc1"
>             },
>             "2": {
>                 "partitionId": 2,
>                 "IODeviceNum": 0,
>                 "active": true,
>                 "activeNodeId": "asterix_nc2",
>                 "nodeId": "asterix_nc2"
>             },
>             "3": {
>                 "partitionId": 3,
>                 "IODeviceNum": 1,
>                 "active": true,
>                 "activeNodeId": "asterix_nc2",
>                 "nodeId": "asterix_nc2"
>             }
>         },
>         "storage.buffercache.size": 33554432,
>         "web.port": 19001,
>         "txn.log.checkpoint.lsnthreshold": 67108864,
>         "txn.log.partitionsize": 268435456,
>         "compiler.groupmemory": 163840,
>         "max.wait.active.cluster": 60,
>         "metadata.registration.timeout.secs": 60,
>         "storage.lsm.bloomfilter.falsepositiverate": 0.01,
>         "plot.activate": false,
>         "storage.memorycomponent.globalbudget": 536870912,
>         "web.queryinterface.port": 19006,
>         "txn.log.buffer.numpages": 8,
>         "web.secondary.port": 19005,
>         "feed.memory.global.budget": 67108864,
>         "compiler.pregelix.home": "~/pregelix",
>         "compiler.framesize": 32768,
>         "feed.memory.available.wait.timeout": 10,
>         "core.dump.paths": {},
>         "txn.log.buffer.pagesize": 131072,
>         "txn.job.recovery.memorysize": 67108864,
>         "storage.metadata.memorycomponent.numpages": 256,
>         "metadata.partition": {
>             "partitionId": 0,
>             "IODeviceNum": 0,
>             "active": true,
>             "activeNodeId": "asterix_nc1",
>             "nodeId": "asterix_nc1"
>         },
>         "txn.log.checkpoint.pollfrequency": 120,
>         "feed.pending.work.threshold": 50,
>         "feed.max.threshold.period": 5,
>         "storage.buffercache.maxopenfiles": 2147483647,
>         "txn.lock.escalationthreshold": 1000,
>         "txn.lock.timeout.waitthreshold": 60000,
>         "storage.memorycomponent.numcomponents": 2,
>         "storage.buffercache.pagesize": 32768,
>         "storage.memorycomponent.numpages": 8,
>         "transaction.log.dirs": {
>             "asterix_nc1": "target/txnLogDir/asterix_nc1",
>             "asterix_nc2": "target/txnLogDir/asterix_nc2"
>         },
>         "txn.commitprofiler.reportinterval": 5,
>         "metadata.callback.port": 0,
>         "txn.lock.timeout.sweepthreshold": 10000,
>         "storage.memorycomponent.pagesize": 131072,
>         "node.stores": {
>             "asterix_nc1": [
>                 "iodevice0",
>                 "iodevice1"
>             ],
>             "asterix_nc2": [
>                 "iodevice0",
>                 "iodevice1"
>             ]
>         }
>     },
>     "diagnosticsUri": "http://127.0.0.1:19002/admin/diagnostics",
>     "ncs": [
>         {
>             "partitions": [
>                 {
>                     "partition_id": "partition_0",
>                     "active": true
>                 },
>                 {
>                     "partition_id": "partition_1",
>                     "active": true
>                 }
>             ],
>             "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/stats",
>             "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/config",
>             "state": "ACTIVE",
>             "node_id": "asterix_nc1",
>             "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc1/threaddump"
>         },
>         {
>             "partitions": [
>                 {
>                     "partition_id": "partition_2",
>                     "active": true
>                 },
>                 {
>                     "partition_id": "partition_3",
>                     "active": true
>                 }
>             ],
>             "statsUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/stats",
>             "configUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/config",
>             "state": "ACTIVE",
>             "node_id": "asterix_nc2",
>             "threadDumpUri": "http://127.0.0.1:19002/admin/cluster/node/asterix_nc2/threaddump"
>         }
>     ]
> }
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)