You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@geode.apache.org by hi...@apache.org on 2016/06/15 00:07:38 UTC
[22/94] [abbrv] [partial] incubator-geode git commit: GEODE-837:
update tests from JUnit3 to JUnit4
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAExpiryDUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAExpiryDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAExpiryDUnitTest.java
index 0e8514f..939e81e 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAExpiryDUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAExpiryDUnitTest.java
@@ -16,8 +16,13 @@
*/
package com.gemstone.gemfire.internal.cache.ha;
+import static org.junit.Assert.*;
+
import java.util.Properties;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
import com.gemstone.gemfire.cache.AttributesFactory;
import com.gemstone.gemfire.cache.Cache;
import com.gemstone.gemfire.cache.CacheException;
@@ -34,13 +39,14 @@ import com.gemstone.gemfire.distributed.DistributedSystem;
import com.gemstone.gemfire.internal.cache.EventID;
import com.gemstone.gemfire.internal.cache.HARegion;
import com.gemstone.gemfire.internal.cache.RegionQueue;
-import com.gemstone.gemfire.test.dunit.DistributedTestCase;
import com.gemstone.gemfire.test.dunit.Host;
import com.gemstone.gemfire.test.dunit.Invoke;
import com.gemstone.gemfire.test.dunit.SerializableRunnable;
import com.gemstone.gemfire.test.dunit.VM;
import com.gemstone.gemfire.test.dunit.Wait;
import com.gemstone.gemfire.test.dunit.WaitCriterion;
+import com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase;
+import com.gemstone.gemfire.test.junit.categories.DistributedTest;
/**
* This test checks Expiration of events in the regionqueue.
@@ -49,11 +55,9 @@ import com.gemstone.gemfire.test.dunit.WaitCriterion;
* 3. Checks size of the regionqueue. Size of the regionqueue should be greater than 0.
* 4. Waits for the period which is slightly greater than expiration period.
* 5. Checks size of the regionqueue. Size of the regionqueue should be equal to 0.
- *
*/
-
-public class HAExpiryDUnitTest extends DistributedTestCase
-{
+@Category(DistributedTest.class)
+public class HAExpiryDUnitTest extends JUnit4DistributedTestCase {
VM vm0 = null;
@@ -73,8 +77,8 @@ public class HAExpiryDUnitTest extends DistributedTestCase
protected static int regionQueueSize = -1;
- public HAExpiryDUnitTest(String name) {
- super(name);
+ public HAExpiryDUnitTest() {
+ super();
}
/**
@@ -103,6 +107,7 @@ public class HAExpiryDUnitTest extends DistributedTestCase
Invoke.invokeInEveryVM(new SerializableRunnable() { public void run() { cache = null; } });
}
+ @Test
public void testExpiryPeriod() throws Exception
{
vm0.invoke(() -> HAExpiryDUnitTest.createRegionQueue(new Boolean(false)));
@@ -132,6 +137,7 @@ public class HAExpiryDUnitTest extends DistributedTestCase
vm3.invoke(() -> HAExpiryDUnitTest.checkSizeAfterExpiration());
}
+ @Test
public void testDurableExpiryPeriod() throws Exception
{
vm0.invoke(() -> HAExpiryDUnitTest.createRegionQueue(new Boolean(true)));
@@ -234,7 +240,7 @@ public class HAExpiryDUnitTest extends DistributedTestCase
public static void createRegionQueue(Boolean isDurable) throws Exception
{
- new HAExpiryDUnitTest("temp").createCache(new Properties());
+ new HAExpiryDUnitTest().createCache(new Properties());
HARegionQueueAttributes hattr = new HARegionQueueAttributes();
// setting expiry time for the regionqueue.
hattr.setExpiryTime(4);
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIBugDUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIBugDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIBugDUnitTest.java
index b6b4071..5e659f5 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIBugDUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIBugDUnitTest.java
@@ -16,10 +16,16 @@
*/
package com.gemstone.gemfire.internal.cache.ha;
+import static org.junit.Assert.*;
+
import java.util.ArrayList;
import java.util.Iterator;
import java.util.Properties;
+import org.junit.Ignore;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
import com.gemstone.gemfire.cache.AttributesFactory;
import com.gemstone.gemfire.cache.Cache;
import com.gemstone.gemfire.cache.CacheException;
@@ -28,9 +34,9 @@ import com.gemstone.gemfire.cache.CacheListener;
import com.gemstone.gemfire.cache.DataPolicy;
import com.gemstone.gemfire.cache.EntryEvent;
import com.gemstone.gemfire.cache.Region;
+import com.gemstone.gemfire.cache.Region.Entry;
import com.gemstone.gemfire.cache.RegionAttributes;
import com.gemstone.gemfire.cache.Scope;
-import com.gemstone.gemfire.cache.Region.Entry;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.cache30.CacheSerializableRunnable;
import com.gemstone.gemfire.distributed.DistributedSystem;
@@ -39,13 +45,14 @@ import com.gemstone.gemfire.internal.cache.HARegion;
import com.gemstone.gemfire.internal.cache.RegionQueue;
import com.gemstone.gemfire.test.dunit.Assert;
import com.gemstone.gemfire.test.dunit.AsyncInvocation;
-import com.gemstone.gemfire.test.dunit.DistributedTestCase;
import com.gemstone.gemfire.test.dunit.Host;
import com.gemstone.gemfire.test.dunit.Invoke;
import com.gemstone.gemfire.test.dunit.LogWriterUtils;
import com.gemstone.gemfire.test.dunit.SerializableRunnable;
import com.gemstone.gemfire.test.dunit.ThreadUtils;
import com.gemstone.gemfire.test.dunit.VM;
+import com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase;
+import com.gemstone.gemfire.test.junit.categories.DistributedTest;
/**
* This test points out the bug when GII of HARegion Queue is happening and at the same time it is receiving puts from peer to peer.
@@ -55,11 +62,9 @@ import com.gemstone.gemfire.test.dunit.VM;
* 4. Start vm1 and create HARegion Queue asynchronously and notify vm0 after its creation
* 5. Put the data from vm0 asynchronously till HARegion Queue gets created in vm1.
* 6. Validate the data. Puts happened during GII should be missed.
- *
*/
-
-public class HAGIIBugDUnitTest extends DistributedTestCase
-{
+@Category(DistributedTest.class)
+public class HAGIIBugDUnitTest extends JUnit4DistributedTestCase {
VM vm0 = null;
@@ -133,15 +138,9 @@ public class HAGIIBugDUnitTest extends DistributedTestCase
assertNotNull(cache);
}
-
- public void testDummy() throws Exception
- {
- LogWriterUtils.getLogWriter().info("This is Dummy test for the GII");
- }
-
-
- public void _testGIIBug() throws Exception
- {
+ @Ignore("TODO")
+ @Test
+ public void testGIIBug() throws Exception {
vm0.invoke(putFromVmBeforeGII("vm0_1"));
populateKeySet("vm0_1");
@@ -336,7 +335,7 @@ public class HAGIIBugDUnitTest extends DistributedTestCase
public static void createRegionQueue() throws Exception
{
- new HAGIIBugDUnitTest("temp").createCache(new Properties());
+ new HAGIIBugDUnitTest().createCache(new Properties());
HARegionQueueAttributes hattr = new HARegionQueueAttributes();
// setting expiry time for the regionqueue.
hattr.setExpiryTime(12000000);
@@ -361,11 +360,6 @@ public class HAGIIBugDUnitTest extends DistributedTestCase
}
}
- public HAGIIBugDUnitTest(String arg0) {
- super(arg0);
-
- }
-
}
/**
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIDUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIDUnitTest.java
index fabe3e6..62375f9 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIDUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HAGIIDUnitTest.java
@@ -14,10 +14,29 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
-
package com.gemstone.gemfire.internal.cache.ha;
-import com.gemstone.gemfire.cache.*;
+import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.*;
+import static com.gemstone.gemfire.test.dunit.Assert.*;
+
+import java.util.Collections;
+import java.util.Iterator;
+import java.util.Map;
+import java.util.Properties;
+
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
+import com.gemstone.gemfire.cache.AttributesFactory;
+import com.gemstone.gemfire.cache.Cache;
+import com.gemstone.gemfire.cache.CacheFactory;
+import com.gemstone.gemfire.cache.DataPolicy;
+import com.gemstone.gemfire.cache.EntryEvent;
+import com.gemstone.gemfire.cache.InterestResultPolicy;
+import com.gemstone.gemfire.cache.Operation;
+import com.gemstone.gemfire.cache.Region;
+import com.gemstone.gemfire.cache.RegionAttributes;
+import com.gemstone.gemfire.cache.Scope;
import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.cache30.ClientServerTestCase;
@@ -32,86 +51,35 @@ import com.gemstone.gemfire.internal.cache.tier.sockets.CacheClientNotifier;
import com.gemstone.gemfire.internal.cache.tier.sockets.ClientTombstoneMessage;
import com.gemstone.gemfire.internal.cache.tier.sockets.ConflationDUnitTest;
import com.gemstone.gemfire.internal.cache.versions.VersionSource;
-import com.gemstone.gemfire.test.dunit.*;
-
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.Map;
-import java.util.Properties;
-
-import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.LOCATORS;
-import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.MCAST_PORT;
+import com.gemstone.gemfire.test.dunit.Assert;
+import com.gemstone.gemfire.test.dunit.Host;
+import com.gemstone.gemfire.test.dunit.Invoke;
+import com.gemstone.gemfire.test.dunit.NetworkUtils;
+import com.gemstone.gemfire.test.dunit.VM;
+import com.gemstone.gemfire.test.dunit.Wait;
+import com.gemstone.gemfire.test.dunit.WaitCriterion;
+import com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase;
+import com.gemstone.gemfire.test.junit.categories.DistributedTest;
/**
* Client is connected to S1 which has a slow dispatcher. Puts are made on S1. Then S2 is started
* and made available for the client. After that , S1 's server is stopped. The client fails over to
* S2. The client should receive all the puts . These puts have arrived on S2 via GII of HARegion.
- *
- *
*/
+@Category(DistributedTest.class)
+public class HAGIIDUnitTest extends JUnit4DistributedTestCase {
-class GIIChecker extends CacheListenerAdapter
-{
- private boolean gotFirst = false;
- private boolean gotSecond = false;
- private boolean gotThird = false;
- private int updates = 0;
-
- public void afterUpdate(EntryEvent event) {
-
- this.updates++;
-
- String key = (String) event.getKey();
- String value = (String) event.getNewValue();
-
- if (key.equals("key-1") && value.equals("value-1")) {
- this.gotFirst = true;
- }
-
- if (key.equals("key-2") && value.equals("value-2")) {
- this.gotSecond = true;
- }
-
- if (key.equals("key-3") && value.equals("value-3")) {
- this.gotThird = true;
- }
- }
-
- public int getUpdates() {
- return this.updates;
- }
-
- public boolean gotFirst() {
- return this.gotFirst;
- }
-
- public boolean gotSecond() {
- return this.gotSecond;
- }
-
- public boolean gotThird() {
- return this.gotThird;
- }
-}
-
-public class HAGIIDUnitTest extends DistributedTestCase
-{
private static Cache cache = null;
//server
private static VM server0 = null;
private static VM server1 = null;
private static VM client0 = null;
- private static final String REGION_NAME = "HAGIIDUnitTest_region";
+ private static final String REGION_NAME = HAGIIDUnitTest.class.getSimpleName() + "_region";
protected static GIIChecker checker = new GIIChecker();
private int PORT2;
- /** constructor */
- public HAGIIDUnitTest(String name) {
- super(name);
- }
-
@Override
public final void postSetUp() throws Exception {
final Host host = Host.getHost(0);
@@ -134,6 +102,7 @@ public class HAGIIDUnitTest extends DistributedTestCase
client0.invoke(() -> HAGIIDUnitTest.createClientCache( NetworkUtils.getServerHostName(host), new Integer(PORT1),new Integer(PORT2)));
}
+ @Test
public void testGIIRegionQueue()
{
client0.invoke(() -> HAGIIDUnitTest.createEntries());
@@ -167,7 +136,7 @@ public class HAGIIDUnitTest extends DistributedTestCase
Properties props = new Properties();
props.setProperty(MCAST_PORT, "0");
props.setProperty(LOCATORS, "");
- new HAGIIDUnitTest("temp").createCache(props);
+ new HAGIIDUnitTest().createCache(props);
AttributesFactory factory = new AttributesFactory();
ClientServerTestCase.configureConnectionPool(factory, host, new int[] {PORT1,PORT2}, true, -1, 2, null, 1000, -1, false, -1);
factory.setScope(Scope.DISTRIBUTED_ACK);
@@ -178,7 +147,7 @@ public class HAGIIDUnitTest extends DistributedTestCase
public static Integer createServer1Cache() throws Exception
{
- new HAGIIDUnitTest("temp").createCache(new Properties());
+ new HAGIIDUnitTest().createCache(new Properties());
AttributesFactory factory = new AttributesFactory();
factory.setScope(Scope.DISTRIBUTED_ACK);
factory.setDataPolicy(DataPolicy.REPLICATE);
@@ -194,7 +163,7 @@ public class HAGIIDUnitTest extends DistributedTestCase
public static void createServer2Cache(Integer port) throws Exception
{
- new HAGIIDUnitTest("temp").createCache(new Properties());
+ new HAGIIDUnitTest().createCache(new Properties());
AttributesFactory factory = new AttributesFactory();
factory.setScope(Scope.DISTRIBUTED_ACK);
factory.setDataPolicy(DataPolicy.REPLICATE);
@@ -219,6 +188,7 @@ public class HAGIIDUnitTest extends DistributedTestCase
Assert.fail("failed while registering keys ", ex);
}
}
+
public static void createEntries()
{
try {
@@ -244,7 +214,7 @@ public class HAGIIDUnitTest extends DistributedTestCase
}
}
catch (Exception e) {
- fail("failed while stopServer()" + e);
+ fail("failed while stopServer()", e);
}
}
@@ -401,7 +371,6 @@ public class HAGIIDUnitTest extends DistributedTestCase
Wait.waitForCriterion(ev, 60 * 1000, 200, true);
// assertIndexDetailsEquals( "key-2",r.getEntry("key-2").getValue());
-
// wait until
// we have a
// dead server
@@ -450,4 +419,49 @@ public class HAGIIDUnitTest extends DistributedTestCase
cache.getDistributedSystem().disconnect();
}
}
+
+ private static class GIIChecker extends CacheListenerAdapter {
+
+ private boolean gotFirst = false;
+ private boolean gotSecond = false;
+ private boolean gotThird = false;
+ private int updates = 0;
+
+ @Override
+ public void afterUpdate(EntryEvent event) {
+
+ this.updates++;
+
+ String key = (String) event.getKey();
+ String value = (String) event.getNewValue();
+
+ if (key.equals("key-1") && value.equals("value-1")) {
+ this.gotFirst = true;
+ }
+
+ if (key.equals("key-2") && value.equals("value-2")) {
+ this.gotSecond = true;
+ }
+
+ if (key.equals("key-3") && value.equals("value-3")) {
+ this.gotThird = true;
+ }
+ }
+
+ public int getUpdates() {
+ return this.updates;
+ }
+
+ public boolean gotFirst() {
+ return this.gotFirst;
+ }
+
+ public boolean gotSecond() {
+ return this.gotSecond;
+ }
+
+ public boolean gotThird() {
+ return this.gotThird;
+ }
+ }
}
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQAddOperationJUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQAddOperationJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQAddOperationJUnitTest.java
index 6cb5b47..4bbc49f 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQAddOperationJUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQAddOperationJUnitTest.java
@@ -16,37 +16,41 @@
*/
package com.gemstone.gemfire.internal.cache.ha;
+import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.*;
+import static org.junit.Assert.*;
+
+import java.io.IOException;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Map;
+
+import org.apache.logging.log4j.Logger;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
import com.gemstone.gemfire.LogWriter;
-import com.gemstone.gemfire.cache.*;
+import com.gemstone.gemfire.cache.AttributesFactory;
+import com.gemstone.gemfire.cache.Cache;
+import com.gemstone.gemfire.cache.CacheException;
+import com.gemstone.gemfire.cache.CacheFactory;
+import com.gemstone.gemfire.cache.CacheListener;
+import com.gemstone.gemfire.cache.DataPolicy;
+import com.gemstone.gemfire.cache.EntryEvent;
+import com.gemstone.gemfire.cache.Scope;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.internal.cache.Conflatable;
import com.gemstone.gemfire.internal.cache.EventID;
import com.gemstone.gemfire.internal.logging.LogService;
import com.gemstone.gemfire.test.dunit.ThreadUtils;
import com.gemstone.gemfire.test.junit.categories.IntegrationTest;
-import junit.framework.Assert;
-import org.apache.logging.log4j.Logger;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
-import org.junit.experimental.categories.Category;
-
-import java.io.IOException;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-
-import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.MCAST_PORT;
-import static org.junit.Assert.*;
/**
* Test to verify Add operation to HARegion Queue with and without conflation.
- *
*/
-
@Category(IntegrationTest.class)
-public class HARQAddOperationJUnitTest
-{
+public class HARQAddOperationJUnitTest {
private static final Logger logger = LogService.getLogger();
/** The cache instance */
@@ -74,78 +78,40 @@ public class HARQAddOperationJUnitTest
volatile static int expiryCount = 0;
- /**
- * Create the cache in setup
- *
- * @throws Exception -
- * thrown if any exception occurs in setUp
- */
@Before
- public void setUp() throws Exception
- {
+ public void setUp() throws Exception {
this.cache = createCache();
this.logWriter = cache.getLogger();
}
- /**
- * Close the cache in tear down *
- *
- * @throws Exception -
- * thrown if any exception occurs in tearDown
- */
@After
- public void tearDown() throws Exception
- {
+ public void tearDown() throws Exception {
this.cache.close();
}
/**
* Creates the cache instance for the test
- *
- * @return the cache instance
- * @throws CacheException -
- * thrown if any exception occurs in cache creation
*/
- private Cache createCache() throws CacheException
- {
+ private Cache createCache() throws CacheException {
return new CacheFactory().set(MCAST_PORT, "0").create();
}
/**
* Creates HA region-queue object
- *
- * @return HA region-queue object
- * @throws IOException
- * @throws ClassNotFoundException
- * @throws CacheException
- * @throws InterruptedException
*/
- protected HARegionQueue createHARegionQueue(String name)
- throws IOException, ClassNotFoundException, CacheException, InterruptedException
- {
+ protected HARegionQueue createHARegionQueue(String name) throws IOException, ClassNotFoundException, CacheException, InterruptedException {
AttributesFactory factory = new AttributesFactory();
factory.setDataPolicy(DataPolicy.REPLICATE);
factory.setScope(Scope.DISTRIBUTED_ACK);
- HARegionQueue regionqueue = HARegionQueue.getHARegionQueueInstance(name,
- cache, HARegionQueue.NON_BLOCKING_HA_QUEUE, false);
+ HARegionQueue regionqueue = HARegionQueue.getHARegionQueueInstance(name, cache, HARegionQueue.NON_BLOCKING_HA_QUEUE, false);
return regionqueue;
}
/**
* Creates HA region-queue object
- *
- * @return HA region-queue object
- * @throws IOException
- * @throws ClassNotFoundException
- * @throws CacheException
- * @throws InterruptedException
*/
- protected HARegionQueue createHARegionQueue(String name,
- HARegionQueueAttributes attrs) throws IOException, ClassNotFoundException, CacheException, InterruptedException
- {
-
- HARegionQueue regionqueue = HARegionQueue.getHARegionQueueInstance(name,
- cache, attrs, HARegionQueue.NON_BLOCKING_HA_QUEUE, false);
+ protected HARegionQueue createHARegionQueue(String name, HARegionQueueAttributes attrs) throws IOException, ClassNotFoundException, CacheException, InterruptedException {
+ HARegionQueue regionqueue = HARegionQueue.getHARegionQueueInstance(name, cache, attrs, HARegionQueue.NON_BLOCKING_HA_QUEUE, false);
return regionqueue;
}
@@ -155,11 +121,9 @@ public class HARQAddOperationJUnitTest
* to value 2 3) Available IDs , Last DispatchedWrapper Set & Conflation Map
* should have size 1. 4) Conflation Map , LastDispatchedWrapper Set &
* Available IDs should have counter corresponding to second operation
- *
*/
@Test
- public void testQueueAddOperationWithConflation() throws Exception
- {
+ public void testQueueAddOperationWithConflation() throws Exception {
this.logWriter
.info("HARegionQueueJUnitTest : testQueueAddOperationWithConflation BEGIN");
this.rq = createHARegionQueue("testQueueAddOperationWithConflation");
@@ -191,11 +155,9 @@ public class HARQAddOperationJUnitTest
* 3) This wrapper should have a set with size 1. 4) The available IDs set
* shoudl have size 1. 5) Put another object by same thread. 6) The wrapper
* set & availableIs List should have size 2 .
- *
*/
@Test
- public void testQueueAddOperationWithoutConflation() throws Exception
- {
+ public void testQueueAddOperationWithoutConflation() throws Exception {
this.logWriter
.info("HARegionQueueJUnitTest : testQueueAddOperationWithoutConflation BEGIN");
this.rq = createHARegionQueue("testQueueAddOperationWithConflation");
@@ -239,12 +201,9 @@ public class HARQAddOperationJUnitTest
* available IDs , LastDispatchedWrapper's Set should have size 0. Events map
* containg should have size 1 ( corresponding to the
* lastDispatchedAndCurrentEvent Wrapper objcet)
- *
- * @throws Exception
*/
@Test
- public void testQueueAddTakeOperationWithoutConflation() throws Exception
- {
+ public void testQueueAddTakeOperationWithoutConflation() throws Exception {
this.logWriter
.info("HARegionQueueJUnitTest : testQueueAddTakeOperationWithoutConflation BEGIN");
@@ -269,12 +228,9 @@ public class HARQAddOperationJUnitTest
* with Threaddentifer as key & sequence as the value for Expiry. Perform a
* take operation. Validate that expiry on ThreadIdentifier removes itself
* from Events Map
- *
- *
*/
@Test
- public void testExpiryOnThreadIdentifier()
- {
+ public void testExpiryOnThreadIdentifier() {
try {
HARegionQueueAttributes attrs = new HARegionQueueAttributes();
attrs.setExpiryTime(2);
@@ -312,12 +268,9 @@ public class HARQAddOperationJUnitTest
* expiry. Validate the data present in Queue experiences expiry. After the
* expiry of the data , AvaialbleIds size should be 0, entry removed from
* Region, LastDispatchedWrapperSet should have size 0.
- *
- *
*/
@Test
- public void testNoExpiryOnThreadIdentifier()
- {
+ public void testNoExpiryOnThreadIdentifier() {
try {
HARegionQueueAttributes hqa = new HARegionQueueAttributes();
hqa.setExpiryTime(8);
@@ -370,11 +323,9 @@ public class HARQAddOperationJUnitTest
* queue contains objects from 1- 10. QRM with sequenceID 5 arrives It should
* remove only remove objects for 1- 5. Then sequenceID 10 come which should
* remove 5-10.
- *
*/
@Test
- public void testMultipleQRMArrival() throws Exception
- {
+ public void testMultipleQRMArrival() throws Exception {
HARegionQueue regionqueue = createHARegionQueue("testNoExpiryOnThreadIdentifier");
EventID[] ids = new EventID[10];
@@ -404,7 +355,6 @@ public class HARQAddOperationJUnitTest
regionqueue.removeDispatchedEvents(ids[9]);
assertEquals(0, regionqueue.getAvalaibleIds().size());
-
}
/**
@@ -414,11 +364,9 @@ public class HARQAddOperationJUnitTest
* before QRM thread acts , the object should be present in the
* lastDispatchedSet & AvailableID. Then the QRM thread gets unblocked , it
* should remove from the available ID.
- *
*/
@Test
- public void testConcurrentPutAndQRM() throws Exception
- {
+ public void testConcurrentPutAndQRM() throws Exception {
testFailed = false;
message = new StringBuffer();
final HARegionQueue regionqueue = createHARegionQueue("testConcurrentPutAndQRM");
@@ -475,8 +423,7 @@ public class HARQAddOperationJUnitTest
* put operation shud remove from region without adding the ID anywhere.
*/
@Test
- public void testConcurrentQRMAndPut() throws Exception
- {
+ public void testConcurrentQRMAndPut() throws Exception {
testFailed = false;
final HARegionQueue regionqueue = createHARegionQueue("testConcurrentQRMAndPut");
final EventID id1 = new EventID(new byte[] { 1 }, 1, 1);
@@ -523,20 +470,15 @@ public class HARQAddOperationJUnitTest
assertEquals(0, regionqueue.getAvalaibleIds().size());
assertEquals(2, regionqueue.getLastDispatchedSequenceId(id2));
-
}
/**
* Two QRMs arriving such that higer sequence number arriving before lower
* sequence number. The lower squnce number should not set itself & also not
* do any checking on the IDs of the LinkedHashSet
- *
- * @throws Exception
*/
-
@Test
- public void testEventMapPopulationForQRM() throws Exception
- {
+ public void testEventMapPopulationForQRM() throws Exception {
HARegionQueue regionqueue = createHARegionQueue("testEventMapPopulationForQRM");
EventID id1 = new EventID(new byte[] { 1 }, 1, 1);
EventID id2 = new EventID(new byte[] { 1 }, 1, 2);
@@ -566,13 +508,9 @@ public class HARQAddOperationJUnitTest
* for that ThreadIdentifier. The ID which gets conflated should not be
* present in the availableID, Region & that ThreadIdentifier's HashSet . The
* conflation map should contain the Old IDs position.
- *
- * @throws Exception
*/
-
@Test
- public void testCleanUpForConflation() throws Exception
- {
+ public void testCleanUpForConflation() throws Exception {
this.logWriter
.info("HARQAddOperationJUnitTest : testCleanUpForConflation BEGIN");
testFailed = false;
@@ -661,10 +599,8 @@ public class HARQAddOperationJUnitTest
* are deleted from the available IDs & the Counters set contained in DACE.
* Conflation is disabled.
*/
-
@Test
- public void testPeekAndRemoveWithoutConflation() throws Exception
- {
+ public void testPeekAndRemoveWithoutConflation() throws Exception {
testFailed = false;
message = null;
final int numOfThreads = 5;
@@ -719,7 +655,6 @@ public class HARQAddOperationJUnitTest
this.logWriter
.info("testPeekAndRemoveWithoutConflation() completed successfully");
-
}
/**
@@ -728,10 +663,8 @@ public class HARQAddOperationJUnitTest
* are deleted from the available IDs & the Counters set contained in DACE.
* Conflation is enabled
*/
-
@Test
- public void testPeekAndRemoveWithConflation() throws Exception
- {
+ public void testPeekAndRemoveWithConflation() throws Exception {
testFailed = false;
message = null;
final int numOfThreads = 5;
@@ -791,7 +724,6 @@ public class HARQAddOperationJUnitTest
this.logWriter
.info("testPeekAndRemoveWithConflation() completed successfully");
-
}
/**
@@ -799,13 +731,9 @@ public class HARQAddOperationJUnitTest
* do a peek of batch size 5, 10 , 15 & 20 respectively. And all of them
* concurrently cal remove. The remove should ensure that the entries are
* deleted from the available IDs & the Counters set contained in DACE.
- *
- * @throws Exception
*/
-
@Test
- public void testPeekForDiffBatchSizeAndRemoveAll() throws Exception
- {
+ public void testPeekForDiffBatchSizeAndRemoveAll() throws Exception {
testFailed = false;
message = null;
barrierCount = 0;
@@ -913,12 +841,9 @@ public class HARQAddOperationJUnitTest
* do a peek of batch size 5, 10 and 15 respectively. And all of them
* concurrently call remove. The remove should ensure that the entries are
* deleted from the available IDs & the Counters set contained in DACE.
- *
- * @throws Exception
*/
@Test
- public void testPeekForDiffBatchSizeAndRemoveSome() throws Exception
- {
+ public void testPeekForDiffBatchSizeAndRemoveSome() throws Exception {
testFailed = false;
barrierCount = 0;
message = null;
@@ -1010,7 +935,6 @@ public class HARQAddOperationJUnitTest
this.logWriter
.info("testPeekForDiffBatchSizeAndRemoveSome() completed successfully");
-
}
/**
@@ -1023,9 +947,8 @@ public class HARQAddOperationJUnitTest
* next expiry should remove the LastDisptachedWrapper
*/
@Test
- public void testAddWithQRMAndExpiry() throws Exception
- {
- try{
+ public void testAddWithQRMAndExpiry() throws Exception {
+ try {
HARegionQueueAttributes attrs = new HARegionQueueAttributes();
attrs.setExpiryTime(10);
final HARegionQueue regionqueue = new HARegionQueue.TestOnlyHARegionQueue("testing", cache, attrs) {
@@ -1067,11 +990,11 @@ public class HARQAddOperationJUnitTest
assertEquals(4, regionqueue.getLastDispatchedSequenceId(new EventID(new byte[] { 1 }, 1, 1)));
// verify 1-5 not in region
for (long i = 1; i < 6; i++) {
- Assert.assertTrue(!regionqueue.getRegion().containsKey(new Long(i)));
+ assertTrue(!regionqueue.getRegion().containsKey(new Long(i)));
}
// verify 6-10 still in region queue
for (long i = 6; i < 11; i++) {
- Assert.assertTrue(regionqueue.getRegion().containsKey(new Long(i)));
+ assertTrue(regionqueue.getRegion().containsKey(new Long(i)));
}
// Perform 5 take operations to remove next 5-9 sequence ids
@@ -1083,7 +1006,7 @@ public class HARQAddOperationJUnitTest
assertEquals(9, regionqueue.getLastDispatchedSequenceId(new EventID(new byte[] { 1 }, 1, 1)));
// verify that sequence ids 1-10 all are removed from the RQ
for (long i = 1; i < 11; i++) {
- Assert.assertTrue(!regionqueue.getRegion().containsKey(new Long(i)));
+ assertTrue(!regionqueue.getRegion().containsKey(new Long(i)));
}
// wait until expiry thread has run once
@@ -1109,11 +1032,11 @@ public class HARQAddOperationJUnitTest
assertNull(regionqueue.getRegion().get(tID));
}
catch (Exception e) {
- fail("Exception occured in test due to " + e);
+ throw new AssertionError("Exception occurred in test due to", e);
}
}
- /**
+ /*
* This test does the following:<br>
* 1)Create a blocking HARegionQueue<br>
* 2)Add some events to the queue with same ThreadIdentifier<br>
@@ -1125,18 +1048,13 @@ public class HARQAddOperationJUnitTest
* 7)Verify that the size of wrapper-map is 1 as all events had same ThreadId<br>
* 8)Verify that the sequenceId against the ThreadId in the wrapper-map is
* same as that of the last event taken<br>
- *
- * @throws Exception -
- * thrown if any exception occurs in test execution
*/
/**
* Behaviour of take() has been changed for relaible messaging feature. Region queue take()
* operation will no longer add to the Dispatch Message Map. Hence disabling the test - SUYOG
- */
-
- public void _testDispatchedMsgsMapUpdateOnTakes() throws Exception
- {
+ */
+ public void _testDispatchedMsgsMapUpdateOnTakes() throws Exception {
this.logWriter
.info("HARQAddOperationJUnitTest : testDispatchedEventsMapUpdateOnTakes BEGIN");
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQueueNewImplDUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQueueNewImplDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQueueNewImplDUnitTest.java
index 2843fc0..340f10b 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQueueNewImplDUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARQueueNewImplDUnitTest.java
@@ -14,17 +14,38 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
-/**
- *
- */
package com.gemstone.gemfire.internal.cache.ha;
+import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.*;
+import static com.gemstone.gemfire.test.dunit.Assert.*;
+import static com.gemstone.gemfire.test.dunit.NetworkUtils.*;
+import static com.gemstone.gemfire.test.dunit.Wait.*;
+
+import java.io.File;
+import java.util.HashMap;
+import java.util.Iterator;
+import java.util.Map;
+import java.util.Properties;
+import java.util.Set;
+
+import org.junit.Ignore;
+import org.junit.Test;
+
import com.gemstone.gemfire.LogWriter;
-import com.gemstone.gemfire.cache.*;
+import com.gemstone.gemfire.cache.AttributesFactory;
+import com.gemstone.gemfire.cache.Cache;
+import com.gemstone.gemfire.cache.CacheFactory;
+import com.gemstone.gemfire.cache.DataPolicy;
+import com.gemstone.gemfire.cache.DiskStoreFactory;
+import com.gemstone.gemfire.cache.EntryEvent;
+import com.gemstone.gemfire.cache.Region;
+import com.gemstone.gemfire.cache.RegionAttributes;
+import com.gemstone.gemfire.cache.Scope;
import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.cache30.ClientServerTestCase;
import com.gemstone.gemfire.distributed.DistributedSystem;
+import com.gemstone.gemfire.distributed.internal.DistributionConfig;
import com.gemstone.gemfire.internal.AvailablePort;
import com.gemstone.gemfire.internal.cache.CacheServerImpl;
import com.gemstone.gemfire.internal.cache.LocalRegion;
@@ -32,71 +53,53 @@ import com.gemstone.gemfire.internal.cache.tier.sockets.CacheServerTestUtil;
import com.gemstone.gemfire.internal.cache.tier.sockets.ClientUpdateMessage;
import com.gemstone.gemfire.internal.cache.tier.sockets.ConflationDUnitTest;
import com.gemstone.gemfire.internal.cache.tier.sockets.HAEventWrapper;
-import com.gemstone.gemfire.test.dunit.*;
-import util.TestException;
-
-import java.io.File;
-import java.util.*;
-
-import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.*;
+import com.gemstone.gemfire.test.dunit.Host;
+import com.gemstone.gemfire.test.dunit.VM;
+import com.gemstone.gemfire.test.dunit.WaitCriterion;
+import com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase;
/**
* This DUnit contains various tests to ensure new implementation of ha region
* queues works as expected.
*
* @since GemFire 5.7
- *
*/
-public class HARQueueNewImplDUnitTest extends DistributedTestCase {
- protected static Cache cache = null;
+public class HARQueueNewImplDUnitTest extends JUnit4DistributedTestCase {
- protected static VM serverVM0 = null;
+ private static final String regionName = HARQueueNewImplDUnitTest.class.getSimpleName();
+ private static final Map map = new HashMap();
+ private static Cache cache = null;
+ private static VM serverVM0 = null;
private static VM serverVM1 = null;
+ private static VM clientVM1 = null;
+ private static VM clientVM2 = null;
- protected static VM clientVM1 = null;
-
- protected static VM clientVM2 = null;
+ private static LogWriter logger = null;
+ private static int numOfCreates = 0;
+ private static int numOfUpdates = 0;
+ private static int numOfInvalidates = 0;
+ private static Object[] deletedValues = null;
private int PORT1;
-
private int PORT2;
- private static final String regionName = "HARQueueNewImplDUnitTest";
-
- private static final Map map = new HashMap();
-
- protected static LogWriter logger = null;
-
- protected static int numOfCreates = 0;
-
- protected static int numOfUpdates = 0;
-
- protected static int numOfInvalidates = 0;
-
- private static Object[] deletedValues = null;
-
- /**
- * @param name
- * name of the test
- */
- public HARQueueNewImplDUnitTest(String name) {
- super(name);
- }
-
/**
* Sets up the test.
*/
@Override
public final void postSetUp() throws Exception {
+ map.clear();
+
final Host host = Host.getHost(0);
serverVM0 = host.getVM(0);
serverVM1 = host.getVM(1);
clientVM1 = host.getVM(2);
clientVM2 = host.getVM(3);
- PORT1 = ((Integer)serverVM0.invoke(() -> HARQueueNewImplDUnitTest.createServerCache( HARegionQueue.HA_EVICTION_POLICY_MEMORY ))).intValue();
- PORT2 = ((Integer)serverVM1.invoke(() -> HARQueueNewImplDUnitTest.createServerCache( HARegionQueue.HA_EVICTION_POLICY_ENTRY ))).intValue();
+ PORT1 = serverVM0.invoke(() -> HARQueueNewImplDUnitTest.createServerCache( HARegionQueue.HA_EVICTION_POLICY_MEMORY ));
+ PORT2 = serverVM1.invoke(() -> HARQueueNewImplDUnitTest.createServerCache( HARegionQueue.HA_EVICTION_POLICY_ENTRY ));
+
numOfCreates = 0;
numOfUpdates = 0;
numOfInvalidates = 0;
@@ -107,6 +110,8 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
*/
@Override
public final void preTearDown() throws Exception {
+ map.clear();
+
closeCache();
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
@@ -114,6 +119,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
// Unset the isSlowStartForTesting flag
serverVM0.invoke(() -> ConflationDUnitTest.unsetIsSlowStart());
serverVM1.invoke(() -> ConflationDUnitTest.unsetIsSlowStart());
+
// then close the servers
serverVM0.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
serverVM1.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
@@ -138,9 +144,8 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
return createServerCache(ePolicy, new Integer(1));
}
- public static Integer createServerCache(String ePolicy, Integer cap)
- throws Exception {
- new HARQueueNewImplDUnitTest("temp").createCache(new Properties());
+ public static Integer createServerCache(String ePolicy, Integer cap) throws Exception {
+ new HARQueueNewImplDUnitTest().createCache(new Properties());
AttributesFactory factory = new AttributesFactory();
factory.setScope(Scope.DISTRIBUTED_ACK);
factory.setDataPolicy(DataPolicy.REPLICATE);
@@ -167,8 +172,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
return new Integer(server1.getPort());
}
- public static Integer createOneMoreBridgeServer(Boolean notifyBySubscription)
- throws Exception {
+ public static Integer createOneMoreBridgeServer(Boolean notifyBySubscription) throws Exception {
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
CacheServer server1 = cache.addCacheServer();
server1.setPort(port);
@@ -180,14 +184,13 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
return new Integer(server1.getPort());
}
- public static void createClientCache(String host, Integer port1, Integer port2,
- String rLevel, Boolean addListener) throws Exception {
+ public static void createClientCache(String host, Integer port1, Integer port2, String rLevel, Boolean addListener) throws Exception {
CacheServerTestUtil.disableShufflingOfEndpoints();
Properties props = new Properties();
props.setProperty(MCAST_PORT, "0");
props.setProperty(LOCATORS, "");
- new HARQueueNewImplDUnitTest("temp").createCache(props);
+ new HARQueueNewImplDUnitTest().createCache(props);
AttributesFactory factory = new AttributesFactory();
ClientServerTestCase.configureConnectionPool(factory, host, port1
.intValue(), port2.intValue(), true, Integer.parseInt(rLevel), 2, null,
@@ -222,8 +225,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
logger = cache.getLogger();
}
- public static void createClientCache(String host, Integer port1, Integer port2,
- String rLevel) throws Exception {
+ public static void createClientCache(String host, Integer port1, Integer port2, String rLevel) throws Exception {
createClientCache(host, port1, port2, rLevel, Boolean.FALSE);
}
@@ -234,7 +236,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
r.registerInterest("ALL_KEYS");
}
catch (Exception ex) {
- Assert.fail("failed in registerInterestListAll", ex);
+ fail("failed in registerInterestListAll", ex);
}
}
@@ -247,7 +249,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
r.registerInterest("k5");
}
catch (Exception ex) {
- Assert.fail("failed while registering keys", ex);
+ fail("failed while registering keys", ex);
}
}
@@ -264,7 +266,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
r.put("k5", "pv5");
}
catch (Exception ex) {
- Assert.fail("failed in putEntries()", ex);
+ fail("failed in putEntries()", ex);
}
}
@@ -280,7 +282,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
r.create("k5", "v5");
}
catch (Exception ex) {
- Assert.fail("failed in createEntries()", ex);
+ fail("failed in createEntries()", ex);
}
}
@@ -293,7 +295,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
catch (Exception ex) {
- Assert.fail("failed in createEntries(Long)", ex);
+ fail("failed in createEntries(Long)", ex);
}
}
@@ -308,7 +310,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
catch (Exception ex) {
- Assert.fail("failed in putHeavyEntries(Long)", ex);
+ fail("failed in putHeavyEntries(Long)", ex);
}
}
@@ -316,18 +318,17 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* This test verifies that the client-messages-region does not store duplicate
* ClientUpdateMessageImpl instances, during a normal put path as well as the
* GII path.
- *
- * @throws Exception
*/
+ @Test
public void testClientMsgsRegionSize() throws Exception {
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
serverVM1.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -348,18 +349,17 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* This test verifies that the ha-region-queues increment the reference count
* of their respective HAEventWrapper instances in the client-messages-region
* correctly, during put as well as GII path.
- *
- * @throws Exception
*/
+ @Test
public void testRefCountForNormalAndGIIPut() throws Exception {
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
serverVM1.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -384,16 +384,15 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* This test verifies that the ha-region-queues decrement the reference count
* of their respective HAEventWrapper instances in the client-messages-region
* correctly, after the events have been peeked and removed from the queue.
- *
- * @throws Exception
*/
+ @Test
public void testRefCountForPeekAndRemove() throws Exception {
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -415,16 +414,15 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* This test verifies that the processing of the QRM messages results in
* decrementing the reference count of corresponding HAEventWrapper instances,
* correctly.
- *
- * @throws Exception
*/
+ @Test
public void testRefCountForQRM() throws Exception {
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -449,18 +447,17 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* proxy/client disconnect), causes the reference count of all HAEventWrapper
* instances belonging to the ha-region-queue to be decremented by one, and
* makes it visible to the client-messages-region.
- *
- * @throws Exception
*/
+ @Test
public void testRefCountForDestroy() throws Exception {
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
serverVM1.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -497,17 +494,16 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* Addresses the bug 39179. If a clientUpdateMessage is dispatched to the
* client while its GII was under way, then it should not be put into the
* HARegionQueue of a client at receiving server side.
- *
- * @throws Exception
*/
+ @Test
public void testConcurrentGIIAndDispatch() throws Exception {
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "40000" ));
serverVM1.invoke(() -> ConflationDUnitTest.setIsSlowStart( "40000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -537,27 +533,21 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
serverVM1.invoke(() -> ConflationDUnitTest.unsetIsSlowStart());
}
- /*
- * public void PtestRefCountForConcurrentPutAndRemove() throws Exception { }
- * public void PtestRefCountForConcurrentGIIPut() throws Exception { }
- */
-
/**
* This test verifies that when two BridgeServerImpl instances are created in
* a single VM, they do share the client-messages-region.
- *
- * @throws Exception
*/
+ @Test
public void testTwoBridgeServersInOneVMDoShareCMR() throws Exception {
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
Integer port3 = (Integer)serverVM0.invoke(() -> HARQueueNewImplDUnitTest.createOneMoreBridgeServer( Boolean.TRUE ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), port3, "0");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), port3, "0");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -574,14 +564,13 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* This test verifies that two clients, connected to two bridge servers with
* different notifyBySubscription values, on a single VM, receive
* updates/invalidates depending upon their notifyBySubscription value.
- *
- * @throws Exception
*/
+ @Test
public void testUpdatesWithTwoBridgeServersInOneVM() throws Exception {
Integer port3 = (Integer)serverVM0.invoke(() -> HARQueueNewImplDUnitTest.createOneMoreBridgeServer( Boolean.FALSE ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1", Boolean.TRUE);
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1", Boolean.TRUE);
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, port3, new Integer(PORT2), "1", Boolean.TRUE ));
registerInterestListAll();
@@ -606,17 +595,16 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* This test verifies that the HAEventWrapper instances present in the
* client-messages-region give up the references to their respective
* ClientUpdateMessageImpl instances.
- *
- * @throws Exception
*/
+ @Test
public void testHAEventWrapperDoesNotHoldCUMOnceInsideCMR() throws Exception {
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -640,9 +628,8 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* servers will have simple HashMap structures. Also, it verifies that such a
* structure (referred to as haContainer, in general) is destroyed when its
* bridge server is stopped.
- *
- * @throws Exception
*/
+ @Test
public void testCMRNotCreatedForNoneEvictionPolicy() throws Exception {
serverVM0.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
serverVM1.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
@@ -653,10 +640,10 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -680,18 +667,17 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
* servers who have eviction policy either as 'mem' or as 'entry'. Also, it
* verifies that such a client-messages-region is destroyed when its bridge
* server is stopped.
- *
- * @throws Exception
*/
+ @Test
public void testCMRCreatedForMemOrEntryEvictionPolicy() throws Exception {
Boolean isRegion = Boolean.TRUE;
// slow start for dispatcher
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "30000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ final String client2Host = getServerHostName(clientVM1.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -713,15 +699,14 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
/**
* This test verifies that the Cache.rootRegions() method does not return the
* client-messages-region of any of the cache's attached bridge servers.
- *
- * @throws Exception
*/
+ @Test
public void testCMRNotReturnedByRootRegionsMethod() throws Exception {
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -735,14 +720,13 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
/**
- * This test verifies that the memory foorprint of the ha region queues is
+ * This test verifies that the memory footprint of the ha region queues is
* less when ha-overflow is enabled (with an appropriate value of haCapacity)
* compared to when it is disabled, for the same amount of data feed.
- *
- * @throws Exception
*/
- public void _testMemoryFootprintOfHARegionQueuesWithAndWithoutOverflow()
- throws Exception {
+ @Ignore("TODO")
+ @Test
+ public void testMemoryFootprintOfHARegionQueuesWithAndWithoutOverflow() throws Exception {
serverVM0.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
serverVM1.invoke(() -> HARQueueNewImplDUnitTest.closeCache());
Thread.sleep(2000);
@@ -756,10 +740,10 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
serverVM0.invoke(() -> ConflationDUnitTest.setIsSlowStart( "60000" ));
serverVM1.invoke(() -> ConflationDUnitTest.setIsSlowStart( "60000" ));
- createClientCache(NetworkUtils.getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
- final String client1Host = NetworkUtils.getServerHostName(clientVM1.getHost());
+ createClientCache(getServerHostName(Host.getHost(0)), new Integer(PORT1), new Integer(PORT2), "1");
+ final String client1Host = getServerHostName(clientVM1.getHost());
clientVM1.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client1Host, new Integer(PORT1), new Integer(PORT2), "1" ));
- final String client2Host = NetworkUtils.getServerHostName(clientVM2.getHost());
+ final String client2Host = getServerHostName(clientVM2.getHost());
clientVM2.invoke(() -> HARQueueNewImplDUnitTest.createClientCache( client2Host, new Integer(PORT1), new Integer(PORT2), "1" ));
registerInterestListAll();
@@ -813,8 +797,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- static Long getUsedMemoryAndVerifyRegionSize(Integer rSize,
- Integer haContainerSize, Integer port) {
+ static Long getUsedMemoryAndVerifyRegionSize(Integer rSize, Integer haContainerSize, Integer port) {
Long retVal = null;
try {
retVal = new Long(Runtime.getRuntime().totalMemory()
@@ -942,22 +925,22 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- public static void verifyNullValuesInCMR(final Integer numOfEntries,
- final Integer port,
- String[] keys) {
+ public static void verifyNullValuesInCMR(final Integer numOfEntries, final Integer port, String[] keys) {
final Region msgsRegion = cache.getRegion(CacheServerImpl
.generateNameForClientMsgsRegion(port.intValue()));
WaitCriterion wc = new WaitCriterion() {
String excuse;
+ @Override
public boolean done() {
int sz = msgsRegion.size();
return sz == numOfEntries.intValue();
}
+ @Override
public String description() {
return excuse;
}
};
- Wait.waitForCriterion(wc, 60 * 1000, 1000, true);
+ waitForCriterion(wc, 60 * 1000, 1000, true);
Set entries = msgsRegion.entrySet();
Iterator iter = entries.iterator();
@@ -972,8 +955,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- public static void makeValuesOfSomeKeysNullInClientMsgsRegion(Integer port,
- String[] keys) {
+ public static void makeValuesOfSomeKeysNullInClientMsgsRegion(Integer port, String[] keys) {
Region msgsRegion = cache.getRegion(CacheServerImpl
.generateNameForClientMsgsRegion(port.intValue()));
assertNotNull(msgsRegion);
@@ -994,8 +976,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- public static void populateValuesOfSomeKeysInClientMsgsRegion(Integer port,
- String[] keys) {
+ public static void populateValuesOfSomeKeysInClientMsgsRegion(Integer port, String[] keys) {
Region msgsRegion = cache.getRegion(CacheServerImpl
.generateNameForClientMsgsRegion(port.intValue()));
assertNotNull(msgsRegion);
@@ -1020,8 +1001,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- public static void verifyQueueData(Integer regionsize,
- Integer msgsRegionsize, Integer port) {
+ public static void verifyQueueData(Integer regionsize, Integer msgsRegionsize, Integer port) {
try {
// Get the clientMessagesRegion and check the size.
Region msgsRegion = cache.getRegion(CacheServerImpl
@@ -1049,10 +1029,10 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- public static void verifyRegionSize(final Integer regionSize,
- final Integer msgsRegionsize, final Integer port) {
+ public static void verifyRegionSize(final Integer regionSize, final Integer msgsRegionsize, final Integer port) {
WaitCriterion wc = new WaitCriterion() {
String excuse;
+ @Override
public boolean done() {
try {
// Get the clientMessagesRegion and check the size.
@@ -1086,17 +1066,18 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
return false;
}
}
+ @Override
public String description() {
return excuse;
}
};
- Wait.waitForCriterion(wc, 120 * 1000, 1000, true);
+ waitForCriterion(wc, 120 * 1000, 1000, true);
}
- public static void verifyRegionSize(final Integer regionSize,
- final Integer msgsRegionsize) {
+ public static void verifyRegionSize(final Integer regionSize, final Integer msgsRegionsize) {
WaitCriterion wc = new WaitCriterion() {
String excuse;
+ @Override
public boolean done() {
try {
// Get the clientMessagesRegion and check the size.
@@ -1124,11 +1105,12 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
return false;
}
}
+ @Override
public String description() {
return excuse;
}
};
- Wait.waitForCriterion(wc, 120 * 1000, 1000, true);
+ waitForCriterion(wc, 120 * 1000, 1000, true);
}
public static void verifyHaContainerType(Boolean isRegion, Integer port) {
@@ -1169,7 +1151,7 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
while (iter.hasNext()) {
rName = ((Region)iter.next()).getName();
if (cmrName.equals(rName)) {
- throw new TestException(
+ throw new AssertionError(
"Cache.rootRegions() method should not return the client_messages_region.");
}
logger.fine("Region name returned from cache.rootRegions(): " + rName);
@@ -1180,30 +1162,33 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
}
- public static void verifyUpdatesReceived(final Integer num, Boolean isUpdate,
- Long waitLimit) {
+ public static void verifyUpdatesReceived(final Integer num, Boolean isUpdate, Long waitLimit) {
try {
if (isUpdate.booleanValue()) {
WaitCriterion ev = new WaitCriterion() {
+ @Override
public boolean done() {
return num.intValue() == numOfUpdates;
}
+ @Override
public String description() {
return null;
}
};
- Wait.waitForCriterion(ev, waitLimit.longValue(), 200, true);
+ waitForCriterion(ev, waitLimit.longValue(), 200, true);
}
else {
WaitCriterion ev = new WaitCriterion() {
+ @Override
public boolean done() {
return num.intValue() == numOfInvalidates;
}
+ @Override
public String description() {
return null;
}
};
- Wait.waitForCriterion(ev, waitLimit.longValue(), 200, true);
+ waitForCriterion(ev, waitLimit.longValue(), 200, true);
}
}
catch (Exception e) {
@@ -1228,14 +1213,16 @@ public class HARQueueNewImplDUnitTest extends DistributedTestCase {
}
final Map m = haContainer;
WaitCriterion ev = new WaitCriterion() {
+ @Override
public boolean done() {
return m.size() == 0;
}
+ @Override
public String description() {
return null;
}
};
- Wait.waitForCriterion(ev, waitLimit.longValue(), 200, true);
+ waitForCriterion(ev, waitLimit.longValue(), 200, true);
}
catch (Exception e) {
fail("failed in waitTillMessagesAreDispatched()" + e);
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionDUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionDUnitTest.java
index 194ce89..3f90b36 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionDUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionDUnitTest.java
@@ -16,9 +16,12 @@
*/
package com.gemstone.gemfire.internal.cache.ha;
+import static org.junit.Assert.*;
+
import java.util.Properties;
-import junit.framework.Assert;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
import com.gemstone.gemfire.cache.AttributesFactory;
import com.gemstone.gemfire.cache.Cache;
@@ -30,20 +33,20 @@ import com.gemstone.gemfire.distributed.DistributedSystem;
import com.gemstone.gemfire.internal.cache.EventID;
import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
import com.gemstone.gemfire.internal.cache.HARegion;
-import com.gemstone.gemfire.test.dunit.DistributedTestCase;
import com.gemstone.gemfire.test.dunit.Host;
import com.gemstone.gemfire.test.dunit.VM;
+import com.gemstone.gemfire.test.dunit.internal.JUnit4DistributedTestCase;
+import com.gemstone.gemfire.test.junit.categories.DistributedTest;
/**
* Test to verify :
*
* 1)put() on a mirrored HARegion does not propagate 2)localDestroy() allowed on
* a mirrored region 3) GII happens normally
- *
- *
*/
-public class HARegionDUnitTest extends DistributedTestCase
-{
+@Category(DistributedTest.class)
+public class HARegionDUnitTest extends JUnit4DistributedTestCase {
+
VM vm0 = null;
VM vm1 = null;
@@ -52,8 +55,8 @@ public class HARegionDUnitTest extends DistributedTestCase
private static final String REGION_NAME = "HARegionDUnitTest_region" ;
/** constructor */
- public HARegionDUnitTest(String name) {
- super(name);
+ public HARegionDUnitTest() {
+ super();
}
/**
@@ -102,6 +105,7 @@ public class HARegionDUnitTest extends DistributedTestCase
* assert put in VM2 was successful by doing a get
*
*/
+ @Test
public void testLocalPut()
{
vm0.invoke(() -> HARegionDUnitTest.createRegion());
@@ -123,6 +127,7 @@ public class HARegionDUnitTest extends DistributedTestCase
* key has not been destroyed in VM2
*
*/
+ @Test
public void testLocalDestroy()
{
vm0.invoke(() -> HARegionDUnitTest.createRegion());
@@ -144,6 +149,7 @@ public class HARegionDUnitTest extends DistributedTestCase
* through GII 6) do a put in VM2 7) assert put in VM2 was successful
*
*/
+ @Test
public void testGII()
{
vm0.invoke(() -> HARegionDUnitTest.createRegion());
@@ -163,6 +169,7 @@ public class HARegionDUnitTest extends DistributedTestCase
* through GII 6) do a put in VM2 7) assert put in VM2 was successful
*
*/
+ @Test
public void testLocalDestroyRegion()
{
vm0.invoke(() -> HARegionDUnitTest.createRegion());
@@ -187,7 +194,7 @@ public class HARegionDUnitTest extends DistributedTestCase
*/
public static void verifyRegionNotDestroyed()
{
- Assert.assertTrue(cache.getRegion(REGION_NAME) != null);
+ assertTrue(cache.getRegion(REGION_NAME) != null);
}
/**
@@ -197,6 +204,7 @@ public class HARegionDUnitTest extends DistributedTestCase
* through GII 6) do a put in VM2 7) assert put in VM2 was successful
*
*/
+ @Test
public void testQRM()
{
vm0.invoke(() -> HARegionDUnitTest.createRegionQueue());
@@ -219,7 +227,7 @@ public class HARegionDUnitTest extends DistributedTestCase
public static void createRegion() throws Exception
{
- HARegionDUnitTest test = new HARegionDUnitTest(REGION_NAME);
+ HARegionDUnitTest test = new HARegionDUnitTest();
cache = test.createCache();
AttributesFactory factory = new AttributesFactory();
factory.setScope(Scope.DISTRIBUTED_ACK);
@@ -239,7 +247,7 @@ public class HARegionDUnitTest extends DistributedTestCase
public static void createRegionQueue() throws Exception
{
- HARegionDUnitTest test = new HARegionDUnitTest(REGION_NAME);
+ HARegionDUnitTest test = new HARegionDUnitTest();
cache = test.createCache();
/*AttributesFactory factory = new AttributesFactory();
factory.setScope(Scope.DISTRIBUTED_ACK);
@@ -258,10 +266,10 @@ public class HARegionDUnitTest extends DistributedTestCase
public static void verifyAddingDispatchMesgs()
{
- Assert.assertTrue(HARegionQueue.getDispatchedMessagesMapForTesting()
+ assertTrue(HARegionQueue.getDispatchedMessagesMapForTesting()
.isEmpty());
hrq.addDispatchedMessage(new ThreadIdentifier(new byte[1],1),1);
- Assert.assertTrue(!HARegionQueue.getDispatchedMessagesMapForTesting()
+ assertTrue(!HARegionQueue.getDispatchedMessagesMapForTesting()
.isEmpty());
}
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/b914df23/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionJUnitTest.java
----------------------------------------------------------------------
diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionJUnitTest.java
index d98b3c7..6c027ea 100755
--- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionJUnitTest.java
+++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ha/HARegionJUnitTest.java
@@ -16,40 +16,56 @@
*/
package com.gemstone.gemfire.internal.cache.ha;
-import com.gemstone.gemfire.cache.*;
-import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
-import com.gemstone.gemfire.internal.cache.EntryEventImpl;
-import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
-import com.gemstone.gemfire.internal.cache.HARegion;
-import com.gemstone.gemfire.test.junit.categories.IntegrationTest;
-import junit.framework.Assert;
+import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.*;
+import static org.junit.Assert.*;
+
+import java.io.IOException;
+
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
import org.junit.experimental.categories.Category;
-import java.io.IOException;
-
-import static com.gemstone.gemfire.distributed.DistributedSystemConfigProperties.MCAST_PORT;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
+import com.gemstone.gemfire.cache.AttributesFactory;
+import com.gemstone.gemfire.cache.Cache;
+import com.gemstone.gemfire.cache.CacheExistsException;
+import com.gemstone.gemfire.cache.CacheFactory;
+import com.gemstone.gemfire.cache.CacheWriterException;
+import com.gemstone.gemfire.cache.DataPolicy;
+import com.gemstone.gemfire.cache.EntryEvent;
+import com.gemstone.gemfire.cache.EvictionAction;
+import com.gemstone.gemfire.cache.EvictionAttributes;
+import com.gemstone.gemfire.cache.ExpirationAction;
+import com.gemstone.gemfire.cache.ExpirationAttributes;
+import com.gemstone.gemfire.cache.GatewayException;
+import com.gemstone.gemfire.cache.Region;
+import com.gemstone.gemfire.cache.RegionAttributes;
+import com.gemstone.gemfire.cache.RegionExistsException;
+import com.gemstone.gemfire.cache.Scope;
+import com.gemstone.gemfire.cache.TimeoutException;
+import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
+import com.gemstone.gemfire.internal.cache.EntryEventImpl;
+import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
+import com.gemstone.gemfire.internal.cache.HARegion;
+import com.gemstone.gemfire.test.junit.categories.IntegrationTest;
/**
* Test verifies the properties of a HARegion which allows localPuts and
* localDestroys on a MirroredRegion
- *
- *
*/
@Category(IntegrationTest.class)
-public class HARegionJUnitTest
-{
+public class HARegionJUnitTest {
+
+ /**
+ * cache
+ */
+ private Cache cache = null;
/**
* create the cache
*/
@Before
- public void setUp() throws Exception
- {
+ public void setUp() throws Exception {
cache = createCache();
}
@@ -57,49 +73,21 @@ public class HARegionJUnitTest
* close the cache in tear down
*/
@After
- public void tearDown() throws Exception
- {
+ public void tearDown() throws Exception {
cache.close();
}
/**
- * cache
- */
- private Cache cache = null;
-
- /**
- *
* create the cache
- *
- * @throws TimeoutException
- * @throws CacheWriterException
- * @throws GatewayException
- * @throws CacheExistsException
- * @throws RegionExistsException
*/
- private Cache createCache() throws TimeoutException, CacheWriterException,
- GatewayException, CacheExistsException,
- RegionExistsException
- {
+ private Cache createCache() throws TimeoutException, CacheWriterException, GatewayException, CacheExistsException, RegionExistsException {
return new CacheFactory().set(MCAST_PORT, "0").create();
}
/**
* create the HARegion
- *
- * @throws TimeoutException
- * @throws CacheWriterException
- * @throws GatewayException
- * @throws CacheExistsException
- * @throws RegionExistsException
- * @throws IOException
- * @throws ClassNotFoundException
*/
- private Region createHARegion() throws TimeoutException,
- CacheWriterException, GatewayException,
- CacheExistsException, RegionExistsException, IOException,
- ClassNotFoundException
- {
+ private Region createHARegion() throws TimeoutException, CacheWriterException, GatewayException, CacheExistsException, RegionExistsException, IOException, ClassNotFoundException {
AttributesFactory factory = new AttributesFactory();
factory.setDataPolicy(DataPolicy.REPLICATE);
factory.setScope(Scope.DISTRIBUTED_ACK);
@@ -108,8 +96,8 @@ public class HARegionJUnitTest
factory.setStatisticsEnabled(true);
;
factory.setCacheListener(new CacheListenerAdapter() {
- public void afterInvalidate(EntryEvent event)
- {
+ @Override
+ public void afterInvalidate(EntryEvent event) {
}
});
RegionAttributes ra = factory.create();
@@ -117,88 +105,56 @@ public class HARegionJUnitTest
null, ra);
region.getAttributesMutator().setEntryTimeToLive(ea);
return region;
-
}
/**
* test no exception being thrown while creating an HARegion
- *
*/
@Test
- public void testRegionCreation()
- {
- try {
- createHARegion();
- }
- catch (Exception e) {
- e.printStackTrace();
- fail("Test failed due to " + e);
- }
+ public void testRegionCreation() throws Exception {
+ createHARegion();
}
/**
* test no exception being thrown while put is being done on an HARegion
- *
*/
@Test
- public void testPut()
- {
- try {
- Region region = createHARegion();
- region.put("key1", "value1");
- Assert.assertEquals(region.get("key1"), "value1");
- }
- catch (Exception e) {
- fail("put failed due to " + e);
- }
+ public void testPut() throws Exception {
+ Region region = createHARegion();
+ region.put("key1", "value1");
+ assertEquals(region.get("key1"), "value1");
}
/**
* test no exception being thrown while doing a localDestroy on a HARegion
- *
*/
@Test
- public void testLocalDestroy()
- {
- try {
- Region region = createHARegion();
- region.put("key1", "value1");
- region.localDestroy("key1");
- Assert.assertEquals(region.get("key1"), null);
- }
- catch (Exception e) {
- e.printStackTrace();
- fail("put failed due to " + e);
- }
+ public void testLocalDestroy() throws Exception {
+ Region region = createHARegion();
+ region.put("key1", "value1");
+ region.localDestroy("key1");
+ assertEquals(region.get("key1"), null);
}
+
/**
* Test to verify event id exists when evict destroy happens.
- *
*/
@Test
- public void testEventIdSetForEvictDestroy()
- {
- try{
- AttributesFactory factory = new AttributesFactory();
-
- factory.setCacheListener(new CacheListenerAdapter(){
- public void afterDestroy(EntryEvent event){
- assertTrue("eventId has not been set for "+ event, ((EntryEventImpl)event).getEventId() != null);
- }
- });
-
- EvictionAttributes evAttr = EvictionAttributes.createLRUEntryAttributes(1,EvictionAction.LOCAL_DESTROY);
- factory.setEvictionAttributes(evAttr);
-
- RegionAttributes attrs = factory.createRegionAttributes();
- Region region = cache.createVMRegion("TEST_REGION", attrs);
- region.put("key1", "value1");
- region.put("key2", "value2");
- }
- catch (Exception e) {
- }
-
-
- }
+ public void testEventIdSetForEvictDestroy() throws Exception {
+ AttributesFactory factory = new AttributesFactory();
+ factory.setCacheListener(new CacheListenerAdapter(){
+ public void afterDestroy(EntryEvent event){
+ assertTrue("eventId has not been set for "+ event, ((EntryEventImpl)event).getEventId() != null);
+ }
+ });
+
+ EvictionAttributes evAttr = EvictionAttributes.createLRUEntryAttributes(1,EvictionAction.LOCAL_DESTROY);
+ factory.setEvictionAttributes(evAttr);
+
+ RegionAttributes attrs = factory.createRegionAttributes();
+ Region region = cache.createVMRegion("TEST_REGION", attrs);
+ region.put("key1", "value1");
+ region.put("key2", "value2");
+ }
}