You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@geode.apache.org by kl...@apache.org on 2015/10/12 20:43:39 UTC
[11/30] incubator-geode git commit: GEODE-243: remove deprecated
Bridge feature
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTestUtil.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTestUtil.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTestUtil.java
index fb233c2..6fd9e1d 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTestUtil.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTestUtil.java
@@ -31,7 +31,7 @@ import com.gemstone.gemfire.cache.client.Pool;
import com.gemstone.gemfire.cache.client.PoolManager;
import com.gemstone.gemfire.cache.client.internal.PoolImpl;
import com.gemstone.gemfire.cache.query.CqEvent;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.cache.util.CqListenerAdapter;
import com.gemstone.gemfire.distributed.DistributedSystem;
@@ -64,42 +64,16 @@ public class CacheServerTestUtil extends DistributedTestCase
super(name);
}
-
-
-// public static void createCacheClient(Properties props, String regionName)
-// throws Exception {
-// createCacheClient(props, regionName, getClientProperties(), Boolean.FALSE);
-// }
public static void createCacheClient(Pool poolAttr, String regionName)
throws Exception {
createCacheClient(poolAttr, regionName, getClientProperties(), Boolean.FALSE);
}
-// public static void createCacheClient(Properties props, String regionName,
-// Properties dsProperties) throws Exception {
-// createCacheClient(props, regionName, dsProperties, Boolean.FALSE);
-// }
public static void createCacheClient(Pool poolAttr, String regionName,
Properties dsProperties) throws Exception {
createCacheClient(poolAttr, regionName, dsProperties, Boolean.FALSE);
}
-// public static void createCacheClient(Properties props, String regionName,
-// Properties dsProperties, Boolean addControlListener) throws Exception {
-// new CacheServerTestUtil("temp").createCache(dsProperties);
-// BridgeWriter writer = new BridgeWriter();
-// writer.init(props);
-// AttributesFactory factory = new AttributesFactory();
-// factory.setScope(Scope.LOCAL);
-// factory.setCacheWriter(writer);
-// if (addControlListener.booleanValue()) {
-// factory.addCacheListener(new ControlListener());
-// }
-// RegionAttributes attrs = factory.create();
-// cache.createRegion(regionName, attrs);
-// pool = (PoolImpl)writer.getConnectionProxy();
-// }
-
public static void createClientCache(Pool poolAttr, String regionName) throws Exception {
createClientCache(poolAttr, regionName, getClientProperties());
}
@@ -299,7 +273,7 @@ public class CacheServerTestUtil extends DistributedTestCase
}
/**
- * Create client regions each with their own BridgeWriter instance.
+ * Create client regions
* @param props
* @param regionName1
* @param regionName2
@@ -356,7 +330,7 @@ public class CacheServerTestUtil extends DistributedTestCase
factory.setDataPolicy(DataPolicy.REPLICATE);
RegionAttributes attrs = factory.create();
cache.createRegion(regionName, attrs);
- BridgeServer server1 = cache.addBridgeServer();
+ CacheServer server1 = cache.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
server1.setPort(port);
server1.setNotifyBySubscription(notifyBySubscription.booleanValue());
@@ -378,7 +352,7 @@ public class CacheServerTestUtil extends DistributedTestCase
factory.setDataPolicy(DataPolicy.REPLICATE);
RegionAttributes attrs = factory.create();
cache.createRegion(regionName, attrs);
- BridgeServer server1 = cache.addBridgeServer();
+ CacheServer server1 = cache.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
server1.setPort(port);
server1.setNotifyBySubscription(notifyBySubscription.booleanValue());
@@ -407,7 +381,7 @@ public class CacheServerTestUtil extends DistributedTestCase
factory.setDataPolicy(DataPolicy.REPLICATE);
RegionAttributes attrs = factory.create();
cache.createRegion(regionName, attrs);
- BridgeServer server = cache.addBridgeServer();
+ CacheServer server = cache.addCacheServer();
server.setPort(serverPort.intValue());
server.setNotifyBySubscription(notifyBySubscription.booleanValue());
server.start();
@@ -428,7 +402,7 @@ public class CacheServerTestUtil extends DistributedTestCase
if (!regionName2.equals("")) {
cache.createRegion(regionName2, attrs);
}
- BridgeServer server1 = cache.addBridgeServer();
+ CacheServer server1 = cache.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
server1.setPort(port);
server1.setNotifyBySubscription(notifyBySubscription.booleanValue());
@@ -504,18 +478,18 @@ public class CacheServerTestUtil extends DistributedTestCase
}
public static void stopCacheServers() {
- Iterator iter = getCache().getBridgeServers().iterator();
+ Iterator iter = getCache().getCacheServers().iterator();
if (iter.hasNext()) {
- BridgeServer server = (BridgeServer) iter.next();
+ CacheServer server = (CacheServer) iter.next();
server.stop();
assertFalse(server.isRunning());
}
}
public static void restartCacheServers() {
- Iterator iter = getCache().getBridgeServers().iterator();
+ Iterator iter = getCache().getCacheServers().iterator();
if (iter.hasNext()) {
- BridgeServer server = (BridgeServer) iter.next();
+ CacheServer server = (CacheServer) iter.next();
try {
server.start();
} catch(Exception e) {
@@ -535,11 +509,6 @@ public class CacheServerTestUtil extends DistributedTestCase
return pool;
}
-// public static BridgeWriter getWriter()
-// {
-// return writer;
-// }
-
/**
* Disables the shuffling of endpoints for a client
*
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTransactionsDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTransactionsDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTransactionsDUnitTest.java
index 9bb6b59..399175d 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTransactionsDUnitTest.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/CacheServerTransactionsDUnitTest.java
@@ -22,7 +22,7 @@ import com.gemstone.gemfire.cache.Region;
import com.gemstone.gemfire.cache.Scope;
import com.gemstone.gemfire.cache.client.PoolManager;
import com.gemstone.gemfire.cache.client.internal.PoolImpl;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.cache30.CacheSerializableRunnable;
import com.gemstone.gemfire.distributed.DistributedSystem;
@@ -772,7 +772,7 @@ public class CacheServerTransactionsDUnitTest extends DistributedTestCase
});
Region r1 = cache.createRegion(REGION_NAME, factory.create());
assertNotNull(r1);
- BridgeServer server1 = cache.addBridgeServer();
+ CacheServer server1 = cache.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
server1.setPort(port);
server1.setMaxThreads(maxThreads.intValue());
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClearPropagationDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClearPropagationDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClearPropagationDUnitTest.java
index dad3b7d..ac345e4 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClearPropagationDUnitTest.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClearPropagationDUnitTest.java
@@ -18,7 +18,7 @@ import com.gemstone.gemfire.cache.Region;
import com.gemstone.gemfire.cache.RegionAttributes;
import com.gemstone.gemfire.cache.RegionEvent;
import com.gemstone.gemfire.cache.Scope;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.cache30.CacheSerializableRunnable;
import com.gemstone.gemfire.distributed.DistributedSystem;
@@ -402,7 +402,7 @@ public class ClearPropagationDUnitTest extends DistributedTestCase
factory.setDataPolicy(DataPolicy.REPLICATE);
RegionAttributes attrs = factory.create();
cache.createRegion(REGION_NAME, attrs);
- BridgeServer server = cache.addBridgeServer();
+ CacheServer server = cache.addCacheServer();
assertNotNull(server);
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
server.setPort(port);
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientConflationDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientConflationDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientConflationDUnitTest.java
index cf95b24..90ec2b5 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientConflationDUnitTest.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientConflationDUnitTest.java
@@ -22,15 +22,15 @@ import com.gemstone.gemfire.cache.client.Pool;
import com.gemstone.gemfire.cache.client.PoolFactory;
import com.gemstone.gemfire.cache.client.PoolManager;
import com.gemstone.gemfire.cache.client.internal.PoolImpl;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.distributed.DistributedSystem;
import com.gemstone.gemfire.distributed.internal.DistributionConfig;
import com.gemstone.gemfire.distributed.internal.InternalDistributedSystem;
import com.gemstone.gemfire.internal.AvailablePort;
-import com.gemstone.gemfire.internal.cache.BridgeObserverAdapter;
-import com.gemstone.gemfire.internal.cache.BridgeObserverHolder;
-import com.gemstone.gemfire.internal.cache.BridgeServerImpl;
+import com.gemstone.gemfire.internal.cache.ClientServerObserverAdapter;
+import com.gemstone.gemfire.internal.cache.ClientServerObserverHolder;
+import com.gemstone.gemfire.internal.cache.CacheServerImpl;
import dunit.DistributedTestCase;
import dunit.Host;
@@ -123,7 +123,7 @@ public class ClientConflationDUnitTest extends DistributedTestCase
createClientCacheFeeder(getServerHostName(Host.getHost(0)), new Integer(PORT));
vm1.invoke(ClientConflationDUnitTest.class, "createClientCache", new Object[] { getServerHostName(vm1.getHost()), new Integer(PORT),
conflation});
- vm1.invoke(ClientConflationDUnitTest.class, "setBridgeObserverForBeforeInterestRecovery");
+ vm1.invoke(ClientConflationDUnitTest.class, "setClientServerObserverForBeforeInterestRecovery");
vm1.invoke(ClientConflationDUnitTest.class, "setAllCountersZero");
vm1.invoke(ClientConflationDUnitTest.class, "assertAllCountersZero");
vm1.invoke(ClientConflationDUnitTest.class, "registerInterest");
@@ -267,10 +267,10 @@ public class ClientConflationDUnitTest extends DistributedTestCase
* reset all counters to zero before interest recovery
*
*/
- public static void setBridgeObserverForBeforeInterestRecovery()
+ public static void setClientServerObserverForBeforeInterestRecovery()
{
PoolImpl.BEFORE_RECOVER_INTEREST_CALLBACK_FLAG = true;
- BridgeObserverHolder.setInstance(new BridgeObserverAdapter() {
+ ClientServerObserverHolder.setInstance(new ClientServerObserverAdapter() {
public void beforeInterestRecovery()
{
setAllCountersZero();
@@ -282,9 +282,9 @@ public class ClientConflationDUnitTest extends DistributedTestCase
* Assert all queues are empty to aid later assertion for listener event counts.
*/
public static void assertAllQueuesEmpty() {
- Iterator servers = cacheServer.getBridgeServers().iterator();
+ Iterator servers = cacheServer.getCacheServers().iterator();
while (servers.hasNext()) {
- Iterator proxies = ((BridgeServerImpl)servers.next()).getAcceptor().
+ Iterator proxies = ((CacheServerImpl)servers.next()).getAcceptor().
getCacheClientNotifier().getClientProxies().iterator();
while (proxies.hasNext()) {
int qsize = ((CacheClientProxy)proxies.next()).getQueueSize();
@@ -402,7 +402,7 @@ public class ClientConflationDUnitTest extends DistributedTestCase
RegionAttributes attrs2 = factory.create();
cacheServer.createRegion(REGION_NAME1, attrs1);
cacheServer.createRegion(REGION_NAME2, attrs2);
- BridgeServer server = cacheServer.addBridgeServer();
+ CacheServer server = cacheServer.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET) ;
server.setPort(port);
server.setNotifyBySubscription(true);
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientHealthMonitorJUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientHealthMonitorJUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientHealthMonitorJUnitTest.java
index 0891524..015bb59 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientHealthMonitorJUnitTest.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientHealthMonitorJUnitTest.java
@@ -28,7 +28,7 @@ import com.gemstone.gemfire.cache.client.ServerConnectivityException;
import com.gemstone.gemfire.cache.client.internal.Connection;
import com.gemstone.gemfire.cache.client.internal.PoolImpl;
import com.gemstone.gemfire.cache.client.internal.ServerRegionProxy;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.distributed.DistributedSystem;
import com.gemstone.gemfire.internal.AvailablePort;
import com.gemstone.gemfire.internal.cache.EventID;
@@ -115,7 +115,7 @@ public class ClientHealthMonitorJUnitTest
*/
private int createServer()
{
- BridgeServer server = null;
+ CacheServer server = null;
try {
Properties p = new Properties();
// make it a loner
@@ -124,7 +124,7 @@ public class ClientHealthMonitorJUnitTest
this.system = DistributedSystem.connect(p);
this.cache = CacheFactory.create(system);
- server = this.cache.addBridgeServer();
+ server = this.cache.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
server.setMaximumTimeBetweenPings(TIME_BETWEEN_PINGS);
server.setMaxThreads(getMaxThreads());
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientInterestNotifyDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientInterestNotifyDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientInterestNotifyDUnitTest.java
index 551bebd..c66291e 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientInterestNotifyDUnitTest.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientInterestNotifyDUnitTest.java
@@ -23,12 +23,12 @@ import com.gemstone.gemfire.cache.InterestResultPolicy;
import com.gemstone.gemfire.cache.client.Pool;
import com.gemstone.gemfire.cache.client.PoolFactory;
import com.gemstone.gemfire.cache.client.PoolManager;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.distributed.DistributedSystem;
import com.gemstone.gemfire.distributed.internal.DistributionConfig;
import com.gemstone.gemfire.internal.AvailablePort;
-import com.gemstone.gemfire.internal.cache.BridgeServerImpl;
+import com.gemstone.gemfire.internal.cache.CacheServerImpl;
import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
import dunit.DistributedTestCase;
@@ -402,10 +402,10 @@ public class ClientInterestNotifyDUnitTest extends DistributedTestCase
// occasional failures in precheckins and cruisecontrol.
public static void assertAllQueuesEmpty() {
- Iterator servers = cacheServer.getBridgeServers().iterator();
+ Iterator servers = cacheServer.getCacheServers().iterator();
assertTrue("No servers found!", servers.hasNext());
while (servers.hasNext()) {
- Iterator proxies = ((BridgeServerImpl)servers.next()).getAcceptor().
+ Iterator proxies = ((CacheServerImpl)servers.next()).getAcceptor().
getCacheClientNotifier().getClientProxies().iterator();
assertTrue("No proxies found!", proxies.hasNext());
while (proxies.hasNext()) {
@@ -420,8 +420,8 @@ public class ClientInterestNotifyDUnitTest extends DistributedTestCase
String excuse;
public boolean done() {
// assume a single cache server as configured in this test
- BridgeServerImpl bridgeServer = (BridgeServerImpl) cacheServer.
- getBridgeServers().iterator().next();
+ CacheServerImpl bridgeServer = (CacheServerImpl) cacheServer.
+ getCacheServers().iterator().next();
if (bridgeServer == null) {
excuse = "No Cache Server";
return false;
@@ -474,7 +474,7 @@ public class ClientInterestNotifyDUnitTest extends DistributedTestCase
cacheServer.createRegion(REGION_NAME1, attrs);
cacheServer.createRegion(REGION_NAME2, attrs);
cacheServer.createRegion(REGION_NAME3, attrs);
- BridgeServer server = cacheServer.addBridgeServer();
+ CacheServer server = cacheServer.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET) ;
server.setPort(port);
server.setNotifyBySubscription(true);
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscDUnitTest.java
new file mode 100755
index 0000000..1813123
--- /dev/null
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscDUnitTest.java
@@ -0,0 +1,1381 @@
+/*=========================================================================
+ * Copyright (c) 2010-2014 Pivotal Software, Inc. All Rights Reserved.
+ * This product is protected by U.S. and international copyright
+ * and intellectual property laws. Pivotal products are covered by
+ * one or more patents listed at http://www.pivotal.io/patents.
+ *=========================================================================
+ */
+package com.gemstone.gemfire.internal.cache.tier.sockets;
+
+import com.gemstone.gemfire.cache.AttributesFactory;
+import com.gemstone.gemfire.cache.Cache;
+import com.gemstone.gemfire.cache.CacheException;
+import com.gemstone.gemfire.cache.CacheWriterException;
+import com.gemstone.gemfire.cache.DataPolicy;
+import com.gemstone.gemfire.cache.EvictionAttributes;
+import com.gemstone.gemfire.cache.Region;
+import com.gemstone.gemfire.cache.RegionAttributes;
+import com.gemstone.gemfire.cache.Scope;
+import com.gemstone.gemfire.cache.client.NoAvailableServersException;
+import com.gemstone.gemfire.cache.client.Pool;
+import com.gemstone.gemfire.cache.client.PoolManager;
+import com.gemstone.gemfire.cache.client.internal.Connection;
+import com.gemstone.gemfire.cache.client.internal.PoolImpl;
+import com.gemstone.gemfire.cache.client.internal.RegisterInterestTracker;
+import com.gemstone.gemfire.cache.server.CacheServer;
+import com.gemstone.gemfire.cache30.CacheSerializableRunnable;
+import com.gemstone.gemfire.cache30.CacheTestCase;
+import com.gemstone.gemfire.cache30.LRUEvictionControllerDUnitTest;
+import com.gemstone.gemfire.distributed.DistributedSystem;
+import com.gemstone.gemfire.distributed.DistributedSystemDisconnectedException;
+import com.gemstone.gemfire.distributed.internal.DistributionConfig;
+import com.gemstone.gemfire.internal.AvailablePort;
+import com.gemstone.gemfire.internal.cache.CacheServerImpl;
+import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
+
+import dunit.DistributedTestCase;
+import dunit.Host;
+import dunit.SerializableCallable;
+import dunit.SerializableRunnable;
+import dunit.VM;
+
+import java.util.Iterator;
+import java.util.Properties;
+import java.util.Set;
+
+import junit.framework.AssertionFailedError;
+
+/**
+ * Tests client server corner cases between Region and Pool
+ *
+ * @author Yogesh Mahajan
+ *
+ */
+public class ClientServerMiscDUnitTest extends CacheTestCase
+{
+// private static Cache cache = null;
+
+ protected static PoolImpl pool = null;
+
+ protected static Connection conn = null;
+
+ private static Cache static_cache;
+
+ private static int PORT1;
+
+ private static final String k1 = "k1";
+
+ private static final String k2 = "k2";
+
+ private static final String server_k1 = "server-k1";
+
+ private static final String server_k2 = "server-k2";
+
+ private static final String REGION_NAME1 = "ClientServerMiscDUnitTest_region1";
+
+ private static final String REGION_NAME2 = "ClientServerMiscDUnitTest_region2";
+
+ private static final String PR_REGION_NAME = "ClientServerMiscDUnitTest_PRregion";
+
+ private static Host host;
+
+ private static VM server1;
+
+ private static VM server2;
+
+ private static RegionAttributes attrs;
+
+
+ // variables for concurrent map API test
+ Properties props = new Properties();
+ final int putRange_1Start = 1;
+ final int putRange_1End = 5;
+ final int putRange_2Start = 6;
+ final int putRange_2End = 10;
+ final int putRange_3Start = 11;
+ final int putRange_3End = 15;
+ final int putRange_4Start = 16;
+ final int putRange_4End = 20;
+ final int removeRange_1Start = 2;
+ final int removeRange_1End = 4;
+ final int removeRange_2Start = 7;
+ final int removeRange_2End = 9;
+
+
+
+ /** constructor */
+ public ClientServerMiscDUnitTest(String name) {
+ super(name);
+ }
+
+ public void setUp() throws Exception
+ {
+ super.setUp();
+ host = Host.getHost(0);
+ server1 = host.getVM(0);
+ server2 = host.getVM(1);
+ }
+
+ private int initServerCache(boolean notifyBySub) {
+ Object[] args = new Object[] { notifyBySub, getMaxThreads()};
+ return ((Integer)server1.invoke(ClientServerMiscDUnitTest.class,
+ "createServerCache",
+ args)).intValue();
+ }
+
+ private int initServerCache2(boolean notifyBySub) {
+ Object[] args = new Object[] {notifyBySub, getMaxThreads()};
+ return ((Integer)server2.invoke(ClientServerMiscDUnitTest.class,
+ "createServerCache",
+ args)).intValue();
+ }
+
+
+ public void testConcurrentOperationsWithDRandPR() throws Exception {
+ int port1 = initServerCache(true); // vm0
+ int port2 = initServerCache2(true); // vm1
+ String serverName = getServerHostName(Host.getHost(0));
+ host.getVM(2).invoke(this.getClass(), "createClientCacheV", new Object[]{serverName, port1});
+ host.getVM(3).invoke(this.getClass(), "createClientCacheV", new Object[]{serverName, port2});
+ getLogWriter().info("Testing concurrent map operations from a client with a distributed region");
+ concurrentMapTest(host.getVM(2), "/" + REGION_NAME1);
+ // TODO add verification in vm3
+ getLogWriter().info("Testing concurrent map operations from a client with a partitioned region");
+ concurrentMapTest(host.getVM(2), "/" + PR_REGION_NAME);
+ // TODO add verification in vm3
+ }
+
+ public void testConcurrentOperationsWithDRandPRandEmptyClient() throws Exception {
+ int port1 = initServerCache(true); // vm0
+ int port2 = initServerCache2(true); // vm1
+ String serverName = getServerHostName(Host.getHost(0));
+ host.getVM(2).invoke(this.getClass(), "createEmptyClientCache", new Object[]{serverName, port1});
+ host.getVM(3).invoke(this.getClass(), "createClientCacheV", new Object[]{serverName, port2});
+ getLogWriter().info("Testing concurrent map operations from a client with a distributed region");
+ concurrentMapTest(host.getVM(2), "/" + REGION_NAME1);
+ // TODO add verification in vm3
+ getLogWriter().info("Testing concurrent map operations from a client with a partitioned region");
+ concurrentMapTest(host.getVM(2), "/" + PR_REGION_NAME);
+ // TODO add verification in vm3
+ }
+
+ /**
+ * Do putIfAbsent(), replace(Object, Object),
+ * replace(Object, Object, Object), remove(Object, Object) operations
+ */
+ public void concurrentMapTest(final VM clientVM, final String rName) {
+
+ //String exceptionStr = "";
+ clientVM.invoke(new CacheSerializableRunnable("doConcurrentMapOperations") {
+ public void run2() throws CacheException {
+ Cache cache = getCache();
+ final Region pr = cache.getRegion(rName);
+ assertNotNull(rName + " not created", pr);
+ boolean isEmpty = pr.getAttributes().getDataPolicy() == DataPolicy.EMPTY;
+
+ // test successful putIfAbsent
+ for (int i = putRange_1Start; i <= putRange_1End; i++) {
+ Object putResult = pr.putIfAbsent(Integer.toString(i),
+ Integer.toString(i));
+ assertNull("Expected null, but got " + putResult + " for key " + i,
+ putResult);
+ }
+ int size;
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test unsuccessful putIfAbsent
+ for (int i = putRange_1Start; i <= putRange_1End; i++) {
+ Object putResult = pr.putIfAbsent(Integer.toString(i),
+ Integer.toString(i + 1));
+ assertEquals("for i=" + i, Integer.toString(i), putResult);
+ assertEquals("for i=" + i, Integer.toString(i), pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test successful replace(key, oldValue, newValue)
+ for (int i = putRange_1Start; i <= putRange_1End; i++) {
+ boolean replaceSucceeded = pr.replace(Integer.toString(i),
+ Integer.toString(i),
+ "replaced" + i);
+ assertTrue("for i=" + i, replaceSucceeded);
+ assertEquals("for i=" + i, "replaced" + i, pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test unsuccessful replace(key, oldValue, newValue)
+ for (int i = putRange_1Start; i <= putRange_2End; i++) {
+ boolean replaceSucceeded = pr.replace(Integer.toString(i),
+ Integer.toString(i), // wrong expected old value
+ "not" + i);
+ assertFalse("for i=" + i, replaceSucceeded);
+ assertEquals("for i=" + i,
+ i <= putRange_1End ? "replaced" + i : null,
+ pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test successful replace(key, value)
+ for (int i = putRange_1Start; i <= putRange_1End; i++) {
+ Object replaceResult = pr.replace(Integer.toString(i),
+ "twice replaced" + i);
+ assertEquals("for i=" + i, "replaced" + i, replaceResult);
+ assertEquals("for i=" + i,
+ "twice replaced" + i,
+ pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test unsuccessful replace(key, value)
+ for (int i = putRange_2Start; i <= putRange_2End; i++) {
+ Object replaceResult = pr.replace(Integer.toString(i),
+ "thrice replaced" + i);
+ assertNull("for i=" + i, replaceResult);
+ assertNull("for i=" + i, pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test unsuccessful remove(key, value)
+ for (int i = putRange_1Start; i <= putRange_2End; i++) {
+ boolean removeResult = pr.remove(Integer.toString(i),
+ Integer.toString(-i));
+ assertFalse("for i=" + i, removeResult);
+ assertEquals("for i=" + i,
+ i <= putRange_1End ? "twice replaced" + i : null,
+ pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", putRange_1End, size);
+ assertFalse("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ // test successful remove(key, value)
+ for (int i = putRange_1Start; i <= putRange_1End; i++) {
+ boolean removeResult = pr.remove(Integer.toString(i),
+ "twice replaced" + i);
+ assertTrue("for i=" + i, removeResult);
+ assertEquals("for i=" + i, null, pr.get(Integer.toString(i)));
+ }
+ if (!isEmpty) {
+ size = pr.size();
+ assertEquals("Size doesn't return expected value", 0, size);
+ pr.localClear();
+ assertTrue("isEmpty doesnt return proper state of the PartitionedRegion",
+ pr.isEmpty());
+ }
+
+ if (!isEmpty) {
+ // bug #42169 - entry not updated on server when locally destroyed on client
+ String key42169 = "key42169";
+ pr.put(key42169, "initialValue42169");
+ pr.localDestroy(key42169);
+ boolean success = pr.replace(key42169, "initialValue42169", "newValue42169");
+ assertTrue("expected replace to succeed", success);
+ pr.destroy(key42169);
+ pr.put(key42169, "secondRound");
+ pr.localDestroy(key42169);
+ Object result = pr.putIfAbsent(key42169, null);
+ assertEquals("expected putIfAbsent to fail", result, "secondRound");
+ pr.destroy(key42169);
+ }
+
+ if (isEmpty) {
+ String key41265 = "key41265";
+ boolean success = pr.remove(key41265, null);
+ assertFalse("expected remove to fail because key does not exist", success);
+ }
+
+ // test null values
+
+ // putIfAbsent with null value creates invalid entry
+ Object oldValue = pr.putIfAbsent("keyForNull", null);
+ assertNull(oldValue);
+ if (!isEmpty) {
+ assertTrue(pr.containsKey("keyForNull"));
+ assertTrue(!pr.containsValueForKey("keyForNull"));
+ }
+
+ // replace allows null value for oldValue, meaning invalidated entry
+ assertTrue(pr.replace("keyForNull", null, "no longer invalid"));
+
+ // replace does not allow null value for new value
+ try {
+ pr.replace("keyForNull", "no longer invalid", null);
+ fail("expected a NullPointerException");
+ } catch (NullPointerException expected) {
+ }
+
+ // other variant of replace does not allow null value for new value
+ try {
+ pr.replace("keyForNull", null);
+ fail ("expected a NullPointerException");
+ } catch (NullPointerException expected) {
+ }
+
+ // replace with null oldvalue matches invalidated entry
+ pr.putIfAbsent("otherKeyForNull", null);
+ int puts = ((GemFireCacheImpl)pr.getCache()).getCachePerfStats().getPuts();
+ boolean success = pr.replace("otherKeyForNull", null, "no longer invalid");
+ assertTrue(success);
+ int newputs = ((GemFireCacheImpl)pr.getCache()).getCachePerfStats().getPuts();
+ assertTrue("stats not updated properly or replace malfunctioned", newputs == puts+1);
+
+ }
+ });
+ }
+
+ /**
+ * Test two regions: notify by subscription is true.
+ * For region1 the interest list is empty , for region 2 the intetest list is all keys.
+ * If an update/create is made on region1 , the client should not receive any.
+ * If the create/update is on region2 , the client should receive the update.
+ */
+ public void testForTwoRegionHavingDifferentInterestList()
+ throws Exception
+ {
+ // start server first
+ PORT1 = initServerCache(true);
+ createClientCache(getServerHostName(Host.getHost(0)), PORT1);
+ populateCache();
+ registerInterest();
+ server1.invoke(ClientServerMiscDUnitTest.class, "put");
+
+// pause(5000 + 5000 + 10000);
+ /*final int maxWaitTime = Integer.getInteger(WAIT_PROPERTY, WAIT_DEFAULT).intValue();
+ try {
+ Thread.yield();
+ Thread.sleep(maxWaitTime);
+ }
+ catch (InterruptedException e) {
+ fail("interrupted");
+ }*/
+ verifyUpdates();
+
+ }
+
+ /**
+ * Test two regions: notify by subscription is true.
+ * Both the regions have registered interest in all the keys.
+ * Now close region1 on the client.
+ * The region1 should get removed from the interest list on CCP at server.
+ * Any update on region1 on server should not get pushed to the client.
+ * Ensure that the message related is not added to the client's queue at all
+ * ( which is diferent from not receiving a callbak on the client).
+ * If an update on region2 is made on the server , then client should receive the calback
+ */
+ public void testForTwoRegionHavingALLKEYSInterest()
+ throws Exception
+ {
+ // start server first
+ PORT1 = initServerCache(true);
+ createClientCache(getServerHostName(Host.getHost(0)), PORT1);
+ populateCache();
+ registerInterestInBothTheRegions();
+ closeRegion1();
+ pause(6000);
+ server1.invoke(ClientServerMiscDUnitTest.class,
+ "verifyInterestListOnServer");
+ server1.invoke(ClientServerMiscDUnitTest.class, "put");
+ //pause(5000);
+ verifyUpdatesOnRegion2();
+ }
+
+ /** Test two regions: notify by subscription is true.
+ * Both the regions have registered interest in all the keys.
+ * Close both the regions. When the last region is closed ,
+ * it should close the ConnectionProxy on the client ,
+ * close all the server connection threads on the server &
+ * remove the CacheClientProxy from the CacheClient notifier
+ */
+ public void testRegionClose() throws Exception
+ {
+ // start server first
+ PORT1 = initServerCache(true);
+ pool = (PoolImpl)createClientCache(getServerHostName(Host.getHost(0)),PORT1);
+ populateCache();
+ registerInterestInBothTheRegions();
+ closeBothRegions();
+ //pause(5000);
+ assertEquals(false, pool.isDestroyed());
+ pool.destroy();
+ assertEquals(true, pool.isDestroyed());
+ server1.invoke(ClientServerMiscDUnitTest.class,
+ "verifyNoCacheClientProxyOnServer");
+
+ }
+
+ /**
+ * Test two regions: notify by
+ * subscription is true. Both the regions have registered interest in all the
+ * keys. Destroy region1 on the client. It should reach the server , kill the
+ * region on the server , propagate it to the interested clients , but it
+ * should keep CacheClient Proxy alive. Destroy Region2 . It should reach
+ * server , close conenction proxy , destroy the region2 on the server ,
+ * remove the cache client proxy from the cache client notifier & propagate it
+ * to the clients. Then create third region and verify that no
+ * CacheClientProxy is created on server
+ */
+ public void testCCPDestroyOnLastDestroyRegion() throws Exception
+ {
+ PORT1 = initServerCache(true);
+ PoolImpl pool = (PoolImpl)createClientCache(getServerHostName(Host.getHost(0)),PORT1);
+ destroyRegion1();
+ // pause(5000);
+ server1.invoke(ClientServerMiscDUnitTest.class,
+ "verifyCacheClientProxyOnServer", new Object[] { new String(
+ REGION_NAME1) });
+ Connection conn = pool.acquireConnection();
+ assertNotNull(conn);
+ assertEquals(1, pool.getConnectedServerCount());
+ assertEquals(false, pool.isDestroyed());
+ destroyRegion2();
+ assertEquals(false, pool.isDestroyed());
+ destroyPRRegion();
+ assertEquals(false, pool.isDestroyed());
+ pool.destroy();
+ assertEquals(true, pool.isDestroyed());
+ // pause(5000);
+ server1.invoke(ClientServerMiscDUnitTest.class,
+ "verifyNoCacheClientProxyOnServer");
+ try {
+ getCache().createRegion(REGION_NAME2, attrs);
+ fail("expected IllegalStateException");
+ } catch (IllegalStateException expected) {
+ }
+ }
+
+ /**
+ * Test two regions:If notify by
+ * subscription is false , both the regions should receive invalidates for the
+ * updates on server in their respective regions
+ *
+ */
+ public void testInvalidatesPropagateOnTwoRegions()
+ throws Exception
+ {
+ // start server first
+ PORT1 = initServerCache(false);
+ createClientCache(getServerHostName(Host.getHost(0)), PORT1);
+ registerInterestForInvalidatesInBothTheRegions();
+ populateCache();
+ server1.invoke(ClientServerMiscDUnitTest.class, "put");
+ //pause(5000);
+ verifyInvalidatesOnBothRegions();
+
+ }
+
+ /**
+ * Test for bug 43407, where LRU in the client caused an entry to be
+ * evicted with DESTROY(), then the client invalidated the entry and
+ * did a get(). After the get() the entry was not seen to be in the
+ * client's cache. This turned out to be expected behavior, but we
+ * now have this test to guarantee that the product behaves as expected.
+ */
+ public void testBug43407()
+ throws Exception
+ {
+ // start server first
+ PORT1 = initServerCache(false);
+ createClientCache(getServerHostName(Host.getHost(0)), PORT1);
+ registerInterestForInvalidatesInBothTheRegions();
+ Region region = static_cache.getRegion(REGION_NAME1);
+ populateCache();
+ region.put("invalidationKey", "invalidationValue");
+ region.localDestroy("invalidationKey");
+ if (region.containsKey("invalidationKey")) {
+ fail("region still contains invalidationKey");
+ }
+ region.invalidate("invalidationKey");
+ if (region.containsKey("invalidationKey")) {
+ fail("this test expects the entry is not created on invalidate() if not there before the operation");
+ }
+ Object value = region.get("invalidationKey");
+ if (value != null) {
+ fail("this test expected a null response to get('invalidationKey')");
+ }
+ if (!region.containsKeyOnServer("invalidationKey")) {
+ fail("expected an entry on the server after invalidation");
+ }
+ // bug 43407 asserts that there should be an entry, but the product does not
+ // do this. This verifies that the product does not behave as asserted in that bug
+ if (region.containsKey("invalidationKey")) {
+ fail("expected no entry after invalidation when entry was not in client but was on server");
+ }
+ }
+
+ /**
+ * Create cache, create pool, notify-by-subscription=false,
+ * create a region and on client and on server.
+ * Do not attach pool to region ,
+ * populate some entries on region both on client and server.
+ * Update the entries on server the client.
+ * The client should not have entry invalidate.
+ *
+ * @throws Exception
+ */
+ public void testInvalidatesPropagateOnRegionHavingNoPool()
+ throws Exception
+ {
+ // start server first
+ PORT1 = initServerCache(false);
+ Properties props = new Properties();
+ props.setProperty(DistributionConfig.MCAST_PORT_NAME, "0");
+ props.setProperty(DistributionConfig.LOCATORS_NAME, "");
+ new ClientServerMiscDUnitTest("temp").createCache(props);
+ String host = getServerHostName(server1.getHost());
+ PoolImpl p = (PoolImpl)PoolManager.createFactory()
+ .addServer(host, PORT1)
+ .setSubscriptionEnabled(true)
+ .setThreadLocalConnections(true)
+ .setReadTimeout(1000)
+ .setSocketBufferSize(32768)
+ .setMinConnections(3)
+ .setSubscriptionRedundancy(-1)
+ .setPingInterval(2000)
+ // .setRetryAttempts(5)
+ // .setRetryInterval(2000)
+ .create("testInvalidatesPropagateOnRegionHavingNoPool");
+
+ AttributesFactory factory = new AttributesFactory();
+ factory.setScope(Scope.DISTRIBUTED_ACK);
+ // factory.setPoolName(p.getName());
+
+ attrs = factory.create();
+ final Region region1 = getCache().createRegion(REGION_NAME1, attrs);
+ final Region region2 = getCache().createRegion(REGION_NAME2, attrs);
+ assertNotNull(region1);
+ assertNotNull(region2);
+ pool = p;
+ conn = pool.acquireConnection();
+ assertNotNull(conn);
+
+ populateCache();
+ server1.invoke(ClientServerMiscDUnitTest.class, "put");
+ // pause(5000);
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = region1.getEntry(k1).getValue();
+ return k1.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(region1.getEntry(k1).getValue(), k1);
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = region1.getEntry(k2).getValue();
+ return k2.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = region2.getEntry(k1).getValue();
+ return k1.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(region1.getEntry(k2).getValue(), k2);
+ // assertEquals(region2.getEntry(k1).getValue(), k1);
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = region2.getEntry(k2).getValue();
+ return k2.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(region2.getEntry(k2).getValue(), k2);
+ }
+
+ /**
+ * Create proxy before cache creation, create cache, create two regions,
+ * attach same bridge writer to both of the regions Region interests AL_KEYS
+ * on both the regions, notify-by-subscription=true . The CCP should have both
+ * the regions in interest list.
+ *
+ * @throws Exception
+ */
+
+ public void testProxyCreationBeforeCacheCreation() throws Exception
+ {
+ Properties props = new Properties();
+ props.setProperty("mcast-port", "0");
+ props.setProperty("locators", "");
+ DistributedSystem ds = getSystem(props);
+ assertNotNull(ds);
+ ds.disconnect();
+ ds = getSystem(props);
+ PORT1 = initServerCache(true);
+ String host = getServerHostName(server1.getHost());
+ Pool p = PoolManager.createFactory()
+ .addServer(host, PORT1)
+ .setSubscriptionEnabled(true)
+ .setSubscriptionRedundancy(-1)
+ // .setRetryAttempts(5)
+ .create("testProxyCreationBeforeCacheCreationPool");
+
+ Cache cache = getCache();
+ assertNotNull(cache);
+
+ AttributesFactory factory = new AttributesFactory();
+ factory.setScope(Scope.DISTRIBUTED_ACK);
+ factory.setPoolName(p.getName());
+
+ RegionAttributes myAttrs = factory.create();
+ Region region1 = cache.createRegion(REGION_NAME1, myAttrs);
+ Region region2 = cache.createRegion(REGION_NAME2, myAttrs);
+ assertNotNull(region1);
+ assertNotNull(region2);
+ //region1.registerInterest(CacheClientProxy.ALL_KEYS);
+ region2.registerInterest("ALL_KEYS");
+ pause(6000);
+ server1.invoke(ClientServerMiscDUnitTest.class,
+ "verifyInterestListOnServer");
+
+ }
+ /**
+ *
+ * Cycling a DistributedSystem with an initialized pool causes interest registration NPE
+ *
+ * Test Scenario:
+ *
+ * Create a DistributedSystem (DS1).
+ * Create a pool, initialize (creates a proxy with DS1 memberid)
+ * Disconnect DS1. Create a DistributedSystem (DS2).
+ * Create a Region with pool, it attempts to register interest using DS2 memberid, gets NPE.
+ *
+ * @throws Exception
+ */
+ public void testBug35380() throws Exception
+ {
+ Properties props = new Properties();
+ props.setProperty("mcast-port", "0");
+ props.setProperty("locators", "");
+ DistributedSystem ds = getSystem(props);
+ assertNotNull(ds);
+
+ PORT1 = initServerCache(true);
+ String host = getServerHostName(server1.getHost());
+ Pool p = PoolManager.createFactory()
+ .addServer(host, PORT1)
+ .setSubscriptionEnabled(true)
+ .setSubscriptionRedundancy(-1)
+ //.setRetryAttempts(5)
+ .create("testBug35380Pool");
+
+ Cache cache = getCache();
+
+ AttributesFactory factory = new AttributesFactory();
+ factory.setScope(Scope.DISTRIBUTED_ACK);
+ factory.setPoolName(p.getName());
+
+ RegionAttributes myAttrs = factory.create();
+ Region region1 = cache.createRegion(REGION_NAME1, myAttrs);
+ Region region2 = cache.createRegion(REGION_NAME2, myAttrs);
+ assertNotNull(region1);
+ assertNotNull(region2);
+
+ region2.registerInterest("ALL_KEYS");
+
+ ds.disconnect();
+ Properties prop = new Properties();
+ prop.setProperty("mcast-port", "0");
+ prop.setProperty("locators", "");
+ ds = getSystem(prop);
+
+ cache = getCache();
+ assertNotNull(cache);
+
+ AttributesFactory factory1 = new AttributesFactory();
+ factory1.setScope(Scope.DISTRIBUTED_ACK);
+ //reuse writer from prev DS
+ factory1.setPoolName(p.getName());
+
+ RegionAttributes attrs1 = factory1.create();
+ try {
+ cache.createRegion(REGION_NAME1, attrs1);
+ fail("expected ShutdownException");
+ }
+ catch(IllegalStateException expected) {
+ }
+ catch (DistributedSystemDisconnectedException expected) {
+ }
+ }
+
+
+ private void createCache(Properties props) throws Exception {
+ createCacheV(props);
+ }
+ private Cache createCacheV(Properties props) throws Exception
+ {
+ DistributedSystem ds = getSystem(props);
+ assertNotNull(ds);
+ ds.disconnect();
+ ds = getSystem(props);
+ Cache cache = getCache();
+ assertNotNull(cache);
+ return cache;
+ }
+
+ public static void createClientCacheV(String h, int port)
+ throws Exception {
+ _createClientCache(h, port, false);
+ }
+
+ public static void createEmptyClientCache(String h, int port)
+ throws Exception {
+ _createClientCache(h, port, true);
+ }
+
+ public static Pool createClientCache(String h, int port)
+ throws Exception {
+ return _createClientCache(h, port, false);
+ }
+
+ public static Pool _createClientCache(String h, int port, boolean empty)
+ throws Exception {
+ Properties props = new Properties();
+ props.setProperty("mcast-port", "0");
+ props.setProperty("locators", "");
+ Cache cache = new ClientServerMiscDUnitTest("temp").createCacheV(props);
+ ClientServerMiscDUnitTest.static_cache = cache;
+ PoolImpl p = (PoolImpl)PoolManager.createFactory()
+ .addServer(h, port)
+ .setSubscriptionEnabled(true)
+ .setThreadLocalConnections(true)
+ .setReadTimeout(1000)
+ .setSocketBufferSize(32768)
+ .setMinConnections(3)
+ .setSubscriptionRedundancy(-1)
+ .setPingInterval(2000)
+ // .setRetryAttempts(5)
+ // .setRetryInterval(2000)
+ .create("ClientServerMiscDUnitTestPool");
+
+ AttributesFactory factory = new AttributesFactory();
+ factory.setScope(Scope.DISTRIBUTED_ACK);
+ if (empty) {
+ factory.setDataPolicy(DataPolicy.EMPTY);
+ }
+ factory.setPoolName(p.getName());
+
+ attrs = factory.create();
+ Region region1 = cache.createRegion(REGION_NAME1, attrs);
+ Region region2 = cache.createRegion(REGION_NAME2, attrs);
+ Region prRegion = cache.createRegion(PR_REGION_NAME, attrs);
+ assertNotNull(region1);
+ assertNotNull(region2);
+ assertNotNull(prRegion);
+ pool = p;
+// conn = pool.acquireConnection();
+// assertNotNull(conn);
+ // TODO does this WaitCriterion actually help?
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ try {
+ conn = pool.acquireConnection();
+ if (conn == null) {
+ excuse = "acquireConnection returned null?";
+ return false;
+ }
+ return true;
+ } catch (NoAvailableServersException e) {
+ excuse = "Cannot find a server: " + e;
+ return false;
+ }
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ return p;
+ }
+
+ public static Integer createServerCache(Boolean notifyBySubscription, Integer maxThreads)
+ throws Exception {
+ Cache cache = new ClientServerMiscDUnitTest("temp").createCacheV(new Properties());
+ unsetSlowDispatcherFlag();
+ AttributesFactory factory = new AttributesFactory();
+ factory.setScope(Scope.DISTRIBUTED_ACK);
+ factory.setEnableConflation(true);
+ factory.setDataPolicy(DataPolicy.REPLICATE);
+ RegionAttributes myAttrs = factory.create();
+ Region r1 = cache.createRegion(REGION_NAME1, myAttrs);
+ Region r2 = cache.createRegion(REGION_NAME2, myAttrs);
+ factory = new AttributesFactory();
+ factory.setDataPolicy(DataPolicy.PARTITION);
+ RegionAttributes prAttrs = factory.create();
+ Region pr = cache.createRegion(PR_REGION_NAME, prAttrs);
+ assertNotNull(r1);
+ assertNotNull(r2);
+ assertNotNull(pr);
+
+ CacheServer server = cache.addCacheServer();
+ int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET);
+ r1.getCache().getDistributedSystem().getLogWriter().info("Starting server on port " + port);
+ server.setPort(port);
+ server.setMaxThreads(maxThreads.intValue());
+ server.setNotifyBySubscription(notifyBySubscription.booleanValue());
+ server.start();
+ r1.getCache().getDistributedSystem().getLogWriter().info("Started server on port " + server.getPort());
+ return new Integer(server.getPort());
+
+ }
+
+ protected int getMaxThreads() {
+ return 0;
+ }
+
+ public static void registerInterest()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r);
+ //r.registerInterestRegex(CacheClientProxy.ALL_KEYS);
+ r.registerInterest("ALL_KEYS");
+ }
+ catch (CacheWriterException e) {
+ e.printStackTrace();
+ fail("Test failed due to CacheWriterException during registerInterest", e);
+ }
+ }
+
+ public static void registerInterestForInvalidatesInBothTheRegions()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ assertNotNull(r1);
+ Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r2);
+ r1.registerInterest("ALL_KEYS", false, false);
+ r2.registerInterest("ALL_KEYS", false, false);
+ }
+ catch (CacheWriterException e) {
+ e.printStackTrace();
+ fail(
+ "Test failed due to CacheWriterException during registerInterestnBothRegions",
+ e);
+ }
+ }
+
+ public static void registerInterestInBothTheRegions()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ assertNotNull(r1);
+ Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r2);
+ r1.registerInterest("ALL_KEYS");
+ r2.registerInterest("ALL_KEYS");
+ }
+ catch (CacheWriterException e) {
+ e.printStackTrace();
+ fail(
+ "Test failed due to CacheWriterException during registerInterestnBothRegions",
+ e);
+ }
+ }
+
+ public static void closeRegion1()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ assertNotNull(r1);
+ r1.close();
+ }
+ catch (Exception e) {
+ e.printStackTrace();
+ fail("Test failed due to Exception during closeRegion1", e);
+ }
+ }
+
+ public static void closeBothRegions()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ Region pr = cache.getRegion(Region.SEPARATOR + PR_REGION_NAME);
+ assertNotNull(r1);
+ assertNotNull(r2);
+ assertNotNull(pr);
+ r1.close();
+ r2.close();
+ pr.close();
+ }
+ catch (Exception e) {
+ e.printStackTrace();
+ fail("Test failed due to Exception during closeBothRegions", e);
+ }
+ }
+
+ public static void destroyRegion1()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ assertNotNull(r1);
+ r1.destroyRegion();
+ }
+ catch (Exception e) {
+ e.printStackTrace();
+ fail("Test failed due to Exception during closeBothRegions", e);
+ }
+ }
+
+ public static void destroyRegion2()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r2);
+ r2.destroyRegion();
+ }
+ catch (Exception e) {
+ e.printStackTrace();
+ fail("Test failed due to Exception during closeBothRegions", e);
+ }
+ }
+
+ public static void destroyPRRegion() {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r2 = cache.getRegion(Region.SEPARATOR + PR_REGION_NAME);
+ assertNotNull(r2);
+ r2.destroyRegion();
+ } catch (Exception e) {
+ // e.printStackTrace();
+ fail("Test failed due to Exception during closeBothRegions", e);
+ }
+ }
+
+ public static void verifyInterestListOnServer()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ assertEquals("More than one BridgeServer", 1, cache.getCacheServers()
+ .size());
+ CacheServerImpl bs = (CacheServerImpl)cache.getCacheServers()
+ .iterator().next();
+ assertNotNull(bs);
+ assertNotNull(bs.getAcceptor());
+ assertNotNull(bs.getAcceptor().getCacheClientNotifier());
+ Iterator iter_prox = bs.getAcceptor().getCacheClientNotifier()
+ .getClientProxies().iterator();
+ while (iter_prox.hasNext()) {
+ CacheClientProxy ccp = (CacheClientProxy)iter_prox.next();
+ // CCP should not contain region1
+ Set akr = ccp.cils[RegisterInterestTracker.interestListIndex].regions;
+ assertNotNull(akr);
+ assertTrue(!akr.contains(Region.SEPARATOR + REGION_NAME1));
+ // CCP should contain region2
+ assertTrue(akr.contains(Region.SEPARATOR + REGION_NAME2));
+ assertEquals(1, akr.size());
+ }
+ }
+ catch (Exception ex) {
+ ex.printStackTrace();
+ fail("while setting verifyInterestListOnServer " + ex);
+ }
+ }
+
+ public static void verifyNoCacheClientProxyOnServer()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ assertEquals("More than one BridgeServer", 1, cache.getCacheServers()
+ .size());
+ CacheServerImpl bs = (CacheServerImpl)cache.getCacheServers()
+ .iterator().next();
+ assertNotNull(bs);
+ assertNotNull(bs.getAcceptor());
+ final CacheClientNotifier ccn = bs.getAcceptor().getCacheClientNotifier();
+
+ assertNotNull(ccn);
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ return ccn.getClientProxies().size() == 0;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 40 * 1000, 1000, true);
+ }
+ catch (Exception ex) {
+ ex.printStackTrace();
+ fail("while setting verifyNoCacheClientProxyOnServer " + ex);
+ }
+ }
+
+ public static void verifyCacheClientProxyOnServer(String regionName)
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ assertNull(cache.getRegion(Region.SEPARATOR + regionName));
+ verifyCacheClientProxyOnServer();
+
+ //assertEquals(1, bs.getAcceptor().getCacheClientNotifier().getClientProxies().size());
+ }
+ catch (Exception ex) {
+ ex.printStackTrace();
+ fail("while setting verifyNoCacheClientProxyOnServer " + ex);
+ }
+ }
+
+ public static void verifyCacheClientProxyOnServer()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ assertEquals("More than one BridgeServer", 1, cache.getCacheServers()
+ .size());
+ CacheServerImpl bs = (CacheServerImpl)cache.getCacheServers()
+ .iterator().next();
+ assertNotNull(bs);
+ assertNotNull(bs.getAcceptor());
+ final CacheClientNotifier ccn = bs.getAcceptor().getCacheClientNotifier();
+
+ assertNotNull(ccn);
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ return ccn.getClientProxies().size() == 1;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 40 * 1000, 1000, true);
+ }
+ catch (Exception ex) {
+ ex.printStackTrace();
+ fail("while setting verifyNoCacheClientProxyOnServer " + ex);
+ }
+ }
+
+ public static void populateCache()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r1);
+ assertNotNull(r2);
+
+ if (!r1.containsKey(k1))
+ r1.create(k1, k1);
+ if (!r1.containsKey(k2))
+ r1.create(k2, k2);
+ if (!r2.containsKey(k1))
+ r2.create(k1, k1);
+ if (!r2.containsKey(k2))
+ r2.create(k2, k2);
+
+ assertEquals(r1.getEntry(k1).getValue(), k1);
+ assertEquals(r1.getEntry(k2).getValue(), k2);
+ assertEquals(r2.getEntry(k1).getValue(), k1);
+ assertEquals(r2.getEntry(k2).getValue(), k2);
+ }
+ catch (Exception ex) {
+ fail("failed while createEntries()", ex);
+ }
+ }
+
+ public static void put()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r1);
+ assertNotNull(r2);
+
+ r1.put(k1, server_k1);
+ r1.put(k2, server_k2);
+
+ r2.put(k1, server_k1);
+ r2.put(k2, server_k2);
+
+ assertEquals(r1.getEntry(k1).getValue(), server_k1);
+ assertEquals(r1.getEntry(k2).getValue(), server_k2);
+ assertEquals(r2.getEntry(k1).getValue(), server_k1);
+ assertEquals(r2.getEntry(k2).getValue(), server_k2);
+ }
+ catch (Exception ex) {
+ fail("failed while put()", ex);
+ }
+ }
+
+ public static void verifyUpdates()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ final Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ final Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r1);
+ assertNotNull(r2);
+ // verify updates
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r1.getEntry(k1).getValue();
+ return k1.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(k1, r1.getEntry(k1).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r1.getEntry(k2).getValue();
+ return k2.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(k2, r1.getEntry(k2).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r2.getEntry(k1).getValue();
+ return server_k1.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(server_k1, r2.getEntry(k1).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r2.getEntry(k2).getValue();
+ return server_k2.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(server_k2, r2.getEntry(k2).getValue());
+ }
+ catch (Exception ex) {
+ fail("failed while verifyUpdates()", ex);
+ }
+ }
+
+ public static void verifyInvalidatesOnBothRegions()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ final Region r1 = cache.getRegion(Region.SEPARATOR + REGION_NAME1);
+ final Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r1);
+ assertNotNull(r2);
+
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r1.getEntry(k1).getValue();
+ return val == null;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 90 * 1000, 1000, true);
+
+ // assertNull(r1.getEntry(k1).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r1.getEntry(k1).getValue();
+ return val == null;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 90 * 1000, 1000, true);
+
+ // assertNull(r1.getEntry(k2).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r1.getEntry(k2).getValue();
+ return val == null;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 90 * 1000, 1000, true);
+
+
+ // assertNull(r2.getEntry(k1).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r2.getEntry(k1).getValue();
+ return val == null;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 90 * 1000, 1000, true);
+
+ // assertNull(r2.getEntry(k2).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r2.getEntry(k2).getValue();
+ return val == null;
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 90 * 1000, 1000, true);
+ }
+ catch (Exception ex) {
+ fail("failed while verifyInvalidatesOnBothRegions()", ex);
+ }
+ }
+
+ public static void verifyUpdatesOnRegion2()
+ {
+ try {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ final Region r2 = cache.getRegion(Region.SEPARATOR + REGION_NAME2);
+ assertNotNull(r2);
+ WaitCriterion wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r2.getEntry(k1).getValue();
+ return server_k1.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(server_k1, r2.getEntry(k1).getValue());
+ wc = new WaitCriterion() {
+ String excuse;
+ public boolean done() {
+ Object val = r2.getEntry(k2).getValue();
+ return server_k2.equals(val);
+ }
+ public String description() {
+ return excuse;
+ }
+ };
+ DistributedTestCase.waitForCriterion(wc, 60 * 1000, 1000, true);
+
+ // assertEquals(server_k2, r2.getEntry(k2).getValue());
+ }
+ catch (Exception ex) {
+ fail("failed while verifyUpdatesOnRegion2()", ex);
+ }
+ }
+
+ public void tearDown2() throws Exception
+ {
+ super.tearDown2();
+ // close the clients first
+ closeCache();
+ // then close the servers
+ server1.invoke(ClientServerMiscDUnitTest.class, "closeCache");
+
+ }
+
+ public static void closeCache()
+ {
+ Cache cache = new ClientServerMiscDUnitTest("temp").getCache();
+ if (cache != null && !cache.isClosed()) {
+ cache.close();
+ cache.getDistributedSystem().disconnect();
+ }
+ }
+
+ /**
+ * set the boolean for starting the dispatcher thread a bit later to FALSE.
+ * This is just a precaution in case any test set it to true and did not unset
+ * it on completion.
+ *
+ */
+ public static void unsetSlowDispatcherFlag()
+ {
+ CacheClientProxy.isSlowStartForTesting = false;
+ }
+
+}
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscSelectorDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscSelectorDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscSelectorDUnitTest.java
new file mode 100644
index 0000000..9b7c13e
--- /dev/null
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ClientServerMiscSelectorDUnitTest.java
@@ -0,0 +1,27 @@
+/*=========================================================================
+ * Copyright (c) 2010-2014 Pivotal Software, Inc. All Rights Reserved.
+ * This product is protected by U.S. and international copyright
+ * and intellectual property laws. Pivotal products are covered by
+ * one or more patents listed at http://www.pivotal.io/patents.
+ *=========================================================================
+ */
+package com.gemstone.gemfire.internal.cache.tier.sockets;
+
+
+/**
+ * Just like parent but enables server thread pool
+ * (ie. selector)
+ *
+ * @author darrel
+ *
+ */
+public class ClientServerMiscSelectorDUnitTest extends ClientServerMiscDUnitTest
+{
+ public ClientServerMiscSelectorDUnitTest(String name) {
+ super(name);
+ }
+
+ protected int getMaxThreads() {
+ return 2;
+ }
+}
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/2eb4e175/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ConflationDUnitTest.java
----------------------------------------------------------------------
diff --git a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ConflationDUnitTest.java b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ConflationDUnitTest.java
index 35eb52d..88ec5d0 100755
--- a/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ConflationDUnitTest.java
+++ b/gemfire-core/src/test/java/com/gemstone/gemfire/internal/cache/tier/sockets/ConflationDUnitTest.java
@@ -19,14 +19,14 @@ import com.gemstone.gemfire.cache.EntryEvent;
import com.gemstone.gemfire.cache.Region;
import com.gemstone.gemfire.cache.RegionAttributes;
import com.gemstone.gemfire.cache.Scope;
-import com.gemstone.gemfire.cache.util.BridgeServer;
+import com.gemstone.gemfire.cache.server.CacheServer;
import com.gemstone.gemfire.cache.util.CacheListenerAdapter;
import com.gemstone.gemfire.distributed.DistributedSystem;
import com.gemstone.gemfire.distributed.internal.DistributionConfig;
import com.gemstone.gemfire.internal.AvailablePort;
-import com.gemstone.gemfire.internal.cache.BridgeObserverAdapter;
-import com.gemstone.gemfire.internal.cache.BridgeObserverHolder;
-import com.gemstone.gemfire.internal.cache.BridgeServerImpl;
+import com.gemstone.gemfire.internal.cache.ClientServerObserverAdapter;
+import com.gemstone.gemfire.internal.cache.ClientServerObserverHolder;
+import com.gemstone.gemfire.internal.cache.CacheServerImpl;
import com.gemstone.gemfire.internal.cache.GemFireCacheImpl;
import com.gemstone.gemfire.internal.cache.HARegion;
import com.gemstone.gemfire.internal.cache.ha.HAHelper;
@@ -131,7 +131,7 @@ public class ConflationDUnitTest extends DistributedTestCase
createClientCache1UniqueWriter ( getServerHostName(Host.getHost(0)), new Integer(PORT));
vm2.invoke(ConflationDUnitTest.class, "createClientCache2UniqueWriter",
new Object[] { getServerHostName(Host.getHost(0)), new Integer(PORT)});
- vm2.invoke(ConflationDUnitTest.class, "setBridgeObserverForBeforeInterestRecovery");
+ vm2.invoke(ConflationDUnitTest.class, "setClientServerObserverForBeforeInterestRecovery");
vm2.invoke(ConflationDUnitTest.class, "setAllCountersZero");
vm2.invoke(ConflationDUnitTest.class, "assertAllCountersZero");
vm2.invoke(ConflationDUnitTest.class, "registerInterest");
@@ -162,7 +162,7 @@ public class ConflationDUnitTest extends DistributedTestCase
createClientCache1CommonWriter( getServerHostName(Host.getHost(0)), new Integer(PORT));
vm2.invoke(ConflationDUnitTest.class, "createClientCache2CommonWriter",
new Object[] { getServerHostName(Host.getHost(0)), new Integer(PORT)});
- vm2.invoke(ConflationDUnitTest.class, "setBridgeObserverForBeforeInterestRecovery");
+ vm2.invoke(ConflationDUnitTest.class, "setClientServerObserverForBeforeInterestRecovery");
vm2.invoke(ConflationDUnitTest.class, "setAllCountersZero");
vm2.invoke(ConflationDUnitTest.class, "assertAllCountersZero");
vm2.invoke(ConflationDUnitTest.class, "registerInterest");
@@ -195,7 +195,7 @@ public class ConflationDUnitTest extends DistributedTestCase
vm2.invoke(ConflationDUnitTest.class,
"createClientCache2CommonWriterTest3", new Object[] {
getServerHostName(Host.getHost(0)), new Integer(PORT) });
- vm2.invoke(ConflationDUnitTest.class, "setBridgeObserverForBeforeInterestRecovery");
+ vm2.invoke(ConflationDUnitTest.class, "setClientServerObserverForBeforeInterestRecovery");
vm2.invoke(ConflationDUnitTest.class, "setAllCountersZero");
vm2.invoke(ConflationDUnitTest.class, "assertAllCountersZero");
vm2.invoke(ConflationDUnitTest.class, "registerInterest");
@@ -484,10 +484,10 @@ public class ConflationDUnitTest extends DistributedTestCase
* reset all counters to zero before interest recovery
*
*/
- public static void setBridgeObserverForBeforeInterestRecovery()
+ public static void setClientServerObserverForBeforeInterestRecovery()
{
PoolImpl.BEFORE_RECOVER_INTEREST_CALLBACK_FLAG = true;
- BridgeObserverHolder.setInstance(new BridgeObserverAdapter() {
+ ClientServerObserverHolder.setInstance(new ClientServerObserverAdapter() {
public void beforeInterestRecovery()
{
setAllCountersZero();
@@ -648,7 +648,7 @@ public class ConflationDUnitTest extends DistributedTestCase
RegionAttributes attrs = factory.create();
cache.createRegion(REGION_NAME1, attrs);
cache.createRegion(REGION_NAME2, attrs);
- BridgeServer server = cache.addBridgeServer();
+ CacheServer server = cache.addCacheServer();
int port = AvailablePort.getRandomAvailablePort(AvailablePort.SOCKET) ;
server.setPort(port);
server.setNotifyBySubscription(true);
@@ -793,8 +793,8 @@ public class ConflationDUnitTest extends DistributedTestCase
public static void getStatsOnServer()
{
Cache cache = GemFireCacheImpl.getInstance();
- Iterator itr = cache.getBridgeServers().iterator();
- BridgeServerImpl server = (BridgeServerImpl)itr.next();
+ Iterator itr = cache.getCacheServers().iterator();
+ CacheServerImpl server = (CacheServerImpl)itr.next();
Iterator iter_prox = server.getAcceptor().getCacheClientNotifier()
.getClientProxies().iterator();
int ccpCount=0;