You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@cloudstack.apache.org by ah...@apache.org on 2013/12/12 22:01:21 UTC

[21/48] All Checkstyle problems corrected

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/SnapshotTestWithFakeData.java
----------------------------------------------------------------------
diff --git a/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/SnapshotTestWithFakeData.java b/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/SnapshotTestWithFakeData.java
index 88fd54d..fcac783 100644
--- a/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/SnapshotTestWithFakeData.java
+++ b/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/SnapshotTestWithFakeData.java
@@ -43,7 +43,6 @@ import org.junit.Before;
 import org.junit.Test;
 import org.junit.runner.RunWith;
 import org.mockito.Matchers;
-import org.mockito.Mockito;
 import org.springframework.test.context.ContextConfiguration;
 import org.springframework.test.context.junit4.SpringJUnit4ClassRunner;
 
@@ -208,7 +207,7 @@ public class SnapshotTestWithFakeData {
         Snapshot.Type snapshotType = Snapshot.Type.RECURRING;
         SnapshotVO snapshotVO =
             new SnapshotVO(dcId, 2, 1, 1L, 1L, UUID.randomUUID().toString(), (short)snapshotType.ordinal(), snapshotType.name(), 100, Hypervisor.HypervisorType.XenServer);
-        return this.snapshotDao.persist(snapshotVO);
+        return snapshotDao.persist(snapshotVO);
     }
 
     private SnapshotVO createSnapshotInDb(Long volumeId) {
@@ -216,7 +215,7 @@ public class SnapshotTestWithFakeData {
         SnapshotVO snapshotVO =
             new SnapshotVO(dcId, 2, 1, volumeId, 1L, UUID.randomUUID().toString(), (short)snapshotType.ordinal(), snapshotType.name(), 100,
                 Hypervisor.HypervisorType.XenServer);
-        return this.snapshotDao.persist(snapshotVO);
+        return snapshotDao.persist(snapshotVO);
     }
 
     private VolumeInfo createVolume(Long templateId, DataStore store) {
@@ -246,8 +245,8 @@ public class SnapshotTestWithFakeData {
         pool.setPodId(podId);
         pool.setScope(ScopeType.CLUSTER);
         pool.setStorageProviderName(DataStoreProvider.DEFAULT_PRIMARY);
-        pool = this.primaryDataStoreDao.persist(pool);
-        DataStore store = this.dataStoreManager.getPrimaryDataStore(pool.getId());
+        pool = primaryDataStoreDao.persist(pool);
+        DataStore store = dataStoreManager.getPrimaryDataStore(pool.getId());
         return store;
     }
 

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/VolumeServiceTest.java
----------------------------------------------------------------------
diff --git a/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/VolumeServiceTest.java b/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/VolumeServiceTest.java
new file mode 100644
index 0000000..2f5b18b
--- /dev/null
+++ b/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/VolumeServiceTest.java
@@ -0,0 +1,470 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.cloudstack.storage.test;
+
+import java.net.URI;
+import java.net.URISyntaxException;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.UUID;
+import java.util.concurrent.ExecutionException;
+
+import javax.inject.Inject;
+
+import org.mockito.Matchers;
+import org.mockito.Mockito;
+import org.springframework.test.context.ContextConfiguration;
+import org.testng.Assert;
+import org.testng.annotations.Test;
+
+import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope;
+import org.apache.cloudstack.engine.subsystem.api.storage.DataObject;
+import org.apache.cloudstack.engine.subsystem.api.storage.DataStore;
+import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreLifeCycle;
+import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreProvider;
+import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreProviderManager;
+import org.apache.cloudstack.engine.subsystem.api.storage.EndPoint;
+import org.apache.cloudstack.engine.subsystem.api.storage.EndPointSelector;
+import org.apache.cloudstack.engine.subsystem.api.storage.TemplateDataFactory;
+import org.apache.cloudstack.engine.subsystem.api.storage.TemplateInfo;
+import org.apache.cloudstack.engine.subsystem.api.storage.TemplateService;
+import org.apache.cloudstack.engine.subsystem.api.storage.VolumeDataFactory;
+import org.apache.cloudstack.engine.subsystem.api.storage.VolumeInfo;
+import org.apache.cloudstack.engine.subsystem.api.storage.VolumeService;
+import org.apache.cloudstack.engine.subsystem.api.storage.VolumeService.VolumeApiResult;
+import org.apache.cloudstack.framework.async.AsyncCallFuture;
+import org.apache.cloudstack.storage.RemoteHostEndPoint;
+import org.apache.cloudstack.storage.datastore.db.ImageStoreDao;
+import org.apache.cloudstack.storage.datastore.db.ImageStoreVO;
+import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao;
+import org.apache.cloudstack.storage.datastore.db.StoragePoolVO;
+
+import com.cloud.agent.AgentManager;
+import com.cloud.dc.ClusterVO;
+import com.cloud.dc.DataCenter.NetworkType;
+import com.cloud.dc.DataCenterVO;
+import com.cloud.dc.HostPodVO;
+import com.cloud.dc.dao.ClusterDao;
+import com.cloud.dc.dao.DataCenterDao;
+import com.cloud.dc.dao.HostPodDao;
+import com.cloud.host.Host;
+import com.cloud.host.HostVO;
+import com.cloud.host.dao.HostDao;
+import com.cloud.hypervisor.Hypervisor.HypervisorType;
+import com.cloud.org.Cluster.ClusterType;
+import com.cloud.org.Managed.ManagedState;
+import com.cloud.resource.ResourceState;
+import com.cloud.storage.DataStoreRole;
+import com.cloud.storage.ScopeType;
+import com.cloud.storage.Storage;
+import com.cloud.storage.Storage.StoragePoolType;
+import com.cloud.storage.Storage.TemplateType;
+import com.cloud.storage.VMTemplateVO;
+import com.cloud.storage.Volume;
+import com.cloud.storage.VolumeVO;
+import com.cloud.storage.dao.VMTemplateDao;
+import com.cloud.storage.dao.VolumeDao;
+import com.cloud.utils.component.ComponentContext;
+
+@ContextConfiguration(locations = {"classpath:/storageContext.xml"})
+public class VolumeServiceTest extends CloudStackTestNGBase {
+    // @Inject
+    // ImageDataStoreProviderManager imageProviderMgr;
+    @Inject
+    TemplateService imageService;
+    @Inject
+    VolumeService volumeService;
+    @Inject
+    VMTemplateDao imageDataDao;
+    @Inject
+    VolumeDao volumeDao;
+    @Inject
+    HostDao hostDao;
+    @Inject
+    HostPodDao podDao;
+    @Inject
+    ClusterDao clusterDao;
+    @Inject
+    DataCenterDao dcDao;
+    @Inject
+    PrimaryDataStoreDao primaryStoreDao;
+    @Inject
+    DataStoreProviderManager dataStoreProviderMgr;
+    @Inject
+    AgentManager agentMgr;
+    @Inject
+    EndPointSelector selector;
+    @Inject
+    TemplateDataFactory imageDataFactory;
+    @Inject
+    VolumeDataFactory volumeFactory;
+    @Inject
+    ImageStoreDao imageStoreDao;
+    ImageStoreVO imageStore;
+    Long dcId;
+    Long clusterId;
+    Long podId;
+    HostVO host;
+    String primaryName = "my primary data store";
+    DataStore primaryStore;
+
+    @Test(priority = -1)
+    public void setUp() {
+        ComponentContext.initComponentsLifeCycle();
+
+        host = hostDao.findByGuid(this.getHostGuid());
+        if (host != null) {
+            dcId = host.getDataCenterId();
+            clusterId = host.getClusterId();
+            podId = host.getPodId();
+            return;
+        }
+        // create data center
+        DataCenterVO dc =
+            new DataCenterVO(UUID.randomUUID().toString(), "test", "8.8.8.8", null, "10.0.0.1", null, "10.0.0.1/24", null, null, NetworkType.Basic, null, null, true,
+                true, null, null);
+        dc = dcDao.persist(dc);
+        dcId = dc.getId();
+        // create pod
+
+        HostPodVO pod = new HostPodVO(UUID.randomUUID().toString(), dc.getId(), this.getHostGateway(), this.getHostCidr(), 8, "test");
+        pod = podDao.persist(pod);
+        podId = pod.getId();
+        // create xen cluster
+        ClusterVO cluster = new ClusterVO(dc.getId(), pod.getId(), "devcloud cluster");
+        cluster.setHypervisorType(HypervisorType.XenServer.toString());
+        cluster.setClusterType(ClusterType.CloudManaged);
+        cluster.setManagedState(ManagedState.Managed);
+        cluster = clusterDao.persist(cluster);
+        clusterId = cluster.getId();
+        // create xen host
+
+        host = new HostVO(this.getHostGuid());
+        host.setName("devcloud xen host");
+        host.setType(Host.Type.Routing);
+        host.setPrivateIpAddress(this.getHostIp());
+        host.setDataCenterId(dc.getId());
+        host.setVersion("6.0.1");
+        host.setAvailable(true);
+        host.setSetup(true);
+        host.setPodId(podId);
+        host.setLastPinged(0);
+        host.setResourceState(ResourceState.Enabled);
+        host.setHypervisorType(HypervisorType.XenServer);
+        host.setClusterId(cluster.getId());
+
+        host = hostDao.persist(host);
+
+        imageStore = new ImageStoreVO();
+        imageStore.setName("test");
+        imageStore.setDataCenterId(dcId);
+        imageStore.setProviderName("CloudStack ImageStore Provider");
+        imageStore.setRole(DataStoreRole.Image);
+        imageStore.setUrl(this.getSecondaryStorage());
+        imageStore.setUuid(UUID.randomUUID().toString());
+        imageStore = imageStoreDao.persist(imageStore);
+
+    }
+
+    @Override
+    protected void injectMockito() {
+        if (host == null) {
+            return;
+        }
+        List<HostVO> results = new ArrayList<HostVO>();
+        results.add(host);
+        Mockito.when(hostDao.listAll()).thenReturn(results);
+        Mockito.when(hostDao.findById(Matchers.anyLong())).thenReturn(host);
+        Mockito.when(hostDao.findHypervisorHostInCluster(Matchers.anyLong())).thenReturn(results);
+        List<EndPoint> eps = new ArrayList<EndPoint>();
+        eps.add(RemoteHostEndPoint.getHypervisorHostEndPoint(host.getId(), host.getPrivateIpAddress(), host.getPublicIpAddress()));
+        Mockito.when(selector.selectAll(Matchers.any(DataStore.class))).thenReturn(eps);
+        Mockito.when(selector.select(Matchers.any(DataObject.class))).thenReturn(eps.get(0));
+        Mockito.when(selector.select(Matchers.any(DataObject.class), Matchers.any(DataObject.class))).thenReturn(eps.get(0));
+    }
+
+    private VMTemplateVO createImageData() {
+        VMTemplateVO image = new VMTemplateVO();
+        image.setTemplateType(TemplateType.USER);
+        image.setUrl(this.getTemplateUrl());
+        image.setUniqueName(UUID.randomUUID().toString());
+        image.setName(UUID.randomUUID().toString());
+        image.setPublicTemplate(true);
+        image.setFeatured(true);
+        image.setRequiresHvm(true);
+        image.setBits(64);
+        image.setFormat(Storage.ImageFormat.VHD);
+        image.setEnablePassword(true);
+        image.setEnableSshKey(true);
+        image.setGuestOSId(1);
+        image.setBootable(true);
+        image.setPrepopulate(true);
+        image.setCrossZones(true);
+        image.setExtractable(true);
+
+        // image.setImageDataStoreId(storeId);
+        image = imageDataDao.persist(image);
+
+        return image;
+    }
+
+    private TemplateInfo createTemplate() {
+        try {
+            DataStore store = createImageStore();
+            VMTemplateVO image = createImageData();
+            TemplateInfo template = imageDataFactory.getTemplate(image.getId(), store);
+            // AsyncCallFuture<TemplateApiResult> future =
+            // imageService.createTemplateAsync(template, store);
+            // future.get();
+            template = imageDataFactory.getTemplate(image.getId(), store);
+            /*
+             * imageProviderMgr.configure("image Provider", new HashMap<String,
+             * Object>()); VMTemplateVO image = createImageData();
+             * ImageDataStoreProvider defaultProvider =
+             * imageProviderMgr.getProvider("DefaultProvider");
+             * ImageDataStoreLifeCycle lifeCycle =
+             * defaultProvider.getLifeCycle(); ImageDataStore store =
+             * lifeCycle.registerDataStore("defaultHttpStore", new
+             * HashMap<String, String>());
+             * imageService.registerTemplate(image.getId(),
+             * store.getImageDataStoreId()); TemplateEntity te =
+             * imageService.getTemplateEntity(image.getId()); return te;
+             */
+            return template;
+        } catch (Exception e) {
+            Assert.fail("failed", e);
+            return null;
+        }
+    }
+
+    // @Test
+    public void createTemplateTest() {
+        createTemplate();
+    }
+
+    @Test
+    public void testCreatePrimaryStorage() {
+        DataStoreProvider provider = dataStoreProviderMgr.getDataStoreProvider("sample primary data store provider");
+        Map<String, Object> params = new HashMap<String, Object>();
+        URI uri = null;
+        try {
+            uri = new URI(this.getPrimaryStorageUrl());
+        } catch (URISyntaxException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        }
+        params.put("url", this.getPrimaryStorageUrl());
+        params.put("server", uri.getHost());
+        params.put("path", uri.getPath());
+        params.put("protocol", StoragePoolType.NetworkFilesystem);
+        params.put("dcId", dcId.toString());
+        params.put("clusterId", clusterId.toString());
+        params.put("name", this.primaryName);
+        params.put("port", "1");
+        params.put("roles", DataStoreRole.Primary.toString());
+        params.put("uuid", UUID.nameUUIDFromBytes(this.getPrimaryStorageUrl().getBytes()).toString());
+        params.put("providerName", String.valueOf(provider.getName()));
+
+        DataStoreLifeCycle lifeCycle = provider.getDataStoreLifeCycle();
+        this.primaryStore = lifeCycle.initialize(params);
+        ClusterScope scope = new ClusterScope(clusterId, podId, dcId);
+        lifeCycle.attachCluster(this.primaryStore, scope);
+    }
+
+    private DataStore createImageStore() {
+        DataStoreProvider provider = dataStoreProviderMgr.getDataStoreProvider("sample image data store provider");
+        Map<String, Object> params = new HashMap<String, Object>();
+        String name = UUID.randomUUID().toString();
+        params.put("name", name);
+        params.put("uuid", name);
+        params.put("protocol", "http");
+        params.put("scope", ScopeType.GLOBAL.toString());
+        params.put("providerName", name);
+        DataStoreLifeCycle lifeCycle = provider.getDataStoreLifeCycle();
+        DataStore store = lifeCycle.initialize(params);
+        return store;
+    }
+
+    // @Test
+    public void testcreateImageStore() {
+        createImageStore();
+    }
+
+    public DataStore createPrimaryDataStore() {
+        try {
+            DataStoreProvider provider = dataStoreProviderMgr.getDataStoreProvider("sample primary data store provider");
+            Map<String, Object> params = new HashMap<String, Object>();
+            URI uri = new URI(this.getPrimaryStorageUrl());
+            params.put("url", this.getPrimaryStorageUrl());
+            params.put("server", uri.getHost());
+            params.put("path", uri.getPath());
+            params.put("protocol", Storage.StoragePoolType.NetworkFilesystem);
+            params.put("dcId", dcId.toString());
+            params.put("clusterId", clusterId.toString());
+            params.put("name", this.primaryName);
+            params.put("port", "1");
+            params.put("roles", DataStoreRole.Primary.toString());
+            params.put("uuid", UUID.nameUUIDFromBytes(this.getPrimaryStorageUrl().getBytes()).toString());
+            params.put("providerName", String.valueOf(provider.getName()));
+
+            DataStoreLifeCycle lifeCycle = provider.getDataStoreLifeCycle();
+            DataStore store = lifeCycle.initialize(params);
+            ClusterScope scope = new ClusterScope(clusterId, podId, dcId);
+            lifeCycle.attachCluster(store, scope);
+
+            /*
+             * PrimaryDataStoreProvider provider =
+             * primaryDataStoreProviderMgr.getDataStoreProvider
+             * ("sample primary data store provider");
+             * primaryDataStoreProviderMgr.configure("primary data store mgr",
+             * new HashMap<String, Object>());
+             *
+             * List<PrimaryDataStoreVO> ds =
+             * primaryStoreDao.findPoolByName(this.primaryName); if (ds.size()
+             * >= 1) { PrimaryDataStoreVO store = ds.get(0); if
+             * (store.getRemoved() == null) { return
+             * provider.getDataStore(store.getId()); } }
+             *
+             *
+             * Map<String, String> params = new HashMap<String, String>();
+             * params.put("url", this.getPrimaryStorageUrl());
+             * params.put("dcId", dcId.toString()); params.put("clusterId",
+             * clusterId.toString()); params.put("name", this.primaryName);
+             * PrimaryDataStoreInfo primaryDataStoreInfo =
+             * provider.registerDataStore(params); PrimaryDataStoreLifeCycle lc
+             * = primaryDataStoreInfo.getLifeCycle(); ClusterScope scope = new
+             * ClusterScope(clusterId, podId, dcId); lc.attachCluster(scope);
+             * return primaryDataStoreInfo;
+             */
+            return store;
+        } catch (Exception e) {
+            return null;
+        }
+    }
+
+    private VolumeVO createVolume(Long templateId, long dataStoreId) {
+        VolumeVO volume = new VolumeVO(Volume.Type.DATADISK, UUID.randomUUID().toString(), this.dcId, 1L, 1L, 1L, 1000, 0L, 0L, "");
+        volume.setPoolId(dataStoreId);
+        volume = volumeDao.persist(volume);
+        return volume;
+    }
+
+    @Test(priority = 2)
+    public void createVolumeFromTemplate() {
+        DataStore primaryStore = this.primaryStore;
+        TemplateInfo te = createTemplate();
+        VolumeVO volume = createVolume(te.getId(), primaryStore.getId());
+        VolumeInfo vol = volumeFactory.getVolume(volume.getId(), primaryStore);
+        // ve.createVolumeFromTemplate(primaryStore.getId(), new VHD(), te);
+        AsyncCallFuture<VolumeApiResult> future = volumeService.createVolumeFromTemplateAsync(vol, primaryStore.getId(), te);
+        try {
+            future.get();
+        } catch (InterruptedException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        } catch (ExecutionException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        }
+    }
+
+    // @Test(priority=3)
+    public void createDataDisk() {
+        DataStore primaryStore = this.primaryStore;
+        VolumeVO volume = createVolume(null, primaryStore.getId());
+        VolumeInfo vol = volumeFactory.getVolume(volume.getId(), primaryStore);
+        AsyncCallFuture<VolumeApiResult> future = volumeService.createVolumeAsync(vol, primaryStore);
+        try {
+            future.get();
+        } catch (InterruptedException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        } catch (ExecutionException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        }
+    }
+
+    // @Test(priority=3)
+    public void createAndDeleteDataDisk() {
+        DataStore primaryStore = this.primaryStore;
+        VolumeVO volume = createVolume(null, primaryStore.getId());
+        VolumeInfo vol = volumeFactory.getVolume(volume.getId(), primaryStore);
+        AsyncCallFuture<VolumeApiResult> future = volumeService.createVolumeAsync(vol, primaryStore);
+        try {
+            future.get();
+        } catch (InterruptedException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        } catch (ExecutionException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        }
+
+        // delete the volume
+        vol = volumeFactory.getVolume(volume.getId(), primaryStore);
+        future = volumeService.expungeVolumeAsync(vol);
+        try {
+            future.get();
+        } catch (InterruptedException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        } catch (ExecutionException e) {
+            // TODO Auto-generated catch block
+            e.printStackTrace();
+        }
+    }
+
+    // @Test(priority=3)
+    public void tearDown() {
+        List<StoragePoolVO> ds = primaryStoreDao.findPoolByName(this.primaryName);
+        for (int i = 0; i < ds.size(); i++) {
+            StoragePoolVO store = ds.get(i);
+            store.setUuid(null);
+            primaryStoreDao.remove(ds.get(i).getId());
+            primaryStoreDao.expunge(ds.get(i).getId());
+        }
+    }
+
+    // @Test
+    // @Test
+    public void test1() {
+        /*
+         * System.out.println(VolumeTypeHelper.getType("Root"));
+         * System.out.println(VolumeDiskTypeHelper.getDiskType("vmdk"));
+         * System.out.println(ImageFormatHelper.getFormat("ova"));
+         * AssertJUnit.assertFalse(new VMDK().equals(new VHD())); VMDK vmdk =
+         * new VMDK(); AssertJUnit.assertTrue(vmdk.equals(vmdk)); VMDK newvmdk =
+         * new VMDK(); AssertJUnit.assertTrue(vmdk.equals(newvmdk));
+         *
+         * ImageFormat ova = new OVA(); ImageFormat iso = new ISO();
+         * AssertJUnit.assertTrue(ova.equals(new OVA()));
+         * AssertJUnit.assertFalse(ova.equals(iso));
+         * AssertJUnit.assertTrue(ImageFormatHelper.getFormat("test").equals(new
+         * Unknown()));
+         *
+         * VolumeDiskType qcow2 = new QCOW2(); ImageFormat qcow2format = new
+         * org.apache.cloudstack.storage.image.format.QCOW2();
+         * AssertJUnit.assertFalse(qcow2.equals(qcow2format));
+         */
+    }
+
+}

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/volumeServiceTest.java
----------------------------------------------------------------------
diff --git a/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/volumeServiceTest.java b/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/volumeServiceTest.java
deleted file mode 100644
index 167dd52..0000000
--- a/engine/storage/integration-test/test/org/apache/cloudstack/storage/test/volumeServiceTest.java
+++ /dev/null
@@ -1,470 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *   http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied.  See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.cloudstack.storage.test;
-
-import java.net.URI;
-import java.net.URISyntaxException;
-import java.util.ArrayList;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.UUID;
-import java.util.concurrent.ExecutionException;
-
-import javax.inject.Inject;
-
-import org.mockito.Matchers;
-import org.mockito.Mockito;
-import org.springframework.test.context.ContextConfiguration;
-import org.testng.Assert;
-import org.testng.annotations.Test;
-
-import org.apache.cloudstack.engine.subsystem.api.storage.ClusterScope;
-import org.apache.cloudstack.engine.subsystem.api.storage.DataObject;
-import org.apache.cloudstack.engine.subsystem.api.storage.DataStore;
-import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreLifeCycle;
-import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreProvider;
-import org.apache.cloudstack.engine.subsystem.api.storage.DataStoreProviderManager;
-import org.apache.cloudstack.engine.subsystem.api.storage.EndPoint;
-import org.apache.cloudstack.engine.subsystem.api.storage.EndPointSelector;
-import org.apache.cloudstack.engine.subsystem.api.storage.TemplateDataFactory;
-import org.apache.cloudstack.engine.subsystem.api.storage.TemplateInfo;
-import org.apache.cloudstack.engine.subsystem.api.storage.TemplateService;
-import org.apache.cloudstack.engine.subsystem.api.storage.VolumeDataFactory;
-import org.apache.cloudstack.engine.subsystem.api.storage.VolumeInfo;
-import org.apache.cloudstack.engine.subsystem.api.storage.VolumeService;
-import org.apache.cloudstack.engine.subsystem.api.storage.VolumeService.VolumeApiResult;
-import org.apache.cloudstack.framework.async.AsyncCallFuture;
-import org.apache.cloudstack.storage.RemoteHostEndPoint;
-import org.apache.cloudstack.storage.datastore.db.ImageStoreDao;
-import org.apache.cloudstack.storage.datastore.db.ImageStoreVO;
-import org.apache.cloudstack.storage.datastore.db.PrimaryDataStoreDao;
-import org.apache.cloudstack.storage.datastore.db.StoragePoolVO;
-
-import com.cloud.agent.AgentManager;
-import com.cloud.dc.ClusterVO;
-import com.cloud.dc.DataCenter.NetworkType;
-import com.cloud.dc.DataCenterVO;
-import com.cloud.dc.HostPodVO;
-import com.cloud.dc.dao.ClusterDao;
-import com.cloud.dc.dao.DataCenterDao;
-import com.cloud.dc.dao.HostPodDao;
-import com.cloud.host.Host;
-import com.cloud.host.HostVO;
-import com.cloud.host.dao.HostDao;
-import com.cloud.hypervisor.Hypervisor.HypervisorType;
-import com.cloud.org.Cluster.ClusterType;
-import com.cloud.org.Managed.ManagedState;
-import com.cloud.resource.ResourceState;
-import com.cloud.storage.DataStoreRole;
-import com.cloud.storage.ScopeType;
-import com.cloud.storage.Storage;
-import com.cloud.storage.Storage.StoragePoolType;
-import com.cloud.storage.Storage.TemplateType;
-import com.cloud.storage.VMTemplateVO;
-import com.cloud.storage.Volume;
-import com.cloud.storage.VolumeVO;
-import com.cloud.storage.dao.VMTemplateDao;
-import com.cloud.storage.dao.VolumeDao;
-import com.cloud.utils.component.ComponentContext;
-
-@ContextConfiguration(locations = {"classpath:/storageContext.xml"})
-public class volumeServiceTest extends CloudStackTestNGBase {
-    // @Inject
-    // ImageDataStoreProviderManager imageProviderMgr;
-    @Inject
-    TemplateService imageService;
-    @Inject
-    VolumeService volumeService;
-    @Inject
-    VMTemplateDao imageDataDao;
-    @Inject
-    VolumeDao volumeDao;
-    @Inject
-    HostDao hostDao;
-    @Inject
-    HostPodDao podDao;
-    @Inject
-    ClusterDao clusterDao;
-    @Inject
-    DataCenterDao dcDao;
-    @Inject
-    PrimaryDataStoreDao primaryStoreDao;
-    @Inject
-    DataStoreProviderManager dataStoreProviderMgr;
-    @Inject
-    AgentManager agentMgr;
-    @Inject
-    EndPointSelector selector;
-    @Inject
-    TemplateDataFactory imageDataFactory;
-    @Inject
-    VolumeDataFactory volumeFactory;
-    @Inject
-    ImageStoreDao imageStoreDao;
-    ImageStoreVO imageStore;
-    Long dcId;
-    Long clusterId;
-    Long podId;
-    HostVO host;
-    String primaryName = "my primary data store";
-    DataStore primaryStore;
-
-    @Test(priority = -1)
-    public void setUp() {
-        ComponentContext.initComponentsLifeCycle();
-
-        host = hostDao.findByGuid(this.getHostGuid());
-        if (host != null) {
-            dcId = host.getDataCenterId();
-            clusterId = host.getClusterId();
-            podId = host.getPodId();
-            return;
-        }
-        // create data center
-        DataCenterVO dc =
-            new DataCenterVO(UUID.randomUUID().toString(), "test", "8.8.8.8", null, "10.0.0.1", null, "10.0.0.1/24", null, null, NetworkType.Basic, null, null, true,
-                true, null, null);
-        dc = dcDao.persist(dc);
-        dcId = dc.getId();
-        // create pod
-
-        HostPodVO pod = new HostPodVO(UUID.randomUUID().toString(), dc.getId(), this.getHostGateway(), this.getHostCidr(), 8, "test");
-        pod = podDao.persist(pod);
-        podId = pod.getId();
-        // create xen cluster
-        ClusterVO cluster = new ClusterVO(dc.getId(), pod.getId(), "devcloud cluster");
-        cluster.setHypervisorType(HypervisorType.XenServer.toString());
-        cluster.setClusterType(ClusterType.CloudManaged);
-        cluster.setManagedState(ManagedState.Managed);
-        cluster = clusterDao.persist(cluster);
-        clusterId = cluster.getId();
-        // create xen host
-
-        host = new HostVO(this.getHostGuid());
-        host.setName("devcloud xen host");
-        host.setType(Host.Type.Routing);
-        host.setPrivateIpAddress(this.getHostIp());
-        host.setDataCenterId(dc.getId());
-        host.setVersion("6.0.1");
-        host.setAvailable(true);
-        host.setSetup(true);
-        host.setPodId(podId);
-        host.setLastPinged(0);
-        host.setResourceState(ResourceState.Enabled);
-        host.setHypervisorType(HypervisorType.XenServer);
-        host.setClusterId(cluster.getId());
-
-        host = hostDao.persist(host);
-
-        imageStore = new ImageStoreVO();
-        imageStore.setName("test");
-        imageStore.setDataCenterId(dcId);
-        imageStore.setProviderName("CloudStack ImageStore Provider");
-        imageStore.setRole(DataStoreRole.Image);
-        imageStore.setUrl(this.getSecondaryStorage());
-        imageStore.setUuid(UUID.randomUUID().toString());
-        imageStore = imageStoreDao.persist(imageStore);
-
-    }
-
-    @Override
-    protected void injectMockito() {
-        if (host == null) {
-            return;
-        }
-        List<HostVO> results = new ArrayList<HostVO>();
-        results.add(host);
-        Mockito.when(hostDao.listAll()).thenReturn(results);
-        Mockito.when(hostDao.findById(Matchers.anyLong())).thenReturn(host);
-        Mockito.when(hostDao.findHypervisorHostInCluster(Matchers.anyLong())).thenReturn(results);
-        List<EndPoint> eps = new ArrayList<EndPoint>();
-        eps.add(RemoteHostEndPoint.getHypervisorHostEndPoint(host.getId(), host.getPrivateIpAddress(), host.getPublicIpAddress()));
-        Mockito.when(selector.selectAll(Matchers.any(DataStore.class))).thenReturn(eps);
-        Mockito.when(selector.select(Matchers.any(DataObject.class))).thenReturn(eps.get(0));
-        Mockito.when(selector.select(Matchers.any(DataObject.class), Matchers.any(DataObject.class))).thenReturn(eps.get(0));
-    }
-
-    private VMTemplateVO createImageData() {
-        VMTemplateVO image = new VMTemplateVO();
-        image.setTemplateType(TemplateType.USER);
-        image.setUrl(this.getTemplateUrl());
-        image.setUniqueName(UUID.randomUUID().toString());
-        image.setName(UUID.randomUUID().toString());
-        image.setPublicTemplate(true);
-        image.setFeatured(true);
-        image.setRequiresHvm(true);
-        image.setBits(64);
-        image.setFormat(Storage.ImageFormat.VHD);
-        image.setEnablePassword(true);
-        image.setEnableSshKey(true);
-        image.setGuestOSId(1);
-        image.setBootable(true);
-        image.setPrepopulate(true);
-        image.setCrossZones(true);
-        image.setExtractable(true);
-
-        // image.setImageDataStoreId(storeId);
-        image = imageDataDao.persist(image);
-
-        return image;
-    }
-
-    private TemplateInfo createTemplate() {
-        try {
-            DataStore store = createImageStore();
-            VMTemplateVO image = createImageData();
-            TemplateInfo template = imageDataFactory.getTemplate(image.getId(), store);
-            // AsyncCallFuture<TemplateApiResult> future =
-            // imageService.createTemplateAsync(template, store);
-            // future.get();
-            template = imageDataFactory.getTemplate(image.getId(), store);
-            /*
-             * imageProviderMgr.configure("image Provider", new HashMap<String,
-             * Object>()); VMTemplateVO image = createImageData();
-             * ImageDataStoreProvider defaultProvider =
-             * imageProviderMgr.getProvider("DefaultProvider");
-             * ImageDataStoreLifeCycle lifeCycle =
-             * defaultProvider.getLifeCycle(); ImageDataStore store =
-             * lifeCycle.registerDataStore("defaultHttpStore", new
-             * HashMap<String, String>());
-             * imageService.registerTemplate(image.getId(),
-             * store.getImageDataStoreId()); TemplateEntity te =
-             * imageService.getTemplateEntity(image.getId()); return te;
-             */
-            return template;
-        } catch (Exception e) {
-            Assert.fail("failed", e);
-            return null;
-        }
-    }
-
-    // @Test
-    public void createTemplateTest() {
-        createTemplate();
-    }
-
-    @Test
-    public void testCreatePrimaryStorage() {
-        DataStoreProvider provider = dataStoreProviderMgr.getDataStoreProvider("sample primary data store provider");
-        Map<String, Object> params = new HashMap<String, Object>();
-        URI uri = null;
-        try {
-            uri = new URI(this.getPrimaryStorageUrl());
-        } catch (URISyntaxException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        }
-        params.put("url", this.getPrimaryStorageUrl());
-        params.put("server", uri.getHost());
-        params.put("path", uri.getPath());
-        params.put("protocol", StoragePoolType.NetworkFilesystem);
-        params.put("dcId", dcId.toString());
-        params.put("clusterId", clusterId.toString());
-        params.put("name", this.primaryName);
-        params.put("port", "1");
-        params.put("roles", DataStoreRole.Primary.toString());
-        params.put("uuid", UUID.nameUUIDFromBytes(this.getPrimaryStorageUrl().getBytes()).toString());
-        params.put("providerName", String.valueOf(provider.getName()));
-
-        DataStoreLifeCycle lifeCycle = provider.getDataStoreLifeCycle();
-        this.primaryStore = lifeCycle.initialize(params);
-        ClusterScope scope = new ClusterScope(clusterId, podId, dcId);
-        lifeCycle.attachCluster(this.primaryStore, scope);
-    }
-
-    private DataStore createImageStore() {
-        DataStoreProvider provider = dataStoreProviderMgr.getDataStoreProvider("sample image data store provider");
-        Map<String, Object> params = new HashMap<String, Object>();
-        String name = UUID.randomUUID().toString();
-        params.put("name", name);
-        params.put("uuid", name);
-        params.put("protocol", "http");
-        params.put("scope", ScopeType.GLOBAL.toString());
-        params.put("providerName", name);
-        DataStoreLifeCycle lifeCycle = provider.getDataStoreLifeCycle();
-        DataStore store = lifeCycle.initialize(params);
-        return store;
-    }
-
-    // @Test
-    public void testcreateImageStore() {
-        createImageStore();
-    }
-
-    public DataStore createPrimaryDataStore() {
-        try {
-            DataStoreProvider provider = dataStoreProviderMgr.getDataStoreProvider("sample primary data store provider");
-            Map<String, Object> params = new HashMap<String, Object>();
-            URI uri = new URI(this.getPrimaryStorageUrl());
-            params.put("url", this.getPrimaryStorageUrl());
-            params.put("server", uri.getHost());
-            params.put("path", uri.getPath());
-            params.put("protocol", Storage.StoragePoolType.NetworkFilesystem);
-            params.put("dcId", dcId.toString());
-            params.put("clusterId", clusterId.toString());
-            params.put("name", this.primaryName);
-            params.put("port", "1");
-            params.put("roles", DataStoreRole.Primary.toString());
-            params.put("uuid", UUID.nameUUIDFromBytes(this.getPrimaryStorageUrl().getBytes()).toString());
-            params.put("providerName", String.valueOf(provider.getName()));
-
-            DataStoreLifeCycle lifeCycle = provider.getDataStoreLifeCycle();
-            DataStore store = lifeCycle.initialize(params);
-            ClusterScope scope = new ClusterScope(clusterId, podId, dcId);
-            lifeCycle.attachCluster(store, scope);
-
-            /*
-             * PrimaryDataStoreProvider provider =
-             * primaryDataStoreProviderMgr.getDataStoreProvider
-             * ("sample primary data store provider");
-             * primaryDataStoreProviderMgr.configure("primary data store mgr",
-             * new HashMap<String, Object>());
-             *
-             * List<PrimaryDataStoreVO> ds =
-             * primaryStoreDao.findPoolByName(this.primaryName); if (ds.size()
-             * >= 1) { PrimaryDataStoreVO store = ds.get(0); if
-             * (store.getRemoved() == null) { return
-             * provider.getDataStore(store.getId()); } }
-             *
-             *
-             * Map<String, String> params = new HashMap<String, String>();
-             * params.put("url", this.getPrimaryStorageUrl());
-             * params.put("dcId", dcId.toString()); params.put("clusterId",
-             * clusterId.toString()); params.put("name", this.primaryName);
-             * PrimaryDataStoreInfo primaryDataStoreInfo =
-             * provider.registerDataStore(params); PrimaryDataStoreLifeCycle lc
-             * = primaryDataStoreInfo.getLifeCycle(); ClusterScope scope = new
-             * ClusterScope(clusterId, podId, dcId); lc.attachCluster(scope);
-             * return primaryDataStoreInfo;
-             */
-            return store;
-        } catch (Exception e) {
-            return null;
-        }
-    }
-
-    private VolumeVO createVolume(Long templateId, long dataStoreId) {
-        VolumeVO volume = new VolumeVO(Volume.Type.DATADISK, UUID.randomUUID().toString(), this.dcId, 1L, 1L, 1L, 1000, 0L, 0L, "");
-        volume.setPoolId(dataStoreId);
-        volume = volumeDao.persist(volume);
-        return volume;
-    }
-
-    @Test(priority = 2)
-    public void createVolumeFromTemplate() {
-        DataStore primaryStore = this.primaryStore;
-        TemplateInfo te = createTemplate();
-        VolumeVO volume = createVolume(te.getId(), primaryStore.getId());
-        VolumeInfo vol = volumeFactory.getVolume(volume.getId(), primaryStore);
-        // ve.createVolumeFromTemplate(primaryStore.getId(), new VHD(), te);
-        AsyncCallFuture<VolumeApiResult> future = volumeService.createVolumeFromTemplateAsync(vol, primaryStore.getId(), te);
-        try {
-            future.get();
-        } catch (InterruptedException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        } catch (ExecutionException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        }
-    }
-
-    // @Test(priority=3)
-    public void createDataDisk() {
-        DataStore primaryStore = this.primaryStore;
-        VolumeVO volume = createVolume(null, primaryStore.getId());
-        VolumeInfo vol = volumeFactory.getVolume(volume.getId(), primaryStore);
-        AsyncCallFuture<VolumeApiResult> future = volumeService.createVolumeAsync(vol, primaryStore);
-        try {
-            future.get();
-        } catch (InterruptedException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        } catch (ExecutionException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        }
-    }
-
-    // @Test(priority=3)
-    public void createAndDeleteDataDisk() {
-        DataStore primaryStore = this.primaryStore;
-        VolumeVO volume = createVolume(null, primaryStore.getId());
-        VolumeInfo vol = volumeFactory.getVolume(volume.getId(), primaryStore);
-        AsyncCallFuture<VolumeApiResult> future = volumeService.createVolumeAsync(vol, primaryStore);
-        try {
-            future.get();
-        } catch (InterruptedException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        } catch (ExecutionException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        }
-
-        // delete the volume
-        vol = volumeFactory.getVolume(volume.getId(), primaryStore);
-        future = volumeService.expungeVolumeAsync(vol);
-        try {
-            future.get();
-        } catch (InterruptedException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        } catch (ExecutionException e) {
-            // TODO Auto-generated catch block
-            e.printStackTrace();
-        }
-    }
-
-    // @Test(priority=3)
-    public void tearDown() {
-        List<StoragePoolVO> ds = primaryStoreDao.findPoolByName(this.primaryName);
-        for (int i = 0; i < ds.size(); i++) {
-            StoragePoolVO store = ds.get(i);
-            store.setUuid(null);
-            primaryStoreDao.remove(ds.get(i).getId());
-            primaryStoreDao.expunge(ds.get(i).getId());
-        }
-    }
-
-    // @Test
-    // @Test
-    public void test1() {
-        /*
-         * System.out.println(VolumeTypeHelper.getType("Root"));
-         * System.out.println(VolumeDiskTypeHelper.getDiskType("vmdk"));
-         * System.out.println(ImageFormatHelper.getFormat("ova"));
-         * AssertJUnit.assertFalse(new VMDK().equals(new VHD())); VMDK vmdk =
-         * new VMDK(); AssertJUnit.assertTrue(vmdk.equals(vmdk)); VMDK newvmdk =
-         * new VMDK(); AssertJUnit.assertTrue(vmdk.equals(newvmdk));
-         *
-         * ImageFormat ova = new OVA(); ImageFormat iso = new ISO();
-         * AssertJUnit.assertTrue(ova.equals(new OVA()));
-         * AssertJUnit.assertFalse(ova.equals(iso));
-         * AssertJUnit.assertTrue(ImageFormatHelper.getFormat("test").equals(new
-         * Unknown()));
-         *
-         * VolumeDiskType qcow2 = new QCOW2(); ImageFormat qcow2format = new
-         * org.apache.cloudstack.storage.image.format.QCOW2();
-         * AssertJUnit.assertFalse(qcow2.equals(qcow2format));
-         */
-    }
-
-}

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/engine/storage/integration-test/test/resources/StorageAllocatorTestContext.xml
----------------------------------------------------------------------
diff --git a/engine/storage/integration-test/test/resources/StorageAllocatorTestContext.xml b/engine/storage/integration-test/test/resources/StorageAllocatorTestContext.xml
index 1f5aa58..72e5c84 100644
--- a/engine/storage/integration-test/test/resources/StorageAllocatorTestContext.xml
+++ b/engine/storage/integration-test/test/resources/StorageAllocatorTestContext.xml
@@ -1,17 +1,16 @@
 <?xml version="1.0" encoding="UTF-8"?>
 <!-- Licensed to the Apache Software Foundation (ASF) under one or more contributor 
   license agreements. See the NOTICE file distributed with this work for additional 
-  information regarding copyright ownership. The ASF licenses this file to 
-  you under the Apache License, Version 2.0 (the "License"); you may not use 
-  this file except in compliance with the License. You may obtain a copy of 
-  the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required 
-  by applicable law or agreed to in writing, software distributed under the 
-  License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS 
-  OF ANY KIND, either express or implied. See the License for the specific 
-  language governing permissions and limitations under the License. -->
-<beans xmlns="http://www.springframework.org/schema/beans"
-  xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:context="http://www.springframework.org/schema/context"
-  xmlns:tx="http://www.springframework.org/schema/tx" xmlns:aop="http://www.springframework.org/schema/aop"
+  information regarding copyright ownership. The ASF licenses this file to you under 
+  the Apache License, Version 2.0 (the "License"); you may not use this file except 
+  in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 
+  Unless required by applicable law or agreed to in writing, software distributed under 
+  the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS 
+  OF ANY KIND, either express or implied. See the License for the specific language 
+  governing permissions and limitations under the License. -->
+<beans xmlns="http://www.springframework.org/schema/beans" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
+  xmlns:context="http://www.springframework.org/schema/context" xmlns:tx="http://www.springframework.org/schema/tx"
+  xmlns:aop="http://www.springframework.org/schema/aop"
   xsi:schemaLocation="http://www.springframework.org/schema/beans
                       http://www.springframework.org/schema/beans/spring-beans-3.0.xsd
                       http://www.springframework.org/schema/tx 
@@ -22,10 +21,10 @@
                       http://www.springframework.org/schema/context/spring-context-3.0.xsd">
 
   <context:annotation-config />
-	<context:component-scan base-package="org.apache.cloudstack.storage" />
+  <context:component-scan base-package="org.apache.cloudstack.storage" />
 
 
-	<!-- @DB support -->
+  <!-- @DB support -->
   <aop:config proxy-target-class="true">
     <aop:aspect id="dbContextBuilder" ref="transactionContextBuilder">
       <aop:pointcut id="captureAnyMethod" expression="execution(* *(..))" />
@@ -36,10 +35,11 @@
   </aop:config>
 
   <bean id="transactionContextBuilder" class="com.cloud.utils.db.TransactionContextBuilder" />
-  <bean id="componentContext" class="com.cloud.utils.component.ComponentContext"/>
+  <bean id="componentContext" class="com.cloud.utils.component.ComponentContext" />
   <bean class="org.apache.cloudstack.storage.test.ChildTestConfiguration" />
 
-  <bean class="org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor">
+  <bean
+    class="org.springframework.beans.factory.annotation.AutowiredAnnotationBeanPostProcessor">
     <property name="requiredParameterValue" value="false" />
   </bean>
 </beans>

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/engine/storage/snapshot/src/org/apache/cloudstack/storage/vmsnapshot/DefaultVMSnapshotStrategy.java
----------------------------------------------------------------------
diff --git a/engine/storage/snapshot/src/org/apache/cloudstack/storage/vmsnapshot/DefaultVMSnapshotStrategy.java b/engine/storage/snapshot/src/org/apache/cloudstack/storage/vmsnapshot/DefaultVMSnapshotStrategy.java
index 42bef7e..e86bd23 100644
--- a/engine/storage/snapshot/src/org/apache/cloudstack/storage/vmsnapshot/DefaultVMSnapshotStrategy.java
+++ b/engine/storage/snapshot/src/org/apache/cloudstack/storage/vmsnapshot/DefaultVMSnapshotStrategy.java
@@ -236,9 +236,9 @@ public class DefaultVMSnapshotStrategy extends ManagerBase implements VMSnapshot
         }
     }
 
-    protected void finalizeDelete(VMSnapshotVO vmSnapshot, List<VolumeObjectTO> VolumeTOs) {
+    protected void finalizeDelete(VMSnapshotVO vmSnapshot, List<VolumeObjectTO> volumeTOs) {
         // update volumes path
-        updateVolumePath(VolumeTOs);
+        updateVolumePath(volumeTOs);
 
         // update children's parent snapshots
         List<VMSnapshotVO> children = vmSnapshotDao.listByParent(vmSnapshot.getId());
@@ -258,9 +258,9 @@ public class DefaultVMSnapshotStrategy extends ManagerBase implements VMSnapshot
         vmSnapshotDao.persist(vmSnapshot);
     }
 
-    protected void finalizeCreate(VMSnapshotVO vmSnapshot, List<VolumeObjectTO> VolumeTOs) {
+    protected void finalizeCreate(VMSnapshotVO vmSnapshot, List<VolumeObjectTO> volumeTOs) {
         // update volumes path
-        updateVolumePath(VolumeTOs);
+        updateVolumePath(volumeTOs);
 
         vmSnapshot.setCurrent(true);
 

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/engine/storage/volume/src/org/apache/cloudstack/storage/volume/VolumeServiceImpl.java
----------------------------------------------------------------------
diff --git a/engine/storage/volume/src/org/apache/cloudstack/storage/volume/VolumeServiceImpl.java b/engine/storage/volume/src/org/apache/cloudstack/storage/volume/VolumeServiceImpl.java
index 37e5bb4..c9d7fab 100644
--- a/engine/storage/volume/src/org/apache/cloudstack/storage/volume/VolumeServiceImpl.java
+++ b/engine/storage/volume/src/org/apache/cloudstack/storage/volume/VolumeServiceImpl.java
@@ -1204,7 +1204,8 @@ public class VolumeServiceImpl implements VolumeService {
                                             com.cloud.configuration.Resource.ResourceType.secondary_storage, volInfo.getSize() - volInfo.getPhysicalSize());
                                     } catch (ResourceAllocationException e) {
                                         s_logger.warn(e.getMessage());
-                                        _alertMgr.sendAlert(AlertManager.AlertType.ALERT_TYPE_RESOURCE_LIMIT_EXCEEDED, volume.getDataCenterId(), volume.getPodId(), e.getMessage(),
+                                        _alertMgr.sendAlert(AlertManager.AlertType.ALERT_TYPE_RESOURCE_LIMIT_EXCEEDED, volume.getDataCenterId(), volume.getPodId(),
+                                            e.getMessage(),
                                             e.getMessage());
                                     } finally {
                                         _resourceLimitMgr.recalculateResourceCount(volume.getAccountId(), volume.getDomainId(),

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/cluster/src/com/cloud/cluster/ClusterManagerImpl.java
----------------------------------------------------------------------
diff --git a/framework/cluster/src/com/cloud/cluster/ClusterManagerImpl.java b/framework/cluster/src/com/cloud/cluster/ClusterManagerImpl.java
index 1058da3..c7c3710 100644
--- a/framework/cluster/src/com/cloud/cluster/ClusterManagerImpl.java
+++ b/framework/cluster/src/com/cloud/cluster/ClusterManagerImpl.java
@@ -255,15 +255,15 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                         try {
                             if (s_logger.isDebugEnabled()) {
                                 s_logger.debug("Cluster PDU " + getSelfPeerName() + " -> " + pdu.getDestPeer() + ". agent: " + pdu.getAgentId() + ", pdu seq: " +
-                                        pdu.getSequenceId() + ", pdu ack seq: " + pdu.getAckSequenceId() + ", json: " + pdu.getJsonPackage());
+                                    pdu.getSequenceId() + ", pdu ack seq: " + pdu.getAckSequenceId() + ", json: " + pdu.getJsonPackage());
                             }
 
                             long startTick = System.currentTimeMillis();
                             String strResult = peerService.execute(pdu);
                             if (s_logger.isDebugEnabled()) {
                                 s_logger.debug("Cluster PDU " + getSelfPeerName() + " -> " + pdu.getDestPeer() + " completed. time: " +
-                                        (System.currentTimeMillis() - startTick) + "ms. agent: " + pdu.getAgentId() + ", pdu seq: " + pdu.getSequenceId() +
-                                        ", pdu ack seq: " + pdu.getAckSequenceId() + ", json: " + pdu.getJsonPackage());
+                                    (System.currentTimeMillis() - startTick) + "ms. agent: " + pdu.getAgentId() + ", pdu seq: " + pdu.getSequenceId() +
+                                    ", pdu ack seq: " + pdu.getAckSequenceId() + ", json: " + pdu.getJsonPackage());
                             }
 
                             if ("true".equals(strResult))
@@ -273,7 +273,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                             invalidatePeerService(pdu.getDestPeer());
                             if (s_logger.isInfoEnabled()) {
                                 s_logger.info("Exception on remote execution, peer: " + pdu.getDestPeer() + ", iteration: " + i + ", exception message :" +
-                                        e.getMessage());
+                                    e.getMessage());
                             }
                         }
                     }
@@ -563,7 +563,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                         if (profiler.getDuration() >= HeartbeatInterval.value()) {
                             if (s_logger.isDebugEnabled())
                                 s_logger.debug("Management server heartbeat takes too long to finish. profiler: " + profiler.toString() + ", profilerHeartbeatUpdate: " +
-                                        profilerHeartbeatUpdate.toString() + ", profilerPeerScan: " + profilerPeerScan.toString());
+                                    profilerHeartbeatUpdate.toString() + ", profilerPeerScan: " + profilerPeerScan.toString());
                         }
                     }
 
@@ -721,7 +721,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                     }
                 }
             }
-            break;
+                break;
 
             case nodeRemoved: {
                 List<ManagementServerHostVO> l = msg.getNodes();
@@ -731,7 +731,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                     }
                 }
             }
-            break;
+                break;
 
             default:
                 break;
@@ -772,7 +772,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                 s_logger.info("Found " + inactiveList.size() + " inactive management server node based on timestamp");
                 for (ManagementServerHostVO host : inactiveList)
                     s_logger.info("management server node msid: " + host.getMsid() + ", name: " + host.getName() + ", service ip: " + host.getServiceIP() +
-                            ", version: " + host.getVersion());
+                        ", version: " + host.getVersion());
             }
 
             List<ManagementServerHostVO> downHostList = new ArrayList<ManagementServerHostVO>();
@@ -810,7 +810,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
 
             if (_mshostPeerDao.countStateSeenInPeers(_mshostId, _runId, ManagementServerHost.State.Down) > 0) {
                 String msg =
-                        "We have detected that at least one management server peer reports that this management server is down, perform active fencing to avoid split-brain situation";
+                    "We have detected that at least one management server peer reports that this management server is down, perform active fencing to avoid split-brain situation";
                 s_logger.error(msg);
                 throw new ActiveFencingException(msg);
             }
@@ -831,7 +831,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                         if (entry.getKey().longValue() != _mshostId.longValue()) {
                             if (s_logger.isDebugEnabled()) {
                                 s_logger.debug("Detected management node left because of invalidated session, id:" + entry.getKey() + ", nodeIP:" +
-                                        entry.getValue().getServiceIP());
+                                    entry.getValue().getServiceIP());
                             }
                             invalidatedNodeList.add(entry.getValue());
                         }
@@ -918,8 +918,8 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
         if (profiler.getDuration() >= HeartbeatInterval.value()) {
             if (s_logger.isDebugEnabled())
                 s_logger.debug("Peer scan takes too long to finish. profiler: " + profiler.toString() + ", profilerQueryActiveList: " +
-                        profilerQueryActiveList.toString() + ", profilerSyncClusterInfo: " + profilerSyncClusterInfo.toString() + ", profilerInvalidatedNodeList: " +
-                        profilerInvalidatedNodeList.toString() + ", profilerRemovedList: " + profilerRemovedList.toString());
+                    profilerQueryActiveList.toString() + ", profilerSyncClusterInfo: " + profilerSyncClusterInfo.toString() + ", profilerInvalidatedNodeList: " +
+                    profilerInvalidatedNodeList.toString() + ", profilerRemovedList: " + profilerRemovedList.toString());
         }
     }
 
@@ -970,7 +970,7 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                     }
 
                     _mshostDao.update(mshost.getId(), getCurrentRunId(), NetUtils.getHostName(), version, _clusterNodeIP, _currentServiceAdapter.getServicePort(),
-                            DateUtil.currentGMTTime());
+                        DateUtil.currentGMTTime());
                 }
 
                 return mshost;
@@ -1165,19 +1165,19 @@ public class ClusterManagerImpl extends ManagerBase implements ClusterManager, C
                         throw new ConfigurationException(msg);
                     } else {
                         String msg =
-                                "Detected another management node with localhost IP is considered as running in DB, however it is not pingable, we will continue cluster initialization with this management server node";
+                            "Detected another management node with localhost IP is considered as running in DB, however it is not pingable, we will continue cluster initialization with this management server node";
                         s_logger.info(msg);
                     }
                 } else {
                     if (pingManagementNode(peer.getMsid())) {
                         String msg =
-                                "Detected that another management node with the same IP " + peer.getServiceIP() +
+                            "Detected that another management node with the same IP " + peer.getServiceIP() +
                                 " is already running, please check your cluster configuration";
                         s_logger.error(msg);
                         throw new ConfigurationException(msg);
                     } else {
                         String msg =
-                                "Detected that another management node with the same IP " + peer.getServiceIP() +
+                            "Detected that another management node with the same IP " + peer.getServiceIP() +
                                 " is considered as running in DB, however it is not pingable, we will continue cluster initialization with this management server node";
                         s_logger.info(msg);
                     }

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/cluster/src/com/cloud/cluster/dao/ManagementServerHostDaoImpl.java
----------------------------------------------------------------------
diff --git a/framework/cluster/src/com/cloud/cluster/dao/ManagementServerHostDaoImpl.java b/framework/cluster/src/com/cloud/cluster/dao/ManagementServerHostDaoImpl.java
index 2287ca7..3d0c3f5 100644
--- a/framework/cluster/src/com/cloud/cluster/dao/ManagementServerHostDaoImpl.java
+++ b/framework/cluster/src/com/cloud/cluster/dao/ManagementServerHostDaoImpl.java
@@ -44,48 +44,48 @@ import com.cloud.utils.exception.CloudRuntimeException;
 @Local(value = {ManagementServerHostDao.class})
 public class ManagementServerHostDaoImpl extends GenericDaoBase<ManagementServerHostVO, Long> implements ManagementServerHostDao {
     private static final Logger s_logger = Logger.getLogger(ManagementServerHostDaoImpl.class);
-    
+
     private final SearchBuilder<ManagementServerHostVO> MsIdSearch;
     private final SearchBuilder<ManagementServerHostVO> ActiveSearch;
     private final SearchBuilder<ManagementServerHostVO> InactiveSearch;
     private final SearchBuilder<ManagementServerHostVO> StateSearch;
 
-	@Override
+    @Override
     public void invalidateRunSession(long id, long runid) {
-	    TransactionLegacy txn = TransactionLegacy.currentTxn();
+        TransactionLegacy txn = TransactionLegacy.currentTxn();
         PreparedStatement pstmt = null;
         try {
             pstmt = txn.prepareAutoCloseStatement("update mshost set runid=0, state='Down' where id=? and runid=?");
             pstmt.setLong(1, id);
             pstmt.setLong(2, runid);
-            
+
             pstmt.executeUpdate();
         } catch (SQLException e) {
-        	throw new CloudRuntimeException("DB exception on " + pstmt.toString(), e);
+            throw new CloudRuntimeException("DB exception on " + pstmt.toString(), e);
         }
-	}
-	
-	@Override
+    }
+
+    @Override
     public ManagementServerHostVO findByMsid(long msid) {
         SearchCriteria<ManagementServerHostVO> sc = MsIdSearch.create();
         sc.setParameters("msid", msid);
-		
-		List<ManagementServerHostVO> l = listIncludingRemovedBy(sc);
+
+        List<ManagementServerHostVO> l = listIncludingRemovedBy(sc);
         if (l != null && l.size() > 0) {
             return l.get(0);
         }
-		 
-		return null;
-	}
-	
-	@Override
+
+        return null;
+    }
+
+    @Override
     @DB
-	public void update(long id, long runid, String name, String version, String serviceIP, int servicePort, Date lastUpdate) {
+    public void update(long id, long runid, String name, String version, String serviceIP, int servicePort, Date lastUpdate) {
         TransactionLegacy txn = TransactionLegacy.currentTxn();
         PreparedStatement pstmt = null;
         try {
             txn.start();
-            
+
             pstmt =
                 txn.prepareAutoCloseStatement("update mshost set name=?, version=?, service_ip=?, service_port=?, last_update=?, removed=null, alert_count=0, runid=?, state=? where id=?");
             pstmt.setString(1, name);
@@ -96,48 +96,48 @@ public class ManagementServerHostDaoImpl extends GenericDaoBase<ManagementServer
             pstmt.setLong(6, runid);
             pstmt.setString(7, ManagementServerHost.State.Up.toString());
             pstmt.setLong(8, id);
-            
+
             pstmt.executeUpdate();
             txn.commit();
         } catch (Exception e) {
             s_logger.warn("Unexpected exception, ", e);
             throw new RuntimeException(e.getMessage(), e);
         }
-	}
-	
-	@Override
+    }
+
+    @Override
     @DB
     public boolean remove(Long id) {
         TransactionLegacy txn = TransactionLegacy.currentTxn();
-    
+
         try {
-        	txn.start();
-        	
-        	ManagementServerHostVO msHost = findById(id);
-        	msHost.setState(ManagementServerHost.State.Down);
-        	super.remove(id);
-        	
-        	txn.commit();
-        	return true;
+            txn.start();
+
+            ManagementServerHostVO msHost = findById(id);
+            msHost.setState(ManagementServerHost.State.Down);
+            super.remove(id);
+
+            txn.commit();
+            return true;
         } catch (Exception e) {
             s_logger.warn("Unexpected exception, ", e);
             throw new RuntimeException(e.getMessage(), e);
         }
     }
 
-	@Override
+    @Override
     @DB
-	public void update(long id, long runid, Date lastUpdate) {
+    public void update(long id, long runid, Date lastUpdate) {
         TransactionLegacy txn = TransactionLegacy.currentTxn();
         PreparedStatement pstmt = null;
         try {
             txn.start();
-            
+
             pstmt = txn.prepareAutoCloseStatement("update mshost set last_update=?, removed=null, alert_count=0 where id=? and runid=?");
             pstmt.setString(1, DateUtil.getDateDisplayString(TimeZone.getTimeZone("GMT"), lastUpdate));
             pstmt.setLong(2, id);
             pstmt.setLong(3, runid);
-            
+
             int count = pstmt.executeUpdate();
             txn.commit();
 
@@ -148,69 +148,69 @@ public class ManagementServerHostDaoImpl extends GenericDaoBase<ManagementServer
             s_logger.warn("Unexpected exception, ", e);
             throw new RuntimeException(e.getMessage(), e);
         }
-	}
-	
-	@Override
+    }
+
+    @Override
     public List<ManagementServerHostVO> getActiveList(Date cutTime) {
-	    SearchCriteria<ManagementServerHostVO> sc = ActiveSearch.create();
-	    sc.setParameters("lastUpdateTime", cutTime);
-	    
-	    return listIncludingRemovedBy(sc);
-	}
+        SearchCriteria<ManagementServerHostVO> sc = ActiveSearch.create();
+        sc.setParameters("lastUpdateTime", cutTime);
 
-	@Override
+        return listIncludingRemovedBy(sc);
+    }
+
+    @Override
     public List<ManagementServerHostVO> getInactiveList(Date cutTime) {
-	    SearchCriteria<ManagementServerHostVO> sc = InactiveSearch.create();
-	    sc.setParameters("lastUpdateTime", cutTime);
-	    
-	    return listIncludingRemovedBy(sc);
-	}
-	
-	@Override
+        SearchCriteria<ManagementServerHostVO> sc = InactiveSearch.create();
+        sc.setParameters("lastUpdateTime", cutTime);
+
+        return listIncludingRemovedBy(sc);
+    }
+
+    @Override
     @DB
-	public int increaseAlertCount(long id) {
+    public int increaseAlertCount(long id) {
         TransactionLegacy txn = TransactionLegacy.currentTxn();
         PreparedStatement pstmt = null;
         int changedRows = 0;
         try {
             txn.start();
-            
+
             pstmt = txn.prepareAutoCloseStatement("update mshost set alert_count=alert_count+1 where id=? and alert_count=0");
             pstmt.setLong(1, id);
-            
+
             changedRows = pstmt.executeUpdate();
             txn.commit();
         } catch (Exception e) {
             s_logger.warn("Unexpected exception, ", e);
             throw new RuntimeException(e.getMessage(), e);
         }
-        
+
         return changedRows;
-	}
-	
-	protected ManagementServerHostDaoImpl() {
-		MsIdSearch = createSearchBuilder();
-		MsIdSearch.and("msid",  MsIdSearch.entity().getMsid(), SearchCriteria.Op.EQ);
-		MsIdSearch.done();
-		
-	    ActiveSearch = createSearchBuilder();
-	    ActiveSearch.and("lastUpdateTime", ActiveSearch.entity().getLastUpdateTime(),  SearchCriteria.Op.GT);
-	    ActiveSearch.and("removed", ActiveSearch.entity().getRemoved(), SearchCriteria.Op.NULL);
-	    ActiveSearch.done();
-
-	    InactiveSearch = createSearchBuilder();
-	    InactiveSearch.and("lastUpdateTime", InactiveSearch.entity().getLastUpdateTime(),  SearchCriteria.Op.LTEQ);
-	    InactiveSearch.and("removed", InactiveSearch.entity().getRemoved(), SearchCriteria.Op.NULL);
-	    InactiveSearch.done();
-	    
-	    StateSearch = createSearchBuilder();
-	    StateSearch.and("state", StateSearch.entity().getState(), SearchCriteria.Op.IN);
-	    StateSearch.done();
-	}
-	
-	@Override
+    }
+
+    protected ManagementServerHostDaoImpl() {
+        MsIdSearch = createSearchBuilder();
+        MsIdSearch.and("msid", MsIdSearch.entity().getMsid(), SearchCriteria.Op.EQ);
+        MsIdSearch.done();
+
+        ActiveSearch = createSearchBuilder();
+        ActiveSearch.and("lastUpdateTime", ActiveSearch.entity().getLastUpdateTime(), SearchCriteria.Op.GT);
+        ActiveSearch.and("removed", ActiveSearch.entity().getRemoved(), SearchCriteria.Op.NULL);
+        ActiveSearch.done();
+
+        InactiveSearch = createSearchBuilder();
+        InactiveSearch.and("lastUpdateTime", InactiveSearch.entity().getLastUpdateTime(), SearchCriteria.Op.LTEQ);
+        InactiveSearch.and("removed", InactiveSearch.entity().getRemoved(), SearchCriteria.Op.NULL);
+        InactiveSearch.done();
+
+        StateSearch = createSearchBuilder();
+        StateSearch.and("state", StateSearch.entity().getState(), SearchCriteria.Op.IN);
+        StateSearch.done();
+    }
+
+    @Override
     public void update(long id, long runId, State state, Date lastUpdate) {
-	    TransactionLegacy txn = TransactionLegacy.currentTxn();
+        TransactionLegacy txn = TransactionLegacy.currentTxn();
         PreparedStatement pstmt = null;
         try {
             pstmt = txn.prepareAutoCloseStatement("update mshost set state=?, last_update=? where id=? and runid=?");
@@ -218,9 +218,9 @@ public class ManagementServerHostDaoImpl extends GenericDaoBase<ManagementServer
             pstmt.setString(2, DateUtil.getDateDisplayString(TimeZone.getTimeZone("GMT"), lastUpdate));
             pstmt.setLong(3, id);
             pstmt.setLong(4, runId);
-            
+
             int count = pstmt.executeUpdate();
-            
+
             if (count < 1) {
                 throw new CloudRuntimeException("Invalid cluster session detected", new ClusterInvalidSessionException("runid " + runId + " is no longer valid"));
             }
@@ -228,21 +228,21 @@ public class ManagementServerHostDaoImpl extends GenericDaoBase<ManagementServer
             throw new CloudRuntimeException("DB exception on " + pstmt.toString(), e);
         }
     }
-	
-	@Override
+
+    @Override
     public List<ManagementServerHostVO> listBy(ManagementServerHost.State... states) {
-	    SearchCriteria<ManagementServerHostVO> sc = StateSearch.create();
+        SearchCriteria<ManagementServerHostVO> sc = StateSearch.create();
 
         sc.setParameters("state", (Object[])states);
-        
+
         return listBy(sc);
-	}
-	
-	@Override
-	public List<Long> listOrphanMsids() {
-		List<Long> orphanList = new ArrayList<Long>();
-		
-	    TransactionLegacy txn = TransactionLegacy.currentTxn();
+    }
+
+    @Override
+    public List<Long> listOrphanMsids() {
+        List<Long> orphanList = new ArrayList<Long>();
+
+        TransactionLegacy txn = TransactionLegacy.currentTxn();
         PreparedStatement pstmt = null;
         try {
             pstmt =
@@ -250,26 +250,26 @@ public class ManagementServerHostDaoImpl extends GenericDaoBase<ManagementServer
 
             ResultSet rs = pstmt.executeQuery();
             while (rs.next()) {
-            	orphanList.add(rs.getLong(1));
+                orphanList.add(rs.getLong(1));
             }
         } catch (SQLException e) {
             throw new CloudRuntimeException("DB exception on " + pstmt.toString(), e);
         }
-        
+
         return orphanList;
-	}
+    }
 
-	@Override
-	public ManagementServerHostVO findOneInUpState(Filter filter) {
-	    SearchCriteria<ManagementServerHostVO> sc = StateSearch.create();
+    @Override
+    public ManagementServerHostVO findOneInUpState(Filter filter) {
+        SearchCriteria<ManagementServerHostVO> sc = StateSearch.create();
 
         sc.setParameters("state", ManagementServerHost.State.Up);
-        
-        List<ManagementServerHostVO> mshosts =  listBy(sc, filter);
+
+        List<ManagementServerHostVO> mshosts = listBy(sc, filter);
         if (mshosts != null && mshosts.size() > 0) {
-        	return mshosts.get(0);	
+            return mshosts.get(0);
         }
         return null;
-	}
-	
+    }
+
 }

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/config/test/org/apache/cloudstack/framework/config/impl/ConfigDepotAdminTest.java
----------------------------------------------------------------------
diff --git a/framework/config/test/org/apache/cloudstack/framework/config/impl/ConfigDepotAdminTest.java b/framework/config/test/org/apache/cloudstack/framework/config/impl/ConfigDepotAdminTest.java
index 245e9a6..da76804 100644
--- a/framework/config/test/org/apache/cloudstack/framework/config/impl/ConfigDepotAdminTest.java
+++ b/framework/config/test/org/apache/cloudstack/framework/config/impl/ConfigDepotAdminTest.java
@@ -39,8 +39,8 @@ import org.apache.cloudstack.framework.config.dao.ConfigurationDao;
 import com.cloud.utils.db.EntityManager;
 
 public class ConfigDepotAdminTest extends TestCase {
-    private final ConfigKey<Integer> DynamicIntCK = new ConfigKey<Integer>(Integer.class, "dynIntKey", "Advance", "10", "Test Key", true);
-    private final ConfigKey<Integer> StaticIntCK = new ConfigKey<Integer>(Integer.class, "statIntKey", "Advance", "10", "Test Key", false);
+    private final static ConfigKey<Integer> DynamicIntCK = new ConfigKey<Integer>(Integer.class, "dynIntKey", "Advance", "10", "Test Key", true);
+    private final static ConfigKey<Integer> StaticIntCK = new ConfigKey<Integer>(Integer.class, "statIntKey", "Advance", "10", "Test Key", false);
 
     @Mock
     Configurable _configurable;

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/db/src/com/cloud/utils/db/GenericDao.java
----------------------------------------------------------------------
diff --git a/framework/db/src/com/cloud/utils/db/GenericDao.java b/framework/db/src/com/cloud/utils/db/GenericDao.java
index 14169ff..cb401cd 100755
--- a/framework/db/src/com/cloud/utils/db/GenericDao.java
+++ b/framework/db/src/com/cloud/utils/db/GenericDao.java
@@ -160,14 +160,14 @@ public interface GenericDao<T, ID extends Serializable> {
      * Search for the entity beans using the sql SQL_CACHE option
      * @param sc
      * @param filter
-     * @param enable_query_cache
+     * @param enableQueryCache
      * @return list of entity beans.
      */
-    List<T> search(SearchCriteria<T> sc, Filter filter, final boolean enable_query_cache);
+    List<T> search(SearchCriteria<T> sc, Filter filter, final boolean enableQueryCache);
 
     List<T> searchIncludingRemoved(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache);
 
-    List<T> searchIncludingRemoved(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache, final boolean enable_query_cache);
+    List<T> searchIncludingRemoved(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache, final boolean enableQueryCache);
 
     /**
      * Customized search with SearchCritiria

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/db/src/com/cloud/utils/db/GenericDaoBase.java
----------------------------------------------------------------------
diff --git a/framework/db/src/com/cloud/utils/db/GenericDaoBase.java b/framework/db/src/com/cloud/utils/db/GenericDaoBase.java
index 8e6f6a4..503d759 100755
--- a/framework/db/src/com/cloud/utils/db/GenericDaoBase.java
+++ b/framework/db/src/com/cloud/utils/db/GenericDaoBase.java
@@ -346,14 +346,14 @@ public abstract class GenericDaoBase<T, ID extends Serializable> extends Compone
     }
 
     @DB()
-    protected List<T> search(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache, final boolean enable_query_cache) {
+    protected List<T> search(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache, final boolean enableQueryCache) {
         if (_removed != null) {
             if (sc == null) {
                 sc = createSearchCriteria();
             }
             sc.addAnd(_removed.second().field.getName(), SearchCriteria.Op.NULL);
         }
-        return searchIncludingRemoved(sc, filter, lock, cache, enable_query_cache);
+        return searchIncludingRemoved(sc, filter, lock, cache, enableQueryCache);
     }
 
     @Override
@@ -362,13 +362,13 @@ public abstract class GenericDaoBase<T, ID extends Serializable> extends Compone
     }
 
     @Override
-    public List<T> searchIncludingRemoved(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache, final boolean enable_query_cache) {
+    public List<T> searchIncludingRemoved(SearchCriteria<T> sc, final Filter filter, final Boolean lock, final boolean cache, final boolean enableQueryCache) {
         String clause = sc != null ? sc.getWhereClause() : null;
         if (clause != null && clause.length() == 0) {
             clause = null;
         }
 
-        final StringBuilder str = createPartialSelectSql(sc, clause != null, enable_query_cache);
+        final StringBuilder str = createPartialSelectSql(sc, clause != null, enableQueryCache);
         if (clause != null) {
             str.append(clause);
         }
@@ -896,11 +896,11 @@ public abstract class GenericDaoBase<T, ID extends Serializable> extends Compone
     }
 
     @DB()
-    protected List<T> listBy(final SearchCriteria<T> sc, final Filter filter, final boolean enable_query_cache) {
+    protected List<T> listBy(final SearchCriteria<T> sc, final Filter filter, final boolean enableQueryCache) {
         if (_removed != null) {
             sc.addAnd(_removed.second().field.getName(), SearchCriteria.Op.NULL);
         }
-        return listIncludingRemovedBy(sc, filter, enable_query_cache);
+        return listIncludingRemovedBy(sc, filter, enableQueryCache);
     }
 
     @DB()
@@ -909,8 +909,8 @@ public abstract class GenericDaoBase<T, ID extends Serializable> extends Compone
     }
 
     @DB()
-    protected List<T> listIncludingRemovedBy(final SearchCriteria<T> sc, final Filter filter, final boolean enable_query_cache) {
-        return searchIncludingRemoved(sc, filter, null, false, enable_query_cache);
+    protected List<T> listIncludingRemovedBy(final SearchCriteria<T> sc, final Filter filter, final boolean enableQueryCache) {
+        return searchIncludingRemoved(sc, filter, null, false, enableQueryCache);
     }
 
     @DB()
@@ -1194,8 +1194,8 @@ public abstract class GenericDaoBase<T, ID extends Serializable> extends Compone
     }
 
     @DB()
-    protected StringBuilder createPartialSelectSql(SearchCriteria<?> sc, final boolean whereClause, final boolean enable_query_cache) {
-        StringBuilder sql = new StringBuilder(enable_query_cache ? _partialQueryCacheSelectSql.first() : _partialSelectSql.first());
+    protected StringBuilder createPartialSelectSql(SearchCriteria<?> sc, final boolean whereClause, final boolean enableQueryCache) {
+        StringBuilder sql = new StringBuilder(enableQueryCache ? _partialQueryCacheSelectSql.first() : _partialSelectSql.first());
         if (sc != null && !sc.isSelectAll()) {
             sql.delete(7, sql.indexOf(" FROM"));
             sc.getSelect(sql, 7);
@@ -1281,8 +1281,8 @@ public abstract class GenericDaoBase<T, ID extends Serializable> extends Compone
 
     @Override
     @DB()
-    public List<T> search(final SearchCriteria<T> sc, final Filter filter, final boolean enable_query_cache) {
-        return search(sc, filter, null, false, enable_query_cache);
+    public List<T> search(final SearchCriteria<T> sc, final Filter filter, final boolean enableQueryCache) {
+        return search(sc, filter, null, false, enableQueryCache);
     }
 
     @Override

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/db/src/com/cloud/utils/db/Merovingian2.java
----------------------------------------------------------------------
diff --git a/framework/db/src/com/cloud/utils/db/Merovingian2.java b/framework/db/src/com/cloud/utils/db/Merovingian2.java
index 59dff2d..6eeea9f 100644
--- a/framework/db/src/com/cloud/utils/db/Merovingian2.java
+++ b/framework/db/src/com/cloud/utils/db/Merovingian2.java
@@ -52,7 +52,7 @@ public class Merovingian2 extends StandardMBean implements MerovingianMBean {
     private static final String SELECT_THREAD_LOCKS_SQL = SELECT_SQL + " WHERE mac=? AND ip=?";
     private static final String CLEANUP_THREAD_LOCKS_SQL = "DELETE FROM op_lock WHERE mac=? AND ip=? AND thread=?";
 
-    TimeZone s_gmtTimeZone = TimeZone.getTimeZone("GMT");
+    TimeZone _gmtTimeZone = TimeZone.getTimeZone("GMT");
 
     private final long _msId;
 
@@ -184,7 +184,7 @@ public class Merovingian2 extends StandardMBean implements MerovingianMBean {
             pstmt.setLong(2, _msId);
             pstmt.setString(3, threadName);
             pstmt.setInt(4, threadId);
-            pstmt.setString(5, DateUtil.getDateDisplayString(s_gmtTimeZone, new Date()));
+            pstmt.setString(5, DateUtil.getDateDisplayString(_gmtTimeZone, new Date()));
             try {
                 int rows = pstmt.executeUpdate();
                 if (rows == 1) {

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/db/src/com/cloud/utils/db/SqlGenerator.java
----------------------------------------------------------------------
diff --git a/framework/db/src/com/cloud/utils/db/SqlGenerator.java b/framework/db/src/com/cloud/utils/db/SqlGenerator.java
index 2c24f62..befe34b 100755
--- a/framework/db/src/com/cloud/utils/db/SqlGenerator.java
+++ b/framework/db/src/com/cloud/utils/db/SqlGenerator.java
@@ -567,10 +567,10 @@ public class SqlGenerator {
         return sql.toString();
     }
 
-    public Pair<StringBuilder, Attribute[]> buildSelectSql(boolean enable_query_cache) {
+    public Pair<StringBuilder, Attribute[]> buildSelectSql(boolean enableQueryCache) {
         StringBuilder sql = new StringBuilder("SELECT ");
 
-        sql.append(enable_query_cache ? "SQL_CACHE " : "");
+        sql.append(enableQueryCache ? "SQL_CACHE " : "");
 
         ArrayList<Attribute> attrs = new ArrayList<Attribute>();
 

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/db/src/com/cloud/utils/db/TransactionLegacy.java
----------------------------------------------------------------------
diff --git a/framework/db/src/com/cloud/utils/db/TransactionLegacy.java b/framework/db/src/com/cloud/utils/db/TransactionLegacy.java
index e5cece1..b36c027 100755
--- a/framework/db/src/com/cloud/utils/db/TransactionLegacy.java
+++ b/framework/db/src/com/cloud/utils/db/TransactionLegacy.java
@@ -1203,7 +1203,7 @@ public class TransactionLegacy {
         final ConnectionFactory connectionFactory = new DriverManagerConnectionFactory("jdbc:mysql://localhost:3306/" + database, "cloud", "cloud");
         final PoolableConnectionFactory poolableConnectionFactory = new PoolableConnectionFactory(connectionFactory, connectionPool, null, null, false, true);
         return new PoolingDataSource(
-        /* connectionPool */poolableConnectionFactory.getPool());
+            /* connectionPool */poolableConnectionFactory.getPool());
     }
 
     /**

http://git-wip-us.apache.org/repos/asf/cloudstack/blob/be5e5cc6/framework/db/test/com/cloud/utils/db/GlobalLockTest.java
----------------------------------------------------------------------
diff --git a/framework/db/test/com/cloud/utils/db/GlobalLockTest.java b/framework/db/test/com/cloud/utils/db/GlobalLockTest.java
index 8b0720b..c26f248 100644
--- a/framework/db/test/com/cloud/utils/db/GlobalLockTest.java
+++ b/framework/db/test/com/cloud/utils/db/GlobalLockTest.java
@@ -28,7 +28,7 @@ import com.cloud.utils.Profiler;
 @ContextConfiguration(locations = "classpath:/testContext.xml")
 public class GlobalLockTest {
     public static final Logger s_logger = Logger.getLogger(GlobalLockTest.class);
-    private final static GlobalLock _workLock = GlobalLock.getInternLock("SecurityGroupWork");
+    private final static GlobalLock WorkLock = GlobalLock.getInternLock("SecurityGroupWork");
 
     public static class Worker implements Runnable {
         int id = 0;
@@ -47,7 +47,7 @@ public class GlobalLockTest {
             try {
                 Profiler p = new Profiler();
                 p.start();
-                locked = _workLock.lock(timeoutSeconds);
+                locked = WorkLock.lock(timeoutSeconds);
                 p.stop();
                 System.out.println("Thread " + id + " waited " + p.getDuration() + " ms, locked=" + locked);
                 if (locked) {
@@ -56,7 +56,7 @@ public class GlobalLockTest {
             } catch (InterruptedException e) {
             } finally {
                 if (locked) {
-                    boolean unlocked = _workLock.unlock();
+                    boolean unlocked = WorkLock.unlock();
                     System.out.println("Thread " + id + "  unlocked=" + unlocked);
                 }
             }