You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by nc...@apache.org on 2016/12/20 21:02:54 UTC
ambari git commit: AMBARI-19149. Code cleanup: apply checks to test
sources (Attila Doroszlai via ncole)
Repository: ambari
Updated Branches:
refs/heads/trunk 51c4f92c1 -> 4fc52a61f
AMBARI-19149. Code cleanup: apply checks to test sources (Attila Doroszlai via ncole)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/4fc52a61
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/4fc52a61
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/4fc52a61
Branch: refs/heads/trunk
Commit: 4fc52a61f5515780a85d54051bfbb08ba7e530b1
Parents: 51c4f92
Author: Nate Cole <nc...@hortonworks.com>
Authored: Tue Dec 20 16:02:32 2016 -0500
Committer: Nate Cole <nc...@hortonworks.com>
Committed: Tue Dec 20 16:02:37 2016 -0500
----------------------------------------------------------------------
ambari-project/pom.xml | 3 +-
ambari-server/checkstyle.xml | 2 +-
.../server/upgrade/UpgradeCatalog250.java | 28 +-
.../server/agent/TestHeartbeatHandler.java | 4 +-
.../AmbariCustomCommandExecutionHelperTest.java | 26 +-
.../internal/MemberResourceProviderTest.java | 1 -
.../UpgradeSummaryResourceProviderTest.java | 2 +-
.../logging/LogQueryResponseTest.java | 188 ++++++------
.../LogSearchDataRetrievalServiceTest.java | 8 -
.../logging/LoggingRequestHelperImplTest.java | 284 +++++++------------
.../server/state/ConfigMergeHelperTest.java | 12 +-
.../QuickLinksProfileParserTest.java | 2 +-
.../server/testing/DeadlockedThreadsTest.java | 14 +-
.../server/topology/BlueprintImplTest.java | 1 -
.../ClusterDeployWithStartOnlyTest.java | 1 -
...InstallWithoutStartOnComponentLevelTest.java | 1 -
.../ClusterInstallWithoutStartTest.java | 1 -
.../server/upgrade/UpgradeCatalog210Test.java | 6 +-
.../server/upgrade/UpgradeCatalog221Test.java | 5 +-
.../server/upgrade/UpgradeCatalog250Test.java | 65 ++---
20 files changed, 263 insertions(+), 391 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-project/pom.xml
----------------------------------------------------------------------
diff --git a/ambari-project/pom.xml b/ambari-project/pom.xml
index f3476d7..16ea2af 100644
--- a/ambari-project/pom.xml
+++ b/ambari-project/pom.xml
@@ -31,7 +31,6 @@
<powermock.version>1.6.3</powermock.version>
<jetty.version>8.1.19.v20160209</jetty.version>
<checkstyle.version>6.19</checkstyle.version> <!-- last version that does not require Java 8 -->
- <checkstyle.skip>false</checkstyle.skip>
</properties>
<profiles>
<profile>
@@ -502,8 +501,8 @@
<encoding>UTF-8</encoding>
<consoleOutput>true</consoleOutput>
<failsOnError>true</failsOnError>
+ <includeTestSourceDirectory>true</includeTestSourceDirectory>
<linkXRef>false</linkXRef>
- <skip>${checkstyle.skip}</skip>
</configuration>
<executions>
<execution>
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/checkstyle.xml
----------------------------------------------------------------------
diff --git a/ambari-server/checkstyle.xml b/ambari-server/checkstyle.xml
index 0742284..bf7698d 100644
--- a/ambari-server/checkstyle.xml
+++ b/ambari-server/checkstyle.xml
@@ -20,7 +20,7 @@
<module name="AvoidStarImport"/>
<module name="IllegalImport"/>
<module name="ImportOrder">
- <property name="groups" value="java,javax,org,com*"/>
+ <property name="groups" value="java,javax,org,com,*"/>
<property name="ordered" value="true"/>
<property name="separated" value="true"/>
<property name="option" value="top"/> <!-- static imports -->
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/main/java/org/apache/ambari/server/upgrade/UpgradeCatalog250.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/java/org/apache/ambari/server/upgrade/UpgradeCatalog250.java b/ambari-server/src/main/java/org/apache/ambari/server/upgrade/UpgradeCatalog250.java
index 9734212..7a385f3 100644
--- a/ambari-server/src/main/java/org/apache/ambari/server/upgrade/UpgradeCatalog250.java
+++ b/ambari-server/src/main/java/org/apache/ambari/server/upgrade/UpgradeCatalog250.java
@@ -17,8 +17,18 @@
*/
package org.apache.ambari.server.upgrade;
-import com.google.inject.Inject;
-import com.google.inject.Injector;
+import java.sql.ResultSet;
+import java.sql.SQLException;
+import java.sql.Statement;
+import java.util.ArrayList;
+import java.util.Arrays;
+import java.util.Collection;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.concurrent.atomic.AtomicLong;
import org.apache.ambari.server.AmbariException;
import org.apache.ambari.server.actionmanager.CommandExecutionType;
@@ -35,18 +45,8 @@ import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.springframework.jdbc.support.JdbcUtils;
-import java.sql.ResultSet;
-import java.sql.SQLException;
-import java.sql.Statement;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-import java.util.concurrent.atomic.AtomicLong;
+import com.google.inject.Inject;
+import com.google.inject.Injector;
/**
* Upgrade catalog for version 2.5.0.
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatHandler.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatHandler.java b/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatHandler.java
index a7f8a21..e7588c7 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatHandler.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/agent/TestHeartbeatHandler.java
@@ -973,9 +973,7 @@ public class TestHeartbeatHandler {
ActionQueue aq = new ActionQueue();
ActionManager am = actionManagerTestHelper.getMockActionManager();
- expect(am.getTasks(EasyMock.<List<Long>>anyObject())).andReturn(
- new ArrayList<HostRoleCommand>() {{
- }});
+ expect(am.getTasks(EasyMock.<List<Long>>anyObject())).andReturn(new ArrayList<HostRoleCommand>());
replay(am);
HeartBeatHandler handler = heartbeatTestHelper.getHeartBeatHandler(am, aq);
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariCustomCommandExecutionHelperTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariCustomCommandExecutionHelperTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariCustomCommandExecutionHelperTest.java
index 360509c..246c628 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariCustomCommandExecutionHelperTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/AmbariCustomCommandExecutionHelperTest.java
@@ -195,11 +195,7 @@ public class AmbariCustomCommandExecutionHelperTest {
new RequestResourceFilter("GANGLIA", "GANGLIA_MONITOR", Collections.singletonList("c1-c6402"))
),
new RequestOperationLevel(Resource.Type.Service, "c1", "GANGLIA", null, null),
- new HashMap<String, String>() {
- {
- }
- },
- false);
+ new HashMap<String, String>(), false);
EasyMock.replay(hostRoleCommand, actionManager);
@@ -234,10 +230,7 @@ public class AmbariCustomCommandExecutionHelperTest {
new RequestResourceFilter("GANGLIA", "GANGLIA_MONITOR", Collections.singletonList("c1-c6401")),
new RequestResourceFilter("GANGLIA", "GANGLIA_MONITOR", Collections.singletonList("c1-c6402"))),
new RequestOperationLevel(Resource.Type.Service, "c1", "GANGLIA", null, null),
- new HashMap<String, String>() {
- {
- }
- }, false);
+ new HashMap<String, String>(), false);
EasyMock.replay(hostRoleCommand, actionManager);
@@ -274,10 +267,7 @@ public class AmbariCustomCommandExecutionHelperTest {
new RequestResourceFilter("GANGLIA", "GANGLIA_MONITOR", Collections.singletonList("c1-c6401")),
new RequestResourceFilter("GANGLIA", "GANGLIA_MONITOR", Collections.singletonList("c1-c6402"))),
new RequestOperationLevel(Resource.Type.Host, "c1", "GANGLIA", null, null),
- new HashMap<String, String>() {
- {
- }
- }, false);
+ new HashMap<String, String>(), false);
EasyMock.replay(hostRoleCommand, actionManager);
@@ -323,10 +313,7 @@ public class AmbariCustomCommandExecutionHelperTest {
Collections.singletonList("c6402"))),
new RequestOperationLevel(Resource.Type.Service, "c1", "ZOOKEEPER", null, null),
- new HashMap<String, String>() {
- {
- }
- }, false);
+ new HashMap<String, String>(), false);
EasyMock.replay(hostRoleCommand, actionManager);
ambariManagementController.createAction(actionRequest, requestProperties);
@@ -366,10 +353,7 @@ public class AmbariCustomCommandExecutionHelperTest {
Collections.singletonList("c6402"))),
new RequestOperationLevel(Resource.Type.Service, "c1", "ZOOKEEPER", null, null),
- new HashMap<String, String>() {
- {
- }
- }, false);
+ new HashMap<String, String>(), false);
EasyMock.replay(hostRoleCommand, actionManager);
ambariManagementController.createAction(actionRequest, requestProperties);
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/MemberResourceProviderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/MemberResourceProviderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/MemberResourceProviderTest.java
index 438e073..0a50844 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/MemberResourceProviderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/MemberResourceProviderTest.java
@@ -18,7 +18,6 @@
package org.apache.ambari.server.controller.internal;
-import static org.easymock.EasyMock.anyObject;
import static org.easymock.EasyMock.createMock;
import static org.easymock.EasyMock.createNiceMock;
import static org.easymock.EasyMock.eq;
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/UpgradeSummaryResourceProviderTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/UpgradeSummaryResourceProviderTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/UpgradeSummaryResourceProviderTest.java
index baec7df..e398a54 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/UpgradeSummaryResourceProviderTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/internal/UpgradeSummaryResourceProviderTest.java
@@ -195,7 +195,7 @@ public class UpgradeSummaryResourceProviderTest {
* @param stageId
*/
@Transactional
- private void createCommands(Cluster cluster, Long upgradeRequestId, Long stageId) {
+ void createCommands(Cluster cluster, Long upgradeRequestId, Long stageId) {
HostEntity h1 = hostDAO.findByName("h1");
ServiceComponentHostEvent event = new ServiceComponentHostOpInProgressEvent("ZOOKEEPER_SERVER", "h1", 1L);
ServiceComponentHostEventWrapper eventWrapper = new ServiceComponentHostEventWrapper(event);
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogQueryResponseTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogQueryResponseTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogQueryResponseTest.java
index 527306c..0f29400 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogQueryResponseTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogQueryResponseTest.java
@@ -127,100 +127,102 @@ public class LogQueryResponseTest {
List<LogLineResult> listOfLineResults =
result.getListOfResults();
- {
- LogLineResult resultOne = listOfLineResults.get(0);
- // verify that all fields in this class are parsed as expected
- assertEquals("Cluster name not parsed properly",
- "clusterone", resultOne.getClusterName());
- assertEquals("Method Name not parsed properly",
- "chooseUnderReplicatedBlocks", resultOne.getLogMethod());
- assertEquals("Log Level not parsed properly",
- "INFO", resultOne.getLogLevel());
- assertEquals("event_count not parsed properly",
- "1", resultOne.getEventCount());
- assertEquals("ip address not parsed properly",
- "192.168.1.1", resultOne.getIpAddress());
- assertEquals("component type not parsed properly",
- "hdfs_namenode", resultOne.getComponentType());
- assertEquals("sequence number not parsed properly",
- "10584", resultOne.getSequenceNumber());
- assertEquals("log file path not parsed properly",
- "/var/log/hadoop/hdfs/hadoop-hdfs-namenode-c6401.ambari.apache.org.log", resultOne.getLogFilePath());
- assertEquals("log src file name not parsed properly",
- "UnderReplicatedBlocks.java", resultOne.getSourceFile());
- assertEquals("log src line number not parsed properly",
- "394", resultOne.getSourceFileLineNumber());
- assertEquals("host name not parsed properly",
- "c6401.ambari.apache.org", resultOne.getHostName());
- assertEquals("log message not parsed properly",
- "chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false", resultOne.getLogMessage());
- assertEquals("logger name not parsed properly",
- "BlockStateChange", resultOne.getLoggerName());
- assertEquals("id not parsed properly",
- "9c5562fb-123f-47c8-aaf5-b5e407326c08", resultOne.getId());
- assertEquals("message MD5 not parsed properly",
- "-3892769501348410581", resultOne.getMessageMD5());
- assertEquals("log time not parsed properly",
- "1458148749036", resultOne.getLogTime());
- assertEquals("event MD5 not parsed properly",
- "1458148749036-2417481968206345035", resultOne.getEventMD5());
- assertEquals("logfile line number not parsed properly",
- "2084", resultOne.getLogFileLineNumber());
- assertEquals("ttl not parsed properly",
- "+7DAYS", resultOne.getTtl());
- assertEquals("expire at not parsed properly",
- "1458753550322", resultOne.getExpirationTime());
- assertEquals("version not parsed properly",
- "1528979784023932928", resultOne.getVersion());
- }
-
- {
- LogLineResult resultTwo = listOfLineResults.get(1);
- // verify second log line record's data is parsed correctly
- assertEquals("Cluster name not parsed properly",
- "clusterone", resultTwo.getClusterName());
- assertEquals("Method Name not parsed properly",
- "putMetrics", resultTwo.getLogMethod());
- assertEquals("Log Level not parsed properly",
- "WARN", resultTwo.getLogLevel());
- assertEquals("event_count not parsed properly",
- "1", resultTwo.getEventCount());
- assertEquals("ip address not parsed properly",
- "192.168.1.1", resultTwo.getIpAddress());
- assertEquals("component type not parsed properly",
- "yarn_resourcemanager", resultTwo.getComponentType());
- assertEquals("sequence number not parsed properly",
- "10583", resultTwo.getSequenceNumber());
- assertEquals("log file path not parsed properly",
- "/var/log/hadoop-yarn/yarn/yarn-yarn-resourcemanager-c6401.ambari.apache.org.log", resultTwo.getLogFilePath());
- assertEquals("log src file name not parsed properly",
- "HadoopTimelineMetricsSink.java", resultTwo.getSourceFile());
- assertEquals("log src line number not parsed properly",
- "262", resultTwo.getSourceFileLineNumber());
- assertEquals("host name not parsed properly",
- "c6401.ambari.apache.org", resultTwo.getHostName());
- assertEquals("log message not parsed properly",
- "Unable to send metrics to collector by address:http://c6401.ambari.apache.org:6188/ws/v1/timeline/metrics", resultTwo.getLogMessage());
- assertEquals("logger name not parsed properly",
- "timeline.HadoopTimelineMetricsSink", resultTwo.getLoggerName());
- assertEquals("id not parsed properly",
- "8361c5a9-5b1c-4f44-bc8f-4c6f07d94228", resultTwo.getId());
- assertEquals("message MD5 not parsed properly",
- "5942185045779825717", resultTwo.getMessageMD5());
- assertEquals("log time not parsed properly",
- "1458148746937", resultTwo.getLogTime());
- assertEquals("event MD5 not parsed properly",
- "14581487469371427138486123628676", resultTwo.getEventMD5());
- assertEquals("logfile line number not parsed properly",
- "549", resultTwo.getLogFileLineNumber());
- assertEquals("ttl not parsed properly",
- "+7DAYS", resultTwo.getTtl());
- assertEquals("expire at not parsed properly",
- "1458753550322", resultTwo.getExpirationTime());
- assertEquals("version not parsed properly",
- "1528979784022884357", resultTwo.getVersion());
- }
+ verifyFirstLine(listOfLineResults);
+ verifySecondLine(listOfLineResults);
+ }
+
+ private static void verifyFirstLine(List<LogLineResult> listOfLineResults) {
+ LogLineResult resultOne = listOfLineResults.get(0);
+ // verify that all fields in this class are parsed as expected
+ assertEquals("Cluster name not parsed properly",
+ "clusterone", resultOne.getClusterName());
+ assertEquals("Method Name not parsed properly",
+ "chooseUnderReplicatedBlocks", resultOne.getLogMethod());
+ assertEquals("Log Level not parsed properly",
+ "INFO", resultOne.getLogLevel());
+ assertEquals("event_count not parsed properly",
+ "1", resultOne.getEventCount());
+ assertEquals("ip address not parsed properly",
+ "192.168.1.1", resultOne.getIpAddress());
+ assertEquals("component type not parsed properly",
+ "hdfs_namenode", resultOne.getComponentType());
+ assertEquals("sequence number not parsed properly",
+ "10584", resultOne.getSequenceNumber());
+ assertEquals("log file path not parsed properly",
+ "/var/log/hadoop/hdfs/hadoop-hdfs-namenode-c6401.ambari.apache.org.log", resultOne.getLogFilePath());
+ assertEquals("log src file name not parsed properly",
+ "UnderReplicatedBlocks.java", resultOne.getSourceFile());
+ assertEquals("log src line number not parsed properly",
+ "394", resultOne.getSourceFileLineNumber());
+ assertEquals("host name not parsed properly",
+ "c6401.ambari.apache.org", resultOne.getHostName());
+ assertEquals("log message not parsed properly",
+ "chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false", resultOne.getLogMessage());
+ assertEquals("logger name not parsed properly",
+ "BlockStateChange", resultOne.getLoggerName());
+ assertEquals("id not parsed properly",
+ "9c5562fb-123f-47c8-aaf5-b5e407326c08", resultOne.getId());
+ assertEquals("message MD5 not parsed properly",
+ "-3892769501348410581", resultOne.getMessageMD5());
+ assertEquals("log time not parsed properly",
+ "1458148749036", resultOne.getLogTime());
+ assertEquals("event MD5 not parsed properly",
+ "1458148749036-2417481968206345035", resultOne.getEventMD5());
+ assertEquals("logfile line number not parsed properly",
+ "2084", resultOne.getLogFileLineNumber());
+ assertEquals("ttl not parsed properly",
+ "+7DAYS", resultOne.getTtl());
+ assertEquals("expire at not parsed properly",
+ "1458753550322", resultOne.getExpirationTime());
+ assertEquals("version not parsed properly",
+ "1528979784023932928", resultOne.getVersion());
+ }
+ private static void verifySecondLine(List<LogLineResult> listOfLineResults) {
+ LogLineResult resultTwo = listOfLineResults.get(1);
+ // verify second log line record's data is parsed correctly
+ assertEquals("Cluster name not parsed properly",
+ "clusterone", resultTwo.getClusterName());
+ assertEquals("Method Name not parsed properly",
+ "putMetrics", resultTwo.getLogMethod());
+ assertEquals("Log Level not parsed properly",
+ "WARN", resultTwo.getLogLevel());
+ assertEquals("event_count not parsed properly",
+ "1", resultTwo.getEventCount());
+ assertEquals("ip address not parsed properly",
+ "192.168.1.1", resultTwo.getIpAddress());
+ assertEquals("component type not parsed properly",
+ "yarn_resourcemanager", resultTwo.getComponentType());
+ assertEquals("sequence number not parsed properly",
+ "10583", resultTwo.getSequenceNumber());
+ assertEquals("log file path not parsed properly",
+ "/var/log/hadoop-yarn/yarn/yarn-yarn-resourcemanager-c6401.ambari.apache.org.log", resultTwo.getLogFilePath());
+ assertEquals("log src file name not parsed properly",
+ "HadoopTimelineMetricsSink.java", resultTwo.getSourceFile());
+ assertEquals("log src line number not parsed properly",
+ "262", resultTwo.getSourceFileLineNumber());
+ assertEquals("host name not parsed properly",
+ "c6401.ambari.apache.org", resultTwo.getHostName());
+ assertEquals("log message not parsed properly",
+ "Unable to send metrics to collector by address:http://c6401.ambari.apache.org:6188/ws/v1/timeline/metrics", resultTwo.getLogMessage());
+ assertEquals("logger name not parsed properly",
+ "timeline.HadoopTimelineMetricsSink", resultTwo.getLoggerName());
+ assertEquals("id not parsed properly",
+ "8361c5a9-5b1c-4f44-bc8f-4c6f07d94228", resultTwo.getId());
+ assertEquals("message MD5 not parsed properly",
+ "5942185045779825717", resultTwo.getMessageMD5());
+ assertEquals("log time not parsed properly",
+ "1458148746937", resultTwo.getLogTime());
+ assertEquals("event MD5 not parsed properly",
+ "14581487469371427138486123628676", resultTwo.getEventMD5());
+ assertEquals("logfile line number not parsed properly",
+ "549", resultTwo.getLogFileLineNumber());
+ assertEquals("ttl not parsed properly",
+ "+7DAYS", resultTwo.getTtl());
+ assertEquals("expire at not parsed properly",
+ "1458753550322", resultTwo.getExpirationTime());
+ assertEquals("version not parsed properly",
+ "1528979784022884357", resultTwo.getVersion());
}
}
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogSearchDataRetrievalServiceTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogSearchDataRetrievalServiceTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogSearchDataRetrievalServiceTest.java
index 1b365e5..033d698 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogSearchDataRetrievalServiceTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LogSearchDataRetrievalServiceTest.java
@@ -36,14 +36,6 @@ import org.junit.Test;
import com.google.common.cache.Cache;
import com.google.inject.Injector;
-
-
-import org.apache.ambari.server.controller.AmbariManagementController;
-import org.easymock.EasyMockSupport;
-import org.junit.Test;
-
-import com.google.common.cache.Cache;
-
/**
* This test verifies the basic behavior of the
* LogSearchDataRetrievalServiceTest, and should
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LoggingRequestHelperImplTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LoggingRequestHelperImplTest.java b/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LoggingRequestHelperImplTest.java
index cb65780..11e56e2 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LoggingRequestHelperImplTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/controller/logging/LoggingRequestHelperImplTest.java
@@ -208,99 +208,8 @@ public class LoggingRequestHelperImplTest {
List<LogLineResult> listOfLineResults =
result.getListOfResults();
- {
- LogLineResult resultOne = listOfLineResults.get(0);
- // verify that all fields in this class are parsed as expected
- assertEquals("Cluster name not parsed properly",
- "clusterone", resultOne.getClusterName());
- assertEquals("Method Name not parsed properly",
- "chooseUnderReplicatedBlocks", resultOne.getLogMethod());
- assertEquals("Log Level not parsed properly",
- "INFO", resultOne.getLogLevel());
- assertEquals("event_count not parsed properly",
- "1", resultOne.getEventCount());
- assertEquals("ip address not parsed properly",
- "192.168.1.1", resultOne.getIpAddress());
- assertEquals("component type not parsed properly",
- "hdfs_namenode", resultOne.getComponentType());
- assertEquals("sequence number not parsed properly",
- "10584", resultOne.getSequenceNumber());
- assertEquals("log file path not parsed properly",
- "/var/log/hadoop/hdfs/hadoop-hdfs-namenode-c6401.ambari.apache.org.log", resultOne.getLogFilePath());
- assertEquals("log src file name not parsed properly",
- "UnderReplicatedBlocks.java", resultOne.getSourceFile());
- assertEquals("log src line number not parsed properly",
- "394", resultOne.getSourceFileLineNumber());
- assertEquals("host name not parsed properly",
- "c6401.ambari.apache.org", resultOne.getHostName());
- assertEquals("log message not parsed properly",
- "chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false", resultOne.getLogMessage());
- assertEquals("logger name not parsed properly",
- "BlockStateChange", resultOne.getLoggerName());
- assertEquals("id not parsed properly",
- "9c5562fb-123f-47c8-aaf5-b5e407326c08", resultOne.getId());
- assertEquals("message MD5 not parsed properly",
- "-3892769501348410581", resultOne.getMessageMD5());
- assertEquals("log time not parsed properly",
- "1458148749036", resultOne.getLogTime());
- assertEquals("event MD5 not parsed properly",
- "1458148749036-2417481968206345035", resultOne.getEventMD5());
- assertEquals("logfile line number not parsed properly",
- "2084", resultOne.getLogFileLineNumber());
- assertEquals("ttl not parsed properly",
- "+7DAYS", resultOne.getTtl());
- assertEquals("expire at not parsed properly",
- "1458753550322", resultOne.getExpirationTime());
- assertEquals("version not parsed properly",
- "1528979784023932928", resultOne.getVersion());
- }
-
- {
- LogLineResult resultTwo = listOfLineResults.get(1);
- // verify second log line record's data is parsed correctly
- assertEquals("Cluster name not parsed properly",
- "clusterone", resultTwo.getClusterName());
- assertEquals("Method Name not parsed properly",
- "putMetrics", resultTwo.getLogMethod());
- assertEquals("Log Level not parsed properly",
- "WARN", resultTwo.getLogLevel());
- assertEquals("event_count not parsed properly",
- "1", resultTwo.getEventCount());
- assertEquals("ip address not parsed properly",
- "192.168.1.1", resultTwo.getIpAddress());
- assertEquals("component type not parsed properly",
- "yarn_resourcemanager", resultTwo.getComponentType());
- assertEquals("sequence number not parsed properly",
- "10583", resultTwo.getSequenceNumber());
- assertEquals("log file path not parsed properly",
- "/var/log/hadoop-yarn/yarn/yarn-yarn-resourcemanager-c6401.ambari.apache.org.log", resultTwo.getLogFilePath());
- assertEquals("log src file name not parsed properly",
- "HadoopTimelineMetricsSink.java", resultTwo.getSourceFile());
- assertEquals("log src line number not parsed properly",
- "262", resultTwo.getSourceFileLineNumber());
- assertEquals("host name not parsed properly",
- "c6401.ambari.apache.org", resultTwo.getHostName());
- assertEquals("log message not parsed properly",
- "Unable to send metrics to collector by address:http://c6401.ambari.apache.org:6188/ws/v1/timeline/metrics", resultTwo.getLogMessage());
- assertEquals("logger name not parsed properly",
- "timeline.HadoopTimelineMetricsSink", resultTwo.getLoggerName());
- assertEquals("id not parsed properly",
- "8361c5a9-5b1c-4f44-bc8f-4c6f07d94228", resultTwo.getId());
- assertEquals("message MD5 not parsed properly",
- "5942185045779825717", resultTwo.getMessageMD5());
- assertEquals("log time not parsed properly",
- "1458148746937", resultTwo.getLogTime());
- assertEquals("event MD5 not parsed properly",
- "14581487469371427138486123628676", resultTwo.getEventMD5());
- assertEquals("logfile line number not parsed properly",
- "549", resultTwo.getLogFileLineNumber());
- assertEquals("ttl not parsed properly",
- "+7DAYS", resultTwo.getTtl());
- assertEquals("expire at not parsed properly",
- "1458753550322", resultTwo.getExpirationTime());
- assertEquals("version not parsed properly",
- "1528979784022884357", resultTwo.getVersion());
- }
+ verifyFirstLine(listOfLineResults);
+ verifySecondLine(listOfLineResults);
mockSupport.verifyAll();
}
@@ -647,103 +556,106 @@ public class LoggingRequestHelperImplTest {
List<LogLineResult> listOfLineResults =
result.getListOfResults();
- {
- LogLineResult resultOne = listOfLineResults.get(0);
- // verify that all fields in this class are parsed as expected
- assertEquals("Cluster name not parsed properly",
- "clusterone", resultOne.getClusterName());
- assertEquals("Method Name not parsed properly",
- "chooseUnderReplicatedBlocks", resultOne.getLogMethod());
- assertEquals("Log Level not parsed properly",
- "INFO", resultOne.getLogLevel());
- assertEquals("event_count not parsed properly",
- "1", resultOne.getEventCount());
- assertEquals("ip address not parsed properly",
- "192.168.1.1", resultOne.getIpAddress());
- assertEquals("component type not parsed properly",
- "hdfs_namenode", resultOne.getComponentType());
- assertEquals("sequence number not parsed properly",
- "10584", resultOne.getSequenceNumber());
- assertEquals("log file path not parsed properly",
- "/var/log/hadoop/hdfs/hadoop-hdfs-namenode-c6401.ambari.apache.org.log", resultOne.getLogFilePath());
- assertEquals("log src file name not parsed properly",
- "UnderReplicatedBlocks.java", resultOne.getSourceFile());
- assertEquals("log src line number not parsed properly",
- "394", resultOne.getSourceFileLineNumber());
- assertEquals("host name not parsed properly",
- "c6401.ambari.apache.org", resultOne.getHostName());
- assertEquals("log message not parsed properly",
- "chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false", resultOne.getLogMessage());
- assertEquals("logger name not parsed properly",
- "BlockStateChange", resultOne.getLoggerName());
- assertEquals("id not parsed properly",
- "9c5562fb-123f-47c8-aaf5-b5e407326c08", resultOne.getId());
- assertEquals("message MD5 not parsed properly",
- "-3892769501348410581", resultOne.getMessageMD5());
- assertEquals("log time not parsed properly",
- "1458148749036", resultOne.getLogTime());
- assertEquals("event MD5 not parsed properly",
- "1458148749036-2417481968206345035", resultOne.getEventMD5());
- assertEquals("logfile line number not parsed properly",
- "2084", resultOne.getLogFileLineNumber());
- assertEquals("ttl not parsed properly",
- "+7DAYS", resultOne.getTtl());
- assertEquals("expire at not parsed properly",
- "1458753550322", resultOne.getExpirationTime());
- assertEquals("version not parsed properly",
- "1528979784023932928", resultOne.getVersion());
- }
-
- {
- LogLineResult resultTwo = listOfLineResults.get(1);
- // verify second log line record's data is parsed correctly
- assertEquals("Cluster name not parsed properly",
- "clusterone", resultTwo.getClusterName());
- assertEquals("Method Name not parsed properly",
- "putMetrics", resultTwo.getLogMethod());
- assertEquals("Log Level not parsed properly",
- "WARN", resultTwo.getLogLevel());
- assertEquals("event_count not parsed properly",
- "1", resultTwo.getEventCount());
- assertEquals("ip address not parsed properly",
- "192.168.1.1", resultTwo.getIpAddress());
- assertEquals("component type not parsed properly",
- "yarn_resourcemanager", resultTwo.getComponentType());
- assertEquals("sequence number not parsed properly",
- "10583", resultTwo.getSequenceNumber());
- assertEquals("log file path not parsed properly",
- "/var/log/hadoop-yarn/yarn/yarn-yarn-resourcemanager-c6401.ambari.apache.org.log", resultTwo.getLogFilePath());
- assertEquals("log src file name not parsed properly",
- "HadoopTimelineMetricsSink.java", resultTwo.getSourceFile());
- assertEquals("log src line number not parsed properly",
- "262", resultTwo.getSourceFileLineNumber());
- assertEquals("host name not parsed properly",
- "c6401.ambari.apache.org", resultTwo.getHostName());
- assertEquals("log message not parsed properly",
- "Unable to send metrics to collector by address:http://c6401.ambari.apache.org:6188/ws/v1/timeline/metrics", resultTwo.getLogMessage());
- assertEquals("logger name not parsed properly",
- "timeline.HadoopTimelineMetricsSink", resultTwo.getLoggerName());
- assertEquals("id not parsed properly",
- "8361c5a9-5b1c-4f44-bc8f-4c6f07d94228", resultTwo.getId());
- assertEquals("message MD5 not parsed properly",
- "5942185045779825717", resultTwo.getMessageMD5());
- assertEquals("log time not parsed properly",
- "1458148746937", resultTwo.getLogTime());
- assertEquals("event MD5 not parsed properly",
- "14581487469371427138486123628676", resultTwo.getEventMD5());
- assertEquals("logfile line number not parsed properly",
- "549", resultTwo.getLogFileLineNumber());
- assertEquals("ttl not parsed properly",
- "+7DAYS", resultTwo.getTtl());
- assertEquals("expire at not parsed properly",
- "1458753550322", resultTwo.getExpirationTime());
- assertEquals("version not parsed properly",
- "1528979784022884357", resultTwo.getVersion());
- }
+ verifyFirstLine(listOfLineResults);
+ verifySecondLine(listOfLineResults);
mockSupport.verifyAll();
}
+ private static void verifyFirstLine(List<LogLineResult> listOfLineResults) {
+ LogLineResult resultOne = listOfLineResults.get(0);
+ // verify that all fields in this class are parsed as expected
+ assertEquals("Cluster name not parsed properly",
+ "clusterone", resultOne.getClusterName());
+ assertEquals("Method Name not parsed properly",
+ "chooseUnderReplicatedBlocks", resultOne.getLogMethod());
+ assertEquals("Log Level not parsed properly",
+ "INFO", resultOne.getLogLevel());
+ assertEquals("event_count not parsed properly",
+ "1", resultOne.getEventCount());
+ assertEquals("ip address not parsed properly",
+ "192.168.1.1", resultOne.getIpAddress());
+ assertEquals("component type not parsed properly",
+ "hdfs_namenode", resultOne.getComponentType());
+ assertEquals("sequence number not parsed properly",
+ "10584", resultOne.getSequenceNumber());
+ assertEquals("log file path not parsed properly",
+ "/var/log/hadoop/hdfs/hadoop-hdfs-namenode-c6401.ambari.apache.org.log", resultOne.getLogFilePath());
+ assertEquals("log src file name not parsed properly",
+ "UnderReplicatedBlocks.java", resultOne.getSourceFile());
+ assertEquals("log src line number not parsed properly",
+ "394", resultOne.getSourceFileLineNumber());
+ assertEquals("host name not parsed properly",
+ "c6401.ambari.apache.org", resultOne.getHostName());
+ assertEquals("log message not parsed properly",
+ "chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false", resultOne.getLogMessage());
+ assertEquals("logger name not parsed properly",
+ "BlockStateChange", resultOne.getLoggerName());
+ assertEquals("id not parsed properly",
+ "9c5562fb-123f-47c8-aaf5-b5e407326c08", resultOne.getId());
+ assertEquals("message MD5 not parsed properly",
+ "-3892769501348410581", resultOne.getMessageMD5());
+ assertEquals("log time not parsed properly",
+ "1458148749036", resultOne.getLogTime());
+ assertEquals("event MD5 not parsed properly",
+ "1458148749036-2417481968206345035", resultOne.getEventMD5());
+ assertEquals("logfile line number not parsed properly",
+ "2084", resultOne.getLogFileLineNumber());
+ assertEquals("ttl not parsed properly",
+ "+7DAYS", resultOne.getTtl());
+ assertEquals("expire at not parsed properly",
+ "1458753550322", resultOne.getExpirationTime());
+ assertEquals("version not parsed properly",
+ "1528979784023932928", resultOne.getVersion());
+ }
+
+ private static void verifySecondLine(List<LogLineResult> listOfLineResults) {
+ LogLineResult resultTwo = listOfLineResults.get(1);
+ // verify second log line record's data is parsed correctly
+ assertEquals("Cluster name not parsed properly",
+ "clusterone", resultTwo.getClusterName());
+ assertEquals("Method Name not parsed properly",
+ "putMetrics", resultTwo.getLogMethod());
+ assertEquals("Log Level not parsed properly",
+ "WARN", resultTwo.getLogLevel());
+ assertEquals("event_count not parsed properly",
+ "1", resultTwo.getEventCount());
+ assertEquals("ip address not parsed properly",
+ "192.168.1.1", resultTwo.getIpAddress());
+ assertEquals("component type not parsed properly",
+ "yarn_resourcemanager", resultTwo.getComponentType());
+ assertEquals("sequence number not parsed properly",
+ "10583", resultTwo.getSequenceNumber());
+ assertEquals("log file path not parsed properly",
+ "/var/log/hadoop-yarn/yarn/yarn-yarn-resourcemanager-c6401.ambari.apache.org.log", resultTwo.getLogFilePath());
+ assertEquals("log src file name not parsed properly",
+ "HadoopTimelineMetricsSink.java", resultTwo.getSourceFile());
+ assertEquals("log src line number not parsed properly",
+ "262", resultTwo.getSourceFileLineNumber());
+ assertEquals("host name not parsed properly",
+ "c6401.ambari.apache.org", resultTwo.getHostName());
+ assertEquals("log message not parsed properly",
+ "Unable to send metrics to collector by address:http://c6401.ambari.apache.org:6188/ws/v1/timeline/metrics", resultTwo.getLogMessage());
+ assertEquals("logger name not parsed properly",
+ "timeline.HadoopTimelineMetricsSink", resultTwo.getLoggerName());
+ assertEquals("id not parsed properly",
+ "8361c5a9-5b1c-4f44-bc8f-4c6f07d94228", resultTwo.getId());
+ assertEquals("message MD5 not parsed properly",
+ "5942185045779825717", resultTwo.getMessageMD5());
+ assertEquals("log time not parsed properly",
+ "1458148746937", resultTwo.getLogTime());
+ assertEquals("event MD5 not parsed properly",
+ "14581487469371427138486123628676", resultTwo.getEventMD5());
+ assertEquals("logfile line number not parsed properly",
+ "549", resultTwo.getLogFileLineNumber());
+ assertEquals("ttl not parsed properly",
+ "+7DAYS", resultTwo.getTtl());
+ assertEquals("expire at not parsed properly",
+ "1458753550322", resultTwo.getExpirationTime());
+ assertEquals("version not parsed properly",
+ "1528979784022884357", resultTwo.getVersion());
+ }
+
@Test
public void testCreateLogFileTailURI() throws Exception {
LoggingCookieStore.INSTANCE.getCookiesMap().clear();
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/state/ConfigMergeHelperTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/state/ConfigMergeHelperTest.java b/ambari-server/src/test/java/org/apache/ambari/server/state/ConfigMergeHelperTest.java
index 84f81e6..112d3c5 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/state/ConfigMergeHelperTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/state/ConfigMergeHelperTest.java
@@ -134,25 +134,19 @@ public class ConfigMergeHelperTest {
newStackId.getStackVersion())).andReturn(newStackProperties);
// desired config of hdfs-env.xml
- Map<String, String> desiredHdfsEnvProperties = new HashMap<String, String>() {{
-
- }};
+ Map<String, String> desiredHdfsEnvProperties = new HashMap<>();
expect(clusterMock.getDesiredConfigByType("hdfs-env.xml")).andReturn(
createConfigMock(desiredHdfsEnvProperties)
);
// desired config of zk-env.xml
- Map<String, String> desiredZkEnvProperties = new HashMap<String, String>() {{
-
- }};
+ Map<String, String> desiredZkEnvProperties = new HashMap<>();
expect(clusterMock.getDesiredConfigByType("hdfs-env.xml")).andReturn(
createConfigMock(desiredZkEnvProperties)
);
// desired config of hadoop-env.xml
- Map<String, String> desiredHadoopEnvProperties = new HashMap<String, String>() {{
-
- }};
+ Map<String, String> desiredHadoopEnvProperties = new HashMap<>();
expect(clusterMock.getDesiredConfigByType("hadoop-env.xml")).andReturn(
createConfigMock(desiredHadoopEnvProperties)
);
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/state/quicklinksprofile/QuickLinksProfileParserTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/state/quicklinksprofile/QuickLinksProfileParserTest.java b/ambari-server/src/test/java/org/apache/ambari/server/state/quicklinksprofile/QuickLinksProfileParserTest.java
index 0644027..5f93475 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/state/quicklinksprofile/QuickLinksProfileParserTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/state/quicklinksprofile/QuickLinksProfileParserTest.java
@@ -20,10 +20,10 @@ package org.apache.ambari.server.state.quicklinksprofile;
import static org.junit.Assert.assertEquals;
-import com.google.common.io.Resources;
import org.codehaus.jackson.JsonParseException;
import org.junit.Test;
+import com.google.common.io.Resources;
public class QuickLinksProfileParserTest {
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/testing/DeadlockedThreadsTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/testing/DeadlockedThreadsTest.java b/ambari-server/src/test/java/org/apache/ambari/server/testing/DeadlockedThreadsTest.java
index 0e99db5..922b334 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/testing/DeadlockedThreadsTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/testing/DeadlockedThreadsTest.java
@@ -172,16 +172,16 @@ public class DeadlockedThreadsTest {
}
}
private void f() {
- w.lock(); {
- g();
- } w.unlock();
+ w.lock();
+ g();
+ w.unlock();
}
private void g() {
- r.lock(); {
- // do some work...
- for (int i = 0; i < 1000 * 1000; i++) ;
- } r.unlock();
+ r.lock();
+ // do some work...
+ for (int i = 0; i < 1000 * 1000; i++) ;
+ r.unlock();
}
}
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java b/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
index 1ffd8f7..8107d09 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
@@ -37,7 +37,6 @@ import java.util.Set;
import org.apache.ambari.server.controller.internal.Stack;
import org.apache.ambari.server.orm.entities.BlueprintEntity;
import org.apache.ambari.server.state.SecurityType;
-import org.easymock.EasyMock;
import org.junit.Before;
import org.junit.Test;
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterDeployWithStartOnlyTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterDeployWithStartOnlyTest.java b/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterDeployWithStartOnlyTest.java
index c205e57..748b4e9 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterDeployWithStartOnlyTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterDeployWithStartOnlyTest.java
@@ -67,7 +67,6 @@ import org.apache.ambari.server.state.Clusters;
import org.apache.ambari.server.state.ComponentInfo;
import org.apache.ambari.server.state.SecurityType;
import org.easymock.Capture;
-import org.easymock.EasyMock;
import org.easymock.EasyMockRule;
import org.easymock.EasyMockSupport;
import org.easymock.Mock;
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartOnComponentLevelTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartOnComponentLevelTest.java b/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartOnComponentLevelTest.java
index b276ecf..a1f3d25 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartOnComponentLevelTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartOnComponentLevelTest.java
@@ -68,7 +68,6 @@ import org.apache.ambari.server.state.Clusters;
import org.apache.ambari.server.state.ComponentInfo;
import org.apache.ambari.server.state.SecurityType;
import org.easymock.Capture;
-import org.easymock.EasyMock;
import org.easymock.EasyMockRule;
import org.easymock.EasyMockSupport;
import org.easymock.Mock;
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartTest.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartTest.java b/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartTest.java
index 58862f1..33f318a 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartTest.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/topology/ClusterInstallWithoutStartTest.java
@@ -68,7 +68,6 @@ import org.apache.ambari.server.state.Clusters;
import org.apache.ambari.server.state.ComponentInfo;
import org.apache.ambari.server.state.SecurityType;
import org.easymock.Capture;
-import org.easymock.EasyMock;
import org.easymock.EasyMockRule;
import org.easymock.EasyMockSupport;
import org.easymock.Mock;
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog210Test.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog210Test.java b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog210Test.java
index 36f66c6..7106021 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog210Test.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog210Test.java
@@ -480,8 +480,7 @@ public class UpgradeCatalog210Test {
final Map<String, String> propertiesExpectedHiveSite = new HashMap<String, String>() {{
put("hive.server2.authentication", "kerberos");
}};
- final Map<String, String> propertiesExpectedHiveServerSite = new HashMap<String, String>() {{
- }};
+ final Map<String, String> propertiesExpectedHiveServerSite = new HashMap<>();
final Map<String, Service> servicesExpected = new HashMap<String, Service>(){{
put("KERBEROS", null);
}};
@@ -555,8 +554,7 @@ public class UpgradeCatalog210Test {
final Map<String, String> propertiesExpectedHiveEnv = new HashMap<String, String>() {{
put("hive_security_authorization", "none");
}};
- final Map<String, String> propertiesExpectedHiveSite = new HashMap<String, String>() {{
- }};
+ final Map<String, String> propertiesExpectedHiveSite = new HashMap<>();
final Map<String, String> propertiesExpectedPluginProperies = new HashMap<String, String>() {{
put("ranger-hive-plugin-enabled", "yes");
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog221Test.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog221Test.java b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog221Test.java
index 27b1174..55ae3ff 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog221Test.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog221Test.java
@@ -478,10 +478,7 @@ public class UpgradeCatalog221Test {
}
};
- Map<String, String> newPropertiesAmsHbaseSecuritySite = new HashMap<String, String>() {
- {
- }
- };
+ Map<String, String> newPropertiesAmsHbaseSecuritySite = new HashMap<>();
EasyMockSupport easyMockSupport = new EasyMockSupport();
Clusters clusters = easyMockSupport.createNiceMock(Clusters.class);
http://git-wip-us.apache.org/repos/asf/ambari/blob/4fc52a61/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog250Test.java
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog250Test.java b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog250Test.java
index 24b4570..8568a19 100644
--- a/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog250Test.java
+++ b/ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog250Test.java
@@ -18,14 +18,31 @@
package org.apache.ambari.server.upgrade;
-import com.google.common.collect.Maps;
-import com.google.gson.Gson;
-import com.google.inject.Binder;
-import com.google.inject.Guice;
-import com.google.inject.Injector;
-import com.google.inject.Module;
-import com.google.inject.Provider;
-import junit.framework.Assert;
+import static org.easymock.EasyMock.anyObject;
+import static org.easymock.EasyMock.anyString;
+import static org.easymock.EasyMock.capture;
+import static org.easymock.EasyMock.createMockBuilder;
+import static org.easymock.EasyMock.eq;
+import static org.easymock.EasyMock.expect;
+import static org.easymock.EasyMock.expectLastCall;
+import static org.easymock.EasyMock.newCapture;
+import static org.easymock.EasyMock.replay;
+import static org.easymock.EasyMock.reset;
+import static org.easymock.EasyMock.verify;
+import static org.junit.Assert.assertTrue;
+
+import java.lang.reflect.Method;
+import java.sql.Connection;
+import java.sql.ResultSet;
+import java.sql.SQLException;
+import java.sql.Statement;
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+import javax.persistence.EntityManager;
import org.apache.ambari.server.AmbariException;
import org.apache.ambari.server.actionmanager.ActionManager;
@@ -53,34 +70,18 @@ import org.easymock.EasyMockSupport;
import org.easymock.Mock;
import org.easymock.MockType;
import org.junit.After;
+import org.junit.Assert;
import org.junit.Before;
import org.junit.Test;
import org.junit.runner.RunWith;
-import javax.persistence.EntityManager;
-import java.lang.reflect.Method;
-import java.sql.Connection;
-import java.sql.ResultSet;
-import java.sql.SQLException;
-import java.sql.Statement;
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-
-import static org.easymock.EasyMock.anyObject;
-import static org.easymock.EasyMock.anyString;
-import static org.easymock.EasyMock.capture;
-import static org.easymock.EasyMock.createMockBuilder;
-import static org.easymock.EasyMock.eq;
-import static org.easymock.EasyMock.expect;
-import static org.easymock.EasyMock.expectLastCall;
-import static org.easymock.EasyMock.newCapture;
-import static org.easymock.EasyMock.replay;
-import static org.easymock.EasyMock.reset;
-import static org.easymock.EasyMock.verify;
-import static org.junit.Assert.assertTrue;
+import com.google.common.collect.Maps;
+import com.google.gson.Gson;
+import com.google.inject.Binder;
+import com.google.inject.Guice;
+import com.google.inject.Injector;
+import com.google.inject.Module;
+import com.google.inject.Provider;
/**
* {@link UpgradeCatalog250} unit tests.