You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ozone.apache.org by ad...@apache.org on 2022/06/13 20:25:48 UTC
[ozone] branch master updated: HDDS-6874. Migrate simple tests in hdds-common to JUnit5 (#3507)
This is an automated email from the ASF dual-hosted git repository.
adoroszlai pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/ozone.git
The following commit(s) were added to refs/heads/master by this push:
new b1fc69e059 HDDS-6874. Migrate simple tests in hdds-common to JUnit5 (#3507)
b1fc69e059 is described below
commit b1fc69e05903af49154115fc503bbfd593d2b114
Author: Kaijie Chen <ch...@kaijie.org>
AuthorDate: Tue Jun 14 04:25:42 2022 +0800
HDDS-6874. Migrate simple tests in hdds-common to JUnit5 (#3507)
---
.../java/org/apache/hadoop/hdds/TestHddsUtils.java | 67 +++++++-------
.../hdds/client/TestECReplicationConfig.java | 32 +++----
.../client/TestReplicationConfigValidator.java | 9 +-
.../conf/TestGeneratedConfigurationOverwrite.java | 18 ++--
.../hdds/fs/TestCachingSpaceUsageSource.java | 2 +-
.../java/org/apache/hadoop/hdds/fs/TestDU.java | 10 +-
.../org/apache/hadoop/hdds/fs/TestDUFactory.java | 2 +-
.../hdds/fs/TestDedicatedDiskSpaceUsage.java | 12 +--
.../fs/TestDedicatedDiskSpaceUsageFactory.java | 2 +-
.../hadoop/hdds/fs/TestSaveSpaceUsageToFile.java | 16 ++--
.../hadoop/hdds/fs/TestSpaceUsageFactory.java | 6 +-
.../hadoop/hdds/protocol/TestDatanodeDetails.java | 8 +-
.../ratis/TestContainerCommandRequestMessage.java | 6 +-
.../apache/hadoop/hdds/ratis/TestRatisHelper.java | 38 ++++----
.../hdds/ratis/TestServerNotLeaderException.java | 16 ++--
.../hdds/scm/container/TestContainerInfo.java | 30 +++---
.../scm/container/TestContainerReplicaInfo.java | 36 ++++----
.../container/TestReplicationManagerReport.java | 102 +++++++++++----------
.../container/common/helpers/TestExcludeList.java | 14 +--
.../apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java | 51 ++++++-----
.../apache/hadoop/hdds/scm/net/TestNetUtils.java | 2 +-
.../hadoop/hdds/scm/pipeline/TestPipeline.java | 42 ++++-----
.../hadoop/hdds/tracing/TestTraceAllMethod.java | 6 +-
.../hadoop/hdds/tracing/TestTracingUtil.java | 4 +-
.../hdds/upgrade/TestHDDSLayoutVersionManager.java | 14 +--
.../hadoop/hdds/utils/TestHddsIdFactory.java | 12 +--
.../hadoop/hdds/utils/TestResourceLimitCache.java | 33 ++++---
.../hadoop/hdds/utils/TestResourceSemaphore.java | 28 +++---
.../hadoop/hdds/utils/TestRetriableTask.java | 4 +-
.../org/apache/hadoop/ozone/TestOzoneConsts.java | 6 +-
.../hadoop/ozone/audit/TestOzoneAuditLogger.java | 24 ++---
.../apache/hadoop/ozone/common/TestChecksum.java | 19 ++--
.../ozone/common/TestChecksumByteBuffer.java | 6 +-
.../common/TestChecksumImplsComputeSameValues.java | 4 +-
.../hadoop/ozone/common/TestChunkBuffer.java | 58 ++++++------
.../TestChunkBufferImplWithByteBufferList.java | 2 +-
.../apache/hadoop/ozone/ha/TestOzoneNetUtils.java | 8 +-
.../apache/hadoop/ozone/lock/TestLockManager.java | 37 ++++----
.../ozone/upgrade/TestBasicUpgradeFinalizer.java | 8 +-
.../TestDefaultUpgradeFinalizationExecutor.java | 2 +-
.../upgrade/TestLayoutVersionInstanceFactory.java | 6 +-
41 files changed, 405 insertions(+), 397 deletions(-)
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java
index 67001010d5..3dbee3efbd 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java
@@ -38,12 +38,12 @@ import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DATANODE_PORT_D
import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DATANODE_PORT_KEY;
import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT;
-import static org.hamcrest.core.Is.is;
-import org.junit.Assert;
-import static org.junit.Assert.assertThat;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
+import static org.junit.jupiter.api.Assertions.fail;
/**
* Testing HddsUtils.
@@ -52,13 +52,13 @@ public class TestHddsUtils {
@Test
public void testGetHostName() {
- Assert.assertEquals(Optional.of("localhost"),
+ assertEquals(Optional.of("localhost"),
HddsUtils.getHostName("localhost:1234"));
- Assert.assertEquals(Optional.of("localhost"),
+ assertEquals(Optional.of("localhost"),
HddsUtils.getHostName("localhost"));
- Assert.assertEquals(Optional.empty(),
+ assertEquals(Optional.empty(),
HddsUtils.getHostName(":1234"));
}
@@ -91,26 +91,26 @@ public class TestHddsUtils {
// Verify valid IP address setup
conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, "1.2.3.4");
addresses = getSCMAddressForDatanodes(conf);
- assertThat(addresses.size(), is(1));
+ assertEquals(1, addresses.size());
addr = addresses.iterator().next();
- assertThat(addr.getHostName(), is("1.2.3.4"));
- assertThat(addr.getPort(), is(OZONE_SCM_DATANODE_PORT_DEFAULT));
+ assertEquals("1.2.3.4", addr.getHostName());
+ assertEquals(OZONE_SCM_DATANODE_PORT_DEFAULT, addr.getPort());
// Verify valid hostname setup
conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, "scm1");
addresses = getSCMAddressForDatanodes(conf);
- assertThat(addresses.size(), is(1));
+ assertEquals(1, addresses.size());
addr = addresses.iterator().next();
- assertThat(addr.getHostName(), is("scm1"));
- assertThat(addr.getPort(), is(OZONE_SCM_DATANODE_PORT_DEFAULT));
+ assertEquals("scm1", addr.getHostName());
+ assertEquals(OZONE_SCM_DATANODE_PORT_DEFAULT, addr.getPort());
// Verify valid hostname and port
conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, "scm1:1234");
addresses = getSCMAddressForDatanodes(conf);
- assertThat(addresses.size(), is(1));
+ assertEquals(1, addresses.size());
addr = addresses.iterator().next();
- assertThat(addr.getHostName(), is("scm1"));
- assertThat(addr.getPort(), is(1234));
+ assertEquals("scm1", addr.getHostName());
+ assertEquals(1234, addr.getPort());
final Map<String, Integer> hostsAndPorts = new HashMap<>();
hostsAndPorts.put("scm1", 1234);
@@ -121,7 +121,7 @@ public class TestHddsUtils {
conf.setStrings(
ScmConfigKeys.OZONE_SCM_NAMES, "scm1:1234,scm2:2345,scm3:3456");
addresses = getSCMAddressForDatanodes(conf);
- assertThat(addresses.size(), is(3));
+ assertEquals(3, addresses.size());
it = addresses.iterator();
HashMap<String, Integer> expected1 = new HashMap<>(hostsAndPorts);
while (it.hasNext()) {
@@ -135,7 +135,7 @@ public class TestHddsUtils {
conf.setStrings(
ScmConfigKeys.OZONE_SCM_NAMES, " scm1:1234, scm2:2345 , scm3:3456 ");
addresses = getSCMAddressForDatanodes(conf);
- assertThat(addresses.size(), is(3));
+ assertEquals(3, addresses.size());
it = addresses.iterator();
HashMap<String, Integer> expected2 = new HashMap<>(hostsAndPorts);
while (it.hasNext()) {
@@ -205,16 +205,14 @@ public class TestHddsUtils {
Collection<InetSocketAddress> scmAddressList =
HddsUtils.getSCMAddressForDatanodes(conf);
- Assert.assertNotNull(scmAddressList);
- Assert.assertEquals(3, scmAddressList.size());
+ Assertions.assertNotNull(scmAddressList);
+ assertEquals(3, scmAddressList.size());
- Iterator<InetSocketAddress> it = scmAddressList.iterator();
- while (it.hasNext()) {
- InetSocketAddress next = it.next();
- expected.remove(next.getHostName() + ":" + next.getPort());
+ for (InetSocketAddress next : scmAddressList) {
+ expected.remove(next.getHostName() + ":" + next.getPort());
}
- Assert.assertTrue(expected.size() == 0);
+ assertEquals(0, expected.size());
}
@@ -228,16 +226,14 @@ public class TestHddsUtils {
OzoneConfiguration conf = new OzoneConfiguration();
conf.set(ScmConfigKeys.OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT,
testnum1);
- Assert.assertTrue(Integer.parseInt(testnum1) ==
- HddsUtils.getNumberFromConfigKeys(
- conf,
+ assertEquals(Integer.parseInt(testnum1),
+ HddsUtils.getNumberFromConfigKeys(conf,
OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT).orElse(0));
/* Test to return first unempty key number from list */
/* first key is absent */
- Assert.assertTrue(Integer.parseInt(testnum1) ==
- HddsUtils.getNumberFromConfigKeys(
- conf,
+ assertEquals(Integer.parseInt(testnum1),
+ HddsUtils.getNumberFromConfigKeys(conf,
ConfUtils.addKeySuffixes(OZONE_SCM_DATANODE_PORT_KEY,
serviceId, nodeId),
OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT).orElse(0));
@@ -246,9 +242,8 @@ public class TestHddsUtils {
conf.set(ConfUtils.addKeySuffixes(OZONE_SCM_DATANODE_PORT_KEY,
serviceId, nodeId),
testnum2);
- Assert.assertTrue(Integer.parseInt(testnum2) ==
- HddsUtils.getNumberFromConfigKeys(
- conf,
+ assertEquals(Integer.parseInt(testnum2),
+ HddsUtils.getNumberFromConfigKeys(conf,
ConfUtils.addKeySuffixes(OZONE_SCM_DATANODE_PORT_KEY,
serviceId, nodeId),
OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT).orElse(0));
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java
index 9440a6e026..333328bd74 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java
@@ -18,15 +18,15 @@
package org.apache.hadoop.hdds.client;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import java.util.HashMap;
import java.util.Map;
import static org.apache.hadoop.hdds.client.ECReplicationConfig.EcCodec.RS;
import static org.apache.hadoop.hdds.client.ECReplicationConfig.EcCodec.XOR;
-import static org.junit.Assert.fail;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
/**
* Unit test for ECReplicationConfig.
@@ -45,10 +45,10 @@ public class TestECReplicationConfig {
for (Map.Entry<String, ECReplicationConfig> e : valid.entrySet()) {
ECReplicationConfig ec = new ECReplicationConfig(e.getKey());
- Assert.assertEquals(e.getValue().getData(), ec.getData());
- Assert.assertEquals(e.getValue().getParity(), ec.getParity());
- Assert.assertEquals(e.getValue().getCodec(), ec.getCodec());
- Assert.assertEquals(e.getValue().getEcChunkSize(), ec.getEcChunkSize());
+ assertEquals(e.getValue().getData(), ec.getData());
+ assertEquals(e.getValue().getParity(), ec.getParity());
+ assertEquals(e.getValue().getCodec(), ec.getCodec());
+ assertEquals(e.getValue().getEcChunkSize(), ec.getEcChunkSize());
}
}
@@ -63,12 +63,8 @@ public class TestECReplicationConfig {
"x3-2"
};
for (String s : invalid) {
- try {
- new ECReplicationConfig(s);
- fail(s + " should not parse correctly");
- } catch (IllegalArgumentException e) {
- // ignore, this expected
- }
+ assertThrows(IllegalArgumentException.class,
+ () -> new ECReplicationConfig(s));
}
}
@@ -81,11 +77,11 @@ public class TestECReplicationConfig {
HddsProtos.ECReplicationConfig proto = orig.toProto();
ECReplicationConfig recovered = new ECReplicationConfig(proto);
- Assert.assertEquals(orig.getData(), recovered.getData());
- Assert.assertEquals(orig.getParity(), recovered.getParity());
- Assert.assertEquals(orig.getCodec(), recovered.getCodec());
- Assert.assertEquals(orig.getEcChunkSize(), recovered.getEcChunkSize());
- Assert.assertTrue(orig.equals(recovered));
+ assertEquals(orig.getData(), recovered.getData());
+ assertEquals(orig.getParity(), recovered.getParity());
+ assertEquals(orig.getCodec(), recovered.getCodec());
+ assertEquals(orig.getEcChunkSize(), recovered.getEcChunkSize());
+ assertEquals(orig, recovered);
}
}
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java
index 96e49e9af0..320069c060 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java
@@ -19,10 +19,11 @@ package org.apache.hadoop.hdds.client;
import org.apache.hadoop.hdds.conf.InMemoryConfiguration;
import org.apache.hadoop.hdds.conf.MutableConfigurationSource;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.ONE;
import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.THREE;
+import static org.junit.jupiter.api.Assertions.assertThrows;
/**
* Test ReplicationConfig validator.
@@ -56,7 +57,7 @@ public class TestReplicationConfigValidator {
}
- @Test(expected = IllegalArgumentException.class)
+ @Test
public void testCustomValidation() {
MutableConfigurationSource config = new InMemoryConfiguration();
config.set("ozone.replication.allowed-configs", "RATIS/THREE");
@@ -66,8 +67,8 @@ public class TestReplicationConfigValidator {
validator.validate(RatisReplicationConfig.getInstance(THREE));
- validator.validate(RatisReplicationConfig.getInstance(ONE));
- //exception is expected
+ assertThrows(IllegalArgumentException.class,
+ () -> validator.validate(RatisReplicationConfig.getInstance(ONE)));
}
}
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java
index f00107b014..03fd2be0b9 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java
@@ -17,10 +17,10 @@
*/
package org.apache.hadoop.hdds.conf;
-import org.junit.After;
-import org.junit.Assert;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import java.nio.file.Files;
import java.nio.file.Path;
@@ -40,13 +40,13 @@ public class TestGeneratedConfigurationOverwrite {
private OzoneConfiguration conf;
- @Before
+ @BeforeEach
public void overwriteConfigFile() throws Exception {
Files.move(generatedConfigurationPath, generatedConfigurationPathBak);
conf = new OzoneConfiguration();
}
- @After
+ @AfterEach
public void recoverConfigFile() throws Exception {
Files.move(generatedConfigurationPathBak, generatedConfigurationPath);
}
@@ -54,13 +54,13 @@ public class TestGeneratedConfigurationOverwrite {
@Test
public void getConfigurationObject() {
// Check Config Type of String
- Assert.assertNotNull(
+ Assertions.assertNotNull(
conf.getObject(SimpleConfiguration.class).getBindHost());
// Check Config Type of Int
- Assert.assertNotEquals(
+ Assertions.assertNotEquals(
conf.getObject(SimpleConfiguration.class).getPort(), 0);
// Check Config Type of Time
- Assert.assertNotEquals(
+ Assertions.assertNotEquals(
conf.getObject(SimpleConfiguration.class).getWaitTime(), 0);
}
}
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
index 28da73b2e7..9c701ca1fc 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
@@ -19,7 +19,7 @@ package org.apache.hadoop.hdds.fs;
import org.apache.commons.lang3.RandomUtils;
import org.apache.hadoop.hdds.fs.MockSpaceUsageCheckParams.Builder;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import org.mockito.stubbing.Answer;
import java.io.File;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java
index c9ed258f24..0d956a18ee 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java
@@ -19,9 +19,9 @@ package org.apache.hadoop.hdds.fs;
import org.apache.hadoop.fs.FileUtil;
import org.apache.hadoop.util.Shell;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import static org.apache.hadoop.ozone.OzoneConsts.KB;
import static org.apache.ozone.test.GenericTestUtils.getTestDir;
@@ -41,14 +41,14 @@ public class TestDU {
private static final File DIR = getTestDir(TestDU.class.getSimpleName());
- @Before
+ @BeforeEach
public void setUp() {
assumeFalse(Shell.WINDOWS);
FileUtil.fullyDelete(DIR);
assertTrue(DIR.mkdirs());
}
- @After
+ @AfterEach
public void tearDown() throws IOException {
FileUtil.fullyDelete(DIR);
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java
index f9b07b47eb..6e603f8ff0 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java
@@ -21,7 +21,7 @@ import java.io.File;
import java.time.Duration;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import static org.apache.ozone.test.GenericTestUtils.getTestDir;
import static org.junit.jupiter.api.Assertions.assertEquals;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java
index 762639f413..3fc7c4eddf 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java
@@ -18,16 +18,16 @@
package org.apache.hadoop.hdds.fs;
import org.apache.hadoop.fs.FileUtil;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import java.io.File;
import java.io.IOException;
import static org.apache.hadoop.hdds.fs.TestDU.createFile;
import static org.apache.ozone.test.GenericTestUtils.getTestDir;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertTrue;
/**
* Tests for {@link DedicatedDiskSpaceUsage}.
@@ -39,13 +39,13 @@ public class TestDedicatedDiskSpaceUsage {
private static final int FILE_SIZE = 1024;
- @Before
+ @BeforeEach
public void setUp() {
FileUtil.fullyDelete(DIR);
assertTrue(DIR.mkdirs());
}
- @After
+ @AfterEach
public void tearDown() throws IOException {
FileUtil.fullyDelete(DIR);
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java
index 4244edd82d..0142ee56af 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java
@@ -24,7 +24,7 @@ import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import static org.apache.hadoop.hdds.fs.DedicatedDiskSpaceUsageFactory.Conf.configKeyForRefreshPeriod;
import static org.apache.ozone.test.GenericTestUtils.getTestDir;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertSame;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java
index 9480992013..f35e697508 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java
@@ -19,9 +19,9 @@ package org.apache.hadoop.hdds.fs;
import org.apache.commons.io.FileUtils;
import org.apache.hadoop.fs.FileUtil;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import java.io.File;
import java.io.IOException;
@@ -32,9 +32,9 @@ import java.util.OptionalLong;
import static org.apache.ozone.test.GenericTestUtils.getTestDir;
import static org.apache.ozone.test.GenericTestUtils.waitFor;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
/**
* Tests for {@link SaveSpaceUsageToFile}.
@@ -51,14 +51,14 @@ public class TestSaveSpaceUsageToFile {
private File file;
- @Before
+ @BeforeEach
public void setup() {
FileUtil.fullyDelete(DIR);
assertTrue(DIR.mkdirs());
file = new File(DIR, "space_usage.txt");
}
- @After
+ @AfterEach
public void cleanup() {
FileUtil.fullyDelete(DIR);
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java
index c3ae128424..5a6e98e229 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java
@@ -24,8 +24,8 @@ import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.ozone.test.GenericTestUtils.LogCapturer;
import static org.apache.hadoop.hdds.fs.SpaceUsageCheckFactory.Conf.configKeyForClassName;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertNotNull;
import static org.junit.jupiter.api.Assertions.assertSame;
@@ -58,7 +58,7 @@ public class TestSpaceUsageFactory {
return factoryClass.cast(factory);
}
- @Before
+ @BeforeEach
public void setUp() {
capturer = LogCapturer.captureLogs(
LoggerFactory.getLogger(SpaceUsageCheckFactory.class));
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java
index d1fcad7d56..e77b97a854 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java
@@ -19,7 +19,7 @@ package org.apache.hadoop.hdds.protocol;
import org.apache.hadoop.hdds.protocol.DatanodeDetails.Port;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import java.util.Set;
@@ -27,9 +27,9 @@ import static org.apache.hadoop.hdds.protocol.DatanodeDetails.Port.Name.ALL_PORT
import static org.apache.hadoop.hdds.protocol.DatanodeDetails.Port.Name.V0_PORTS;
import static org.apache.hadoop.ozone.ClientVersion.DEFAULT_VERSION;
import static org.apache.hadoop.ozone.ClientVersion.VERSION_HANDLES_UNKNOWN_DN_PORTS;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
+import static org.junit.jupiter.api.Assertions.fail;
/**
* Test for {@link DatanodeDetails}.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java
index d3ddbe0ef3..75e19f5d45 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java
@@ -36,8 +36,8 @@ import org.apache.hadoop.ozone.common.ChecksumData;
import org.apache.hadoop.ozone.common.OzoneChecksumException;
import org.apache.ratis.thirdparty.com.google.protobuf.ByteString;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
/** Testing {@link ContainerCommandRequestMessage}. */
public class TestContainerCommandRequestMessage {
@@ -150,6 +150,6 @@ public class TestContainerCommandRequestMessage {
= ContainerCommandRequestMessage.toMessage(original, null);
final ContainerCommandRequestProto computed
= ContainerCommandRequestMessage.toProto(message.getContent(), null);
- Assert.assertEquals(original, computed);
+ Assertions.assertEquals(original, computed);
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java
index 68fb0ffc47..d8f626165e 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hdds.ratis;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.ratis.conf.RaftProperties;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
/**
* Test RatisHelper class.
@@ -42,11 +42,11 @@ public class TestRatisHelper {
RaftProperties raftProperties = new RaftProperties();
RatisHelper.createRaftClientProperties(ozoneConfiguration, raftProperties);
- Assert.assertEquals("30s",
+ Assertions.assertEquals("30s",
raftProperties.get("raft.client.rpc.watch.request.timeout"));
- Assert.assertEquals("30s",
+ Assertions.assertEquals("30s",
raftProperties.get("raft.client.rpc.request.timeout"));
- Assert.assertNull(
+ Assertions.assertNull(
raftProperties.get("raft.server.rpc.watch.request.timeout"));
}
@@ -67,16 +67,17 @@ public class TestRatisHelper {
RaftProperties raftProperties = new RaftProperties();
RatisHelper.createRaftClientProperties(ozoneConfiguration, raftProperties);
- Assert.assertEquals("30MB",
+ Assertions.assertEquals("30MB",
raftProperties.get("raft.grpc.message.size.max"));
- Assert.assertEquals("1MB",
+ Assertions.assertEquals("1MB",
raftProperties.get("raft.grpc.flow.control.window"));
// As we dont match tls and server raft.grpc properties. So they should
// be null.
- Assert.assertNull(raftProperties.get("raft.grpc.tls.set"));
- Assert.assertNull(raftProperties.get("raft.grpc.tls.mutual_authn.enabled"));
- Assert.assertNull(raftProperties.get("raft.grpc.server.port"));
+ Assertions.assertNull(raftProperties.get("raft.grpc.tls.set"));
+ Assertions.assertNull(
+ raftProperties.get("raft.grpc.tls.mutual_authn.enabled"));
+ Assertions.assertNull(raftProperties.get("raft.grpc.server.port"));
}
@@ -100,15 +101,15 @@ public class TestRatisHelper {
RatisHelper.createRaftServerProperties(ozoneConfiguration,
raftProperties);
- Assert.assertEquals("30MB",
+ Assertions.assertEquals("30MB",
raftProperties.get("raft.grpc.message.size.max"));
- Assert.assertEquals("1MB",
+ Assertions.assertEquals("1MB",
raftProperties.get("raft.grpc.flow.control.window"));
- Assert.assertEquals("true",
+ Assertions.assertEquals("true",
raftProperties.get("raft.grpc.tls.enabled"));
- Assert.assertEquals("true",
+ Assertions.assertEquals("true",
raftProperties.get("raft.grpc.tls.mutual_authn.enabled"));
- Assert.assertEquals("100",
+ Assertions.assertEquals("100",
raftProperties.get("raft.grpc.server.port"));
}
@@ -127,11 +128,12 @@ public class TestRatisHelper {
RaftProperties raftProperties = new RaftProperties();
RatisHelper.createRaftServerProperties(ozoneConfiguration, raftProperties);
- Assert.assertEquals("30s",
+ Assertions.assertEquals("30s",
raftProperties.get("raft.server.rpc.watch.request.timeout"));
- Assert.assertEquals("30s",
+ Assertions.assertEquals("30s",
raftProperties.get("raft.server.rpc.request.timeout"));
- Assert.assertNull(raftProperties.get("raft.client.rpc.request.timeout"));
+ Assertions.assertNull(
+ raftProperties.get("raft.client.rpc.request.timeout"));
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java
index 3bdcf85cdf..7d44d060fd 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java
@@ -19,8 +19,8 @@
package org.apache.hadoop.hdds.ratis;
import edu.umd.cs.findbugs.annotations.SuppressFBWarnings;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
/** Class to test {@link ServerNotLeaderException} parsing. **/
@@ -34,7 +34,7 @@ public class TestServerNotLeaderException {
"Server:cf0bc565-a41b-4784-a24d-3048d5a5b013 is not the leader. "
+ "Suggested leader is Server:scm5-3.scm5.root.hwx.site:9863";
ServerNotLeaderException snle = new ServerNotLeaderException(msg);
- Assert.assertEquals(snle.getSuggestedLeader(), "scm5-3.scm5.root.hwx" +
+ Assertions.assertEquals(snle.getSuggestedLeader(), "scm5-3.scm5.root.hwx" +
".site:9863");
String message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -42,7 +42,7 @@ public class TestServerNotLeaderException {
"at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
".convertToNotLeaderException(ServerNotLeaderException.java:96)";
snle = new ServerNotLeaderException(message);
- Assert.assertEquals("scm5-3.scm5.root.hwx.site:9863",
+ Assertions.assertEquals("scm5-3.scm5.root.hwx.site:9863",
snle.getSuggestedLeader());
// Test hostname with out "."
@@ -51,7 +51,7 @@ public class TestServerNotLeaderException {
"at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
".convertToNotLeaderException(ServerNotLeaderException.java:96)";
snle = new ServerNotLeaderException(message);
- Assert.assertEquals("localhost:98634",
+ Assertions.assertEquals("localhost:98634",
snle.getSuggestedLeader());
message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -59,7 +59,7 @@ public class TestServerNotLeaderException {
"at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
".convertToNotLeaderException(ServerNotLeaderException.java:96)";
snle = new ServerNotLeaderException(message);
- Assert.assertEquals(null,
+ Assertions.assertEquals(null,
snle.getSuggestedLeader());
message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -67,7 +67,7 @@ public class TestServerNotLeaderException {
"at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
".convertToNotLeaderException(ServerNotLeaderException.java:96)";
snle = new ServerNotLeaderException(message);
- Assert.assertEquals("localhost:98634",
+ Assertions.assertEquals("localhost:98634",
snle.getSuggestedLeader());
message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -75,7 +75,7 @@ public class TestServerNotLeaderException {
"at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
".convertToNotLeaderException(ServerNotLeaderException.java)";
snle = new ServerNotLeaderException(message);
- Assert.assertEquals(null,
+ Assertions.assertEquals(null,
snle.getSuggestedLeader());
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java
index 9ec19bf327..dfc5a0c7a9 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java
@@ -22,8 +22,8 @@ import org.apache.hadoop.hdds.client.RatisReplicationConfig;
import org.apache.hadoop.hdds.client.ReplicationConfig;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
import org.apache.hadoop.hdds.scm.pipeline.PipelineID;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
import java.io.IOException;
@@ -44,34 +44,34 @@ public class TestContainerInfo {
HddsProtos.ContainerInfoProto proto = container.getProtobuf();
// No EC Config
- Assert.assertFalse(proto.hasEcReplicationConfig());
- Assert.assertEquals(THREE, proto.getReplicationFactor());
- Assert.assertEquals(RATIS, proto.getReplicationType());
+ Assertions.assertFalse(proto.hasEcReplicationConfig());
+ Assertions.assertEquals(THREE, proto.getReplicationFactor());
+ Assertions.assertEquals(RATIS, proto.getReplicationType());
// Reconstruct object from Proto
ContainerInfo recovered = ContainerInfo.fromProtobuf(proto);
- Assert.assertEquals(RATIS, recovered.getReplicationType());
- Assert.assertTrue(
+ Assertions.assertEquals(RATIS, recovered.getReplicationType());
+ Assertions.assertTrue(
recovered.getReplicationConfig() instanceof RatisReplicationConfig);
// EC Config
container = createContainerInfo(new ECReplicationConfig(3, 2));
proto = container.getProtobuf();
- Assert.assertEquals(3, proto.getEcReplicationConfig().getData());
- Assert.assertEquals(2, proto.getEcReplicationConfig().getParity());
- Assert.assertFalse(proto.hasReplicationFactor());
- Assert.assertEquals(EC, proto.getReplicationType());
+ Assertions.assertEquals(3, proto.getEcReplicationConfig().getData());
+ Assertions.assertEquals(2, proto.getEcReplicationConfig().getParity());
+ Assertions.assertFalse(proto.hasReplicationFactor());
+ Assertions.assertEquals(EC, proto.getReplicationType());
// Reconstruct object from Proto
recovered = ContainerInfo.fromProtobuf(proto);
- Assert.assertEquals(EC, recovered.getReplicationType());
- Assert.assertTrue(
+ Assertions.assertEquals(EC, recovered.getReplicationType());
+ Assertions.assertTrue(
recovered.getReplicationConfig() instanceof ECReplicationConfig);
ECReplicationConfig config =
(ECReplicationConfig)recovered.getReplicationConfig();
- Assert.assertEquals(3, config.getData());
- Assert.assertEquals(2, config.getParity());
+ Assertions.assertEquals(3, config.getData());
+ Assertions.assertEquals(2, config.getParity());
}
private ContainerInfo createContainerInfo(ReplicationConfig repConfig) {
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java
index 4f63ea7240..5203b98ef3 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hdds.scm.container;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.MockDatanodeDetails;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
import java.util.UUID;
@@ -46,17 +46,17 @@ public class TestContainerReplicaInfo {
ContainerReplicaInfo info = ContainerReplicaInfo.fromProto(proto);
- Assert.assertEquals(proto.getContainerID(), info.getContainerID());
- Assert.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
- Assert.assertEquals(proto.getKeyCount(), info.getKeyCount());
- Assert.assertEquals(proto.getPlaceOfBirth(),
+ Assertions.assertEquals(proto.getContainerID(), info.getContainerID());
+ Assertions.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
+ Assertions.assertEquals(proto.getKeyCount(), info.getKeyCount());
+ Assertions.assertEquals(proto.getPlaceOfBirth(),
info.getPlaceOfBirth().toString());
- Assert.assertEquals(DatanodeDetails.getFromProtoBuf(
+ Assertions.assertEquals(DatanodeDetails.getFromProtoBuf(
proto.getDatanodeDetails()), info.getDatanodeDetails());
- Assert.assertEquals(proto.getSequenceID(), info.getSequenceId());
- Assert.assertEquals(proto.getState(), info.getState());
+ Assertions.assertEquals(proto.getSequenceID(), info.getSequenceId());
+ Assertions.assertEquals(proto.getState(), info.getState());
// If replicaIndex is not in the proto, then -1 should be returned
- Assert.assertEquals(-1, info.getReplicaIndex());
+ Assertions.assertEquals(-1, info.getReplicaIndex());
}
@Test
@@ -76,15 +76,15 @@ public class TestContainerReplicaInfo {
ContainerReplicaInfo info = ContainerReplicaInfo.fromProto(proto);
- Assert.assertEquals(proto.getContainerID(), info.getContainerID());
- Assert.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
- Assert.assertEquals(proto.getKeyCount(), info.getKeyCount());
- Assert.assertEquals(proto.getPlaceOfBirth(),
+ Assertions.assertEquals(proto.getContainerID(), info.getContainerID());
+ Assertions.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
+ Assertions.assertEquals(proto.getKeyCount(), info.getKeyCount());
+ Assertions.assertEquals(proto.getPlaceOfBirth(),
info.getPlaceOfBirth().toString());
- Assert.assertEquals(DatanodeDetails.getFromProtoBuf(
+ Assertions.assertEquals(DatanodeDetails.getFromProtoBuf(
proto.getDatanodeDetails()), info.getDatanodeDetails());
- Assert.assertEquals(proto.getSequenceID(), info.getSequenceId());
- Assert.assertEquals(proto.getState(), info.getState());
- Assert.assertEquals(4, info.getReplicaIndex());
+ Assertions.assertEquals(proto.getSequenceID(), info.getSequenceId());
+ Assertions.assertEquals(proto.getState(), info.getState());
+ Assertions.assertEquals(4, info.getReplicaIndex());
}
}
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java
index a68f70c871..ada60397ef 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java
@@ -21,9 +21,8 @@ import com.fasterxml.jackson.databind.JsonNode;
import com.fasterxml.jackson.databind.ObjectMapper;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
import org.apache.hadoop.hdds.server.JsonUtils;
-import org.junit.Assert;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import java.io.IOException;
import java.util.ArrayList;
@@ -32,6 +31,9 @@ import java.util.Random;
import java.util.concurrent.ThreadLocalRandom;
import static com.fasterxml.jackson.databind.node.JsonNodeType.ARRAY;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
+import static org.junit.jupiter.api.Assertions.assertTrue;
/**
* Tests for the ReplicationManagerReport class.
@@ -40,7 +42,7 @@ public class TestReplicationManagerReport {
private ReplicationManagerReport report;
- @Before
+ @BeforeEach
public void setup() {
report = new ReplicationManagerReport();
}
@@ -55,18 +57,18 @@ public class TestReplicationManagerReport {
report.increment(HddsProtos.LifeCycleState.CLOSED);
report.increment(HddsProtos.LifeCycleState.CLOSED);
- Assert.assertEquals(2,
+ assertEquals(2,
report.getStat(ReplicationManagerReport.HealthState.UNDER_REPLICATED));
- Assert.assertEquals(1,
+ assertEquals(1,
report.getStat(ReplicationManagerReport.HealthState.OVER_REPLICATED));
- Assert.assertEquals(0,
+ assertEquals(0,
report.getStat(ReplicationManagerReport.HealthState.MIS_REPLICATED));
- Assert.assertEquals(1,
+ assertEquals(1,
report.getStat(HddsProtos.LifeCycleState.OPEN));
- Assert.assertEquals(2,
+ assertEquals(2,
report.getStat(HddsProtos.LifeCycleState.CLOSED));
- Assert.assertEquals(0,
+ assertEquals(0,
report.getStat(HddsProtos.LifeCycleState.QUASI_CLOSED));
}
@@ -93,29 +95,29 @@ public class TestReplicationManagerReport {
ObjectMapper mapper = new ObjectMapper();
JsonNode json = mapper.readTree(jsonString);
- Assert.assertTrue(json.get("reportTimeStamp").longValue() > 0);
+ assertTrue(json.get("reportTimeStamp").longValue() > 0);
JsonNode stats = json.get("stats");
- Assert.assertEquals(1, stats.get("OPEN").longValue());
- Assert.assertEquals(0, stats.get("CLOSING").longValue());
- Assert.assertEquals(0, stats.get("QUASI_CLOSED").longValue());
- Assert.assertEquals(2, stats.get("CLOSED").longValue());
- Assert.assertEquals(0, stats.get("DELETING").longValue());
- Assert.assertEquals(0, stats.get("DELETED").longValue());
-
- Assert.assertEquals(2, stats.get("UNDER_REPLICATED").longValue());
- Assert.assertEquals(1, stats.get("OVER_REPLICATED").longValue());
- Assert.assertEquals(0, stats.get("MIS_REPLICATED").longValue());
- Assert.assertEquals(0, stats.get("MISSING").longValue());
- Assert.assertEquals(0, stats.get("UNHEALTHY").longValue());
- Assert.assertEquals(0, stats.get("EMPTY").longValue());
- Assert.assertEquals(0, stats.get("OPEN_UNHEALTHY").longValue());
- Assert.assertEquals(0, stats.get("QUASI_CLOSED_STUCK").longValue());
+ assertEquals(1, stats.get("OPEN").longValue());
+ assertEquals(0, stats.get("CLOSING").longValue());
+ assertEquals(0, stats.get("QUASI_CLOSED").longValue());
+ assertEquals(2, stats.get("CLOSED").longValue());
+ assertEquals(0, stats.get("DELETING").longValue());
+ assertEquals(0, stats.get("DELETED").longValue());
+
+ assertEquals(2, stats.get("UNDER_REPLICATED").longValue());
+ assertEquals(1, stats.get("OVER_REPLICATED").longValue());
+ assertEquals(0, stats.get("MIS_REPLICATED").longValue());
+ assertEquals(0, stats.get("MISSING").longValue());
+ assertEquals(0, stats.get("UNHEALTHY").longValue());
+ assertEquals(0, stats.get("EMPTY").longValue());
+ assertEquals(0, stats.get("OPEN_UNHEALTHY").longValue());
+ assertEquals(0, stats.get("QUASI_CLOSED_STUCK").longValue());
JsonNode samples = json.get("samples");
- Assert.assertEquals(ARRAY, samples.get("UNDER_REPLICATED").getNodeType());
- Assert.assertEquals(1, samples.get("UNDER_REPLICATED").get(0).longValue());
- Assert.assertEquals(2, samples.get("UNDER_REPLICATED").get(1).longValue());
- Assert.assertEquals(3, samples.get("OVER_REPLICATED").get(0).longValue());
+ assertEquals(ARRAY, samples.get("UNDER_REPLICATED").getNodeType());
+ assertEquals(1, samples.get("UNDER_REPLICATED").get(0).longValue());
+ assertEquals(2, samples.get("UNDER_REPLICATED").get(1).longValue());
+ assertEquals(3, samples.get("OVER_REPLICATED").get(0).longValue());
}
@Test
@@ -130,27 +132,27 @@ public class TestReplicationManagerReport {
ReplicationManagerReport.HealthState.OVER_REPLICATED,
new ContainerID(3));
- Assert.assertEquals(2,
+ assertEquals(2,
report.getStat(ReplicationManagerReport.HealthState.UNDER_REPLICATED));
- Assert.assertEquals(1,
+ assertEquals(1,
report.getStat(ReplicationManagerReport.HealthState.OVER_REPLICATED));
- Assert.assertEquals(0,
+ assertEquals(0,
report.getStat(ReplicationManagerReport.HealthState.MIS_REPLICATED));
List<ContainerID> sample =
report.getSample(ReplicationManagerReport.HealthState.UNDER_REPLICATED);
- Assert.assertEquals(new ContainerID(1), sample.get(0));
- Assert.assertEquals(new ContainerID(2), sample.get(1));
- Assert.assertEquals(2, sample.size());
+ assertEquals(new ContainerID(1), sample.get(0));
+ assertEquals(new ContainerID(2), sample.get(1));
+ assertEquals(2, sample.size());
sample =
report.getSample(ReplicationManagerReport.HealthState.OVER_REPLICATED);
- Assert.assertEquals(new ContainerID(3), sample.get(0));
- Assert.assertEquals(1, sample.size());
+ assertEquals(new ContainerID(3), sample.get(0));
+ assertEquals(1, sample.size());
sample =
report.getSample(ReplicationManagerReport.HealthState.MIS_REPLICATED);
- Assert.assertEquals(0, sample.size());
+ assertEquals(0, sample.size());
}
@Test
@@ -162,9 +164,9 @@ public class TestReplicationManagerReport {
}
List<ContainerID> sample =
report.getSample(ReplicationManagerReport.HealthState.UNDER_REPLICATED);
- Assert.assertEquals(ReplicationManagerReport.SAMPLE_LIMIT, sample.size());
+ assertEquals(ReplicationManagerReport.SAMPLE_LIMIT, sample.size());
for (int i = 0; i < ReplicationManagerReport.SAMPLE_LIMIT; i++) {
- Assert.assertEquals(new ContainerID(i), sample.get(i));
+ assertEquals(new ContainerID(i), sample.get(i));
}
}
@@ -187,16 +189,16 @@ public class TestReplicationManagerReport {
HddsProtos.ReplicationManagerReportProto proto = report.toProtobuf();
ReplicationManagerReport newReport
= ReplicationManagerReport.fromProtobuf(proto);
- Assert.assertEquals(report.getReportTimeStamp(),
+ assertEquals(report.getReportTimeStamp(),
newReport.getReportTimeStamp());
for (HddsProtos.LifeCycleState s : HddsProtos.LifeCycleState.values()) {
- Assert.assertEquals(report.getStat(s), newReport.getStat(s));
+ assertEquals(report.getStat(s), newReport.getStat(s));
}
for (ReplicationManagerReport.HealthState s :
ReplicationManagerReport.HealthState.values()) {
- Assert.assertTrue(report.getSample(s).equals(newReport.getSample(s)));
+ assertEquals(report.getSample(s), newReport.getSample(s));
}
}
@@ -225,21 +227,23 @@ public class TestReplicationManagerReport {
// Ensure no exception is thrown
ReplicationManagerReport newReport =
ReplicationManagerReport.fromProtobuf(proto.build());
- Assert.assertEquals(20, newReport.getStat(
+ assertEquals(20, newReport.getStat(
ReplicationManagerReport.HealthState.UNDER_REPLICATED));
}
- @Test(expected = IllegalStateException.class)
+ @Test
public void testStatCannotBeSetTwice() {
report.setStat(HddsProtos.LifeCycleState.CLOSED.toString(), 10);
- report.setStat(HddsProtos.LifeCycleState.CLOSED.toString(), 10);
+ assertThrows(IllegalStateException.class, () -> report
+ .setStat(HddsProtos.LifeCycleState.CLOSED.toString(), 10));
}
- @Test(expected = IllegalStateException.class)
+ @Test
public void testSampleCannotBeSetTwice() {
List<ContainerID> containers = new ArrayList<>();
containers.add(ContainerID.valueOf(1));
report.setSample(HddsProtos.LifeCycleState.CLOSED.toString(), containers);
- report.setSample(HddsProtos.LifeCycleState.CLOSED.toString(), containers);
+ assertThrows(IllegalStateException.class, () -> report
+ .setSample(HddsProtos.LifeCycleState.CLOSED.toString(), containers));
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java
index d3fb49b3a2..d5330749ae 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.hdds.scm.container.common.helpers;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.ozone.test.TestClock;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
import java.time.Instant;
import java.time.ZoneOffset;
@@ -39,9 +39,9 @@ public class TestExcludeList {
.setIpAddress("127.0.0.1").setHostName("localhost").addPort(
DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
.build());
- Assert.assertTrue(list.getDatanodes().size() == 1);
+ Assertions.assertTrue(list.getDatanodes().size() == 1);
clock.fastForward(11);
- Assert.assertTrue(list.getDatanodes().size() == 0);
+ Assertions.assertTrue(list.getDatanodes().size() == 0);
list.addDatanode(DatanodeDetails.newBuilder().setUuid(UUID.randomUUID())
.setIpAddress("127.0.0.2").setHostName("localhost").addPort(
DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
@@ -50,7 +50,7 @@ public class TestExcludeList {
.setIpAddress("127.0.0.3").setHostName("localhost").addPort(
DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
.build());
- Assert.assertTrue(list.getDatanodes().size() == 2);
+ Assertions.assertTrue(list.getDatanodes().size() == 2);
}
@Test
@@ -60,8 +60,8 @@ public class TestExcludeList {
.setIpAddress("127.0.0.1").setHostName("localhost").addPort(
DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
.build());
- Assert.assertTrue(list.getDatanodes().size() == 1);
+ Assertions.assertTrue(list.getDatanodes().size() == 1);
clock.fastForward(1);
- Assert.assertTrue(list.getDatanodes().size() == 1);
+ Assertions.assertTrue(list.getDatanodes().size() == 1);
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java
index e561bb7ccc..12129f4d3f 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java
@@ -22,9 +22,8 @@ import org.apache.hadoop.hdds.conf.ConfigurationException;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.hdds.scm.ScmConfigKeys;
import org.apache.hadoop.ozone.ha.ConfUtils;
-import org.junit.Assert;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import java.util.List;
@@ -41,6 +40,9 @@ import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DATANODE_PORT_K
import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_SECURITY_SERVICE_ADDRESS_KEY;
import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT;
import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_SECURITY_SERVICE_PORT_KEY;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertNotNull;
+import static org.junit.jupiter.api.Assertions.assertThrows;
/**
* Tests for {@link SCMNodeInfo}.
@@ -51,7 +53,7 @@ public class TestSCMNodeInfo {
private String scmServiceId = "scmserviceId";
private String[] nodes = new String[]{"scm1", "scm2", "scm3"};
- @Before
+ @BeforeEach
public void setup() {
conf.set(ScmConfigKeys.OZONE_SCM_SERVICE_IDS_KEY, scmServiceId);
conf.set(ScmConfigKeys.OZONE_SCM_NODES_KEY + "." + scmServiceId,
@@ -92,15 +94,15 @@ public class TestSCMNodeInfo {
int count = 1;
for (SCMNodeInfo scmNodeInfo : scmNodeInfos) {
- Assert.assertEquals(scmServiceId, scmNodeInfo.getServiceId());
- Assert.assertEquals("scm" + count++, scmNodeInfo.getNodeId());
- Assert.assertEquals("localhost:" + ++port,
+ assertEquals(scmServiceId, scmNodeInfo.getServiceId());
+ assertEquals("scm" + count++, scmNodeInfo.getNodeId());
+ assertEquals("localhost:" + ++port,
scmNodeInfo.getBlockClientAddress());
- Assert.assertEquals("localhost:" + ++port,
+ assertEquals("localhost:" + ++port,
scmNodeInfo.getScmSecurityAddress());
- Assert.assertEquals("localhost:" + ++port,
+ assertEquals("localhost:" + ++port,
scmNodeInfo.getScmClientAddress());
- Assert.assertEquals("localhost:" + ++port,
+ assertEquals("localhost:" + ++port,
scmNodeInfo.getScmDatanodeAddress());
}
}
@@ -116,30 +118,31 @@ public class TestSCMNodeInfo {
int count = 1;
for (SCMNodeInfo scmNodeInfo : scmNodeInfos) {
- Assert.assertEquals(scmServiceId, scmNodeInfo.getServiceId());
- Assert.assertEquals("scm" + count++, scmNodeInfo.getNodeId());
- Assert.assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
+ assertEquals(scmServiceId, scmNodeInfo.getServiceId());
+ assertEquals("scm" + count++, scmNodeInfo.getNodeId());
+ assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
scmNodeInfo.getBlockClientAddress());
- Assert.assertEquals("localhost:" +
+ assertEquals("localhost:" +
OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT,
scmNodeInfo.getScmSecurityAddress());
- Assert.assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
+ assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
scmNodeInfo.getScmClientAddress());
- Assert.assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
+ assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
scmNodeInfo.getScmDatanodeAddress());
}
}
- @Test(expected = ConfigurationException.class)
+ @Test
public void testSCMHANodeInfoWithMissingSCMAddress() {
conf.set(ConfUtils.addKeySuffixes(OZONE_SCM_ADDRESS_KEY,
scmServiceId, "scm1"), "localhost");
conf.set(ConfUtils.addKeySuffixes(OZONE_SCM_ADDRESS_KEY,
scmServiceId, "scm1"), "localhost");
- SCMNodeInfo.buildNodeInfo(conf);
+ assertThrows(ConfigurationException.class,
+ () -> SCMNodeInfo.buildNodeInfo(conf));
}
@Test
@@ -150,15 +153,15 @@ public class TestSCMNodeInfo {
List< SCMNodeInfo > scmNodeInfos = SCMNodeInfo.buildNodeInfo(config);
- Assert.assertNotNull(scmNodeInfos);
- Assert.assertTrue(scmNodeInfos.size() == 1);
- Assert.assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
+ assertNotNull(scmNodeInfos);
+ assertEquals(1, scmNodeInfos.size());
+ assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
scmNodeInfos.get(0).getBlockClientAddress());
- Assert.assertEquals("localhost:" + OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT,
+ assertEquals("localhost:" + OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT,
scmNodeInfos.get(0).getScmSecurityAddress());
- Assert.assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
+ assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
scmNodeInfos.get(0).getScmClientAddress());
- Assert.assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
+ assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
scmNodeInfos.get(0).getScmDatanodeAddress());
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java
index e040e489ce..f3bf37ab28 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java
@@ -17,7 +17,7 @@
*/
package org.apache.hadoop.hdds.scm.net;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertThrows;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java
index b0f9875ece..ac8920991c 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java
@@ -19,8 +19,7 @@ package org.apache.hadoop.hdds.scm.pipeline;
import org.apache.hadoop.hdds.protocol.DatanodeDetails;
import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import java.io.IOException;
@@ -29,6 +28,8 @@ import static org.apache.hadoop.hdds.protocol.DatanodeDetails.Port.Name.V0_PORTS
import static org.apache.hadoop.hdds.protocol.TestDatanodeDetails.assertPorts;
import static org.apache.hadoop.ozone.ClientVersion.DEFAULT_VERSION;
import static org.apache.hadoop.ozone.ClientVersion.VERSION_HANDLES_UNKNOWN_DN_PORTS;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
/**
* Test for {@link Pipeline}.
@@ -58,16 +59,15 @@ public class TestPipeline {
//when EC config is empty/null
HddsProtos.Pipeline protobufMessage = subject.getProtobufMessage(1);
- Assert.assertEquals(0, protobufMessage.getEcReplicationConfig().getData());
+ assertEquals(0, protobufMessage.getEcReplicationConfig().getData());
//when EC config is NOT empty
subject = MockPipeline.createEcPipeline();
protobufMessage = subject.getProtobufMessage(1);
- Assert.assertEquals(3, protobufMessage.getEcReplicationConfig().getData());
- Assert
- .assertEquals(2, protobufMessage.getEcReplicationConfig().getParity());
+ assertEquals(3, protobufMessage.getEcReplicationConfig().getData());
+ assertEquals(2, protobufMessage.getEcReplicationConfig().getParity());
}
@@ -78,37 +78,37 @@ public class TestPipeline {
Pipeline reloadedPipeline = Pipeline.getFromProtobuf(protobufMessage);
for (DatanodeDetails dn : pipeline.getNodes()) {
- Assert.assertEquals(pipeline.getReplicaIndex(dn),
+ assertEquals(pipeline.getReplicaIndex(dn),
reloadedPipeline.getReplicaIndex(dn));
}
}
@Test
- public void testECPipelineIsAlwaysHealthy() throws IOException {
+ public void testECPipelineIsAlwaysHealthy() {
Pipeline pipeline = MockPipeline.createEcPipeline();
- Assert.assertTrue(pipeline.isHealthy());
+ assertTrue(pipeline.isHealthy());
}
@Test
public void testBuilderCopiesAllFieldsFromOtherPipeline() {
Pipeline original = MockPipeline.createEcPipeline();
Pipeline copied = Pipeline.newBuilder(original).build();
- Assert.assertEquals(original.getId(), copied.getId());
- Assert.assertEquals(original.getReplicationConfig(),
+ assertEquals(original.getId(), copied.getId());
+ assertEquals(original.getReplicationConfig(),
copied.getReplicationConfig());
- Assert.assertEquals(original.getPipelineState(), copied.getPipelineState());
- Assert.assertEquals(original.getId(), copied.getId());
- Assert.assertEquals(original.getId(), copied.getId());
- Assert.assertEquals(original.getId(), copied.getId());
- Assert.assertEquals(original.getNodeSet(), copied.getNodeSet());
- Assert.assertEquals(original.getNodesInOrder(), copied.getNodesInOrder());
- Assert.assertEquals(original.getLeaderId(), copied.getLeaderId());
- Assert.assertEquals(original.getCreationTimestamp(),
+ assertEquals(original.getPipelineState(), copied.getPipelineState());
+ assertEquals(original.getId(), copied.getId());
+ assertEquals(original.getId(), copied.getId());
+ assertEquals(original.getId(), copied.getId());
+ assertEquals(original.getNodeSet(), copied.getNodeSet());
+ assertEquals(original.getNodesInOrder(), copied.getNodesInOrder());
+ assertEquals(original.getLeaderId(), copied.getLeaderId());
+ assertEquals(original.getCreationTimestamp(),
copied.getCreationTimestamp());
- Assert.assertEquals(original.getSuggestedLeaderId(),
+ assertEquals(original.getSuggestedLeaderId(),
copied.getSuggestedLeaderId());
for (DatanodeDetails dn : original.getNodes()) {
- Assert.assertEquals(original.getReplicaIndex(dn),
+ assertEquals(original.getReplicaIndex(dn),
copied.getReplicaIndex(dn));
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java
index 36c7a6e8a1..036f2586f5 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java
@@ -17,10 +17,10 @@
*/
package org.apache.hadoop.hdds.tracing;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertThrows;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
/**
* Test for {@link TraceAllMethod}.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java
index 135b237a7a..378ac6d4df 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java
@@ -22,10 +22,10 @@ import org.apache.hadoop.hdds.conf.MutableConfigurationSource;
import org.apache.hadoop.hdds.scm.ScmConfigKeys;
import org.apache.hadoop.hdds.tracing.TestTraceAllMethod.Service;
import org.apache.hadoop.hdds.tracing.TestTraceAllMethod.ServiceImpl;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import static org.apache.hadoop.hdds.tracing.TracingUtil.createProxy;
-import static org.junit.Assert.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
/**
* Test for {@link TracingUtil}.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java
index 4a09f3fa94..8a81aea168 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java
@@ -23,9 +23,9 @@ import static org.apache.hadoop.hdds.upgrade.HDDSLayoutFeature.INITIAL_VERSION;
import static org.apache.hadoop.hdds.upgrade.HDDSLayoutVersionManager.maxLayoutVersion;
import static org.apache.hadoop.ozone.upgrade.LayoutFeature.UpgradeActionType.ON_FINALIZE;
import static org.apache.hadoop.ozone.upgrade.LayoutFeature.UpgradeActionType.ON_FIRST_UPGRADE_START;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
import static org.mockito.ArgumentMatchers.any;
import static org.mockito.Mockito.doCallRealMethod;
import static org.mockito.Mockito.mock;
@@ -38,8 +38,8 @@ import java.util.Optional;
import org.apache.hadoop.hdds.upgrade.test.MockComponent;
import org.apache.hadoop.hdds.upgrade.test.MockComponent.MockDnUpgradeAction;
import org.apache.hadoop.hdds.upgrade.test.MockComponent.MockScmUpgradeAction;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
/**
* Class to test HDDS upgrade action registrations.
@@ -58,9 +58,9 @@ public class TestHDDSLayoutVersionManager {
//Cluster is finalized, hence should not register.
Optional<HDDSUpgradeAction> action = INITIAL_VERSION.scmAction(ON_FINALIZE);
- Assert.assertFalse(action.isPresent());
+ Assertions.assertFalse(action.isPresent());
action = DATANODE_SCHEMA_V2.datanodeAction(ON_FIRST_UPGRADE_START);
- Assert.assertFalse(action.isPresent());
+ Assertions.assertFalse(action.isPresent());
// Start from an unfinalized version manager.
lvm = mock(HDDSLayoutVersionManager.class);
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java
index 11d0fad55e..cc27a3b4a7 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java
@@ -26,11 +26,11 @@ import java.util.concurrent.ExecutorService;
import java.util.concurrent.Executors;
import java.util.concurrent.Future;
import org.apache.hadoop.hdds.HddsIdFactory;
-import org.junit.After;
-import static org.junit.Assert.assertEquals;
+import org.junit.jupiter.api.AfterEach;
+import static org.junit.jupiter.api.Assertions.assertEquals;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
/**
* Test the JMX interface for the rocksdb metastore implementation.
@@ -41,7 +41,7 @@ public class TestHddsIdFactory {
private static final int IDS_PER_THREAD = 10000;
private static final int NUM_OF_THREADS = 5;
- @After
+ @AfterEach
public void cleanup() {
ID_SET.clear();
}
@@ -65,7 +65,7 @@ public class TestHddsIdFactory {
for (int idNum = 0; idNum < IDS_PER_THREAD; idNum++) {
long var = HddsIdFactory.getLongId();
if (ID_SET.contains(var)) {
- Assert.fail("Duplicate id found");
+ Assertions.fail("Duplicate id found");
}
ID_SET.add(var);
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java
index 2f54922edc..4dd12fe37f 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java
@@ -17,8 +17,9 @@
package org.apache.hadoop.hdds.utils;
import org.apache.ozone.test.GenericTestUtils;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
import java.util.concurrent.CompletableFuture;
import java.util.concurrent.ConcurrentHashMap;
@@ -48,7 +49,7 @@ public class TestResourceLimitCache {
// Create a future which blocks to put 1. Currently map has acquired 10
// permits out of 10
- CompletableFuture future = CompletableFuture.supplyAsync(() -> {
+ CompletableFuture<String> future = CompletableFuture.supplyAsync(() -> {
try {
return resourceCache.put(1, "a");
} catch (InterruptedException e) {
@@ -56,9 +57,9 @@ public class TestResourceLimitCache {
}
return null;
});
- Assert.assertTrue(!future.isDone());
+ Assertions.assertFalse(future.isDone());
Thread.sleep(100);
- Assert.assertTrue(!future.isDone());
+ Assertions.assertFalse(future.isDone());
// remove 4 so that permits are released for key 1 to be put. Currently map
// has acquired 6 permits out of 10
@@ -66,8 +67,9 @@ public class TestResourceLimitCache {
GenericTestUtils.waitFor(future::isDone, 100, 1000);
// map has the key 1
- Assert.assertTrue(future.isDone() && !future.isCompletedExceptionally());
- Assert.assertNotNull(resourceCache.get(1));
+ Assertions.assertTrue(future.isDone());
+ Assertions.assertFalse(future.isCompletedExceptionally());
+ Assertions.assertNotNull(resourceCache.get(1));
// Create a future which blocks to put 4. Currently map has acquired 7
// permits out of 10
@@ -79,9 +81,9 @@ public class TestResourceLimitCache {
return null;
}
}, pool);
- Assert.assertTrue(!future.isDone());
+ Assertions.assertFalse(future.isDone());
Thread.sleep(100);
- Assert.assertTrue(!future.isDone());
+ Assertions.assertFalse(future.isDone());
// Shutdown the thread pool for putting key 4
pool.shutdownNow();
@@ -89,20 +91,23 @@ public class TestResourceLimitCache {
future.cancel(true);
// remove key 1 so currently map has acquired 6 permits out of 10
resourceCache.remove(1);
- Assert.assertNull(resourceCache.get(4));
+ Assertions.assertNull(resourceCache.get(4));
}
- @Test(timeout = 5000)
+ @Test
+ @Timeout(5)
public void testRemove() throws Exception {
testRemove(cache -> cache.remove(2), 2);
}
- @Test(timeout = 5000)
+ @Test
+ @Timeout(5)
public void testRemoveIf() throws Exception {
testRemove(cache -> cache.removeIf(k -> k <= 2), 1, 2);
}
- @Test(timeout = 5000)
+ @Test
+ @Timeout(5)
public void testClear() throws Exception {
testRemove(Cache::clear, 1, 2, 3);
}
@@ -124,7 +129,7 @@ public class TestResourceLimitCache {
// THEN
for (Integer k : removedKeys) {
- Assert.assertNull(resourceCache.get(k));
+ Assertions.assertNull(resourceCache.get(k));
}
// can put new entries
for (int i = 1; i <= removedKeys.length; ++i) {
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java
index f9c194d45c..43e74e687f 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java
@@ -17,14 +17,18 @@
*/
package org.apache.hadoop.hdds.utils;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
+
+import static org.junit.jupiter.api.Assertions.assertThrows;
/**
* Test for ResourceSemaphore.
*/
public class TestResourceSemaphore {
- @Test(timeout = 1000)
+ @Test
+ @Timeout(1)
public void testGroup() {
final ResourceSemaphore.Group g = new ResourceSemaphore.Group(3, 1);
@@ -49,28 +53,20 @@ public class TestResourceSemaphore {
g.release(0, 0);
assertUsed(g, 0, 0);
- try {
- g.release(1, 0);
- Assert.fail("Should have failed.");
- } catch (IllegalStateException e) {
- }
- try {
- g.release(0, 1);
- Assert.fail("Should have failed.");
- } catch (IllegalStateException e) {
- }
+ assertThrows(IllegalStateException.class, () -> g.release(1, 0));
+ assertThrows(IllegalStateException.class, () -> g.release(0, 1));
}
static void assertUsed(ResourceSemaphore.Group g, int... expected) {
- Assert.assertEquals(expected.length, g.resourceSize());
+ Assertions.assertEquals(expected.length, g.resourceSize());
for (int i = 0; i < expected.length; i++) {
- Assert.assertEquals(expected[i], g.get(i).used());
+ Assertions.assertEquals(expected[i], g.get(i).used());
}
}
static void assertAcquire(ResourceSemaphore.Group g, boolean expected,
int... permits) {
final boolean computed = g.tryAcquire(permits);
- Assert.assertEquals(expected, computed);
+ Assertions.assertEquals(expected, computed);
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java
index 148ccf94a1..29e75aa853 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java
@@ -17,12 +17,12 @@
*/
package org.apache.hadoop.hdds.utils;
-import static org.junit.Assert.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertThrows;
import org.apache.hadoop.io.retry.RetryPolicies;
import org.apache.hadoop.io.retry.RetryPolicy;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import java.io.IOException;
import java.util.concurrent.TimeUnit;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java
index 3ba76d651d..1d36b662f9 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java
@@ -17,8 +17,8 @@
*/
package org.apache.hadoop.ozone;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
/**
* Test Ozone Constants (e.g. for compatibility).
@@ -28,7 +28,7 @@ public class TestOzoneConsts {
@Test
public void testOzoneTenantPolicyLabelCompatibility() {
// Value of the policy label should not be changed
- Assert.assertEquals(
+ Assertions.assertEquals(
"OzoneTenant", OzoneConsts.OZONE_TENANT_RANGER_POLICY_LABEL);
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java
index 1b8e744554..d42795ede5 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java
@@ -19,9 +19,9 @@ package org.apache.hadoop.ozone.audit;
import org.apache.commons.io.FileUtils;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
-import org.junit.AfterClass;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
@@ -37,8 +37,8 @@ import static org.apache.hadoop.ozone.audit.AuditEventStatus.FAILURE;
import static org.apache.hadoop.ozone.audit.AuditEventStatus.SUCCESS;
import static org.hamcrest.MatcherAssert.assertThat;
import static org.hamcrest.core.StringContains.containsString;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
import org.hamcrest.Matcher;
import org.hamcrest.collection.IsIterableContainingInOrder;
@@ -102,7 +102,7 @@ public class TestOzoneAuditLogger {
.withResult(SUCCESS)
.withException(null).build();
- @AfterClass
+ @AfterAll
public static void tearDown() {
File file = new File("audit.log");
if (FileUtils.deleteQuietly(file)) {
@@ -113,7 +113,7 @@ public class TestOzoneAuditLogger {
}
}
- @Before
+ @BeforeEach
public void init() {
AUDIT.refreshDebugCmdSet();
}
@@ -166,11 +166,11 @@ public class TestOzoneAuditLogger {
@Test
public void messageIncludesAllParts() {
String message = WRITE_FAIL_MSG.getFormattedMessage();
- assertTrue(message, message.contains(USER));
- assertTrue(message, message.contains(IP_ADDRESS));
- assertTrue(message, message.contains(DummyAction.CREATE_VOLUME.name()));
- assertTrue(message, message.contains(PARAMS.toString()));
- assertTrue(message, message.contains(FAILURE.getStatus()));
+ assertTrue(message.contains(USER), message);
+ assertTrue(message.contains(IP_ADDRESS), message);
+ assertTrue(message.contains(DummyAction.CREATE_VOLUME.name()), message);
+ assertTrue(message.contains(PARAMS.toString()), message);
+ assertTrue(message.contains(FAILURE.getStatus()), message);
}
/**
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java
index a61ff9054b..e01b767b10 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.ozone.common;
import org.apache.commons.lang3.RandomStringUtils;
import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
import static java.nio.charset.StandardCharsets.UTF_8;
@@ -54,11 +54,11 @@ public class TestChecksum {
// A checksum is calculate for each bytesPerChecksum number of bytes in
// the data. Since that value is 10 here and the data length is 55, we
// should have 6 checksums in checksumData.
- Assert.assertEquals(6, checksumData.getChecksums().size());
+ Assertions.assertEquals(6, checksumData.getChecksums().size());
// Checksum verification should pass
- Assert.assertTrue("Checksum mismatch",
- Checksum.verifyChecksum(data, checksumData));
+ Assertions.assertTrue(Checksum.verifyChecksum(data, checksumData),
+ "Checksum mismatch");
}
/**
@@ -75,8 +75,8 @@ public class TestChecksum {
// mismatch
data[50] = (byte) (data[50] + 1);
ChecksumData newChecksumData = checksum.computeChecksum(data);
- Assert.assertNotEquals("Checksums should not match for different data",
- originalChecksumData, newChecksumData);
+ Assertions.assertNotEquals(originalChecksumData, newChecksumData,
+ "Checksums should not match for different data");
}
/**
@@ -92,8 +92,7 @@ public class TestChecksum {
Checksum checksum2 = getChecksum(ContainerProtos.ChecksumType.CRC32);
// The two checksums should not match as they have different types
- Assert.assertNotEquals(
- "Checksums should not match for different checksum types",
- checksum1, checksum2);
+ Assertions.assertNotEquals(checksum1, checksum2,
+ "Checksums should not match for different checksum types");
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
index 4aac0b4847..6c9236d949 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.ozone.common;
import org.apache.hadoop.util.PureJavaCrc32;
import org.apache.hadoop.util.PureJavaCrc32C;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
import java.nio.charset.StandardCharsets;
import java.util.Random;
@@ -96,7 +96,7 @@ public class TestChecksumByteBuffer {
}
private void checkSame() {
- Assert.assertEquals(expected.getValue(), testee.getValue());
+ Assertions.assertEquals(expected.getValue(), testee.getValue());
}
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
index b15f59f2a5..5e02ceaf08 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
@@ -21,14 +21,14 @@ import org.apache.commons.lang3.RandomUtils;
import org.apache.hadoop.util.NativeCRC32Wrapper;
import org.apache.hadoop.util.PureJavaCrc32;
import org.apache.hadoop.util.PureJavaCrc32C;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.List;
import java.util.zip.CRC32;
-import static junit.framework.TestCase.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
/**
* Tests to verify that different checksum implementations compute the same
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java
index 1e85099179..c405b83012 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java
@@ -30,9 +30,9 @@ import java.util.concurrent.ThreadLocalRandom;
import org.apache.hadoop.hdds.utils.MockGatheringChannel;
import org.apache.ratis.thirdparty.com.google.protobuf.ByteString;
-import org.junit.Assert;
-import org.junit.Test;
import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
/**
* Test {@link ChunkBuffer} implementations.
@@ -42,7 +42,8 @@ public class TestChunkBuffer {
return ThreadLocalRandom.current().nextInt(n);
}
- @Test(timeout = 1_000)
+ @Test
+ @Timeout(1)
public void testImplWithByteBuffer() {
runTestImplWithByteBuffer(1);
runTestImplWithByteBuffer(1 << 10);
@@ -57,7 +58,8 @@ public class TestChunkBuffer {
runTestImpl(expected, 0, ChunkBuffer.allocate(n));
}
- @Test(timeout = 1_000)
+ @Test
+ @Timeout(1)
public void testIncrementalChunkBuffer() {
runTestIncrementalChunkBuffer(1, 1);
runTestIncrementalChunkBuffer(4, 8);
@@ -76,7 +78,8 @@ public class TestChunkBuffer {
new IncrementalChunkBuffer(n, increment, false));
}
- @Test(timeout = 1_000)
+ @Test
+ @Timeout(1)
public void testImplWithList() {
runTestImplWithList(4, 8);
runTestImplWithList(16, 1 << 10);
@@ -114,14 +117,14 @@ public class TestChunkBuffer {
System.out.println("n=" + n + ", impl=" + impl);
// check position, remaining
- Assert.assertEquals(0, impl.position());
- Assert.assertEquals(n, impl.remaining());
- Assert.assertEquals(n, impl.limit());
+ Assertions.assertEquals(0, impl.position());
+ Assertions.assertEquals(n, impl.remaining());
+ Assertions.assertEquals(n, impl.limit());
impl.put(expected);
- Assert.assertEquals(n, impl.position());
- Assert.assertEquals(0, impl.remaining());
- Assert.assertEquals(n, impl.limit());
+ Assertions.assertEquals(n, impl.position());
+ Assertions.assertEquals(0, impl.remaining());
+ Assertions.assertEquals(n, impl.limit());
// duplicate
assertDuplicate(expected, impl);
@@ -159,8 +162,8 @@ public class TestChunkBuffer {
byte[] expected, ChunkBuffer impl, int bpc) {
final int n = expected.length;
final ChunkBuffer duplicated = impl.duplicate(0, n);
- Assert.assertEquals(0, duplicated.position());
- Assert.assertEquals(n, duplicated.remaining());
+ Assertions.assertEquals(0, duplicated.position());
+ Assertions.assertEquals(n, duplicated.remaining());
final int numChecksums = (n + bpc - 1) / bpc;
final Iterator<ByteBuffer> i = duplicated.iterate(bpc).iterator();
@@ -169,57 +172,58 @@ public class TestChunkBuffer {
final ByteBuffer b = i.next();
final int expectedRemaining = j < numChecksums - 1 ?
bpc : n - bpc * (numChecksums - 1);
- Assert.assertEquals(expectedRemaining, b.remaining());
+ Assertions.assertEquals(expectedRemaining, b.remaining());
final int offset = j * bpc;
for (int k = 0; k < expectedRemaining; k++) {
- Assert.assertEquals(expected[offset + k], b.get());
+ Assertions.assertEquals(expected[offset + k], b.get());
count++;
}
}
- Assert.assertEquals(n, count);
- Assert.assertFalse(i.hasNext());
+ Assertions.assertEquals(n, count);
+ Assertions.assertFalse(i.hasNext());
Assertions.assertThrows(NoSuchElementException.class, i::next);
}
private static void assertToByteString(
byte[] expected, int offset, int length, ChunkBuffer impl) {
final ChunkBuffer duplicated = impl.duplicate(offset, offset + length);
- Assert.assertEquals(offset, duplicated.position());
- Assert.assertEquals(length, duplicated.remaining());
+ Assertions.assertEquals(offset, duplicated.position());
+ Assertions.assertEquals(length, duplicated.remaining());
final ByteString computed = duplicated.toByteString(buffer -> {
buffer.mark();
final ByteString string = ByteString.copyFrom(buffer);
buffer.reset();
return string;
});
- Assert.assertEquals(offset, duplicated.position());
- Assert.assertEquals(length, duplicated.remaining());
+ Assertions.assertEquals(offset, duplicated.position());
+ Assertions.assertEquals(length, duplicated.remaining());
assertEquals("offset=" + offset + ", length=" + length,
ByteString.copyFrom(expected, offset, length), computed);
}
private static void assertWrite(byte[] expected, ChunkBuffer impl) {
impl.rewind();
- Assert.assertEquals(0, impl.position());
+ Assertions.assertEquals(0, impl.position());
ByteArrayOutputStream output = new ByteArrayOutputStream(expected.length);
try {
impl.writeTo(new MockGatheringChannel(Channels.newChannel(output)));
} catch (IOException e) {
- Assert.fail("Unexpected error: " + e);
+ Assertions.fail("Unexpected error: " + e);
}
- Assert.assertArrayEquals(expected, output.toByteArray());
- Assert.assertFalse(impl.hasRemaining());
+ Assertions.assertArrayEquals(expected, output.toByteArray());
+ Assertions.assertFalse(impl.hasRemaining());
}
private static void assertEquals(String message,
ByteString expected, ByteString actual) {
- Assert.assertEquals(message,
+ Assertions.assertEquals(
toString(expected.toByteArray()),
- toString(actual.toByteArray()));
+ toString(actual.toByteArray()),
+ message);
}
private static String toString(byte[] arr) {
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java
index b867fbe79b..3da43166e7 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java
@@ -18,7 +18,7 @@
package org.apache.hadoop.ozone.common;
import com.google.common.collect.ImmutableList;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import java.nio.BufferOverflowException;
import java.nio.ByteBuffer;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java
index 2d14abd3f5..5ec42b9117 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java
@@ -18,8 +18,8 @@ package org.apache.hadoop.ozone.ha;
import org.apache.hadoop.net.NetUtils;
import org.apache.hadoop.ozone.util.OzoneNetUtils;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
import java.net.InetSocketAddress;
@@ -35,7 +35,7 @@ public class TestOzoneNetUtils {
InetSocketAddress addr0 = NetUtils.createSocketAddr(fqdn, port);
InetSocketAddress addr1 = OzoneNetUtils.getAddressWithHostNameLocal(
addr0);
- Assert.assertEquals("pod0", addr1.getHostName());
- Assert.assertEquals(port, addr1.getPort());
+ Assertions.assertEquals("pod0", addr1.getHostName());
+ Assertions.assertEquals(port, addr1.getPort());
}
}
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java
index 1b992dd51c..dfdbba1d89 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java
@@ -20,8 +20,9 @@ package org.apache.hadoop.ozone.lock;
import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.ozone.test.GenericTestUtils;
import org.apache.hadoop.util.Daemon;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
import java.util.concurrent.atomic.AtomicBoolean;
import java.util.concurrent.atomic.AtomicInteger;
@@ -31,7 +32,8 @@ import java.util.concurrent.atomic.AtomicInteger;
*/
public class TestLockManager {
- @Test(timeout = 1000)
+ @Test
+ @Timeout(1)
public void testWriteLockWithDifferentResource() {
final LockManager<String> manager =
new LockManager<>(new OzoneConfiguration());
@@ -40,7 +42,7 @@ public class TestLockManager {
manager.writeLock("/resourceTwo");
manager.writeUnlock("/resourceOne");
manager.writeUnlock("/resourceTwo");
- Assert.assertTrue(true);
+ Assertions.assertTrue(true);
}
@Test
@@ -58,16 +60,17 @@ public class TestLockManager {
Thread.sleep(100);
// Since the other thread is trying to get write lock on same object,
// it will wait.
- Assert.assertFalse(gotLock.get());
+ Assertions.assertFalse(gotLock.get());
manager.writeUnlock("/resourceOne");
// Since we have released the write lock, the other thread should have
// the lock now
// Let's give some time for the other thread to run
Thread.sleep(100);
- Assert.assertTrue(gotLock.get());
+ Assertions.assertTrue(gotLock.get());
}
- @Test(timeout = 1000)
+ @Test
+ @Timeout(1)
public void testReadLockWithDifferentResource() {
final LockManager<String> manager =
new LockManager<>(new OzoneConfiguration());
@@ -75,7 +78,7 @@ public class TestLockManager {
manager.readLock("/resourceTwo");
manager.readUnlock("/resourceOne");
manager.readUnlock("/resourceTwo");
- Assert.assertTrue(true);
+ Assertions.assertTrue(true);
}
@Test
@@ -92,7 +95,7 @@ public class TestLockManager {
// Let's give some time for the other thread to run
Thread.sleep(100);
// Since the new thread is trying to get read lock, it should work.
- Assert.assertTrue(gotLock.get());
+ Assertions.assertTrue(gotLock.get());
manager.readUnlock("/resourceOne");
}
@@ -111,13 +114,13 @@ public class TestLockManager {
Thread.sleep(100);
// Since the other thread is trying to get read lock on same object,
// it will wait.
- Assert.assertFalse(gotLock.get());
+ Assertions.assertFalse(gotLock.get());
manager.writeUnlock("/resourceOne");
// Since we have released the write lock, the other thread should have
// the lock now
// Let's give some time for the other thread to run
Thread.sleep(100);
- Assert.assertTrue(gotLock.get());
+ Assertions.assertTrue(gotLock.get());
}
@Test
@@ -135,13 +138,13 @@ public class TestLockManager {
Thread.sleep(100);
// Since the other thread is trying to get write lock on same object,
// it will wait.
- Assert.assertFalse(gotLock.get());
+ Assertions.assertFalse(gotLock.get());
manager.readUnlock("/resourceOne");
// Since we have released the read lock, the other thread should have
// the lock now
// Let's give some time for the other thread to run
Thread.sleep(100);
- Assert.assertTrue(gotLock.get());
+ Assertions.assertTrue(gotLock.get());
}
@Test
@@ -160,17 +163,17 @@ public class TestLockManager {
Thread.sleep(100);
// Since the other thread is trying to get write lock on same object,
// it will wait.
- Assert.assertFalse(gotLock.get());
+ Assertions.assertFalse(gotLock.get());
manager.readUnlock("/resourceOne");
//We have only released one read lock, we still hold another read lock.
Thread.sleep(100);
- Assert.assertFalse(gotLock.get());
+ Assertions.assertFalse(gotLock.get());
manager.readUnlock("/resourceOne");
// Since we have released the read lock, the other thread should have
// the lock now
// Let's give some time for the other thread to run
Thread.sleep(100);
- Assert.assertTrue(gotLock.get());
+ Assertions.assertTrue(gotLock.get());
}
@Test
@@ -198,6 +201,6 @@ public class TestLockManager {
}
GenericTestUtils.waitFor(() -> done.get() == count, 100,
10 * count * sleep);
- Assert.assertEquals(count, done.get());
+ Assertions.assertEquals(count, done.get());
}
}
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java
index e63246ce88..5b83f3552e 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java
@@ -22,9 +22,9 @@ import static org.apache.hadoop.ozone.upgrade.TestUpgradeFinalizerActions.MockLa
import static org.apache.hadoop.ozone.upgrade.TestUpgradeFinalizerActions.MockLayoutFeature.VERSION_3;
import static org.apache.hadoop.ozone.upgrade.UpgradeFinalizer.Status.ALREADY_FINALIZED;
import static org.apache.hadoop.ozone.upgrade.UpgradeFinalizer.Status.FINALIZATION_DONE;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
import static org.mockito.ArgumentMatchers.anyString;
import static org.mockito.Mockito.doCallRealMethod;
import static org.mockito.Mockito.inOrder;
@@ -36,7 +36,7 @@ import java.io.IOException;
import org.apache.hadoop.ozone.common.Storage;
import org.apache.hadoop.ozone.upgrade.TestUpgradeFinalizerActions.MockLayoutVersionManager;
import org.apache.hadoop.ozone.upgrade.UpgradeFinalizer.StatusAndMessages;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import org.mockito.ArgumentMatchers;
import org.mockito.InOrder;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java
index c3eea3aebd..9bc8ded519 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java
@@ -25,7 +25,7 @@ import java.io.IOException;
import org.apache.hadoop.ozone.common.Storage;
import org.apache.ozone.test.LambdaTestUtils;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
/**
* Test for DefaultUpgradeFinalizationExecutor.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java
index d7ac8ad8c2..ab82d0265b 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java
@@ -18,15 +18,15 @@
package org.apache.hadoop.ozone.upgrade;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.when;
import java.util.function.Supplier;
import org.apache.ozone.test.LambdaTestUtils;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
/**
* Test out APIs of VersionSpecificInstanceFactory.
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@ozone.apache.org
For additional commands, e-mail: commits-help@ozone.apache.org