You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ozone.apache.org by ad...@apache.org on 2022/06/13 20:25:48 UTC

[ozone] branch master updated: HDDS-6874. Migrate simple tests in hdds-common to JUnit5 (#3507)

This is an automated email from the ASF dual-hosted git repository.

adoroszlai pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/ozone.git


The following commit(s) were added to refs/heads/master by this push:
     new b1fc69e059 HDDS-6874. Migrate simple tests in hdds-common to JUnit5 (#3507)
b1fc69e059 is described below

commit b1fc69e05903af49154115fc503bbfd593d2b114
Author: Kaijie Chen <ch...@kaijie.org>
AuthorDate: Tue Jun 14 04:25:42 2022 +0800

    HDDS-6874. Migrate simple tests in hdds-common to JUnit5 (#3507)
---
 .../java/org/apache/hadoop/hdds/TestHddsUtils.java |  67 +++++++-------
 .../hdds/client/TestECReplicationConfig.java       |  32 +++----
 .../client/TestReplicationConfigValidator.java     |   9 +-
 .../conf/TestGeneratedConfigurationOverwrite.java  |  18 ++--
 .../hdds/fs/TestCachingSpaceUsageSource.java       |   2 +-
 .../java/org/apache/hadoop/hdds/fs/TestDU.java     |  10 +-
 .../org/apache/hadoop/hdds/fs/TestDUFactory.java   |   2 +-
 .../hdds/fs/TestDedicatedDiskSpaceUsage.java       |  12 +--
 .../fs/TestDedicatedDiskSpaceUsageFactory.java     |   2 +-
 .../hadoop/hdds/fs/TestSaveSpaceUsageToFile.java   |  16 ++--
 .../hadoop/hdds/fs/TestSpaceUsageFactory.java      |   6 +-
 .../hadoop/hdds/protocol/TestDatanodeDetails.java  |   8 +-
 .../ratis/TestContainerCommandRequestMessage.java  |   6 +-
 .../apache/hadoop/hdds/ratis/TestRatisHelper.java  |  38 ++++----
 .../hdds/ratis/TestServerNotLeaderException.java   |  16 ++--
 .../hdds/scm/container/TestContainerInfo.java      |  30 +++---
 .../scm/container/TestContainerReplicaInfo.java    |  36 ++++----
 .../container/TestReplicationManagerReport.java    | 102 +++++++++++----------
 .../container/common/helpers/TestExcludeList.java  |  14 +--
 .../apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java |  51 ++++++-----
 .../apache/hadoop/hdds/scm/net/TestNetUtils.java   |   2 +-
 .../hadoop/hdds/scm/pipeline/TestPipeline.java     |  42 ++++-----
 .../hadoop/hdds/tracing/TestTraceAllMethod.java    |   6 +-
 .../hadoop/hdds/tracing/TestTracingUtil.java       |   4 +-
 .../hdds/upgrade/TestHDDSLayoutVersionManager.java |  14 +--
 .../hadoop/hdds/utils/TestHddsIdFactory.java       |  12 +--
 .../hadoop/hdds/utils/TestResourceLimitCache.java  |  33 ++++---
 .../hadoop/hdds/utils/TestResourceSemaphore.java   |  28 +++---
 .../hadoop/hdds/utils/TestRetriableTask.java       |   4 +-
 .../org/apache/hadoop/ozone/TestOzoneConsts.java   |   6 +-
 .../hadoop/ozone/audit/TestOzoneAuditLogger.java   |  24 ++---
 .../apache/hadoop/ozone/common/TestChecksum.java   |  19 ++--
 .../ozone/common/TestChecksumByteBuffer.java       |   6 +-
 .../common/TestChecksumImplsComputeSameValues.java |   4 +-
 .../hadoop/ozone/common/TestChunkBuffer.java       |  58 ++++++------
 .../TestChunkBufferImplWithByteBufferList.java     |   2 +-
 .../apache/hadoop/ozone/ha/TestOzoneNetUtils.java  |   8 +-
 .../apache/hadoop/ozone/lock/TestLockManager.java  |  37 ++++----
 .../ozone/upgrade/TestBasicUpgradeFinalizer.java   |   8 +-
 .../TestDefaultUpgradeFinalizationExecutor.java    |   2 +-
 .../upgrade/TestLayoutVersionInstanceFactory.java  |   6 +-
 41 files changed, 405 insertions(+), 397 deletions(-)

diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java
index 67001010d5..3dbee3efbd 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/TestHddsUtils.java
@@ -38,12 +38,12 @@ import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DATANODE_PORT_D
 import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DATANODE_PORT_KEY;
 import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT;
 
-import static org.hamcrest.core.Is.is;
-import org.junit.Assert;
-import static org.junit.Assert.assertThat;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
+import static org.junit.jupiter.api.Assertions.fail;
 
 /**
  * Testing HddsUtils.
@@ -52,13 +52,13 @@ public class TestHddsUtils {
 
   @Test
   public void testGetHostName() {
-    Assert.assertEquals(Optional.of("localhost"),
+    assertEquals(Optional.of("localhost"),
         HddsUtils.getHostName("localhost:1234"));
 
-    Assert.assertEquals(Optional.of("localhost"),
+    assertEquals(Optional.of("localhost"),
         HddsUtils.getHostName("localhost"));
 
-    Assert.assertEquals(Optional.empty(),
+    assertEquals(Optional.empty(),
         HddsUtils.getHostName(":1234"));
   }
 
@@ -91,26 +91,26 @@ public class TestHddsUtils {
     // Verify valid IP address setup
     conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, "1.2.3.4");
     addresses = getSCMAddressForDatanodes(conf);
-    assertThat(addresses.size(), is(1));
+    assertEquals(1, addresses.size());
     addr = addresses.iterator().next();
-    assertThat(addr.getHostName(), is("1.2.3.4"));
-    assertThat(addr.getPort(), is(OZONE_SCM_DATANODE_PORT_DEFAULT));
+    assertEquals("1.2.3.4", addr.getHostName());
+    assertEquals(OZONE_SCM_DATANODE_PORT_DEFAULT, addr.getPort());
 
     // Verify valid hostname setup
     conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, "scm1");
     addresses = getSCMAddressForDatanodes(conf);
-    assertThat(addresses.size(), is(1));
+    assertEquals(1, addresses.size());
     addr = addresses.iterator().next();
-    assertThat(addr.getHostName(), is("scm1"));
-    assertThat(addr.getPort(), is(OZONE_SCM_DATANODE_PORT_DEFAULT));
+    assertEquals("scm1", addr.getHostName());
+    assertEquals(OZONE_SCM_DATANODE_PORT_DEFAULT, addr.getPort());
 
     // Verify valid hostname and port
     conf.setStrings(ScmConfigKeys.OZONE_SCM_NAMES, "scm1:1234");
     addresses = getSCMAddressForDatanodes(conf);
-    assertThat(addresses.size(), is(1));
+    assertEquals(1, addresses.size());
     addr = addresses.iterator().next();
-    assertThat(addr.getHostName(), is("scm1"));
-    assertThat(addr.getPort(), is(1234));
+    assertEquals("scm1", addr.getHostName());
+    assertEquals(1234, addr.getPort());
 
     final Map<String, Integer> hostsAndPorts = new HashMap<>();
     hostsAndPorts.put("scm1", 1234);
@@ -121,7 +121,7 @@ public class TestHddsUtils {
     conf.setStrings(
         ScmConfigKeys.OZONE_SCM_NAMES, "scm1:1234,scm2:2345,scm3:3456");
     addresses = getSCMAddressForDatanodes(conf);
-    assertThat(addresses.size(), is(3));
+    assertEquals(3, addresses.size());
     it = addresses.iterator();
     HashMap<String, Integer> expected1 = new HashMap<>(hostsAndPorts);
     while (it.hasNext()) {
@@ -135,7 +135,7 @@ public class TestHddsUtils {
     conf.setStrings(
         ScmConfigKeys.OZONE_SCM_NAMES, " scm1:1234, scm2:2345 , scm3:3456 ");
     addresses = getSCMAddressForDatanodes(conf);
-    assertThat(addresses.size(), is(3));
+    assertEquals(3, addresses.size());
     it = addresses.iterator();
     HashMap<String, Integer> expected2 = new HashMap<>(hostsAndPorts);
     while (it.hasNext()) {
@@ -205,16 +205,14 @@ public class TestHddsUtils {
     Collection<InetSocketAddress> scmAddressList =
         HddsUtils.getSCMAddressForDatanodes(conf);
 
-    Assert.assertNotNull(scmAddressList);
-    Assert.assertEquals(3, scmAddressList.size());
+    Assertions.assertNotNull(scmAddressList);
+    assertEquals(3, scmAddressList.size());
 
-    Iterator<InetSocketAddress> it = scmAddressList.iterator();
-    while (it.hasNext()) {
-      InetSocketAddress next = it.next();
-      expected.remove(next.getHostName()  + ":" + next.getPort());
+    for (InetSocketAddress next : scmAddressList) {
+      expected.remove(next.getHostName() + ":" + next.getPort());
     }
 
-    Assert.assertTrue(expected.size() == 0);
+    assertEquals(0, expected.size());
 
   }
 
@@ -228,16 +226,14 @@ public class TestHddsUtils {
     OzoneConfiguration conf = new OzoneConfiguration();
     conf.set(ScmConfigKeys.OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT,
         testnum1);
-    Assert.assertTrue(Integer.parseInt(testnum1) ==
-        HddsUtils.getNumberFromConfigKeys(
-            conf,
+    assertEquals(Integer.parseInt(testnum1),
+        HddsUtils.getNumberFromConfigKeys(conf,
             OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT).orElse(0));
 
     /* Test to return first unempty key number from list */
     /* first key is absent */
-    Assert.assertTrue(Integer.parseInt(testnum1) ==
-        HddsUtils.getNumberFromConfigKeys(
-            conf,
+    assertEquals(Integer.parseInt(testnum1),
+        HddsUtils.getNumberFromConfigKeys(conf,
             ConfUtils.addKeySuffixes(OZONE_SCM_DATANODE_PORT_KEY,
                 serviceId, nodeId),
             OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT).orElse(0));
@@ -246,9 +242,8 @@ public class TestHddsUtils {
     conf.set(ConfUtils.addKeySuffixes(OZONE_SCM_DATANODE_PORT_KEY,
             serviceId, nodeId),
         testnum2);
-    Assert.assertTrue(Integer.parseInt(testnum2) ==
-        HddsUtils.getNumberFromConfigKeys(
-            conf,
+    assertEquals(Integer.parseInt(testnum2),
+        HddsUtils.getNumberFromConfigKeys(conf,
             ConfUtils.addKeySuffixes(OZONE_SCM_DATANODE_PORT_KEY,
                 serviceId, nodeId),
             OZONE_SCM_PIPELINE_OWNER_CONTAINER_COUNT).orElse(0));
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java
index 9440a6e026..333328bd74 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestECReplicationConfig.java
@@ -18,15 +18,15 @@
 package org.apache.hadoop.hdds.client;
 
 import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import java.util.HashMap;
 import java.util.Map;
 
 import static org.apache.hadoop.hdds.client.ECReplicationConfig.EcCodec.RS;
 import static org.apache.hadoop.hdds.client.ECReplicationConfig.EcCodec.XOR;
-import static org.junit.Assert.fail;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
 
 /**
  * Unit test for ECReplicationConfig.
@@ -45,10 +45,10 @@ public class TestECReplicationConfig {
 
     for (Map.Entry<String, ECReplicationConfig> e : valid.entrySet()) {
       ECReplicationConfig ec = new ECReplicationConfig(e.getKey());
-      Assert.assertEquals(e.getValue().getData(), ec.getData());
-      Assert.assertEquals(e.getValue().getParity(), ec.getParity());
-      Assert.assertEquals(e.getValue().getCodec(), ec.getCodec());
-      Assert.assertEquals(e.getValue().getEcChunkSize(), ec.getEcChunkSize());
+      assertEquals(e.getValue().getData(), ec.getData());
+      assertEquals(e.getValue().getParity(), ec.getParity());
+      assertEquals(e.getValue().getCodec(), ec.getCodec());
+      assertEquals(e.getValue().getEcChunkSize(), ec.getEcChunkSize());
     }
   }
 
@@ -63,12 +63,8 @@ public class TestECReplicationConfig {
         "x3-2"
     };
     for (String s : invalid) {
-      try {
-        new ECReplicationConfig(s);
-        fail(s + " should not parse correctly");
-      } catch (IllegalArgumentException e) {
-        // ignore, this expected
-      }
+      assertThrows(IllegalArgumentException.class,
+          () -> new ECReplicationConfig(s));
     }
   }
 
@@ -81,11 +77,11 @@ public class TestECReplicationConfig {
     HddsProtos.ECReplicationConfig proto = orig.toProto();
 
     ECReplicationConfig recovered = new ECReplicationConfig(proto);
-    Assert.assertEquals(orig.getData(), recovered.getData());
-    Assert.assertEquals(orig.getParity(), recovered.getParity());
-    Assert.assertEquals(orig.getCodec(), recovered.getCodec());
-    Assert.assertEquals(orig.getEcChunkSize(), recovered.getEcChunkSize());
-    Assert.assertTrue(orig.equals(recovered));
+    assertEquals(orig.getData(), recovered.getData());
+    assertEquals(orig.getParity(), recovered.getParity());
+    assertEquals(orig.getCodec(), recovered.getCodec());
+    assertEquals(orig.getEcChunkSize(), recovered.getEcChunkSize());
+    assertEquals(orig, recovered);
   }
 
 }
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java
index 96e49e9af0..320069c060 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/client/TestReplicationConfigValidator.java
@@ -19,10 +19,11 @@ package org.apache.hadoop.hdds.client;
 
 import org.apache.hadoop.hdds.conf.InMemoryConfiguration;
 import org.apache.hadoop.hdds.conf.MutableConfigurationSource;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.ONE;
 import static org.apache.hadoop.hdds.protocol.proto.HddsProtos.ReplicationFactor.THREE;
+import static org.junit.jupiter.api.Assertions.assertThrows;
 
 /**
  * Test ReplicationConfig validator.
@@ -56,7 +57,7 @@ public class TestReplicationConfigValidator {
 
   }
 
-  @Test(expected = IllegalArgumentException.class)
+  @Test
   public void testCustomValidation() {
     MutableConfigurationSource config = new InMemoryConfiguration();
     config.set("ozone.replication.allowed-configs", "RATIS/THREE");
@@ -66,8 +67,8 @@ public class TestReplicationConfigValidator {
 
     validator.validate(RatisReplicationConfig.getInstance(THREE));
 
-    validator.validate(RatisReplicationConfig.getInstance(ONE));
-    //exception is expected
+    assertThrows(IllegalArgumentException.class,
+        () -> validator.validate(RatisReplicationConfig.getInstance(ONE)));
 
   }
 }
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java
index f00107b014..03fd2be0b9 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/conf/TestGeneratedConfigurationOverwrite.java
@@ -17,10 +17,10 @@
  */
 package org.apache.hadoop.hdds.conf;
 
-import org.junit.After;
-import org.junit.Assert;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 
 import java.nio.file.Files;
 import java.nio.file.Path;
@@ -40,13 +40,13 @@ public class TestGeneratedConfigurationOverwrite {
 
   private OzoneConfiguration conf;
 
-  @Before
+  @BeforeEach
   public void overwriteConfigFile() throws Exception {
     Files.move(generatedConfigurationPath, generatedConfigurationPathBak);
     conf = new OzoneConfiguration();
   }
 
-  @After
+  @AfterEach
   public void recoverConfigFile() throws Exception {
     Files.move(generatedConfigurationPathBak, generatedConfigurationPath);
   }
@@ -54,13 +54,13 @@ public class TestGeneratedConfigurationOverwrite {
   @Test
   public void getConfigurationObject() {
     // Check Config Type of String
-    Assert.assertNotNull(
+    Assertions.assertNotNull(
         conf.getObject(SimpleConfiguration.class).getBindHost());
     // Check Config Type of Int
-    Assert.assertNotEquals(
+    Assertions.assertNotEquals(
         conf.getObject(SimpleConfiguration.class).getPort(), 0);
     // Check Config Type of Time
-    Assert.assertNotEquals(
+    Assertions.assertNotEquals(
         conf.getObject(SimpleConfiguration.class).getWaitTime(), 0);
   }
 }
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
index 28da73b2e7..9c701ca1fc 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
@@ -19,7 +19,7 @@ package org.apache.hadoop.hdds.fs;
 
 import org.apache.commons.lang3.RandomUtils;
 import org.apache.hadoop.hdds.fs.MockSpaceUsageCheckParams.Builder;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 import org.mockito.stubbing.Answer;
 
 import java.io.File;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java
index c9ed258f24..0d956a18ee 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDU.java
@@ -19,9 +19,9 @@ package org.apache.hadoop.hdds.fs;
 
 import org.apache.hadoop.fs.FileUtil;
 import org.apache.hadoop.util.Shell;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 
 import static org.apache.hadoop.ozone.OzoneConsts.KB;
 import static org.apache.ozone.test.GenericTestUtils.getTestDir;
@@ -41,14 +41,14 @@ public class TestDU {
 
   private static final File DIR = getTestDir(TestDU.class.getSimpleName());
 
-  @Before
+  @BeforeEach
   public void setUp() {
     assumeFalse(Shell.WINDOWS);
     FileUtil.fullyDelete(DIR);
     assertTrue(DIR.mkdirs());
   }
 
-  @After
+  @AfterEach
   public void tearDown() throws IOException {
     FileUtil.fullyDelete(DIR);
   }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java
index f9b07b47eb..6e603f8ff0 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDUFactory.java
@@ -21,7 +21,7 @@ import java.io.File;
 import java.time.Duration;
 
 import org.apache.hadoop.hdds.conf.OzoneConfiguration;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import static org.apache.ozone.test.GenericTestUtils.getTestDir;
 import static org.junit.jupiter.api.Assertions.assertEquals;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java
index 762639f413..3fc7c4eddf 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsage.java
@@ -18,16 +18,16 @@
 package org.apache.hadoop.hdds.fs;
 
 import org.apache.hadoop.fs.FileUtil;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 
 import java.io.File;
 import java.io.IOException;
 
 import static org.apache.hadoop.hdds.fs.TestDU.createFile;
 import static org.apache.ozone.test.GenericTestUtils.getTestDir;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 
 /**
  * Tests for {@link DedicatedDiskSpaceUsage}.
@@ -39,13 +39,13 @@ public class TestDedicatedDiskSpaceUsage {
 
   private static final int FILE_SIZE = 1024;
 
-  @Before
+  @BeforeEach
   public void setUp() {
     FileUtil.fullyDelete(DIR);
     assertTrue(DIR.mkdirs());
   }
 
-  @After
+  @AfterEach
   public void tearDown() throws IOException {
     FileUtil.fullyDelete(DIR);
   }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java
index 4244edd82d..0142ee56af 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestDedicatedDiskSpaceUsageFactory.java
@@ -24,7 +24,7 @@ import org.apache.hadoop.hdds.conf.OzoneConfiguration;
 
 import static org.apache.hadoop.hdds.fs.DedicatedDiskSpaceUsageFactory.Conf.configKeyForRefreshPeriod;
 import static org.apache.ozone.test.GenericTestUtils.getTestDir;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 import static org.junit.jupiter.api.Assertions.assertEquals;
 import static org.junit.jupiter.api.Assertions.assertSame;
 
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java
index 9480992013..f35e697508 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSaveSpaceUsageToFile.java
@@ -19,9 +19,9 @@ package org.apache.hadoop.hdds.fs;
 
 import org.apache.commons.io.FileUtils;
 import org.apache.hadoop.fs.FileUtil;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 
 import java.io.File;
 import java.io.IOException;
@@ -32,9 +32,9 @@ import java.util.OptionalLong;
 
 import static org.apache.ozone.test.GenericTestUtils.getTestDir;
 import static org.apache.ozone.test.GenericTestUtils.waitFor;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 
 /**
  * Tests for {@link SaveSpaceUsageToFile}.
@@ -51,14 +51,14 @@ public class TestSaveSpaceUsageToFile {
 
   private File file;
 
-  @Before
+  @BeforeEach
   public void setup() {
     FileUtil.fullyDelete(DIR);
     assertTrue(DIR.mkdirs());
     file = new File(DIR, "space_usage.txt");
   }
 
-  @After
+  @AfterEach
   public void cleanup() {
     FileUtil.fullyDelete(DIR);
   }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java
index c3ae128424..5a6e98e229 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/fs/TestSpaceUsageFactory.java
@@ -24,8 +24,8 @@ import org.apache.hadoop.hdds.conf.OzoneConfiguration;
 import org.apache.ozone.test.GenericTestUtils.LogCapturer;
 
 import static org.apache.hadoop.hdds.fs.SpaceUsageCheckFactory.Conf.configKeyForClassName;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 import static org.junit.jupiter.api.Assertions.assertEquals;
 import static org.junit.jupiter.api.Assertions.assertNotNull;
 import static org.junit.jupiter.api.Assertions.assertSame;
@@ -58,7 +58,7 @@ public class TestSpaceUsageFactory {
     return factoryClass.cast(factory);
   }
 
-  @Before
+  @BeforeEach
   public void setUp() {
     capturer = LogCapturer.captureLogs(
         LoggerFactory.getLogger(SpaceUsageCheckFactory.class));
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java
index d1fcad7d56..e77b97a854 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/protocol/TestDatanodeDetails.java
@@ -19,7 +19,7 @@ package org.apache.hadoop.hdds.protocol;
 
 import org.apache.hadoop.hdds.protocol.DatanodeDetails.Port;
 import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import java.util.Set;
 
@@ -27,9 +27,9 @@ import static org.apache.hadoop.hdds.protocol.DatanodeDetails.Port.Name.ALL_PORT
 import static org.apache.hadoop.hdds.protocol.DatanodeDetails.Port.Name.V0_PORTS;
 import static org.apache.hadoop.ozone.ClientVersion.DEFAULT_VERSION;
 import static org.apache.hadoop.ozone.ClientVersion.VERSION_HANDLES_UNKNOWN_DN_PORTS;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
+import static org.junit.jupiter.api.Assertions.fail;
 
 /**
  * Test for {@link DatanodeDetails}.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java
index d3ddbe0ef3..75e19f5d45 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestContainerCommandRequestMessage.java
@@ -36,8 +36,8 @@ import org.apache.hadoop.ozone.common.ChecksumData;
 import org.apache.hadoop.ozone.common.OzoneChecksumException;
 
 import org.apache.ratis.thirdparty.com.google.protobuf.ByteString;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 /** Testing {@link ContainerCommandRequestMessage}. */
 public class TestContainerCommandRequestMessage {
@@ -150,6 +150,6 @@ public class TestContainerCommandRequestMessage {
         = ContainerCommandRequestMessage.toMessage(original, null);
     final ContainerCommandRequestProto computed
         = ContainerCommandRequestMessage.toProto(message.getContent(), null);
-    Assert.assertEquals(original, computed);
+    Assertions.assertEquals(original, computed);
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java
index 68fb0ffc47..d8f626165e 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestRatisHelper.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hdds.ratis;
 
 import org.apache.hadoop.hdds.conf.OzoneConfiguration;
 import org.apache.ratis.conf.RaftProperties;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 /**
  * Test RatisHelper class.
@@ -42,11 +42,11 @@ public class TestRatisHelper {
     RaftProperties raftProperties = new RaftProperties();
     RatisHelper.createRaftClientProperties(ozoneConfiguration, raftProperties);
 
-    Assert.assertEquals("30s",
+    Assertions.assertEquals("30s",
         raftProperties.get("raft.client.rpc.watch.request.timeout"));
-    Assert.assertEquals("30s",
+    Assertions.assertEquals("30s",
         raftProperties.get("raft.client.rpc.request.timeout"));
-    Assert.assertNull(
+    Assertions.assertNull(
         raftProperties.get("raft.server.rpc.watch.request.timeout"));
 
   }
@@ -67,16 +67,17 @@ public class TestRatisHelper {
     RaftProperties raftProperties = new RaftProperties();
     RatisHelper.createRaftClientProperties(ozoneConfiguration, raftProperties);
 
-    Assert.assertEquals("30MB",
+    Assertions.assertEquals("30MB",
         raftProperties.get("raft.grpc.message.size.max"));
-    Assert.assertEquals("1MB",
+    Assertions.assertEquals("1MB",
         raftProperties.get("raft.grpc.flow.control.window"));
 
     // As we dont match tls and server raft.grpc properties. So they should
     // be null.
-    Assert.assertNull(raftProperties.get("raft.grpc.tls.set"));
-    Assert.assertNull(raftProperties.get("raft.grpc.tls.mutual_authn.enabled"));
-    Assert.assertNull(raftProperties.get("raft.grpc.server.port"));
+    Assertions.assertNull(raftProperties.get("raft.grpc.tls.set"));
+    Assertions.assertNull(
+        raftProperties.get("raft.grpc.tls.mutual_authn.enabled"));
+    Assertions.assertNull(raftProperties.get("raft.grpc.server.port"));
 
   }
 
@@ -100,15 +101,15 @@ public class TestRatisHelper {
     RatisHelper.createRaftServerProperties(ozoneConfiguration,
         raftProperties);
 
-    Assert.assertEquals("30MB",
+    Assertions.assertEquals("30MB",
         raftProperties.get("raft.grpc.message.size.max"));
-    Assert.assertEquals("1MB",
+    Assertions.assertEquals("1MB",
         raftProperties.get("raft.grpc.flow.control.window"));
-    Assert.assertEquals("true",
+    Assertions.assertEquals("true",
         raftProperties.get("raft.grpc.tls.enabled"));
-    Assert.assertEquals("true",
+    Assertions.assertEquals("true",
         raftProperties.get("raft.grpc.tls.mutual_authn.enabled"));
-    Assert.assertEquals("100",
+    Assertions.assertEquals("100",
         raftProperties.get("raft.grpc.server.port"));
 
   }
@@ -127,11 +128,12 @@ public class TestRatisHelper {
     RaftProperties raftProperties = new RaftProperties();
     RatisHelper.createRaftServerProperties(ozoneConfiguration, raftProperties);
 
-    Assert.assertEquals("30s",
+    Assertions.assertEquals("30s",
         raftProperties.get("raft.server.rpc.watch.request.timeout"));
-    Assert.assertEquals("30s",
+    Assertions.assertEquals("30s",
         raftProperties.get("raft.server.rpc.request.timeout"));
-    Assert.assertNull(raftProperties.get("raft.client.rpc.request.timeout"));
+    Assertions.assertNull(
+        raftProperties.get("raft.client.rpc.request.timeout"));
 
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java
index 3bdcf85cdf..7d44d060fd 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/ratis/TestServerNotLeaderException.java
@@ -19,8 +19,8 @@
 package org.apache.hadoop.hdds.ratis;
 
 import edu.umd.cs.findbugs.annotations.SuppressFBWarnings;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 /** Class to test {@link ServerNotLeaderException} parsing. **/
 
@@ -34,7 +34,7 @@ public class TestServerNotLeaderException {
         "Server:cf0bc565-a41b-4784-a24d-3048d5a5b013 is not the leader. "
             + "Suggested leader is Server:scm5-3.scm5.root.hwx.site:9863";
     ServerNotLeaderException snle = new ServerNotLeaderException(msg);
-    Assert.assertEquals(snle.getSuggestedLeader(), "scm5-3.scm5.root.hwx" +
+    Assertions.assertEquals(snle.getSuggestedLeader(), "scm5-3.scm5.root.hwx" +
         ".site:9863");
 
     String message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -42,7 +42,7 @@ public class TestServerNotLeaderException {
         "at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
         ".convertToNotLeaderException(ServerNotLeaderException.java:96)";
     snle = new ServerNotLeaderException(message);
-    Assert.assertEquals("scm5-3.scm5.root.hwx.site:9863",
+    Assertions.assertEquals("scm5-3.scm5.root.hwx.site:9863",
         snle.getSuggestedLeader());
 
     // Test hostname with out "."
@@ -51,7 +51,7 @@ public class TestServerNotLeaderException {
         "at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
         ".convertToNotLeaderException(ServerNotLeaderException.java:96)";
     snle = new ServerNotLeaderException(message);
-    Assert.assertEquals("localhost:98634",
+    Assertions.assertEquals("localhost:98634",
         snle.getSuggestedLeader());
 
     message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -59,7 +59,7 @@ public class TestServerNotLeaderException {
         "at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
         ".convertToNotLeaderException(ServerNotLeaderException.java:96)";
     snle = new ServerNotLeaderException(message);
-    Assert.assertEquals(null,
+    Assertions.assertEquals(null,
         snle.getSuggestedLeader());
 
     message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -67,7 +67,7 @@ public class TestServerNotLeaderException {
         "at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
         ".convertToNotLeaderException(ServerNotLeaderException.java:96)";
     snle = new ServerNotLeaderException(message);
-    Assert.assertEquals("localhost:98634",
+    Assertions.assertEquals("localhost:98634",
         snle.getSuggestedLeader());
 
     message = "Server:7fdd7170-75cc-4e11-b343-c2657c2f2f39 is not the " +
@@ -75,7 +75,7 @@ public class TestServerNotLeaderException {
         "at org.apache.hadoop.hdds.ratis.ServerNotLeaderException" +
         ".convertToNotLeaderException(ServerNotLeaderException.java)";
     snle = new ServerNotLeaderException(message);
-    Assert.assertEquals(null,
+    Assertions.assertEquals(null,
         snle.getSuggestedLeader());
   }
 
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java
index 9ec19bf327..dfc5a0c7a9 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerInfo.java
@@ -22,8 +22,8 @@ import org.apache.hadoop.hdds.client.RatisReplicationConfig;
 import org.apache.hadoop.hdds.client.ReplicationConfig;
 import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
 import org.apache.hadoop.hdds.scm.pipeline.PipelineID;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 import java.io.IOException;
 
@@ -44,34 +44,34 @@ public class TestContainerInfo {
     HddsProtos.ContainerInfoProto proto = container.getProtobuf();
 
     // No EC Config
-    Assert.assertFalse(proto.hasEcReplicationConfig());
-    Assert.assertEquals(THREE, proto.getReplicationFactor());
-    Assert.assertEquals(RATIS, proto.getReplicationType());
+    Assertions.assertFalse(proto.hasEcReplicationConfig());
+    Assertions.assertEquals(THREE, proto.getReplicationFactor());
+    Assertions.assertEquals(RATIS, proto.getReplicationType());
 
     // Reconstruct object from Proto
     ContainerInfo recovered = ContainerInfo.fromProtobuf(proto);
-    Assert.assertEquals(RATIS, recovered.getReplicationType());
-    Assert.assertTrue(
+    Assertions.assertEquals(RATIS, recovered.getReplicationType());
+    Assertions.assertTrue(
         recovered.getReplicationConfig() instanceof RatisReplicationConfig);
 
     // EC Config
     container = createContainerInfo(new ECReplicationConfig(3, 2));
     proto = container.getProtobuf();
 
-    Assert.assertEquals(3, proto.getEcReplicationConfig().getData());
-    Assert.assertEquals(2, proto.getEcReplicationConfig().getParity());
-    Assert.assertFalse(proto.hasReplicationFactor());
-    Assert.assertEquals(EC, proto.getReplicationType());
+    Assertions.assertEquals(3, proto.getEcReplicationConfig().getData());
+    Assertions.assertEquals(2, proto.getEcReplicationConfig().getParity());
+    Assertions.assertFalse(proto.hasReplicationFactor());
+    Assertions.assertEquals(EC, proto.getReplicationType());
 
     // Reconstruct object from Proto
     recovered = ContainerInfo.fromProtobuf(proto);
-    Assert.assertEquals(EC, recovered.getReplicationType());
-    Assert.assertTrue(
+    Assertions.assertEquals(EC, recovered.getReplicationType());
+    Assertions.assertTrue(
         recovered.getReplicationConfig() instanceof ECReplicationConfig);
     ECReplicationConfig config =
         (ECReplicationConfig)recovered.getReplicationConfig();
-    Assert.assertEquals(3, config.getData());
-    Assert.assertEquals(2, config.getParity());
+    Assertions.assertEquals(3, config.getData());
+    Assertions.assertEquals(2, config.getParity());
   }
 
   private ContainerInfo createContainerInfo(ReplicationConfig repConfig) {
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java
index 4f63ea7240..5203b98ef3 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerReplicaInfo.java
@@ -20,8 +20,8 @@ package org.apache.hadoop.hdds.scm.container;
 import org.apache.hadoop.hdds.protocol.DatanodeDetails;
 import org.apache.hadoop.hdds.protocol.MockDatanodeDetails;
 import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 import java.util.UUID;
 
@@ -46,17 +46,17 @@ public class TestContainerReplicaInfo {
 
     ContainerReplicaInfo info = ContainerReplicaInfo.fromProto(proto);
 
-    Assert.assertEquals(proto.getContainerID(), info.getContainerID());
-    Assert.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
-    Assert.assertEquals(proto.getKeyCount(), info.getKeyCount());
-    Assert.assertEquals(proto.getPlaceOfBirth(),
+    Assertions.assertEquals(proto.getContainerID(), info.getContainerID());
+    Assertions.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
+    Assertions.assertEquals(proto.getKeyCount(), info.getKeyCount());
+    Assertions.assertEquals(proto.getPlaceOfBirth(),
         info.getPlaceOfBirth().toString());
-    Assert.assertEquals(DatanodeDetails.getFromProtoBuf(
+    Assertions.assertEquals(DatanodeDetails.getFromProtoBuf(
         proto.getDatanodeDetails()), info.getDatanodeDetails());
-    Assert.assertEquals(proto.getSequenceID(), info.getSequenceId());
-    Assert.assertEquals(proto.getState(), info.getState());
+    Assertions.assertEquals(proto.getSequenceID(), info.getSequenceId());
+    Assertions.assertEquals(proto.getState(), info.getState());
     // If replicaIndex is not in the proto, then -1 should be returned
-    Assert.assertEquals(-1, info.getReplicaIndex());
+    Assertions.assertEquals(-1, info.getReplicaIndex());
   }
 
   @Test
@@ -76,15 +76,15 @@ public class TestContainerReplicaInfo {
 
     ContainerReplicaInfo info = ContainerReplicaInfo.fromProto(proto);
 
-    Assert.assertEquals(proto.getContainerID(), info.getContainerID());
-    Assert.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
-    Assert.assertEquals(proto.getKeyCount(), info.getKeyCount());
-    Assert.assertEquals(proto.getPlaceOfBirth(),
+    Assertions.assertEquals(proto.getContainerID(), info.getContainerID());
+    Assertions.assertEquals(proto.getBytesUsed(), info.getBytesUsed());
+    Assertions.assertEquals(proto.getKeyCount(), info.getKeyCount());
+    Assertions.assertEquals(proto.getPlaceOfBirth(),
         info.getPlaceOfBirth().toString());
-    Assert.assertEquals(DatanodeDetails.getFromProtoBuf(
+    Assertions.assertEquals(DatanodeDetails.getFromProtoBuf(
         proto.getDatanodeDetails()), info.getDatanodeDetails());
-    Assert.assertEquals(proto.getSequenceID(), info.getSequenceId());
-    Assert.assertEquals(proto.getState(), info.getState());
-    Assert.assertEquals(4, info.getReplicaIndex());
+    Assertions.assertEquals(proto.getSequenceID(), info.getSequenceId());
+    Assertions.assertEquals(proto.getState(), info.getState());
+    Assertions.assertEquals(4, info.getReplicaIndex());
   }
 }
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java
index a68f70c871..ada60397ef 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/TestReplicationManagerReport.java
@@ -21,9 +21,8 @@ import com.fasterxml.jackson.databind.JsonNode;
 import com.fasterxml.jackson.databind.ObjectMapper;
 import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
 import org.apache.hadoop.hdds.server.JsonUtils;
-import org.junit.Assert;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -32,6 +31,9 @@ import java.util.Random;
 import java.util.concurrent.ThreadLocalRandom;
 
 import static com.fasterxml.jackson.databind.node.JsonNodeType.ARRAY;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 
 /**
  * Tests for the ReplicationManagerReport class.
@@ -40,7 +42,7 @@ public class TestReplicationManagerReport {
 
   private ReplicationManagerReport report;
 
-  @Before
+  @BeforeEach
   public void setup() {
     report = new ReplicationManagerReport();
   }
@@ -55,18 +57,18 @@ public class TestReplicationManagerReport {
     report.increment(HddsProtos.LifeCycleState.CLOSED);
     report.increment(HddsProtos.LifeCycleState.CLOSED);
 
-    Assert.assertEquals(2,
+    assertEquals(2,
         report.getStat(ReplicationManagerReport.HealthState.UNDER_REPLICATED));
-    Assert.assertEquals(1,
+    assertEquals(1,
         report.getStat(ReplicationManagerReport.HealthState.OVER_REPLICATED));
-    Assert.assertEquals(0,
+    assertEquals(0,
         report.getStat(ReplicationManagerReport.HealthState.MIS_REPLICATED));
 
-    Assert.assertEquals(1,
+    assertEquals(1,
         report.getStat(HddsProtos.LifeCycleState.OPEN));
-    Assert.assertEquals(2,
+    assertEquals(2,
         report.getStat(HddsProtos.LifeCycleState.CLOSED));
-    Assert.assertEquals(0,
+    assertEquals(0,
         report.getStat(HddsProtos.LifeCycleState.QUASI_CLOSED));
   }
 
@@ -93,29 +95,29 @@ public class TestReplicationManagerReport {
     ObjectMapper mapper = new ObjectMapper();
     JsonNode json = mapper.readTree(jsonString);
 
-    Assert.assertTrue(json.get("reportTimeStamp").longValue() > 0);
+    assertTrue(json.get("reportTimeStamp").longValue() > 0);
     JsonNode stats = json.get("stats");
-    Assert.assertEquals(1, stats.get("OPEN").longValue());
-    Assert.assertEquals(0, stats.get("CLOSING").longValue());
-    Assert.assertEquals(0, stats.get("QUASI_CLOSED").longValue());
-    Assert.assertEquals(2, stats.get("CLOSED").longValue());
-    Assert.assertEquals(0, stats.get("DELETING").longValue());
-    Assert.assertEquals(0, stats.get("DELETED").longValue());
-
-    Assert.assertEquals(2, stats.get("UNDER_REPLICATED").longValue());
-    Assert.assertEquals(1, stats.get("OVER_REPLICATED").longValue());
-    Assert.assertEquals(0, stats.get("MIS_REPLICATED").longValue());
-    Assert.assertEquals(0, stats.get("MISSING").longValue());
-    Assert.assertEquals(0, stats.get("UNHEALTHY").longValue());
-    Assert.assertEquals(0, stats.get("EMPTY").longValue());
-    Assert.assertEquals(0, stats.get("OPEN_UNHEALTHY").longValue());
-    Assert.assertEquals(0, stats.get("QUASI_CLOSED_STUCK").longValue());
+    assertEquals(1, stats.get("OPEN").longValue());
+    assertEquals(0, stats.get("CLOSING").longValue());
+    assertEquals(0, stats.get("QUASI_CLOSED").longValue());
+    assertEquals(2, stats.get("CLOSED").longValue());
+    assertEquals(0, stats.get("DELETING").longValue());
+    assertEquals(0, stats.get("DELETED").longValue());
+
+    assertEquals(2, stats.get("UNDER_REPLICATED").longValue());
+    assertEquals(1, stats.get("OVER_REPLICATED").longValue());
+    assertEquals(0, stats.get("MIS_REPLICATED").longValue());
+    assertEquals(0, stats.get("MISSING").longValue());
+    assertEquals(0, stats.get("UNHEALTHY").longValue());
+    assertEquals(0, stats.get("EMPTY").longValue());
+    assertEquals(0, stats.get("OPEN_UNHEALTHY").longValue());
+    assertEquals(0, stats.get("QUASI_CLOSED_STUCK").longValue());
 
     JsonNode samples = json.get("samples");
-    Assert.assertEquals(ARRAY, samples.get("UNDER_REPLICATED").getNodeType());
-    Assert.assertEquals(1, samples.get("UNDER_REPLICATED").get(0).longValue());
-    Assert.assertEquals(2, samples.get("UNDER_REPLICATED").get(1).longValue());
-    Assert.assertEquals(3, samples.get("OVER_REPLICATED").get(0).longValue());
+    assertEquals(ARRAY, samples.get("UNDER_REPLICATED").getNodeType());
+    assertEquals(1, samples.get("UNDER_REPLICATED").get(0).longValue());
+    assertEquals(2, samples.get("UNDER_REPLICATED").get(1).longValue());
+    assertEquals(3, samples.get("OVER_REPLICATED").get(0).longValue());
   }
 
   @Test
@@ -130,27 +132,27 @@ public class TestReplicationManagerReport {
         ReplicationManagerReport.HealthState.OVER_REPLICATED,
         new ContainerID(3));
 
-    Assert.assertEquals(2,
+    assertEquals(2,
         report.getStat(ReplicationManagerReport.HealthState.UNDER_REPLICATED));
-    Assert.assertEquals(1,
+    assertEquals(1,
         report.getStat(ReplicationManagerReport.HealthState.OVER_REPLICATED));
-    Assert.assertEquals(0,
+    assertEquals(0,
         report.getStat(ReplicationManagerReport.HealthState.MIS_REPLICATED));
 
     List<ContainerID> sample =
         report.getSample(ReplicationManagerReport.HealthState.UNDER_REPLICATED);
-    Assert.assertEquals(new ContainerID(1), sample.get(0));
-    Assert.assertEquals(new ContainerID(2), sample.get(1));
-    Assert.assertEquals(2, sample.size());
+    assertEquals(new ContainerID(1), sample.get(0));
+    assertEquals(new ContainerID(2), sample.get(1));
+    assertEquals(2, sample.size());
 
     sample =
         report.getSample(ReplicationManagerReport.HealthState.OVER_REPLICATED);
-    Assert.assertEquals(new ContainerID(3), sample.get(0));
-    Assert.assertEquals(1, sample.size());
+    assertEquals(new ContainerID(3), sample.get(0));
+    assertEquals(1, sample.size());
 
     sample =
         report.getSample(ReplicationManagerReport.HealthState.MIS_REPLICATED);
-    Assert.assertEquals(0, sample.size());
+    assertEquals(0, sample.size());
   }
 
   @Test
@@ -162,9 +164,9 @@ public class TestReplicationManagerReport {
     }
     List<ContainerID> sample =
         report.getSample(ReplicationManagerReport.HealthState.UNDER_REPLICATED);
-    Assert.assertEquals(ReplicationManagerReport.SAMPLE_LIMIT, sample.size());
+    assertEquals(ReplicationManagerReport.SAMPLE_LIMIT, sample.size());
     for (int i = 0; i < ReplicationManagerReport.SAMPLE_LIMIT; i++) {
-      Assert.assertEquals(new ContainerID(i), sample.get(i));
+      assertEquals(new ContainerID(i), sample.get(i));
     }
   }
 
@@ -187,16 +189,16 @@ public class TestReplicationManagerReport {
     HddsProtos.ReplicationManagerReportProto proto = report.toProtobuf();
     ReplicationManagerReport newReport
         = ReplicationManagerReport.fromProtobuf(proto);
-    Assert.assertEquals(report.getReportTimeStamp(),
+    assertEquals(report.getReportTimeStamp(),
         newReport.getReportTimeStamp());
 
     for (HddsProtos.LifeCycleState s : HddsProtos.LifeCycleState.values()) {
-      Assert.assertEquals(report.getStat(s), newReport.getStat(s));
+      assertEquals(report.getStat(s), newReport.getStat(s));
     }
 
     for (ReplicationManagerReport.HealthState s :
         ReplicationManagerReport.HealthState.values()) {
-      Assert.assertTrue(report.getSample(s).equals(newReport.getSample(s)));
+      assertEquals(report.getSample(s), newReport.getSample(s));
     }
   }
 
@@ -225,21 +227,23 @@ public class TestReplicationManagerReport {
     // Ensure no exception is thrown
     ReplicationManagerReport newReport =
         ReplicationManagerReport.fromProtobuf(proto.build());
-    Assert.assertEquals(20, newReport.getStat(
+    assertEquals(20, newReport.getStat(
         ReplicationManagerReport.HealthState.UNDER_REPLICATED));
   }
 
-  @Test(expected = IllegalStateException.class)
+  @Test
   public void testStatCannotBeSetTwice() {
     report.setStat(HddsProtos.LifeCycleState.CLOSED.toString(), 10);
-    report.setStat(HddsProtos.LifeCycleState.CLOSED.toString(), 10);
+    assertThrows(IllegalStateException.class, () -> report
+        .setStat(HddsProtos.LifeCycleState.CLOSED.toString(), 10));
   }
 
-  @Test(expected = IllegalStateException.class)
+  @Test
   public void testSampleCannotBeSetTwice() {
     List<ContainerID> containers = new ArrayList<>();
     containers.add(ContainerID.valueOf(1));
     report.setSample(HddsProtos.LifeCycleState.CLOSED.toString(), containers);
-    report.setSample(HddsProtos.LifeCycleState.CLOSED.toString(), containers);
+    assertThrows(IllegalStateException.class, () -> report
+        .setSample(HddsProtos.LifeCycleState.CLOSED.toString(), containers));
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java
index d3fb49b3a2..d5330749ae 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/container/common/helpers/TestExcludeList.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.hdds.scm.container.common.helpers;
 
 import org.apache.hadoop.hdds.protocol.DatanodeDetails;
 import org.apache.ozone.test.TestClock;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 import java.time.Instant;
 import java.time.ZoneOffset;
@@ -39,9 +39,9 @@ public class TestExcludeList {
         .setIpAddress("127.0.0.1").setHostName("localhost").addPort(
             DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
         .build());
-    Assert.assertTrue(list.getDatanodes().size() == 1);
+    Assertions.assertTrue(list.getDatanodes().size() == 1);
     clock.fastForward(11);
-    Assert.assertTrue(list.getDatanodes().size() == 0);
+    Assertions.assertTrue(list.getDatanodes().size() == 0);
     list.addDatanode(DatanodeDetails.newBuilder().setUuid(UUID.randomUUID())
         .setIpAddress("127.0.0.2").setHostName("localhost").addPort(
             DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
@@ -50,7 +50,7 @@ public class TestExcludeList {
         .setIpAddress("127.0.0.3").setHostName("localhost").addPort(
             DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
         .build());
-    Assert.assertTrue(list.getDatanodes().size() == 2);
+    Assertions.assertTrue(list.getDatanodes().size() == 2);
   }
 
   @Test
@@ -60,8 +60,8 @@ public class TestExcludeList {
         .setIpAddress("127.0.0.1").setHostName("localhost").addPort(
             DatanodeDetails.newPort(DatanodeDetails.Port.Name.STANDALONE, 2001))
         .build());
-    Assert.assertTrue(list.getDatanodes().size() == 1);
+    Assertions.assertTrue(list.getDatanodes().size() == 1);
     clock.fastForward(1);
-    Assert.assertTrue(list.getDatanodes().size() == 1);
+    Assertions.assertTrue(list.getDatanodes().size() == 1);
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java
index e561bb7ccc..12129f4d3f 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/ha/TestSCMNodeInfo.java
@@ -22,9 +22,8 @@ import org.apache.hadoop.hdds.conf.ConfigurationException;
 import org.apache.hadoop.hdds.conf.OzoneConfiguration;
 import org.apache.hadoop.hdds.scm.ScmConfigKeys;
 import org.apache.hadoop.ozone.ha.ConfUtils;
-import org.junit.Assert;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 
 import java.util.List;
 
@@ -41,6 +40,9 @@ import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_DATANODE_PORT_K
 import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_SECURITY_SERVICE_ADDRESS_KEY;
 import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT;
 import static org.apache.hadoop.hdds.scm.ScmConfigKeys.OZONE_SCM_SECURITY_SERVICE_PORT_KEY;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertNotNull;
+import static org.junit.jupiter.api.Assertions.assertThrows;
 
 /**
  * Tests for {@link SCMNodeInfo}.
@@ -51,7 +53,7 @@ public class TestSCMNodeInfo {
   private String scmServiceId = "scmserviceId";
   private String[] nodes = new String[]{"scm1", "scm2", "scm3"};
 
-  @Before
+  @BeforeEach
   public void setup() {
     conf.set(ScmConfigKeys.OZONE_SCM_SERVICE_IDS_KEY, scmServiceId);
     conf.set(ScmConfigKeys.OZONE_SCM_NODES_KEY + "." + scmServiceId,
@@ -92,15 +94,15 @@ public class TestSCMNodeInfo {
 
     int count = 1;
     for (SCMNodeInfo scmNodeInfo : scmNodeInfos) {
-      Assert.assertEquals(scmServiceId, scmNodeInfo.getServiceId());
-      Assert.assertEquals("scm" + count++, scmNodeInfo.getNodeId());
-      Assert.assertEquals("localhost:" + ++port,
+      assertEquals(scmServiceId, scmNodeInfo.getServiceId());
+      assertEquals("scm" + count++, scmNodeInfo.getNodeId());
+      assertEquals("localhost:" + ++port,
           scmNodeInfo.getBlockClientAddress());
-      Assert.assertEquals("localhost:" + ++port,
+      assertEquals("localhost:" + ++port,
           scmNodeInfo.getScmSecurityAddress());
-      Assert.assertEquals("localhost:" + ++port,
+      assertEquals("localhost:" + ++port,
           scmNodeInfo.getScmClientAddress());
-      Assert.assertEquals("localhost:" + ++port,
+      assertEquals("localhost:" + ++port,
           scmNodeInfo.getScmDatanodeAddress());
     }
   }
@@ -116,30 +118,31 @@ public class TestSCMNodeInfo {
 
     int count = 1;
     for (SCMNodeInfo scmNodeInfo : scmNodeInfos) {
-      Assert.assertEquals(scmServiceId, scmNodeInfo.getServiceId());
-      Assert.assertEquals("scm" + count++, scmNodeInfo.getNodeId());
-      Assert.assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
+      assertEquals(scmServiceId, scmNodeInfo.getServiceId());
+      assertEquals("scm" + count++, scmNodeInfo.getNodeId());
+      assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
           scmNodeInfo.getBlockClientAddress());
-      Assert.assertEquals("localhost:" +
+      assertEquals("localhost:" +
               OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT,
           scmNodeInfo.getScmSecurityAddress());
-      Assert.assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
+      assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
           scmNodeInfo.getScmClientAddress());
-      Assert.assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
+      assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
           scmNodeInfo.getScmDatanodeAddress());
     }
 
 
   }
 
-  @Test(expected = ConfigurationException.class)
+  @Test
   public void testSCMHANodeInfoWithMissingSCMAddress() {
     conf.set(ConfUtils.addKeySuffixes(OZONE_SCM_ADDRESS_KEY,
         scmServiceId, "scm1"), "localhost");
     conf.set(ConfUtils.addKeySuffixes(OZONE_SCM_ADDRESS_KEY,
         scmServiceId, "scm1"), "localhost");
 
-    SCMNodeInfo.buildNodeInfo(conf);
+    assertThrows(ConfigurationException.class,
+        () -> SCMNodeInfo.buildNodeInfo(conf));
   }
 
   @Test
@@ -150,15 +153,15 @@ public class TestSCMNodeInfo {
 
     List< SCMNodeInfo > scmNodeInfos = SCMNodeInfo.buildNodeInfo(config);
 
-    Assert.assertNotNull(scmNodeInfos);
-    Assert.assertTrue(scmNodeInfos.size() == 1);
-    Assert.assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
+    assertNotNull(scmNodeInfos);
+    assertEquals(1, scmNodeInfos.size());
+    assertEquals("localhost:" + OZONE_SCM_BLOCK_CLIENT_PORT_DEFAULT,
         scmNodeInfos.get(0).getBlockClientAddress());
-    Assert.assertEquals("localhost:" + OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT,
+    assertEquals("localhost:" + OZONE_SCM_SECURITY_SERVICE_PORT_DEFAULT,
         scmNodeInfos.get(0).getScmSecurityAddress());
-    Assert.assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
+    assertEquals("localhost:" + OZONE_SCM_CLIENT_PORT_DEFAULT,
         scmNodeInfos.get(0).getScmClientAddress());
-    Assert.assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
+    assertEquals("localhost:" + OZONE_SCM_DATANODE_PORT_DEFAULT,
         scmNodeInfos.get(0).getScmDatanodeAddress());
   }
 
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java
index e040e489ce..f3bf37ab28 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/net/TestNetUtils.java
@@ -17,7 +17,7 @@
  */
 package org.apache.hadoop.hdds.scm.net;
 
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import static org.junit.jupiter.api.Assertions.assertEquals;
 import static org.junit.jupiter.api.Assertions.assertThrows;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java
index b0f9875ece..ac8920991c 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestPipeline.java
@@ -19,8 +19,7 @@ package org.apache.hadoop.hdds.scm.pipeline;
 
 import org.apache.hadoop.hdds.protocol.DatanodeDetails;
 import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import java.io.IOException;
 
@@ -29,6 +28,8 @@ import static org.apache.hadoop.hdds.protocol.DatanodeDetails.Port.Name.V0_PORTS
 import static org.apache.hadoop.hdds.protocol.TestDatanodeDetails.assertPorts;
 import static org.apache.hadoop.ozone.ClientVersion.DEFAULT_VERSION;
 import static org.apache.hadoop.ozone.ClientVersion.VERSION_HANDLES_UNKNOWN_DN_PORTS;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 
 /**
  * Test for {@link Pipeline}.
@@ -58,16 +59,15 @@ public class TestPipeline {
 
     //when EC config is empty/null
     HddsProtos.Pipeline protobufMessage = subject.getProtobufMessage(1);
-    Assert.assertEquals(0, protobufMessage.getEcReplicationConfig().getData());
+    assertEquals(0, protobufMessage.getEcReplicationConfig().getData());
 
 
     //when EC config is NOT empty
     subject = MockPipeline.createEcPipeline();
 
     protobufMessage = subject.getProtobufMessage(1);
-    Assert.assertEquals(3, protobufMessage.getEcReplicationConfig().getData());
-    Assert
-        .assertEquals(2, protobufMessage.getEcReplicationConfig().getParity());
+    assertEquals(3, protobufMessage.getEcReplicationConfig().getData());
+    assertEquals(2, protobufMessage.getEcReplicationConfig().getParity());
 
   }
 
@@ -78,37 +78,37 @@ public class TestPipeline {
     Pipeline reloadedPipeline = Pipeline.getFromProtobuf(protobufMessage);
 
     for (DatanodeDetails dn : pipeline.getNodes()) {
-      Assert.assertEquals(pipeline.getReplicaIndex(dn),
+      assertEquals(pipeline.getReplicaIndex(dn),
           reloadedPipeline.getReplicaIndex(dn));
     }
   }
 
   @Test
-  public void testECPipelineIsAlwaysHealthy() throws IOException {
+  public void testECPipelineIsAlwaysHealthy() {
     Pipeline pipeline = MockPipeline.createEcPipeline();
-    Assert.assertTrue(pipeline.isHealthy());
+    assertTrue(pipeline.isHealthy());
   }
 
   @Test
   public void testBuilderCopiesAllFieldsFromOtherPipeline() {
     Pipeline original = MockPipeline.createEcPipeline();
     Pipeline copied = Pipeline.newBuilder(original).build();
-    Assert.assertEquals(original.getId(), copied.getId());
-    Assert.assertEquals(original.getReplicationConfig(),
+    assertEquals(original.getId(), copied.getId());
+    assertEquals(original.getReplicationConfig(),
         copied.getReplicationConfig());
-    Assert.assertEquals(original.getPipelineState(), copied.getPipelineState());
-    Assert.assertEquals(original.getId(), copied.getId());
-    Assert.assertEquals(original.getId(), copied.getId());
-    Assert.assertEquals(original.getId(), copied.getId());
-    Assert.assertEquals(original.getNodeSet(), copied.getNodeSet());
-    Assert.assertEquals(original.getNodesInOrder(), copied.getNodesInOrder());
-    Assert.assertEquals(original.getLeaderId(), copied.getLeaderId());
-    Assert.assertEquals(original.getCreationTimestamp(),
+    assertEquals(original.getPipelineState(), copied.getPipelineState());
+    assertEquals(original.getId(), copied.getId());
+    assertEquals(original.getId(), copied.getId());
+    assertEquals(original.getId(), copied.getId());
+    assertEquals(original.getNodeSet(), copied.getNodeSet());
+    assertEquals(original.getNodesInOrder(), copied.getNodesInOrder());
+    assertEquals(original.getLeaderId(), copied.getLeaderId());
+    assertEquals(original.getCreationTimestamp(),
         copied.getCreationTimestamp());
-    Assert.assertEquals(original.getSuggestedLeaderId(),
+    assertEquals(original.getSuggestedLeaderId(),
         copied.getSuggestedLeaderId());
     for (DatanodeDetails dn : original.getNodes()) {
-      Assert.assertEquals(original.getReplicaIndex(dn),
+      assertEquals(original.getReplicaIndex(dn),
           copied.getReplicaIndex(dn));
     }
   }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java
index 36c7a6e8a1..036f2586f5 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTraceAllMethod.java
@@ -17,10 +17,10 @@
  */
 package org.apache.hadoop.hdds.tracing;
 
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertThrows;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertThrows;
 
 /**
  * Test for {@link TraceAllMethod}.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java
index 135b237a7a..378ac6d4df 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/tracing/TestTracingUtil.java
@@ -22,10 +22,10 @@ import org.apache.hadoop.hdds.conf.MutableConfigurationSource;
 import org.apache.hadoop.hdds.scm.ScmConfigKeys;
 import org.apache.hadoop.hdds.tracing.TestTraceAllMethod.Service;
 import org.apache.hadoop.hdds.tracing.TestTraceAllMethod.ServiceImpl;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import static org.apache.hadoop.hdds.tracing.TracingUtil.createProxy;
-import static org.junit.Assert.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
 
 /**
  * Test for {@link TracingUtil}.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java
index 4a09f3fa94..8a81aea168 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/upgrade/TestHDDSLayoutVersionManager.java
@@ -23,9 +23,9 @@ import static org.apache.hadoop.hdds.upgrade.HDDSLayoutFeature.INITIAL_VERSION;
 import static org.apache.hadoop.hdds.upgrade.HDDSLayoutVersionManager.maxLayoutVersion;
 import static org.apache.hadoop.ozone.upgrade.LayoutFeature.UpgradeActionType.ON_FINALIZE;
 import static org.apache.hadoop.ozone.upgrade.LayoutFeature.UpgradeActionType.ON_FIRST_UPGRADE_START;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 import static org.mockito.ArgumentMatchers.any;
 import static org.mockito.Mockito.doCallRealMethod;
 import static org.mockito.Mockito.mock;
@@ -38,8 +38,8 @@ import java.util.Optional;
 import org.apache.hadoop.hdds.upgrade.test.MockComponent;
 import org.apache.hadoop.hdds.upgrade.test.MockComponent.MockDnUpgradeAction;
 import org.apache.hadoop.hdds.upgrade.test.MockComponent.MockScmUpgradeAction;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 /**
  * Class to test HDDS upgrade action registrations.
@@ -58,9 +58,9 @@ public class TestHDDSLayoutVersionManager {
 
     //Cluster is finalized, hence should not register.
     Optional<HDDSUpgradeAction> action = INITIAL_VERSION.scmAction(ON_FINALIZE);
-    Assert.assertFalse(action.isPresent());
+    Assertions.assertFalse(action.isPresent());
     action = DATANODE_SCHEMA_V2.datanodeAction(ON_FIRST_UPGRADE_START);
-    Assert.assertFalse(action.isPresent());
+    Assertions.assertFalse(action.isPresent());
 
     // Start from an unfinalized version manager.
     lvm = mock(HDDSLayoutVersionManager.class);
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java
index 11d0fad55e..cc27a3b4a7 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestHddsIdFactory.java
@@ -26,11 +26,11 @@ import java.util.concurrent.ExecutorService;
 import java.util.concurrent.Executors;
 import java.util.concurrent.Future;
 import org.apache.hadoop.hdds.HddsIdFactory;
-import org.junit.After;
-import static org.junit.Assert.assertEquals;
+import org.junit.jupiter.api.AfterEach;
+import static org.junit.jupiter.api.Assertions.assertEquals;
 
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 /**
  * Test the JMX interface for the rocksdb metastore implementation.
@@ -41,7 +41,7 @@ public class TestHddsIdFactory {
   private static final int IDS_PER_THREAD = 10000;
   private static final int NUM_OF_THREADS = 5;
 
-  @After
+  @AfterEach
   public void cleanup() {
     ID_SET.clear();
   }
@@ -65,7 +65,7 @@ public class TestHddsIdFactory {
         for (int idNum = 0; idNum < IDS_PER_THREAD; idNum++) {
           long var = HddsIdFactory.getLongId();
           if (ID_SET.contains(var)) {
-            Assert.fail("Duplicate id found");
+            Assertions.fail("Duplicate id found");
           }
           ID_SET.add(var);
         }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java
index 2f54922edc..4dd12fe37f 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceLimitCache.java
@@ -17,8 +17,9 @@
 package org.apache.hadoop.hdds.utils;
 
 import org.apache.ozone.test.GenericTestUtils;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
 
 import java.util.concurrent.CompletableFuture;
 import java.util.concurrent.ConcurrentHashMap;
@@ -48,7 +49,7 @@ public class TestResourceLimitCache {
 
     // Create a future which blocks to put 1. Currently map has acquired 10
     // permits out of 10
-    CompletableFuture future = CompletableFuture.supplyAsync(() -> {
+    CompletableFuture<String> future = CompletableFuture.supplyAsync(() -> {
       try {
         return resourceCache.put(1, "a");
       } catch (InterruptedException e) {
@@ -56,9 +57,9 @@ public class TestResourceLimitCache {
       }
       return null;
     });
-    Assert.assertTrue(!future.isDone());
+    Assertions.assertFalse(future.isDone());
     Thread.sleep(100);
-    Assert.assertTrue(!future.isDone());
+    Assertions.assertFalse(future.isDone());
 
     // remove 4 so that permits are released for key 1 to be put. Currently map
     // has acquired 6 permits out of 10
@@ -66,8 +67,9 @@ public class TestResourceLimitCache {
 
     GenericTestUtils.waitFor(future::isDone, 100, 1000);
     // map has the key 1
-    Assert.assertTrue(future.isDone() && !future.isCompletedExceptionally());
-    Assert.assertNotNull(resourceCache.get(1));
+    Assertions.assertTrue(future.isDone());
+    Assertions.assertFalse(future.isCompletedExceptionally());
+    Assertions.assertNotNull(resourceCache.get(1));
 
     // Create a future which blocks to put 4. Currently map has acquired 7
     // permits out of 10
@@ -79,9 +81,9 @@ public class TestResourceLimitCache {
         return null;
       }
     }, pool);
-    Assert.assertTrue(!future.isDone());
+    Assertions.assertFalse(future.isDone());
     Thread.sleep(100);
-    Assert.assertTrue(!future.isDone());
+    Assertions.assertFalse(future.isDone());
 
     // Shutdown the thread pool for putting key 4
     pool.shutdownNow();
@@ -89,20 +91,23 @@ public class TestResourceLimitCache {
     future.cancel(true);
     // remove key 1 so currently map has acquired 6 permits out of 10
     resourceCache.remove(1);
-    Assert.assertNull(resourceCache.get(4));
+    Assertions.assertNull(resourceCache.get(4));
   }
 
-  @Test(timeout = 5000)
+  @Test
+  @Timeout(5)
   public void testRemove() throws Exception {
     testRemove(cache -> cache.remove(2), 2);
   }
 
-  @Test(timeout = 5000)
+  @Test
+  @Timeout(5)
   public void testRemoveIf() throws Exception {
     testRemove(cache -> cache.removeIf(k -> k <= 2), 1, 2);
   }
 
-  @Test(timeout = 5000)
+  @Test
+  @Timeout(5)
   public void testClear() throws Exception {
     testRemove(Cache::clear, 1, 2, 3);
   }
@@ -124,7 +129,7 @@ public class TestResourceLimitCache {
 
     // THEN
     for (Integer k : removedKeys) {
-      Assert.assertNull(resourceCache.get(k));
+      Assertions.assertNull(resourceCache.get(k));
     }
     // can put new entries
     for (int i = 1; i <= removedKeys.length; ++i) {
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java
index f9c194d45c..43e74e687f 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestResourceSemaphore.java
@@ -17,14 +17,18 @@
  */
 package org.apache.hadoop.hdds.utils;
 
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
+
+import static org.junit.jupiter.api.Assertions.assertThrows;
 
 /**
  * Test for ResourceSemaphore.
  */
 public class TestResourceSemaphore {
-  @Test(timeout = 1000)
+  @Test
+  @Timeout(1)
   public void testGroup() {
     final ResourceSemaphore.Group g = new ResourceSemaphore.Group(3, 1);
 
@@ -49,28 +53,20 @@ public class TestResourceSemaphore {
     g.release(0, 0);
     assertUsed(g, 0, 0);
 
-    try {
-      g.release(1, 0);
-      Assert.fail("Should have failed.");
-    } catch (IllegalStateException e) {
-    }
-    try {
-      g.release(0, 1);
-      Assert.fail("Should have failed.");
-    } catch (IllegalStateException e) {
-    }
+    assertThrows(IllegalStateException.class, () -> g.release(1, 0));
+    assertThrows(IllegalStateException.class, () -> g.release(0, 1));
   }
 
   static void assertUsed(ResourceSemaphore.Group g, int... expected) {
-    Assert.assertEquals(expected.length, g.resourceSize());
+    Assertions.assertEquals(expected.length, g.resourceSize());
     for (int i = 0; i < expected.length; i++) {
-      Assert.assertEquals(expected[i], g.get(i).used());
+      Assertions.assertEquals(expected[i], g.get(i).used());
     }
   }
 
   static void assertAcquire(ResourceSemaphore.Group g, boolean expected,
       int... permits) {
     final boolean computed = g.tryAcquire(permits);
-    Assert.assertEquals(expected, computed);
+    Assertions.assertEquals(expected, computed);
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java
index 148ccf94a1..29e75aa853 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/hdds/utils/TestRetriableTask.java
@@ -17,12 +17,12 @@
  */
 package org.apache.hadoop.hdds.utils;
 
-import static org.junit.Assert.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
 import static org.junit.jupiter.api.Assertions.assertThrows;
 
 import org.apache.hadoop.io.retry.RetryPolicies;
 import org.apache.hadoop.io.retry.RetryPolicy;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import java.io.IOException;
 import java.util.concurrent.TimeUnit;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java
index 3ba76d651d..1d36b662f9 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/TestOzoneConsts.java
@@ -17,8 +17,8 @@
  */
 package org.apache.hadoop.ozone;
 
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 /**
  * Test Ozone Constants (e.g. for compatibility).
@@ -28,7 +28,7 @@ public class TestOzoneConsts {
   @Test
   public void testOzoneTenantPolicyLabelCompatibility() {
     // Value of the policy label should not be changed
-    Assert.assertEquals(
+    Assertions.assertEquals(
         "OzoneTenant", OzoneConsts.OZONE_TENANT_RANGER_POLICY_LABEL);
   }
 
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java
index 1b8e744554..d42795ede5 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/audit/TestOzoneAuditLogger.java
@@ -19,9 +19,9 @@ package org.apache.hadoop.ozone.audit;
 
 import org.apache.commons.io.FileUtils;
 import org.apache.hadoop.hdds.conf.OzoneConfiguration;
-import org.junit.AfterClass;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
@@ -37,8 +37,8 @@ import static org.apache.hadoop.ozone.audit.AuditEventStatus.FAILURE;
 import static org.apache.hadoop.ozone.audit.AuditEventStatus.SUCCESS;
 import static org.hamcrest.MatcherAssert.assertThat;
 import static org.hamcrest.core.StringContains.containsString;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 import org.hamcrest.Matcher;
 import org.hamcrest.collection.IsIterableContainingInOrder;
 
@@ -102,7 +102,7 @@ public class TestOzoneAuditLogger {
           .withResult(SUCCESS)
           .withException(null).build();
 
-  @AfterClass
+  @AfterAll
   public static void tearDown() {
     File file = new File("audit.log");
     if (FileUtils.deleteQuietly(file)) {
@@ -113,7 +113,7 @@ public class TestOzoneAuditLogger {
     }
   }
 
-  @Before
+  @BeforeEach
   public void init() {
     AUDIT.refreshDebugCmdSet();
   }
@@ -166,11 +166,11 @@ public class TestOzoneAuditLogger {
   @Test
   public void messageIncludesAllParts() {
     String message = WRITE_FAIL_MSG.getFormattedMessage();
-    assertTrue(message, message.contains(USER));
-    assertTrue(message, message.contains(IP_ADDRESS));
-    assertTrue(message, message.contains(DummyAction.CREATE_VOLUME.name()));
-    assertTrue(message, message.contains(PARAMS.toString()));
-    assertTrue(message, message.contains(FAILURE.getStatus()));
+    assertTrue(message.contains(USER), message);
+    assertTrue(message.contains(IP_ADDRESS), message);
+    assertTrue(message.contains(DummyAction.CREATE_VOLUME.name()), message);
+    assertTrue(message.contains(PARAMS.toString()), message);
+    assertTrue(message.contains(FAILURE.getStatus()), message);
   }
 
   /**
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java
index a61ff9054b..e01b767b10 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksum.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.ozone.common;
 
 import org.apache.commons.lang3.RandomStringUtils;
 import org.apache.hadoop.hdds.protocol.datanode.proto.ContainerProtos;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 import static java.nio.charset.StandardCharsets.UTF_8;
 
@@ -54,11 +54,11 @@ public class TestChecksum {
     // A checksum is calculate for each bytesPerChecksum number of bytes in
     // the data. Since that value is 10 here and the data length is 55, we
     // should have 6 checksums in checksumData.
-    Assert.assertEquals(6, checksumData.getChecksums().size());
+    Assertions.assertEquals(6, checksumData.getChecksums().size());
 
     // Checksum verification should pass
-    Assert.assertTrue("Checksum mismatch",
-        Checksum.verifyChecksum(data, checksumData));
+    Assertions.assertTrue(Checksum.verifyChecksum(data, checksumData),
+        "Checksum mismatch");
   }
 
   /**
@@ -75,8 +75,8 @@ public class TestChecksum {
     // mismatch
     data[50] = (byte) (data[50] + 1);
     ChecksumData newChecksumData = checksum.computeChecksum(data);
-    Assert.assertNotEquals("Checksums should not match for different data",
-        originalChecksumData, newChecksumData);
+    Assertions.assertNotEquals(originalChecksumData, newChecksumData,
+        "Checksums should not match for different data");
   }
 
   /**
@@ -92,8 +92,7 @@ public class TestChecksum {
     Checksum checksum2 = getChecksum(ContainerProtos.ChecksumType.CRC32);
 
     // The two checksums should not match as they have different types
-    Assert.assertNotEquals(
-        "Checksums should not match for different checksum types",
-        checksum1, checksum2);
+    Assertions.assertNotEquals(checksum1, checksum2,
+        "Checksums should not match for different checksum types");
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
index 4aac0b4847..6c9236d949 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
@@ -19,8 +19,8 @@ package org.apache.hadoop.ozone.common;
 
 import org.apache.hadoop.util.PureJavaCrc32;
 import org.apache.hadoop.util.PureJavaCrc32C;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 import java.nio.charset.StandardCharsets;
 import java.util.Random;
@@ -96,7 +96,7 @@ public class TestChecksumByteBuffer {
     }
 
     private void checkSame() {
-      Assert.assertEquals(expected.getValue(), testee.getValue());
+      Assertions.assertEquals(expected.getValue(), testee.getValue());
     }
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
index b15f59f2a5..5e02ceaf08 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
@@ -21,14 +21,14 @@ import org.apache.commons.lang3.RandomUtils;
 import org.apache.hadoop.util.NativeCRC32Wrapper;
 import org.apache.hadoop.util.PureJavaCrc32;
 import org.apache.hadoop.util.PureJavaCrc32C;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import java.nio.ByteBuffer;
 import java.util.ArrayList;
 import java.util.List;
 import java.util.zip.CRC32;
 
-import static junit.framework.TestCase.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
 
 /**
  * Tests to verify that different checksum implementations compute the same
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java
index 1e85099179..c405b83012 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBuffer.java
@@ -30,9 +30,9 @@ import java.util.concurrent.ThreadLocalRandom;
 import org.apache.hadoop.hdds.utils.MockGatheringChannel;
 
 import org.apache.ratis.thirdparty.com.google.protobuf.ByteString;
-import org.junit.Assert;
-import org.junit.Test;
 import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
 
 /**
  * Test {@link ChunkBuffer} implementations.
@@ -42,7 +42,8 @@ public class TestChunkBuffer {
     return ThreadLocalRandom.current().nextInt(n);
   }
 
-  @Test(timeout = 1_000)
+  @Test
+  @Timeout(1)
   public void testImplWithByteBuffer() {
     runTestImplWithByteBuffer(1);
     runTestImplWithByteBuffer(1 << 10);
@@ -57,7 +58,8 @@ public class TestChunkBuffer {
     runTestImpl(expected, 0, ChunkBuffer.allocate(n));
   }
 
-  @Test(timeout = 1_000)
+  @Test
+  @Timeout(1)
   public void testIncrementalChunkBuffer() {
     runTestIncrementalChunkBuffer(1, 1);
     runTestIncrementalChunkBuffer(4, 8);
@@ -76,7 +78,8 @@ public class TestChunkBuffer {
         new IncrementalChunkBuffer(n, increment, false));
   }
 
-  @Test(timeout = 1_000)
+  @Test
+  @Timeout(1)
   public void testImplWithList() {
     runTestImplWithList(4, 8);
     runTestImplWithList(16, 1 << 10);
@@ -114,14 +117,14 @@ public class TestChunkBuffer {
     System.out.println("n=" + n + ", impl=" + impl);
 
     // check position, remaining
-    Assert.assertEquals(0, impl.position());
-    Assert.assertEquals(n, impl.remaining());
-    Assert.assertEquals(n, impl.limit());
+    Assertions.assertEquals(0, impl.position());
+    Assertions.assertEquals(n, impl.remaining());
+    Assertions.assertEquals(n, impl.limit());
 
     impl.put(expected);
-    Assert.assertEquals(n, impl.position());
-    Assert.assertEquals(0, impl.remaining());
-    Assert.assertEquals(n, impl.limit());
+    Assertions.assertEquals(n, impl.position());
+    Assertions.assertEquals(0, impl.remaining());
+    Assertions.assertEquals(n, impl.limit());
 
     // duplicate
     assertDuplicate(expected, impl);
@@ -159,8 +162,8 @@ public class TestChunkBuffer {
       byte[] expected, ChunkBuffer impl, int bpc) {
     final int n = expected.length;
     final ChunkBuffer duplicated = impl.duplicate(0, n);
-    Assert.assertEquals(0, duplicated.position());
-    Assert.assertEquals(n, duplicated.remaining());
+    Assertions.assertEquals(0, duplicated.position());
+    Assertions.assertEquals(n, duplicated.remaining());
 
     final int numChecksums = (n + bpc - 1) / bpc;
     final Iterator<ByteBuffer> i = duplicated.iterate(bpc).iterator();
@@ -169,57 +172,58 @@ public class TestChunkBuffer {
       final ByteBuffer b = i.next();
       final int expectedRemaining = j < numChecksums - 1 ?
           bpc : n - bpc * (numChecksums - 1);
-      Assert.assertEquals(expectedRemaining, b.remaining());
+      Assertions.assertEquals(expectedRemaining, b.remaining());
 
       final int offset = j * bpc;
       for (int k = 0; k < expectedRemaining; k++) {
-        Assert.assertEquals(expected[offset + k], b.get());
+        Assertions.assertEquals(expected[offset + k], b.get());
         count++;
       }
     }
-    Assert.assertEquals(n, count);
-    Assert.assertFalse(i.hasNext());
+    Assertions.assertEquals(n, count);
+    Assertions.assertFalse(i.hasNext());
     Assertions.assertThrows(NoSuchElementException.class, i::next);
   }
 
   private static void assertToByteString(
       byte[] expected, int offset, int length, ChunkBuffer impl) {
     final ChunkBuffer duplicated = impl.duplicate(offset, offset + length);
-    Assert.assertEquals(offset, duplicated.position());
-    Assert.assertEquals(length, duplicated.remaining());
+    Assertions.assertEquals(offset, duplicated.position());
+    Assertions.assertEquals(length, duplicated.remaining());
     final ByteString computed = duplicated.toByteString(buffer -> {
       buffer.mark();
       final ByteString string = ByteString.copyFrom(buffer);
       buffer.reset();
       return string;
     });
-    Assert.assertEquals(offset, duplicated.position());
-    Assert.assertEquals(length, duplicated.remaining());
+    Assertions.assertEquals(offset, duplicated.position());
+    Assertions.assertEquals(length, duplicated.remaining());
     assertEquals("offset=" + offset + ", length=" + length,
         ByteString.copyFrom(expected, offset, length), computed);
   }
 
   private static void assertWrite(byte[] expected, ChunkBuffer impl) {
     impl.rewind();
-    Assert.assertEquals(0, impl.position());
+    Assertions.assertEquals(0, impl.position());
 
     ByteArrayOutputStream output = new ByteArrayOutputStream(expected.length);
 
     try {
       impl.writeTo(new MockGatheringChannel(Channels.newChannel(output)));
     } catch (IOException e) {
-      Assert.fail("Unexpected error: " + e);
+      Assertions.fail("Unexpected error: " + e);
     }
 
-    Assert.assertArrayEquals(expected, output.toByteArray());
-    Assert.assertFalse(impl.hasRemaining());
+    Assertions.assertArrayEquals(expected, output.toByteArray());
+    Assertions.assertFalse(impl.hasRemaining());
   }
 
   private static void assertEquals(String message,
       ByteString expected, ByteString actual) {
-    Assert.assertEquals(message,
+    Assertions.assertEquals(
         toString(expected.toByteArray()),
-        toString(actual.toByteArray()));
+        toString(actual.toByteArray()),
+        message);
   }
 
   private static String toString(byte[] arr) {
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java
index b867fbe79b..3da43166e7 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChunkBufferImplWithByteBufferList.java
@@ -18,7 +18,7 @@
 package org.apache.hadoop.ozone.common;
 
 import com.google.common.collect.ImmutableList;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 import java.nio.BufferOverflowException;
 import java.nio.ByteBuffer;
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java
index 2d14abd3f5..5ec42b9117 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/ha/TestOzoneNetUtils.java
@@ -18,8 +18,8 @@ package org.apache.hadoop.ozone.ha;
 
 import org.apache.hadoop.net.NetUtils;
 import org.apache.hadoop.ozone.util.OzoneNetUtils;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
 
 import java.net.InetSocketAddress;
 
@@ -35,7 +35,7 @@ public class TestOzoneNetUtils {
     InetSocketAddress addr0 = NetUtils.createSocketAddr(fqdn, port);
     InetSocketAddress addr1 = OzoneNetUtils.getAddressWithHostNameLocal(
             addr0);
-    Assert.assertEquals("pod0", addr1.getHostName());
-    Assert.assertEquals(port, addr1.getPort());
+    Assertions.assertEquals("pod0", addr1.getHostName());
+    Assertions.assertEquals(port, addr1.getPort());
   }
 }
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java
index 1b992dd51c..dfdbba1d89 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/lock/TestLockManager.java
@@ -20,8 +20,9 @@ package org.apache.hadoop.ozone.lock;
 import org.apache.hadoop.hdds.conf.OzoneConfiguration;
 import org.apache.ozone.test.GenericTestUtils;
 import org.apache.hadoop.util.Daemon;
-import org.junit.Assert;
-import org.junit.Test;
+import org.junit.jupiter.api.Assertions;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.Timeout;
 
 import java.util.concurrent.atomic.AtomicBoolean;
 import java.util.concurrent.atomic.AtomicInteger;
@@ -31,7 +32,8 @@ import java.util.concurrent.atomic.AtomicInteger;
  */
 public class TestLockManager {
 
-  @Test(timeout = 1000)
+  @Test
+  @Timeout(1)
   public void testWriteLockWithDifferentResource() {
     final LockManager<String> manager =
         new LockManager<>(new OzoneConfiguration());
@@ -40,7 +42,7 @@ public class TestLockManager {
     manager.writeLock("/resourceTwo");
     manager.writeUnlock("/resourceOne");
     manager.writeUnlock("/resourceTwo");
-    Assert.assertTrue(true);
+    Assertions.assertTrue(true);
   }
 
   @Test
@@ -58,16 +60,17 @@ public class TestLockManager {
     Thread.sleep(100);
     // Since the other thread is trying to get write lock on same object,
     // it will wait.
-    Assert.assertFalse(gotLock.get());
+    Assertions.assertFalse(gotLock.get());
     manager.writeUnlock("/resourceOne");
     // Since we have released the write lock, the other thread should have
     // the lock now
     // Let's give some time for the other thread to run
     Thread.sleep(100);
-    Assert.assertTrue(gotLock.get());
+    Assertions.assertTrue(gotLock.get());
   }
 
-  @Test(timeout = 1000)
+  @Test
+  @Timeout(1)
   public void testReadLockWithDifferentResource() {
     final LockManager<String> manager =
         new LockManager<>(new OzoneConfiguration());
@@ -75,7 +78,7 @@ public class TestLockManager {
     manager.readLock("/resourceTwo");
     manager.readUnlock("/resourceOne");
     manager.readUnlock("/resourceTwo");
-    Assert.assertTrue(true);
+    Assertions.assertTrue(true);
   }
 
   @Test
@@ -92,7 +95,7 @@ public class TestLockManager {
     // Let's give some time for the other thread to run
     Thread.sleep(100);
     // Since the new thread is trying to get read lock, it should work.
-    Assert.assertTrue(gotLock.get());
+    Assertions.assertTrue(gotLock.get());
     manager.readUnlock("/resourceOne");
   }
 
@@ -111,13 +114,13 @@ public class TestLockManager {
     Thread.sleep(100);
     // Since the other thread is trying to get read lock on same object,
     // it will wait.
-    Assert.assertFalse(gotLock.get());
+    Assertions.assertFalse(gotLock.get());
     manager.writeUnlock("/resourceOne");
     // Since we have released the write lock, the other thread should have
     // the lock now
     // Let's give some time for the other thread to run
     Thread.sleep(100);
-    Assert.assertTrue(gotLock.get());
+    Assertions.assertTrue(gotLock.get());
   }
 
   @Test
@@ -135,13 +138,13 @@ public class TestLockManager {
     Thread.sleep(100);
     // Since the other thread is trying to get write lock on same object,
     // it will wait.
-    Assert.assertFalse(gotLock.get());
+    Assertions.assertFalse(gotLock.get());
     manager.readUnlock("/resourceOne");
     // Since we have released the read lock, the other thread should have
     // the lock now
     // Let's give some time for the other thread to run
     Thread.sleep(100);
-    Assert.assertTrue(gotLock.get());
+    Assertions.assertTrue(gotLock.get());
   }
 
   @Test
@@ -160,17 +163,17 @@ public class TestLockManager {
     Thread.sleep(100);
     // Since the other thread is trying to get write lock on same object,
     // it will wait.
-    Assert.assertFalse(gotLock.get());
+    Assertions.assertFalse(gotLock.get());
     manager.readUnlock("/resourceOne");
     //We have only released one read lock, we still hold another read lock.
     Thread.sleep(100);
-    Assert.assertFalse(gotLock.get());
+    Assertions.assertFalse(gotLock.get());
     manager.readUnlock("/resourceOne");
     // Since we have released the read lock, the other thread should have
     // the lock now
     // Let's give some time for the other thread to run
     Thread.sleep(100);
-    Assert.assertTrue(gotLock.get());
+    Assertions.assertTrue(gotLock.get());
   }
 
   @Test
@@ -198,6 +201,6 @@ public class TestLockManager {
     }
     GenericTestUtils.waitFor(() -> done.get() == count, 100,
         10 * count * sleep);
-    Assert.assertEquals(count, done.get());
+    Assertions.assertEquals(count, done.get());
   }
 }
\ No newline at end of file
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java
index e63246ce88..5b83f3552e 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestBasicUpgradeFinalizer.java
@@ -22,9 +22,9 @@ import static org.apache.hadoop.ozone.upgrade.TestUpgradeFinalizerActions.MockLa
 import static org.apache.hadoop.ozone.upgrade.TestUpgradeFinalizerActions.MockLayoutFeature.VERSION_3;
 import static org.apache.hadoop.ozone.upgrade.UpgradeFinalizer.Status.ALREADY_FINALIZED;
 import static org.apache.hadoop.ozone.upgrade.UpgradeFinalizer.Status.FINALIZATION_DONE;
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 import static org.mockito.ArgumentMatchers.anyString;
 import static org.mockito.Mockito.doCallRealMethod;
 import static org.mockito.Mockito.inOrder;
@@ -36,7 +36,7 @@ import java.io.IOException;
 import org.apache.hadoop.ozone.common.Storage;
 import org.apache.hadoop.ozone.upgrade.TestUpgradeFinalizerActions.MockLayoutVersionManager;
 import org.apache.hadoop.ozone.upgrade.UpgradeFinalizer.StatusAndMessages;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 import org.mockito.ArgumentMatchers;
 import org.mockito.InOrder;
 
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java
index c3eea3aebd..9bc8ded519 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestDefaultUpgradeFinalizationExecutor.java
@@ -25,7 +25,7 @@ import java.io.IOException;
 
 import org.apache.hadoop.ozone.common.Storage;
 import org.apache.ozone.test.LambdaTestUtils;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 /**
  * Test for DefaultUpgradeFinalizationExecutor.
diff --git a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java
index d7ac8ad8c2..ab82d0265b 100644
--- a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java
+++ b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/upgrade/TestLayoutVersionInstanceFactory.java
@@ -18,15 +18,15 @@
 
 package org.apache.hadoop.ozone.upgrade;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertTrue;
 import static org.mockito.Mockito.mock;
 import static org.mockito.Mockito.when;
 
 import java.util.function.Supplier;
 
 import org.apache.ozone.test.LambdaTestUtils;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
 
 /**
  * Test out APIs of VersionSpecificInstanceFactory.


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@ozone.apache.org
For additional commands, e-mail: commits-help@ozone.apache.org