You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@camel.apache.org by gn...@apache.org on 2020/07/09 20:02:05 UTC
[camel] 14/19: [CAMEL-11807] Upgrade camel-kafka to junit5
This is an automated email from the ASF dual-hosted git repository.
gnodet pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/camel.git
commit 9e0d5c23e8b96051432af31730004bae671552d1
Author: Guillaume Nodet <gn...@gmail.com>
AuthorDate: Thu Jul 9 14:07:56 2020 +0200
[CAMEL-11807] Upgrade camel-kafka to junit5
---
components/camel-kafka/pom.xml | 11 ++-
.../component/kafka/BaseEmbeddedKafkaTest.java | 6 +-
.../camel/component/kafka/KafkaComponentTest.java | 39 +++++-----
.../kafka/KafkaConsumerBatchSizeTest.java | 10 +--
.../component/kafka/KafkaConsumerFullTest.java | 29 ++++---
.../kafka/KafkaConsumerLastRecordHeaderTest.java | 21 ++---
.../kafka/KafkaConsumerManualCommitTest.java | 17 ++--
.../kafka/KafkaConsumerRebalanceTest.java | 10 ++-
.../camel/component/kafka/KafkaConsumerTest.java | 8 +-
.../kafka/KafkaConsumerTopicIsPatternTest.java | 12 +--
.../camel/component/kafka/KafkaEndpointTest.java | 12 +--
.../component/kafka/KafkaProducerFullTest.java | 90 ++++++++++++----------
.../camel/component/kafka/KafkaProducerTest.java | 10 +--
.../serde/DefaultKafkaHeaderDeserializerTest.java | 2 +-
.../serde/DefaultKafkaHeaderSerializerTest.java | 22 ++----
.../kafka/KafkaIdempotentRepositoryEagerTest.java | 2 +-
.../KafkaIdempotentRepositoryNonEagerTest.java | 2 +-
.../kafka/clients/consumer/KafkaConsumerTest.java | 10 ++-
18 files changed, 168 insertions(+), 145 deletions(-)
diff --git a/components/camel-kafka/pom.xml b/components/camel-kafka/pom.xml
index b271c76..e9afe60 100644
--- a/components/camel-kafka/pom.xml
+++ b/components/camel-kafka/pom.xml
@@ -51,7 +51,7 @@
<!-- test -->
<dependency>
<groupId>org.apache.camel</groupId>
- <artifactId>camel-testcontainers</artifactId>
+ <artifactId>camel-testcontainers-junit5</artifactId>
<scope>test</scope>
</dependency>
<dependency>
@@ -62,12 +62,17 @@
</dependency>
<dependency>
<groupId>org.apache.camel</groupId>
- <artifactId>camel-test</artifactId>
+ <artifactId>camel-test-junit5</artifactId>
+ <scope>test</scope>
+ </dependency>
+ <dependency>
+ <groupId>org.junit.jupiter</groupId>
+ <artifactId>junit-jupiter-params</artifactId>
<scope>test</scope>
</dependency>
<dependency>
<groupId>org.mockito</groupId>
- <artifactId>mockito-core</artifactId>
+ <artifactId>mockito-junit-jupiter</artifactId>
<scope>test</scope>
</dependency>
<dependency>
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/BaseEmbeddedKafkaTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/BaseEmbeddedKafkaTest.java
index 3c1f2ef..c7f71ff 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/BaseEmbeddedKafkaTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/BaseEmbeddedKafkaTest.java
@@ -19,11 +19,11 @@ package org.apache.camel.component.kafka;
import java.util.Properties;
import org.apache.camel.CamelContext;
-import org.apache.camel.test.junit4.CamelTestSupport;
+import org.apache.camel.test.junit5.CamelTestSupport;
import org.apache.kafka.clients.admin.AdminClient;
import org.apache.kafka.clients.admin.KafkaAdminClient;
import org.apache.kafka.clients.producer.ProducerConfig;
-import org.junit.BeforeClass;
+import org.junit.jupiter.api.BeforeAll;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.testcontainers.containers.KafkaContainer;
@@ -45,7 +45,7 @@ public abstract class BaseEmbeddedKafkaTest extends CamelTestSupport {
kafkaAdminClient = createAdminClient();
}
- @BeforeClass
+ @BeforeAll
public static void beforeClass() {
LOG.info("### Embedded Kafka cluster broker list: " + kafkaBroker.getBootstrapServers());
}
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaComponentTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaComponentTest.java
index 376b680..695c2ea 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaComponentTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaComponentTest.java
@@ -23,12 +23,15 @@ import java.util.Properties;
import org.apache.camel.support.jsse.KeyStoreParameters;
import org.apache.camel.support.jsse.SSLContextParameters;
import org.apache.camel.support.jsse.TrustManagersParameters;
-import org.apache.camel.test.junit4.CamelTestSupport;
+import org.apache.camel.test.junit5.CamelTestSupport;
import org.apache.kafka.clients.CommonClientConfigs;
import org.apache.kafka.clients.producer.ProducerConfig;
import org.apache.kafka.common.config.SaslConfigs;
import org.apache.kafka.common.config.SslConfigs;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
+
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertNull;
public class KafkaComponentTest extends CamelTestSupport {
@@ -103,27 +106,27 @@ public class KafkaComponentTest extends CamelTestSupport {
assertEquals("mytopic", endpoint.getConfiguration().getTopic());
assertEquals("1", endpoint.getConfiguration().getRequestRequiredAcks());
- assertEquals(new Integer(1), endpoint.getConfiguration().getBufferMemorySize());
- assertEquals(new Integer(10), endpoint.getConfiguration().getProducerBatchSize());
- assertEquals(new Integer(12), endpoint.getConfiguration().getConnectionMaxIdleMs());
- assertEquals(new Integer(1), endpoint.getConfiguration().getMaxBlockMs());
- assertEquals(new Integer(1), endpoint.getConfiguration().getBufferMemorySize());
+ assertEquals(Integer.valueOf(1), endpoint.getConfiguration().getBufferMemorySize());
+ assertEquals(Integer.valueOf(10), endpoint.getConfiguration().getProducerBatchSize());
+ assertEquals(Integer.valueOf(12), endpoint.getConfiguration().getConnectionMaxIdleMs());
+ assertEquals(Integer.valueOf(1), endpoint.getConfiguration().getMaxBlockMs());
+ assertEquals(Integer.valueOf(1), endpoint.getConfiguration().getBufferMemorySize());
assertEquals("testing", endpoint.getConfiguration().getClientId());
assertEquals("none", endpoint.getConfiguration().getCompressionCodec());
- assertEquals(new Integer(1), endpoint.getConfiguration().getLingerMs());
- assertEquals(new Integer(100), endpoint.getConfiguration().getMaxRequestSize());
+ assertEquals(Integer.valueOf(1), endpoint.getConfiguration().getLingerMs());
+ assertEquals(Integer.valueOf(100), endpoint.getConfiguration().getMaxRequestSize());
assertEquals(100, endpoint.getConfiguration().getRequestTimeoutMs().intValue());
- assertEquals(new Integer(1029), endpoint.getConfiguration().getMetadataMaxAgeMs());
- assertEquals(new Integer(23), endpoint.getConfiguration().getReceiveBufferBytes());
- assertEquals(new Integer(234), endpoint.getConfiguration().getReconnectBackoffMs());
- assertEquals(new Integer(234), endpoint.getConfiguration().getReconnectBackoffMaxMs());
- assertEquals(new Integer(0), endpoint.getConfiguration().getRetries());
+ assertEquals(Integer.valueOf(1029), endpoint.getConfiguration().getMetadataMaxAgeMs());
+ assertEquals(Integer.valueOf(23), endpoint.getConfiguration().getReceiveBufferBytes());
+ assertEquals(Integer.valueOf(234), endpoint.getConfiguration().getReconnectBackoffMs());
+ assertEquals(Integer.valueOf(234), endpoint.getConfiguration().getReconnectBackoffMaxMs());
+ assertEquals(Integer.valueOf(0), endpoint.getConfiguration().getRetries());
assertEquals(3782, endpoint.getConfiguration().getRetryBackoffMs().intValue());
assertEquals(765, endpoint.getConfiguration().getSendBufferBytes().intValue());
- assertEquals(new Integer(1), endpoint.getConfiguration().getMaxInFlightRequest());
+ assertEquals(Integer.valueOf(1), endpoint.getConfiguration().getMaxInFlightRequest());
assertEquals("org.apache.camel.reporters.TestReport,org.apache.camel.reporters.SampleReport", endpoint.getConfiguration().getMetricReporters());
- assertEquals(new Integer(3), endpoint.getConfiguration().getNoOfMetricsSample());
- assertEquals(new Integer(12344), endpoint.getConfiguration().getMetricsSampleWindowMs());
+ assertEquals(Integer.valueOf(3), endpoint.getConfiguration().getNoOfMetricsSample());
+ assertEquals(Integer.valueOf(12344), endpoint.getConfiguration().getMetricsSampleWindowMs());
assertEquals(KafkaConstants.KAFKA_DEFAULT_SERIALIZER, endpoint.getConfiguration().getSerializerClass());
assertEquals(KafkaConstants.KAFKA_DEFAULT_SERIALIZER, endpoint.getConfiguration().getKeySerializerClass());
assertEquals("testing", endpoint.getConfiguration().getSslKeyPassword());
@@ -139,7 +142,7 @@ public class KafkaComponentTest extends CamelTestSupport {
assertEquals("test", endpoint.getConfiguration().getSslProvider());
assertEquals("JKS", endpoint.getConfiguration().getSslTruststoreType());
assertEquals("/usr/bin/kinit", endpoint.getConfiguration().getKerberosInitCmd());
- assertEquals(new Integer(60000), endpoint.getConfiguration().getKerberosBeforeReloginMinTime());
+ assertEquals(Integer.valueOf(60000), endpoint.getConfiguration().getKerberosBeforeReloginMinTime());
assertEquals(new Double(0.05), endpoint.getConfiguration().getKerberosRenewJitter());
assertEquals(new Double(0.8), endpoint.getConfiguration().getKerberosRenewWindowFactor());
assertEquals("MAC", endpoint.getConfiguration().getSslCipherSuites());
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerBatchSizeTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerBatchSizeTest.java
index 4588d8f..2ecaade 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerBatchSizeTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerBatchSizeTest.java
@@ -24,9 +24,9 @@ import org.apache.camel.EndpointInject;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.mock.MockEndpoint;
import org.apache.kafka.clients.producer.ProducerRecord;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
public class KafkaConsumerBatchSizeTest extends BaseEmbeddedKafkaTest {
@@ -40,13 +40,13 @@ public class KafkaConsumerBatchSizeTest extends BaseEmbeddedKafkaTest {
private org.apache.kafka.clients.producer.KafkaProducer<String, String> producer;
- @Before
+ @BeforeEach
public void before() {
Properties props = getDefaultProperties();
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(props);
}
- @After
+ @AfterEach
public void after() {
if (producer != null) {
producer.close();
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerFullTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerFullTest.java
index e26a324..be592f4 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerFullTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerFullTest.java
@@ -30,10 +30,15 @@ import org.apache.camel.component.kafka.serde.DefaultKafkaHeaderDeserializer;
import org.apache.camel.component.mock.MockEndpoint;
import org.apache.kafka.clients.producer.ProducerRecord;
import org.apache.kafka.common.header.internals.RecordHeader;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Ignore;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Disabled;
+import org.junit.jupiter.api.Test;
+
+import static org.apache.camel.test.junit5.TestSupport.assertIsInstanceOf;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertFalse;
+import static org.junit.jupiter.api.Assertions.assertTrue;
public class KafkaConsumerFullTest extends BaseEmbeddedKafkaTest {
@@ -52,13 +57,13 @@ public class KafkaConsumerFullTest extends BaseEmbeddedKafkaTest {
private org.apache.kafka.clients.producer.KafkaProducer<String, String> producer;
- @Before
+ @BeforeEach
public void before() {
Properties props = getDefaultProperties();
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(props);
}
- @After
+ @AfterEach
public void after() {
if (producer != null) {
producer.close();
@@ -103,8 +108,8 @@ public class KafkaConsumerFullTest extends BaseEmbeddedKafkaTest {
assertEquals(5, StreamSupport.stream(MockConsumerInterceptor.recordsCaptured.get(0).records(TOPIC).spliterator(), false).count());
Map<String, Object> headers = to.getExchanges().get(0).getIn().getHeaders();
- assertFalse("Should not receive skipped header", headers.containsKey(skippedHeaderKey));
- assertTrue("Should receive propagated header", headers.containsKey(propagatedHeaderKey));
+ assertFalse(headers.containsKey(skippedHeaderKey), "Should not receive skipped header");
+ assertTrue(headers.containsKey(propagatedHeaderKey), "Should receive propagated header");
}
@Test
@@ -120,12 +125,12 @@ public class KafkaConsumerFullTest extends BaseEmbeddedKafkaTest {
to.assertIsSatisfied(3000);
Map<String, Object> headers = to.getExchanges().get(0).getIn().getHeaders();
- assertTrue("Should receive KafkaEndpoint populated kafka.TOPIC header", headers.containsKey(KafkaConstants.TOPIC));
- assertEquals("Topic name received", TOPIC, headers.get(KafkaConstants.TOPIC));
+ assertTrue(headers.containsKey(KafkaConstants.TOPIC), "Should receive KafkaEndpoint populated kafka.TOPIC header");
+ assertEquals(TOPIC, headers.get(KafkaConstants.TOPIC), "Topic name received");
}
@Test
- @Ignore("Currently there is a bug in kafka which leads to an uninterruptable thread so a resub take too long (works manually)")
+ @Disabled("Currently there is a bug in kafka which leads to an uninterruptable thread so a resub take too long (works manually)")
public void kafkaMessageIsConsumedByCamelSeekedToBeginning() throws Exception {
to.expectedMessageCount(5);
to.expectedBodiesReceivedInAnyOrder("message-0", "message-1", "message-2", "message-3", "message-4");
@@ -154,7 +159,7 @@ public class KafkaConsumerFullTest extends BaseEmbeddedKafkaTest {
}
@Test
- @Ignore("Currently there is a bug in kafka which leads to an uninterruptable thread so a resub take too long (works manually)")
+ @Disabled("Currently there is a bug in kafka which leads to an uninterruptable thread so a resub take too long (works manually)")
public void kafkaMessageIsConsumedByCamelSeekedToEnd() throws Exception {
to.expectedMessageCount(5);
to.expectedBodiesReceivedInAnyOrder("message-0", "message-1", "message-2", "message-3", "message-4");
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerLastRecordHeaderTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerLastRecordHeaderTest.java
index a4eb580..574be9b 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerLastRecordHeaderTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerLastRecordHeaderTest.java
@@ -25,9 +25,12 @@ import org.apache.camel.Exchange;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.mock.MockEndpoint;
import org.apache.kafka.clients.producer.ProducerRecord;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertNotNull;
public class KafkaConsumerLastRecordHeaderTest extends BaseEmbeddedKafkaTest {
private static final String TOPIC = "last-record";
@@ -37,13 +40,13 @@ public class KafkaConsumerLastRecordHeaderTest extends BaseEmbeddedKafkaTest {
private org.apache.kafka.clients.producer.KafkaProducer<String, String> producer;
- @Before
+ @BeforeEach
public void before() {
Properties props = getDefaultProperties();
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(props);
}
- @After
+ @AfterEach
public void after() {
if (producer != null) {
producer.close();
@@ -71,12 +74,12 @@ public class KafkaConsumerLastRecordHeaderTest extends BaseEmbeddedKafkaTest {
List<Exchange> exchanges = result.getExchanges();
for (int i = 0; i < exchanges.size(); i++) {
Boolean header = exchanges.get(i).getIn().getHeader(KafkaConstants.LAST_RECORD_BEFORE_COMMIT, Boolean.class);
- assertNotNull("Header not set for #" + i, header);
- assertEquals("Header invalid for #" + i, header, i == exchanges.size() - 1);
+ assertNotNull(header, "Header not set for #" + i);
+ assertEquals(header, i == exchanges.size() - 1, "Header invalid for #" + i);
// as long as the partitions count is 1 on topic:
header = exchanges.get(i).getIn().getHeader(KafkaConstants.LAST_POLL_RECORD, Boolean.class);
- assertNotNull("Last record header not set for #" + i, header);
- assertEquals("Last record header invalid for #" + i, header, i == exchanges.size() - 1);
+ assertNotNull(header, "Last record header not set for #" + i);
+ assertEquals(header, i == exchanges.size() - 1, "Last record header invalid for #" + i);
}
}
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerManualCommitTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerManualCommitTest.java
index 0e6524d..284d73f 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerManualCommitTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerManualCommitTest.java
@@ -26,12 +26,15 @@ import org.apache.camel.EndpointInject;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.mock.MockEndpoint;
import org.apache.kafka.clients.producer.ProducerRecord;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Ignore;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Disabled;
+import org.junit.jupiter.api.Test;
-@Ignore
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertNotNull;
+
+@Disabled
public class KafkaConsumerManualCommitTest extends BaseEmbeddedKafkaTest {
public static final String TOPIC = "test";
@@ -45,13 +48,13 @@ public class KafkaConsumerManualCommitTest extends BaseEmbeddedKafkaTest {
private org.apache.kafka.clients.producer.KafkaProducer<String, String> producer;
- @Before
+ @BeforeEach
public void before() {
Properties props = getDefaultProperties();
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(props);
}
- @After
+ @AfterEach
public void after() {
if (producer != null) {
producer.close();
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerRebalanceTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerRebalanceTest.java
index ca19ce0..0c6ac84 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerRebalanceTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerRebalanceTest.java
@@ -25,8 +25,10 @@ import org.apache.camel.EndpointInject;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.mock.MockEndpoint;
import org.apache.camel.spi.StateRepository;
-import org.junit.After;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.Test;
+
+import static org.junit.jupiter.api.Assertions.assertTrue;
public class KafkaConsumerRebalanceTest extends BaseEmbeddedKafkaTest {
private static final String TOPIC = "offset-rebalance";
@@ -47,10 +49,10 @@ public class KafkaConsumerRebalanceTest extends BaseEmbeddedKafkaTest {
@Test
public void offsetGetStateMustHaveBeenCalledTwice() throws Exception {
boolean offsetGetStateCalled = messagesLatch.await(30000, TimeUnit.MILLISECONDS);
- assertTrue("StateRepository.getState should have been called twice for topic " + TOPIC + ". Remaining count : " + messagesLatch.getCount(), offsetGetStateCalled);
+ assertTrue(offsetGetStateCalled, "StateRepository.getState should have been called twice for topic " + TOPIC + ". Remaining count : " + messagesLatch.getCount());
}
- @After
+ @AfterEach
public void after() {
// clean all test topics
kafkaAdminClient.deleteTopics(Collections.singletonList(TOPIC));
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTest.java
index 98be7e8..1c22fd1 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTest.java
@@ -17,8 +17,9 @@
package org.apache.camel.component.kafka;
import org.apache.camel.Processor;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
+import static org.junit.jupiter.api.Assertions.assertThrows;
import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.when;
@@ -29,12 +30,13 @@ public class KafkaConsumerTest {
private KafkaEndpoint endpoint = mock(KafkaEndpoint.class);
private Processor processor = mock(Processor.class);
- @Test(expected = IllegalArgumentException.class)
+ @Test
public void consumerRequiresBootstrapServers() throws Exception {
when(endpoint.getComponent()).thenReturn(component);
when(endpoint.getConfiguration()).thenReturn(configuration);
when(endpoint.getConfiguration().getGroupId()).thenReturn("groupOne");
- new KafkaConsumer(endpoint, processor);
+ assertThrows(IllegalArgumentException.class,
+ () -> new KafkaConsumer(endpoint, processor));
}
@Test
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTopicIsPatternTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTopicIsPatternTest.java
index 65325aa..61b0ce1 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTopicIsPatternTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaConsumerTopicIsPatternTest.java
@@ -25,9 +25,11 @@ import org.apache.camel.EndpointInject;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.mock.MockEndpoint;
import org.apache.kafka.clients.producer.ProducerRecord;
-import org.junit.After;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+
+import static org.junit.jupiter.api.Assertions.assertEquals;
public class KafkaConsumerTopicIsPatternTest extends BaseEmbeddedKafkaTest {
@@ -43,14 +45,14 @@ public class KafkaConsumerTopicIsPatternTest extends BaseEmbeddedKafkaTest {
private org.apache.kafka.clients.producer.KafkaProducer<String, String> producer;
- @Before
+ @BeforeEach
public void before() {
Properties props = getDefaultProperties();
producer = new org.apache.kafka.clients.producer.KafkaProducer<>(props);
}
- @After
+ @AfterEach
public void after() {
if (producer != null) {
producer.close();
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaEndpointTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaEndpointTest.java
index e952fe8..7bfb0df 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaEndpointTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaEndpointTest.java
@@ -20,18 +20,18 @@ import org.apache.camel.Exchange;
import org.apache.camel.Message;
import org.apache.camel.impl.DefaultCamelContext;
import org.apache.kafka.clients.consumer.ConsumerRecord;
-import org.junit.Before;
-import org.junit.Test;
-import org.junit.runner.RunWith;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.extension.ExtendWith;
import org.mockito.Mock;
-import org.mockito.junit.MockitoJUnitRunner;
+import org.mockito.junit.jupiter.MockitoExtension;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertTrue;
import static org.mockito.Mockito.when;
-@RunWith(MockitoJUnitRunner.class)
+@ExtendWith(MockitoExtension.class)
public class KafkaEndpointTest {
private KafkaEndpoint endpoint;
@@ -42,7 +42,7 @@ public class KafkaEndpointTest {
@Mock
private KafkaComponent mockKafkaComponent;
- @Before
+ @BeforeEach
public void setup() {
KafkaComponent kafka = new KafkaComponent(new DefaultCamelContext());
kafka.init();
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerFullTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerFullTest.java
index a953c48..862846c 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerFullTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerFullTest.java
@@ -45,9 +45,15 @@ import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.apache.kafka.clients.producer.RecordMetadata;
import org.apache.kafka.common.header.Header;
import org.apache.kafka.common.header.Headers;
-import org.junit.AfterClass;
-import org.junit.BeforeClass;
-import org.junit.Test;
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.Test;
+
+import static org.apache.camel.test.junit5.TestSupport.assertIsInstanceOf;
+import static org.junit.jupiter.api.Assertions.assertArrayEquals;
+import static org.junit.jupiter.api.Assertions.assertEquals;
+import static org.junit.jupiter.api.Assertions.assertNotNull;
+import static org.junit.jupiter.api.Assertions.assertTrue;
public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
@@ -109,13 +115,13 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
@BindToRegistry("myHeaderSerializer")
private MyKafkaHeadersSerializer serializer = new MyKafkaHeadersSerializer();
- @BeforeClass
+ @BeforeAll
public static void before() {
stringsConsumerConn = createStringKafkaConsumer("DemoConsumer");
bytesConsumerConn = createByteKafkaConsumer(GROUP_BYTES);
}
- @AfterClass
+ @AfterAll
public static void after() {
// clean all test topics
final List<String> topics = new ArrayList<>();
@@ -162,16 +168,16 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
boolean allMessagesReceived = messagesLatch.await(200, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
List<Exchange> exchangeList = mockEndpoint.getExchanges();
- assertEquals("Fifteen Exchanges are expected", exchangeList.size(), 15);
+ assertEquals(exchangeList.size(), 15, "Fifteen Exchanges are expected");
for (Exchange exchange : exchangeList) {
@SuppressWarnings("unchecked")
List<RecordMetadata> recordMetaData1 = (List<RecordMetadata>)(exchange.getIn().getHeader(KafkaConstants.KAFKA_RECORDMETA));
- assertEquals("One RecordMetadata is expected.", recordMetaData1.size(), 1);
- assertTrue("Offset is positive", recordMetaData1.get(0).offset() >= 0);
- assertTrue("Topic Name start with 'test'", recordMetaData1.get(0).topic().startsWith("test"));
+ assertEquals(recordMetaData1.size(), 1, "One RecordMetadata is expected.");
+ assertTrue(recordMetaData1.get(0).offset() >= 0, "Offset is positive");
+ assertTrue(recordMetaData1.get(0).topic().startsWith("test"), "Topic Name start with 'test'");
}
}
@@ -190,16 +196,16 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
boolean allMessagesReceived = messagesLatch.await(200, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
List<Exchange> exchangeList = mockEndpoint.getExchanges();
- assertEquals("Fifteen Exchanges are expected", exchangeList.size(), 15);
+ assertEquals(exchangeList.size(), 15, "Fifteen Exchanges are expected");
for (Exchange exchange : exchangeList) {
@SuppressWarnings("unchecked")
List<RecordMetadata> recordMetaData1 = (List<RecordMetadata>)(exchange.getIn().getHeader(KafkaConstants.KAFKA_RECORDMETA));
- assertEquals("One RecordMetadata is expected.", recordMetaData1.size(), 1);
- assertTrue("Offset is positive", recordMetaData1.get(0).offset() >= 0);
- assertTrue("Topic Name start with 'test'", recordMetaData1.get(0).topic().startsWith("test"));
+ assertEquals(1, recordMetaData1.size(), "One RecordMetadata is expected.");
+ assertTrue(recordMetaData1.get(0).offset() >= 0, "Offset is positive");
+ assertTrue(recordMetaData1.get(0).topic().startsWith("test"), "Topic Name start with 'test'");
}
}
@@ -217,7 +223,7 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
boolean allMessagesReceived = messagesLatch.await(200, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
assertEquals(messageInTopic + messageInOtherTopic, MockProducerInterceptor.recordsCaptured.size());
}
@@ -245,24 +251,24 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
boolean allMessagesReceived = messagesLatch.await(200, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
List<Exchange> exchangeList = mockEndpoint.getExchanges();
- assertEquals("Two Exchanges are expected", exchangeList.size(), 2);
+ assertEquals(2, exchangeList.size(), "Two Exchanges are expected");
Exchange e1 = exchangeList.get(0);
@SuppressWarnings("unchecked")
List<RecordMetadata> recordMetaData1 = (List<RecordMetadata>)(e1.getIn().getHeader(KafkaConstants.KAFKA_RECORDMETA));
- assertEquals("Ten RecordMetadata is expected.", recordMetaData1.size(), 10);
+ assertEquals(10, recordMetaData1.size(), "Ten RecordMetadata is expected.");
for (RecordMetadata recordMeta : recordMetaData1) {
- assertTrue("Offset is positive", recordMeta.offset() >= 0);
- assertTrue("Topic Name start with 'test'", recordMeta.topic().startsWith("test"));
+ assertTrue(recordMeta.offset() >= 0, "Offset is positive");
+ assertTrue(recordMeta.topic().startsWith("test"), "Topic Name start with 'test'");
}
Exchange e2 = exchangeList.get(1);
@SuppressWarnings("unchecked")
List<RecordMetadata> recordMetaData2 = (List<RecordMetadata>)(e2.getIn().getHeader(KafkaConstants.KAFKA_RECORDMETA));
- assertEquals("Five RecordMetadata is expected.", recordMetaData2.size(), 5);
+ assertEquals(5, recordMetaData2.size(), "Five RecordMetadata is expected.");
for (RecordMetadata recordMeta : recordMetaData2) {
- assertTrue("Offset is positive", recordMeta.offset() >= 0);
- assertTrue("Topic Name start with 'test'", recordMeta.topic().startsWith("test"));
+ assertTrue(recordMeta.offset() >= 0, "Offset is positive");
+ assertTrue(recordMeta.topic().startsWith("test"), "Topic Name start with 'test'");
}
}
@@ -286,16 +292,16 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
boolean allMessagesReceived = messagesLatch.await(200, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
List<Exchange> exchangeList = mockEndpoint.getExchanges();
- assertEquals("Fifteen Exchanges are expected", exchangeList.size(), 15);
+ assertEquals(exchangeList.size(), 15, "Fifteen Exchanges are expected");
for (Exchange exchange : exchangeList) {
@SuppressWarnings("unchecked")
List<RecordMetadata> recordMetaData1 = (List<RecordMetadata>)(exchange.getIn().getHeader(KafkaConstants.KAFKA_RECORDMETA));
- assertEquals("One RecordMetadata is expected.", recordMetaData1.size(), 1);
- assertTrue("Offset is positive", recordMetaData1.get(0).offset() >= 0);
- assertTrue("Topic Name start with 'test'", recordMetaData1.get(0).topic().startsWith("test"));
+ assertEquals(1, recordMetaData1.size(), "One RecordMetadata is expected.");
+ assertTrue(recordMetaData1.get(0).offset() >= 0, "Offset is positive");
+ assertTrue(recordMetaData1.get(0).topic().startsWith("test"), "Topic Name start with 'test'");
}
}
@@ -337,19 +343,19 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
List<ConsumerRecord<String, String>> records = pollForRecords(createStringKafkaConsumer("propagatedHeaderConsumer"), TOPIC_PROPAGATED_HEADERS, messagesLatch);
boolean allMessagesReceived = messagesLatch.await(10_000, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
ConsumerRecord<String, String> record = records.get(0);
Headers headers = record.headers();
- assertNotNull("Kafka Headers should not be null.", headers);
+ assertNotNull(headers, "Kafka Headers should not be null.");
// we have 6 headers
- assertEquals("6 propagated header is expected.", 6, headers.toArray().length);
- assertEquals("Propagated string value received", propagatedStringHeaderValue, new String(getHeaderValue(propagatedStringHeaderKey, headers)));
- assertEquals("Propagated integer value received", propagatedIntegerHeaderValue, new Integer(ByteBuffer.wrap(getHeaderValue(propagatedIntegerHeaderKey, headers)).getInt()));
- assertEquals("Propagated long value received", propagatedLongHeaderValue, new Long(ByteBuffer.wrap(getHeaderValue(propagatedLongHeaderKey, headers)).getLong()));
- assertEquals("Propagated double value received", propagatedDoubleHeaderValue, new Double(ByteBuffer.wrap(getHeaderValue(propagatedDoubleHeaderKey, headers)).getDouble()));
- assertArrayEquals("Propagated byte array value received", propagatedBytesHeaderValue, getHeaderValue(propagatedBytesHeaderKey, headers));
- assertEquals("Propagated boolean value received", propagatedBooleanHeaderValue, Boolean.valueOf(new String(getHeaderValue(propagatedBooleanHeaderKey, headers))));
+ assertEquals(6, headers.toArray().length, "6 propagated header is expected.");
+ assertEquals(propagatedStringHeaderValue, new String(getHeaderValue(propagatedStringHeaderKey, headers)), "Propagated string value received");
+ assertEquals(propagatedIntegerHeaderValue, Integer.valueOf(ByteBuffer.wrap(getHeaderValue(propagatedIntegerHeaderKey, headers)).getInt()), "Propagated integer value received");
+ assertEquals(propagatedLongHeaderValue, Long.valueOf(ByteBuffer.wrap(getHeaderValue(propagatedLongHeaderKey, headers)).getLong()), "Propagated long value received");
+ assertEquals(propagatedDoubleHeaderValue, Double.valueOf(ByteBuffer.wrap(getHeaderValue(propagatedDoubleHeaderKey, headers)).getDouble()), "Propagated double value received");
+ assertArrayEquals(propagatedBytesHeaderValue, getHeaderValue(propagatedBytesHeaderKey, headers), "Propagated byte array value received");
+ assertEquals(propagatedBooleanHeaderValue, Boolean.valueOf(new String(getHeaderValue(propagatedBooleanHeaderKey, headers))), "Propagated boolean value received");
}
@Test
@@ -366,13 +372,13 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
List<ConsumerRecord<String, String>> records = pollForRecords(createStringKafkaConsumer("noRecordSpecificHeadersConsumer"), TOPIC_NO_RECORD_SPECIFIC_HEADERS, messagesLatch);
boolean allMessagesReceived = messagesLatch.await(10_000, TimeUnit.MILLISECONDS);
- assertTrue("Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount(), allMessagesReceived);
+ assertTrue(allMessagesReceived, "Not all messages were published to the kafka topics. Not received: " + messagesLatch.getCount());
ConsumerRecord<String, String> record = records.get(0);
Headers headers = record.headers();
- assertNotNull("Kafka Headers should not be null.", headers);
+ assertNotNull(headers, "Kafka Headers should not be null.");
// we have 0 headers
- assertEquals("0 propagated headers are expected", 0, headers.toArray().length);
+ assertEquals(0, headers.toArray().length, "0 propagated headers are expected");
}
@Test
@@ -389,7 +395,7 @@ public class KafkaProducerFullTest extends BaseEmbeddedKafkaTest {
private byte[] getHeaderValue(String headerKey, Headers headers) {
Header foundHeader = StreamSupport.stream(headers.spliterator(), false).filter(header -> header.key().equals(headerKey)).findFirst().orElse(null);
- assertNotNull("Header should be sent", foundHeader);
+ assertNotNull(foundHeader, "Header should be sent");
return foundHeader.value();
}
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerTest.java
index 57ee16c..be5df93 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/KafkaProducerTest.java
@@ -43,7 +43,7 @@ import org.apache.kafka.clients.producer.ProducerConfig;
import org.apache.kafka.clients.producer.ProducerRecord;
import org.apache.kafka.clients.producer.RecordMetadata;
import org.apache.kafka.common.errors.ApiException;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import org.mockito.ArgumentCaptor;
import org.mockito.Mockito;
@@ -51,6 +51,7 @@ import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotNull;
import static org.junit.Assert.assertNull;
import static org.junit.Assert.assertTrue;
+import static org.junit.jupiter.api.Assertions.assertThrows;
import static org.mockito.ArgumentMatchers.any;
import static org.mockito.ArgumentMatchers.eq;
import static org.mockito.ArgumentMatchers.isA;
@@ -117,7 +118,7 @@ public class KafkaProducerTest {
assertRecordMetadataExists();
}
- @Test(expected = Exception.class)
+ @Test
@SuppressWarnings({"unchecked"})
public void processSendsMessageWithException() throws Exception {
endpoint.getConfiguration().setTopic("sometopic");
@@ -127,9 +128,8 @@ public class KafkaProducerTest {
Mockito.when(exchange.getIn()).thenReturn(in);
in.setHeader(KafkaConstants.PARTITION_KEY, 4);
- producer.process(exchange);
-
- assertRecordMetadataExists();
+ assertThrows(Exception.class,
+ () -> producer.process(exchange));
}
@Test
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderDeserializerTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderDeserializerTest.java
index 334fcf5..b54959d 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderDeserializerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderDeserializerTest.java
@@ -17,7 +17,7 @@
package org.apache.camel.component.kafka.serde;
import org.hamcrest.CoreMatchers;
-import org.junit.Test;
+import org.junit.jupiter.api.Test;
import static org.junit.Assert.assertArrayEquals;
import static org.junit.Assert.assertThat;
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderSerializerTest.java b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderSerializerTest.java
index cda893e..9962302 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderSerializerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/component/kafka/serde/DefaultKafkaHeaderSerializerTest.java
@@ -19,34 +19,24 @@ package org.apache.camel.component.kafka.serde;
import java.util.Arrays;
import java.util.Collection;
-import org.junit.Test;
-import org.junit.runner.RunWith;
-import org.junit.runners.Parameterized;
+import org.junit.jupiter.params.ParameterizedTest;
+import org.junit.jupiter.params.provider.MethodSource;
import static org.junit.Assert.assertArrayEquals;
-@RunWith(Parameterized.class)
public class DefaultKafkaHeaderSerializerTest {
private KafkaHeaderSerializer serializer = new DefaultKafkaHeaderSerializer();
- private Object value;
- private byte[] expectedResult;
-
- public DefaultKafkaHeaderSerializerTest(Object value, byte[] expectedResult) {
- this.value = value;
- this.expectedResult = expectedResult;
- }
-
- @Test
- public void serialize() {
+ @ParameterizedTest
+ @MethodSource("primeNumbers")
+ public void serialize(Object value, byte[] expectedResult) {
byte[] result = serializer.serialize("someKey", value);
assertArrayEquals(expectedResult, result);
}
- @Parameterized.Parameters
- public static Collection primeNumbers() {
+ public static Collection<Object[]> primeNumbers() {
return Arrays.asList(new Object[][] {{Boolean.TRUE, "true".getBytes()}, // boolean
{-12, new byte[] {-1, -1, -1, -12}}, // integer
{19L, new byte[] {0, 0, 0, 0, 0, 0, 0, 19}}, // long
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryEagerTest.java b/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryEagerTest.java
index 3974db6..7c96898 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryEagerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryEagerTest.java
@@ -23,7 +23,7 @@ import org.apache.camel.RoutesBuilder;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.kafka.BaseEmbeddedKafkaTest;
import org.apache.camel.component.mock.MockEndpoint;
-import org.junit.Test;
+import org.junit.jupiter.api.*; import static org.junit.jupiter.api.Assertions.*;
/**
* Test for eager idempotentRepository usage.
diff --git a/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryNonEagerTest.java b/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryNonEagerTest.java
index 66e1f44..6633f39 100644
--- a/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryNonEagerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/camel/processor/idempotent/kafka/KafkaIdempotentRepositoryNonEagerTest.java
@@ -23,7 +23,7 @@ import org.apache.camel.RoutesBuilder;
import org.apache.camel.builder.RouteBuilder;
import org.apache.camel.component.kafka.BaseEmbeddedKafkaTest;
import org.apache.camel.component.mock.MockEndpoint;
-import org.junit.Test;
+import org.junit.jupiter.api.*; import static org.junit.jupiter.api.Assertions.*;
/**
* Test for non-eager idempotentRepository usage.
diff --git a/components/camel-kafka/src/test/java/org/apache/kafka/clients/consumer/KafkaConsumerTest.java b/components/camel-kafka/src/test/java/org/apache/kafka/clients/consumer/KafkaConsumerTest.java
index 134787f..528346d 100644
--- a/components/camel-kafka/src/test/java/org/apache/kafka/clients/consumer/KafkaConsumerTest.java
+++ b/components/camel-kafka/src/test/java/org/apache/kafka/clients/consumer/KafkaConsumerTest.java
@@ -17,23 +17,25 @@
package org.apache.kafka.clients.consumer;
import org.hamcrest.core.IsNot;
-import org.junit.Before;
-import org.junit.Test;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.*;
+import org.junit.jupiter.api.extension.ExtendWith;
import org.junit.runner.RunWith;
import org.mockito.Mock;
import org.mockito.junit.MockitoJUnitRunner;
+import org.mockito.junit.jupiter.MockitoExtension;
import static org.hamcrest.core.IsNull.nullValue;
import static org.junit.Assert.assertThat;
import static org.mockito.Mockito.when;
-@RunWith(MockitoJUnitRunner.class)
+@ExtendWith(MockitoExtension.class)
public class KafkaConsumerTest {
@Mock
private KafkaConsumer<Object, Object> kafkaConsumer;
- @Before
+ @BeforeEach
public void init() {
when(kafkaConsumer.poll(1000)).thenReturn(ConsumerRecords.empty());
}