You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Praveen (JIRA)" <ji...@apache.org> on 2019/04/15 15:12:00 UTC
[jira] [Created] (SPARK-27465) Kafka Client 0.11.0.0 is not
Supporting the kafkatestutils package
Praveen created SPARK-27465:
-------------------------------
Summary: Kafka Client 0.11.0.0 is not Supporting the kafkatestutils package
Key: SPARK-27465
URL: https://issues.apache.org/jira/browse/SPARK-27465
Project: Spark
Issue Type: Bug
Components: Java API
Affects Versions: 2.3.3
Reporter: Praveen
Hi Team,
We are getting the below exceptions with Kafka Client Version 0.11.0.0 for KafkaTestUtils Package. But its working fine when we use the Kafka Client Version 0.10.0.1. Please suggest the way forwards. We are using the package "
import org.apache.spark.streaming.kafka010.KafkaTestUtils;"
ERROR:
java.lang.NoSuchMethodError: kafka.server.KafkaServer$.$lessinit$greater$default$2()Lkafka/utils/Time;
at org.apache.spark.streaming.kafka010.KafkaTestUtils$$anonfun$setupEmbeddedKafkaServer$2.apply(KafkaTestUtils.scala:110)
at org.apache.spark.streaming.kafka010.KafkaTestUtils$$anonfun$setupEmbeddedKafkaServer$2.apply(KafkaTestUtils.scala:107)
at org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2234)
at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2226)
at org.apache.spark.streaming.kafka010.KafkaTestUtils.setupEmbeddedKafkaServer(KafkaTestUtils.scala:107)
at org.apache.spark.streaming.kafka010.KafkaTestUtils.setup(KafkaTestUtils.scala:122)
at com.netcracker.rms.smart.esp.ESPTestEnv.prepareKafkaTestUtils(ESPTestEnv.java:203)
at com.netcracker.rms.smart.esp.ESPTestEnv.setUp(ESPTestEnv.java:157)
at com.netcracker.rms.smart.esp.TestEventStreamProcessor.setUp(TestEventStreamProcessor.java:58)
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org