You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by Lehuede sebastien <le...@gmail.com> on 2021/03/30 13:44:37 UTC

Scala : ClassCastException with Kafka Connector and ObjectNode

Hi all,

I’m currently trying to use Scala to setup a simple Kafka consumer that receive JSON formatted events and then just send them to Elasticsearch. This is the first step and after I want to add some processing logic. 

My code works well but interesting fields form my JSON formatted events are under the key « value ». So I tried to use a map function to get fields under « value » and send them to ES.

But when I try to use this map function I always get a ClassCastException error. 

Additional Information :

• My Job run on a Flink Kubernetes Application cluster with a standalone job.
• If I remove the map function, everything is working fine and I can find my events in Elasticsearch
• I tried to replace to ‘JSONKeyValueDeserializationSchema’ by ’SimpleStringSchema’ and then use a mapper in my map function but I still have the same issue.
• I’m using Scala version 2.12.12

Here is my code : 



// Create the Flink Kafka connector
val flinkKafkaConsumer = new FlinkKafkaConsumer(topics, new JSONKeyValueDeserializationSchema(false), kafkaProperties)
flinkKafkaConsumer.setStartFromEarliest()

val eventsStream: DataStream[ObjectNode] = env.addSource(flinkKafkaConsumer)
  .name("Event Stream : Kafka consumer")
  .map(event => event.get("value").deepCopy())

eventsStream.map(_.toString).addSink(esSinkBuilder.build()).name("Event Stream : Elasticsearch Stream")

// execute program
env.execute(« Kafka to ES")



Here is the error : 

[nybble-jobmanager] org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy
[nybble-jobmanager]    at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:118) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:80) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:233) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:224) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:215) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:665) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:89) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:447) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
[nybble-jobmanager]    at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) ~[?:?]
[nybble-jobmanager]    at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) ~[?:?]
[nybble-jobmanager]    at java.lang.reflect.Method.invoke(Unknown Source) ~[?:?]
[nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:306) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:213) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:77) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:159) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at scala.PartialFunction.applyOrElse(PartialFunction.scala:127) [scala-library-2.12.12.jar:?]
[nybble-jobmanager]    at scala.PartialFunction.applyOrElse$(PartialFunction.scala:126) [scala-library-2.12.12.jar:?]
[nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:175) [scala-library-2.12.12.jar:?]
[nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:176) [scala-library-2.12.12.jar:?]
[nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:176) [scala-library-2.12.12.jar:?]
[nybble-jobmanager]    at akka.actor.Actor.aroundReceive(Actor.scala:517) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.actor.Actor.aroundReceive$(Actor.scala:515) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.actor.ActorCell.invoke(ActorCell.scala:561) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.Mailbox.run(Mailbox.scala:225) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.Mailbox.exec(Mailbox.scala:235) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) [akka-actor_2.12-2.5.21.jar:2.5.21]
[nybble-jobmanager] Caused by: java.lang.ClassCastException: class org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.node.ObjectNode cannot be cast to class scala.runtime.Nothing$ (org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.node.ObjectNode and scala.runtime.Nothing$ are in unnamed module of loader 'app')
[nybble-jobmanager]    at io.nybble.NybbleMain$.$anonfun$main$4(NybbleMain.scala:138) ~[?:?]
[nybble-jobmanager]    at org.apache.flink.streaming.api.scala.DataStream$$anon$4.map(DataStream.scala:625) ~[flink-streaming-scala_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:38) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollectWithTimestamp(StreamSourceContexts.java:322) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collectWithTimestamp(StreamSourceContexts.java:426) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:365) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:183) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.runFetchLoop(KafkaFetcher.java:142) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:826) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:66) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
[nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:241) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]


Does anyone has already had this kind of issue ? I think I’ve missed something even if it’s a really basic use-case and code, I’m beginner in Scala. 

Thanks in advance for your help !
Sebastien

Re: Scala : ClassCastException with Kafka Connector and ObjectNode

Posted by Lehuede sebastien <le...@gmail.com>.
Hi Till,

That solved my issue ! Many many thanks for the solution and for the useful StackOverflow link ! ☺️

Cheers,
Sébastien 

> Le 30 mars 2021 à 18:16, Till Rohrmann <tr...@apache.org> a écrit :
> 
> Hi Sebastien,
> 
> I think the Scala compiler infers the most specific type for deepCopy() which is Nothing (Nothing is the subtype of every type) [1] because you haven't specified a type here. In order to make it work you have to specify the concrete type:
> 
> event.get("value").deepCopy[ObjectNode]()
> 
> [1] https://stackoverflow.com/a/18008591/4815083 <https://stackoverflow.com/a/18008591/4815083>
> 
> Cheers,
> Till
> 
> On Tue, Mar 30, 2021 at 3:45 PM Lehuede sebastien <lehuede.s@gmail.com <ma...@gmail.com>> wrote:
> Hi all,
> 
> I’m currently trying to use Scala to setup a simple Kafka consumer that receive JSON formatted events and then just send them to Elasticsearch. This is the first step and after I want to add some processing logic. 
> 
> My code works well but interesting fields form my JSON formatted events are under the key « value ». So I tried to use a map function to get fields under « value » and send them to ES.
> 
> But when I try to use this map function I always get a ClassCastException error. 
> 
> Additional Information :
> 
> • My Job run on a Flink Kubernetes Application cluster with a standalone job.
> • If I remove the map function, everything is working fine and I can find my events in Elasticsearch
> • I tried to replace to ‘JSONKeyValueDeserializationSchema’ by ’SimpleStringSchema’ and then use a mapper in my map function but I still have the same issue.
> • I’m using Scala version 2.12.12
> 
> Here is my code : 
> 
> 
> 
> // Create the Flink Kafka connector
> val flinkKafkaConsumer = new FlinkKafkaConsumer(topics, new JSONKeyValueDeserializationSchema(false), kafkaProperties)
> flinkKafkaConsumer.setStartFromEarliest()
> 
> val eventsStream: DataStream[ObjectNode] = env.addSource(flinkKafkaConsumer)
>   .name("Event Stream : Kafka consumer")
>   .map(event => event.get("value").deepCopy())
> 
> eventsStream.map(_.toString).addSink(esSinkBuilder.build()).name("Event Stream : Elasticsearch Stream")
> 
> // execute program
> env.execute(« Kafka to ES")
> 
> 
> 
> Here is the error : 
> 
> [nybble-jobmanager] org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy
> [nybble-jobmanager]    at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:118) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:80) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:233) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:224) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:215) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:665) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:89) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:447) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
> [nybble-jobmanager]    at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) ~[?:?]
> [nybble-jobmanager]    at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) ~[?:?]
> [nybble-jobmanager]    at java.lang.reflect.Method.invoke(Unknown Source) ~[?:?]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:306) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:213) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:77) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:159) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at scala.PartialFunction.applyOrElse(PartialFunction.scala:127) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at scala.PartialFunction.applyOrElse$(PartialFunction.scala:126) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:175) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:176) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:176) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at akka.actor.Actor.aroundReceive(Actor.scala:517) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.Actor.aroundReceive$(Actor.scala:515) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.ActorCell.invoke(ActorCell.scala:561) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.Mailbox.run(Mailbox.scala:225) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.Mailbox.exec(Mailbox.scala:235) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager] Caused by: java.lang.ClassCastException: class org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.node.ObjectNode cannot be cast to class scala.runtime.Nothing$ (org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.node.ObjectNode and scala.runtime.Nothing$ are in unnamed module of loader 'app')
> [nybble-jobmanager]    at io.nybble.NybbleMain$.$anonfun$main$4(NybbleMain.scala:138) ~[?:?]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.scala.DataStream$$anon$4.map(DataStream.scala:625) ~[flink-streaming-scala_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:38) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollectWithTimestamp(StreamSourceContexts.java:322) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collectWithTimestamp(StreamSourceContexts.java:426) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:365) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:183) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.runFetchLoop(KafkaFetcher.java:142) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:826) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:66) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:241) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> 
> 
> Does anyone has already had this kind of issue ? I think I’ve missed something even if it’s a really basic use-case and code, I’m beginner in Scala. 
> 
> Thanks in advance for your help !
> Sebastien


Re: Scala : ClassCastException with Kafka Connector and ObjectNode

Posted by Till Rohrmann <tr...@apache.org>.
Hi Sebastien,

I think the Scala compiler infers the most specific type for
deepCopy() which is Nothing (Nothing is the subtype of every type) [1]
because you haven't specified a type here. In order to make it work you
have to specify the concrete type:

event.get("value").deepCopy[ObjectNode]()

[1] https://stackoverflow.com/a/18008591/4815083

Cheers,
Till

On Tue, Mar 30, 2021 at 3:45 PM Lehuede sebastien <le...@gmail.com>
wrote:

> Hi all,
>
> I’m currently trying to use Scala to setup a simple Kafka consumer that
> receive JSON formatted events and then just send them to Elasticsearch.
> This is the first step and after I want to add some processing logic.
>
> My code works well but interesting fields form my JSON formatted events
> are under the key « value ». So I tried to use a map function to get fields
> under « value » and send them to ES.
>
> But when I try to use this map function I always get a ClassCastException
> error.
>
> Additional Information :
>
> • My Job run on a Flink Kubernetes Application cluster with a standalone
> job.
> • If I remove the map function, everything is working fine and I can find
> my events in Elasticsearch
> • I tried to replace to ‘JSONKeyValueDeserializationSchema’ by
> ’SimpleStringSchema’ and then use a mapper in my map function but I still
> have the same issue.
> • I’m using Scala version 2.12.12
>
> Here is my code :
>
>
>
> // Create the Flink Kafka connector
> val flinkKafkaConsumer = new FlinkKafkaConsumer(topics, new JSONKeyValueDeserializationSchema(false), kafkaProperties)
> flinkKafkaConsumer.setStartFromEarliest()
>
> val eventsStream: DataStream[ObjectNode] = env.addSource(flinkKafkaConsumer)
>   .name("Event Stream : Kafka consumer")
>   .map(event => event.get("value").deepCopy())
>
> eventsStream.map(_.toString).addSink(esSinkBuilder.build()).name("Event Stream : Elasticsearch Stream")
>
> // execute program
> env.execute(« Kafka to ES")
>
>
>
> Here is the error :
>
> [nybble-jobmanager] org.apache.flink.runtime.JobException: Recovery is suppressed by NoRestartBackoffTimeStrategy
> [nybble-jobmanager]    at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:118) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:80) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:233) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:224) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:215) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:665) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:89) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:447) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
> [nybble-jobmanager]    at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) ~[?:?]
> [nybble-jobmanager]    at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) ~[?:?]
> [nybble-jobmanager]    at java.lang.reflect.Method.invoke(Unknown Source) ~[?:?]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:306) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:213) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:77) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:159) ~[flink-runtime_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at scala.PartialFunction.applyOrElse(PartialFunction.scala:127) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at scala.PartialFunction.applyOrElse$(PartialFunction.scala:126) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:175) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:176) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:176) [scala-library-2.12.12.jar:?]
> [nybble-jobmanager]    at akka.actor.Actor.aroundReceive(Actor.scala:517) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.Actor.aroundReceive$(Actor.scala:515) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.actor.ActorCell.invoke(ActorCell.scala:561) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.Mailbox.run(Mailbox.scala:225) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.Mailbox.exec(Mailbox.scala:235) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager]    at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) [akka-actor_2.12-2.5.21.jar:2.5.21]
> [nybble-jobmanager] Caused by: java.lang.ClassCastException: class org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.node.ObjectNode cannot be cast to class scala.runtime.Nothing$ (org.apache.flink.shaded.jackson2.com.fasterxml.jackson.databind.node.ObjectNode and scala.runtime.Nothing$ are in unnamed module of loader 'app')
> [nybble-jobmanager]    at io.nybble.NybbleMain$.$anonfun$main$4(NybbleMain.scala:138) ~[?:?]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.scala.DataStream$$anon$4.map(DataStream.scala:625) ~[flink-streaming-scala_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:38) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollectWithTimestamp(StreamSourceContexts.java:322) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collectWithTimestamp(StreamSourceContexts.java:426) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:365) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:183) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.runFetchLoop(KafkaFetcher.java:142) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:826) ~[flink-connector-kafka_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:66) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
> [nybble-jobmanager]    at org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:241) ~[flink-streaming-java_2.12-1.12.1.jar:1.12.1]
>
>
>
> Does anyone has already had this kind of issue ? I think I’ve missed
> something even if it’s a really basic use-case and code, I’m beginner in
> Scala.
>
> Thanks in advance for your help !
> Sebastien
>