You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by Vijayendra Yadav <co...@gmail.com> on 2020/08/24 19:53:56 UTC

Flink Couchbase

Hi Team,

Running a flink job on Yarn, I am trying to make connections to
couchbase DB in one of my map functions in Flink Streaming job. But my task
manager containers keep failing
and keep assigning new containers and not giving me an opportunity to get
any useful logs.

 val cluster = Cluster.connect("host", "user", "pwd")
 val bucket = cluster.bucket("bucket")
 val collection = bucket.defaultCollection

Only thing I see is yarn exception:

java.lang.Exception: Container released on a *lost* node
    at org.apache.flink.yarn.YarnResourceManager
.lambda$onContainersCompleted$0(YarnResourceManager.java:343)
    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(
AkkaRpcActor.java:397)
    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(
AkkaRpcActor.java:190)
    at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor
.handleRpcMessage(FencedAkkaRpcActor.java:74)
    at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(
AkkaRpcActor.java:152)
    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
    at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
    at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
    at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
    at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
    at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
    at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
    at akka.actor.ActorCell.invoke(ActorCell.scala:561)
    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
    at akka.dispatch.Mailbox.run(Mailbox.scala:225)
    at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
    at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
    at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool
.java:1339)
    at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
    at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread
.java:107)



Could you please provide any insight on how to get logs. And why a simple
connection will not work.

Note: it works in my local system yarn.

Regards,
Vijay

Re: Flink Couchbase

Posted by Vijayendra Yadav <co...@gmail.com>.
Thanks Yang that helped.

Sent from my iPhone

> On Aug 24, 2020, at 8:44 PM, Yang Wang <da...@gmail.com> wrote:
> 
> 
> I think at least you have two different exceptions.
> 
> > java.lang.Exception: Container released on a *lost* node
> This usually means a Yarn nodemanager is down. So all the containers running on this node will be 
> released and rescheduled to a new one. If you want to figure out the root cause, you need to check
> the Yarn nodemanager logs.
> 
> > java.lang.OutOfMemoryError: Metaspace
> Could you check the value of flink configuration "taskmanager.memory.jvm-metaspace.size"? If it is
> too small, increasing it will help. Usually, 256m is enough for most cases.
> 
> 
> Best,
> Yang
> 
> Vijayendra Yadav <co...@gmail.com> 于2020年8月25日周二 上午4:51写道:
>> Another one -
>> 
>> Exception in thread "FileCache shutdown hook"
>> Exception: java.lang.OutOfMemoryError thrown from the UncaughtExceptionHandler in thread "FileCache shutdown hook"
>> 
>> Regards,
>> Vijay
>> 
>>> On Mon, Aug 24, 2020 at 1:04 PM Vijayendra Yadav <co...@gmail.com> wrote:
>>> Actually got this message in rolledover container logs: 
>>> 
>>> [org.slf4j.impl.Log4jLoggerFactory]
>>> Exception in thread "cb-timer-1-1" java.lang.OutOfMemoryError: Metaspace
>>> Exception in thread "Thread-16" java.lang.OutOfMemoryError: Metaspace
>>> Exception in thread "TransientBlobCache shutdown hook" java.lang.OutOfMemoryError: Metaspace
>>> Exception in thread "FileChannelManagerImpl-io shutdown hook" java.lang.OutOfMemoryError: Metaspace
>>> Exception in thread "Kafka Fetcher for Source: flink-kafka-consumer -> Map -> Filter -> Map -> Sink: s3-sink-raw (2/3)" java.lang.OutOfMemoryError: Metaspace
>>> Exception in thread "FileCache shutdown hook" java.lang.OutOfMemoryError: Metaspace
>>> Any suggestions on how to fix it ?
>>> 
>>> 
>>>> On Mon, Aug 24, 2020 at 12:53 PM Vijayendra Yadav <co...@gmail.com> wrote:
>>>> Hi Team,
>>>> 
>>>> Running a flink job on Yarn, I am trying to make connections to couchbase DB in one of my map functions in Flink Streaming job. But my task manager containers keep failing
>>>> and keep assigning new containers and not giving me an opportunity to get any useful logs. 
>>>> 
>>>>  val cluster = Cluster.connect("host", "user", "pwd")
>>>>  val bucket = cluster.bucket("bucket")
>>>>  val collection = bucket.defaultCollection
>>>> 
>>>> Only thing I see is yarn exception: 
>>>> 
>>>> java.lang.Exception: Container released on a *lost* node
>>>>     at org.apache.flink.yarn.YarnResourceManager.lambda$onContainersCompleted$0(YarnResourceManager.java:343)
>>>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:397)
>>>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:190)
>>>>     at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
>>>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
>>>>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
>>>>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
>>>>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
>>>>     at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
>>>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
>>>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>>>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>>>>     at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
>>>>     at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
>>>>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
>>>>     at akka.actor.ActorCell.invoke(ActorCell.scala:561)
>>>>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
>>>>     at akka.dispatch.Mailbox.run(Mailbox.scala:225)
>>>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
>>>>     at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>>>     at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>>>>     at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>>>>     at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>>>> 
>>>> 
>>>> 
>>>> Could you please provide any insight on how to get logs. And why a simple connection will not work.
>>>> 
>>>> Note: it works in my local system yarn.
>>>> 
>>>> Regards,
>>>> Vijay

Re: Flink Couchbase

Posted by Yang Wang <da...@gmail.com>.
I think at least you have two different exceptions.

> java.lang.Exception: Container released on a *lost* node
This usually means a Yarn nodemanager is down. So all the containers
running on this node will be
released and rescheduled to a new one. If you want to figure out the root
cause, you need to check
the Yarn nodemanager logs.

> java.lang.OutOfMemoryError: Metaspace
Could you check the value of flink configuration
"taskmanager.memory.jvm-metaspace.size"? If it is
too small, increasing it will help. Usually, 256m is enough for most cases.


Best,
Yang

Vijayendra Yadav <co...@gmail.com> 于2020年8月25日周二 上午4:51写道:

> Another one -
>
> Exception in thread "FileCache shutdown hook"
> Exception: java.lang.OutOfMemoryError thrown from the
> UncaughtExceptionHandler in thread "FileCache shutdown hook"
>
> Regards,
> Vijay
>
> On Mon, Aug 24, 2020 at 1:04 PM Vijayendra Yadav <co...@gmail.com>
> wrote:
>
>> Actually got this message in rolledover container logs:
>>
>> [org.slf4j.impl.Log4jLoggerFactory]
>> Exception in thread "cb-timer-1-1" java.lang.OutOfMemoryError: Metaspace
>> Exception in thread "Thread-16" java.lang.OutOfMemoryError: Metaspace
>> Exception in thread "TransientBlobCache shutdown hook" java.lang.OutOfMemoryError: Metaspace
>> Exception in thread "FileChannelManagerImpl-io shutdown hook" java.lang.OutOfMemoryError: Metaspace
>> Exception in thread "Kafka Fetcher for Source: flink-kafka-consumer -> Map -> Filter -> Map -> Sink: s3-sink-raw (2/3)" java.lang.OutOfMemoryError: Metaspace
>> Exception in thread "FileCache shutdown hook" java.lang.OutOfMemoryError: Metaspace
>>
>> Any suggestions on how to fix it ?
>>
>>
>>
>> On Mon, Aug 24, 2020 at 12:53 PM Vijayendra Yadav <co...@gmail.com>
>> wrote:
>>
>>> Hi Team,
>>>
>>> Running a flink job on Yarn, I am trying to make connections to
>>> couchbase DB in one of my map functions in Flink Streaming job. But my task
>>> manager containers keep failing
>>> and keep assigning new containers and not giving me an opportunity to
>>> get any useful logs.
>>>
>>>  val cluster = Cluster.connect("host", "user", "pwd")
>>>  val bucket = cluster.bucket("bucket")
>>>  val collection = bucket.defaultCollection
>>>
>>> Only thing I see is yarn exception:
>>>
>>> java.lang.Exception: Container released on a *lost* node
>>>     at org.apache.flink.yarn.YarnResourceManager
>>> .lambda$onContainersCompleted$0(YarnResourceManager.java:343)
>>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(
>>> AkkaRpcActor.java:397)
>>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(
>>> AkkaRpcActor.java:190)
>>>     at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor
>>> .handleRpcMessage(FencedAkkaRpcActor.java:74)
>>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(
>>> AkkaRpcActor.java:152)
>>>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
>>>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
>>>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123
>>> )
>>>     at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:
>>> 21)
>>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:
>>> 170)
>>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:
>>> 171)
>>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:
>>> 171)
>>>     at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
>>>     at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
>>>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
>>>     at akka.actor.ActorCell.invoke(ActorCell.scala:561)
>>>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
>>>     at akka.dispatch.Mailbox.run(Mailbox.scala:225)
>>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
>>>     at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>>     at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(
>>> ForkJoinPool.java:1339)
>>>     at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:
>>> 1979)
>>>     at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(
>>> ForkJoinWorkerThread.java:107)
>>>
>>>
>>>
>>> Could you please provide any insight on how to get logs. And why a
>>> simple connection will not work.
>>>
>>> Note: it works in my local system yarn.
>>>
>>> Regards,
>>> Vijay
>>>
>>

Re: Flink Couchbase

Posted by Vijayendra Yadav <co...@gmail.com>.
Another one -

Exception in thread "FileCache shutdown hook"
Exception: java.lang.OutOfMemoryError thrown from the
UncaughtExceptionHandler in thread "FileCache shutdown hook"

Regards,
Vijay

On Mon, Aug 24, 2020 at 1:04 PM Vijayendra Yadav <co...@gmail.com>
wrote:

> Actually got this message in rolledover container logs:
>
> [org.slf4j.impl.Log4jLoggerFactory]
> Exception in thread "cb-timer-1-1" java.lang.OutOfMemoryError: Metaspace
> Exception in thread "Thread-16" java.lang.OutOfMemoryError: Metaspace
> Exception in thread "TransientBlobCache shutdown hook" java.lang.OutOfMemoryError: Metaspace
> Exception in thread "FileChannelManagerImpl-io shutdown hook" java.lang.OutOfMemoryError: Metaspace
> Exception in thread "Kafka Fetcher for Source: flink-kafka-consumer -> Map -> Filter -> Map -> Sink: s3-sink-raw (2/3)" java.lang.OutOfMemoryError: Metaspace
> Exception in thread "FileCache shutdown hook" java.lang.OutOfMemoryError: Metaspace
>
> Any suggestions on how to fix it ?
>
>
>
> On Mon, Aug 24, 2020 at 12:53 PM Vijayendra Yadav <co...@gmail.com>
> wrote:
>
>> Hi Team,
>>
>> Running a flink job on Yarn, I am trying to make connections to
>> couchbase DB in one of my map functions in Flink Streaming job. But my task
>> manager containers keep failing
>> and keep assigning new containers and not giving me an opportunity to get
>> any useful logs.
>>
>>  val cluster = Cluster.connect("host", "user", "pwd")
>>  val bucket = cluster.bucket("bucket")
>>  val collection = bucket.defaultCollection
>>
>> Only thing I see is yarn exception:
>>
>> java.lang.Exception: Container released on a *lost* node
>>     at org.apache.flink.yarn.YarnResourceManager
>> .lambda$onContainersCompleted$0(YarnResourceManager.java:343)
>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(
>> AkkaRpcActor.java:397)
>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(
>> AkkaRpcActor.java:190)
>>     at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor
>> .handleRpcMessage(FencedAkkaRpcActor.java:74)
>>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(
>> AkkaRpcActor.java:152)
>>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
>>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
>>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
>>     at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21
>> )
>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170
>> )
>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171
>> )
>>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171
>> )
>>     at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
>>     at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
>>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
>>     at akka.actor.ActorCell.invoke(ActorCell.scala:561)
>>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
>>     at akka.dispatch.Mailbox.run(Mailbox.scala:225)
>>     at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
>>     at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>>     at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool
>> .java:1339)
>>     at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:
>> 1979)
>>     at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(
>> ForkJoinWorkerThread.java:107)
>>
>>
>>
>> Could you please provide any insight on how to get logs. And why a simple
>> connection will not work.
>>
>> Note: it works in my local system yarn.
>>
>> Regards,
>> Vijay
>>
>

Re: Flink Couchbase

Posted by Vijayendra Yadav <co...@gmail.com>.
Actually got this message in rolledover container logs:

[org.slf4j.impl.Log4jLoggerFactory]
Exception in thread "cb-timer-1-1" java.lang.OutOfMemoryError: Metaspace
Exception in thread "Thread-16" java.lang.OutOfMemoryError: Metaspace
Exception in thread "TransientBlobCache shutdown hook"
java.lang.OutOfMemoryError: Metaspace
Exception in thread "FileChannelManagerImpl-io shutdown hook"
java.lang.OutOfMemoryError: Metaspace
Exception in thread "Kafka Fetcher for Source: flink-kafka-consumer ->
Map -> Filter -> Map -> Sink: s3-sink-raw (2/3)"
java.lang.OutOfMemoryError: Metaspace
Exception in thread "FileCache shutdown hook"
java.lang.OutOfMemoryError: Metaspace

Any suggestions on how to fix it ?



On Mon, Aug 24, 2020 at 12:53 PM Vijayendra Yadav <co...@gmail.com>
wrote:

> Hi Team,
>
> Running a flink job on Yarn, I am trying to make connections to
> couchbase DB in one of my map functions in Flink Streaming job. But my task
> manager containers keep failing
> and keep assigning new containers and not giving me an opportunity to get
> any useful logs.
>
>  val cluster = Cluster.connect("host", "user", "pwd")
>  val bucket = cluster.bucket("bucket")
>  val collection = bucket.defaultCollection
>
> Only thing I see is yarn exception:
>
> java.lang.Exception: Container released on a *lost* node
>     at org.apache.flink.yarn.YarnResourceManager
> .lambda$onContainersCompleted$0(YarnResourceManager.java:343)
>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(
> AkkaRpcActor.java:397)
>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(
> AkkaRpcActor.java:190)
>     at org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor
> .handleRpcMessage(FencedAkkaRpcActor.java:74)
>     at org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(
> AkkaRpcActor.java:152)
>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
>     at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
>     at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
>     at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>     at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>     at akka.actor.Actor$class.aroundReceive(Actor.scala:517)
>     at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
>     at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
>     at akka.actor.ActorCell.invoke(ActorCell.scala:561)
>     at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
>     at akka.dispatch.Mailbox.run(Mailbox.scala:225)
>     at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
>     at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>     at akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool
> .java:1339)
>     at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:
> 1979)
>     at akka.dispatch.forkjoin.ForkJoinWorkerThread.run(
> ForkJoinWorkerThread.java:107)
>
>
>
> Could you please provide any insight on how to get logs. And why a simple
> connection will not work.
>
> Note: it works in my local system yarn.
>
> Regards,
> Vijay
>