You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by Fotis P <fo...@gmail.com> on 2015/04/14 20:30:10 UTC
JobTimeoutException: Lost connection to JobManager
Hello everyone,
I am getting this weird exception while running some simple counting jobs
in Flink.
Exception in thread "main"
org.apache.flink.runtime.client.JobTimeoutException: Lost connection to
JobManager
at
org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
at
org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
at
org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
at
org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
at
org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
at
trackers.preprocessing.ExtractInfoFromLogs.main(ExtractInfoFromLogs.java:133)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after
[100000 milliseconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
at scala.concurrent.Await$.result(package.scala:107)
at scala.concurrent.Await.result(package.scala)
at
org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
... 10 more
The only call above which comes from my code is
ExtractInfoFromLogs.java:133 which is the environment.execute() method.
This exception comes when dealing with largish files (>10GB). No exception
is thrown when I am working with a smaller subset of my data.
Also I would swear that it was working fine until a few days ago, and the
code has not been changed :S Only change was a re-import of maven
dependencies.
I am unsure what other information I could provide that would help you help
me :)
I am running everything locally through the intelij IDE. Maven dependency
is set to 0.9-SNAPSHOT.
I have an 8-core Ubuntu 14.04 machine.
Thanks in advance :D
Re: JobTimeoutException: Lost connection to JobManager
Posted by Fotis P <fo...@gmail.com>.
Hello all,
I am glad to report that the problem has been resolved.
The new version appears to have been pushed in maven already, since all I
did was "re-import dependencies" from InteliJ menu.
Thank you very much for the very quick response.
Fotis
2015-04-15 15:02 GMT+02:00 Ufuk Celebi <uc...@apache.org>:
> On 15 Apr 2015, at 14:18, Maximilian Michels <mx...@apache.org> wrote:
>
> > The exception indicates that you're still using the old version. It
> takes some time for the new Maven artifact to get deployed to the snapshot
> repository. Apparently, a artifact has already been deployed this morning.
> Did you delete the jar files in your .m2 folder?
>
> I think that's what he meant.
>
> The problem is that the snapshot repositories take some time to
> synchronize.
>
> Please
> 1. git clone https://github.com/apache/flink.git
> 2. cd flink
> 3. mvn clean install -DskipTests
>
> This way you build Flink yourself and are guaranteed to work on a version
> with the fix.
>
> Sorry for the inconvenience. Does this solve it?
>
> – Ufuk
Re: JobTimeoutException: Lost connection to JobManager
Posted by Ufuk Celebi <uc...@apache.org>.
On 15 Apr 2015, at 14:18, Maximilian Michels <mx...@apache.org> wrote:
> The exception indicates that you're still using the old version. It takes some time for the new Maven artifact to get deployed to the snapshot repository. Apparently, a artifact has already been deployed this morning. Did you delete the jar files in your .m2 folder?
I think that's what he meant.
The problem is that the snapshot repositories take some time to synchronize.
Please
1. git clone https://github.com/apache/flink.git
2. cd flink
3. mvn clean install -DskipTests
This way you build Flink yourself and are guaranteed to work on a version with the fix.
Sorry for the inconvenience. Does this solve it?
– Ufuk
Re: JobTimeoutException: Lost connection to JobManager
Posted by Maximilian Michels <mx...@apache.org>.
The exception indicates that you're still using the old version. It takes
some time for the new Maven artifact to get deployed to the snapshot
repository. Apparently, a artifact has already been deployed this morning.
Did you delete the jar files in your .m2 folder?
On Wed, Apr 15, 2015 at 1:38 PM, Mohamed Nadjib MAMI <ma...@iai.uni-bonn.de>
wrote:
> Hello,
>
> I'm still facing the problem with 0.9-SNAPSHOT version. Tried to remove
> the libraries and download them again but same issue.
>
> Greetings,
> Mohamed
>
>
> Exception in thread "main"
> org.apache.flink.runtime.client.JobTimeoutException: Lost connection to
> JobManager
> at
> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
> at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
> at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
> at
> org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
> at
> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
> at Main.main(Main.java:142)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
> [100000 milliseconds]
> at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
> at
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
> at
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
> at scala.concurrent.Await$.result(package.scala:107)
> at scala.concurrent.Await.result(package.scala)
> at
> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
> ... 5 more
>
>
> On 15.04.2015 01:02, Stephan Ewen wrote:
>
> I pushed a fix to the master. The problem should now be gone.
>
> Please let us know if you experience other issues!
>
> Greetings,
> Stephan
>
>
> On Tue, Apr 14, 2015 at 9:57 PM, Mohamed Nadjib MAMI <mami@iai.uni-bonn.de
> > wrote:
>
>> Hello,
>>
>> Once I got the message, few seconds, I received your email. Well, this
>> just to cast a need for a fix.
>>
>> Happy to feel the dynamism of the work. Great work.
>>
>>
>> On 14.04.2015 21:50, Stephan Ewen wrote:
>>
>> You are on the latest snapshot version? I think there is an inconsistency
>> in there. Will try to fix that toning.
>>
>> Can you actually use the milestone1 version? That one should be good.
>>
>> Greetings,
>> Stephan
>> Am 14.04.2015 20:31 schrieb "Fotis P" <fo...@gmail.com>:
>>
>>> Hello everyone,
>>>
>>> I am getting this weird exception while running some simple counting
>>> jobs in Flink.
>>>
>>> Exception in thread "main"
>>> org.apache.flink.runtime.client.JobTimeoutException: Lost connection to
>>> JobManager
>>> at
>>> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
>>> at
>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
>>> at
>>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
>>> at
>>> org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
>>> at
>>> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
>>> at
>>> trackers.preprocessing.ExtractInfoFromLogs.main(ExtractInfoFromLogs.java:133)
>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>> at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> at java.lang.reflect.Method.invoke(Method.java:606)
>>> at
>>> com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
>>> Caused by: java.util.concurrent.TimeoutException: Futures timed out
>>> after [100000 milliseconds]
>>> at
>>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>> at
>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>> at
>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>> at scala.concurrent.Await$.result(package.scala:107)
>>> at scala.concurrent.Await.result(package.scala)
>>> at
>>> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
>>> ... 10 more
>>>
>>>
>>> The only call above which comes from my code is
>>> ExtractInfoFromLogs.java:133 which is the environment.execute() method.
>>>
>>> This exception comes when dealing with largish files (>10GB). No
>>> exception is thrown when I am working with a smaller subset of my data.
>>> Also I would swear that it was working fine until a few days ago, and
>>> the code has not been changed :S Only change was a re-import of maven
>>> dependencies.
>>>
>>> I am unsure what other information I could provide that would help you
>>> help me :)
>>>
>>> I am running everything locally through the intelij IDE. Maven
>>> dependency is set to 0.9-SNAPSHOT.
>>> I have an 8-core Ubuntu 14.04 machine.
>>>
>>> Thanks in advance :D
>>>
>>
>> --
>> Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候, تحياتي.
>> Mohamed Nadjib Mami
>> PhD Student - EIS Department - Bonn University, Germany.
>> About me! <http://www.strikingly.com/mohamed-nadjib-mami>
>> LinkedIn
>>
>
>
>
> --
> Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候, تحياتي.
> Mohamed Nadjib Mami
> PhD Student - EIS Department - Bonn University, Germany.
> About me! <http://www.strikingly.com/mohamed-nadjib-mami>
> LinkedIn
>
Re: JobTimeoutException: Lost connection to JobManager
Posted by Mohamed Nadjib MAMI <ma...@iai.uni-bonn.de>.
Hello,
I'm still facing the problem with 0.9-SNAPSHOT version. Tried to remove
the libraries and download them again but same issue.
Greetings,
Mohamed
Exception in thread "main"
org.apache.flink.runtime.client.JobTimeoutException: Lost connection to
JobManager
at
org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
at
org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
at
org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
at
org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
at
org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
at Main.main(Main.java:142)
Caused by: java.util.concurrent.TimeoutException: Futures timed out
after [100000 milliseconds]
at
scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
at scala.concurrent.Await$.result(package.scala:107)
at scala.concurrent.Await.result(package.scala)
at
org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
... 5 more
On 15.04.2015 01:02, Stephan Ewen wrote:
> I pushed a fix to the master. The problem should now be gone.
>
> Please let us know if you experience other issues!
>
> Greetings,
> Stephan
>
>
> On Tue, Apr 14, 2015 at 9:57 PM, Mohamed Nadjib MAMI
> <mami@iai.uni-bonn.de <ma...@iai.uni-bonn.de>> wrote:
>
> Hello,
>
> Once I got the message, few seconds, I received your email. Well,
> this just to cast a need for a fix.
>
> Happy to feel the dynamism of the work. Great work.
>
>
> On 14.04.2015 21:50, Stephan Ewen wrote:
>>
>> You are on the latest snapshot version? I think there is an
>> inconsistency in there. Will try to fix that toning.
>>
>> Can you actually use the milestone1 version? That one should be good.
>>
>> Greetings,
>> Stephan
>>
>> Am 14.04.2015 20:31 schrieb "Fotis P" <fotisp87@gmail.com
>> <ma...@gmail.com>>:
>>
>> Hello everyone,
>>
>> I am getting this weird exception while running some simple
>> counting jobs in Flink.
>>
>> Exception in thread "main"
>> org.apache.flink.runtime.client.JobTimeoutException: Lost
>> connection to JobManager
>> at
>> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
>> at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
>> at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
>> at
>> org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
>> at
>> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
>> at
>> trackers.preprocessing.ExtractInfoFromLogs.main(ExtractInfoFromLogs.java:133)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:606)
>> at
>> com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
>> Caused by: java.util.concurrent.TimeoutException: Futures
>> timed out after [100000 milliseconds]
>> at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>> at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>> at
>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>> at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>> at scala.concurrent.Await$.result(package.scala:107)
>> at scala.concurrent.Await.result(package.scala)
>> at
>> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
>> ... 10 more
>>
>>
>> The only call above which comes from my code is
>> ExtractInfoFromLogs.java:133 which is the
>> environment.execute() method.
>>
>> This exception comes when dealing with largish files (>10GB).
>> No exception is thrown when I am working with a smaller
>> subset of my data.
>> Also I would swear that it was working fine until a few days
>> ago, and the code has not been changed :S Only change was a
>> re-import of maven dependencies.
>>
>> I am unsure what other information I could provide that would
>> help you help me :)
>>
>> I am running everything locally through the intelij IDE.
>> Maven dependency is set to 0.9-SNAPSHOT.
>> I have an 8-core Ubuntu 14.04 machine.
>>
>> Thanks in advance :D
>>
>
> --
> Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候,
> تحياتي. Mohamed Nadjib Mami
> PhD Student - EIS Department - Bonn University, Germany.
> About me! <http://www.strikingly.com/mohamed-nadjib-mami>
> LinkedIn
>
>
--
Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候,
تحياتي. Mohamed Nadjib Mami
PhD Student - EIS Department - Bonn University, Germany.
About me! <http://www.strikingly.com/mohamed-nadjib-mami>
LinkedIn
Re: JobTimeoutException: Lost connection to JobManager
Posted by Stephan Ewen <se...@apache.org>.
I pushed a fix to the master. The problem should now be gone.
Please let us know if you experience other issues!
Greetings,
Stephan
On Tue, Apr 14, 2015 at 9:57 PM, Mohamed Nadjib MAMI <ma...@iai.uni-bonn.de>
wrote:
> Hello,
>
> Once I got the message, few seconds, I received your email. Well, this
> just to cast a need for a fix.
>
> Happy to feel the dynamism of the work. Great work.
>
>
> On 14.04.2015 21:50, Stephan Ewen wrote:
>
> You are on the latest snapshot version? I think there is an inconsistency
> in there. Will try to fix that toning.
>
> Can you actually use the milestone1 version? That one should be good.
>
> Greetings,
> Stephan
> Am 14.04.2015 20:31 schrieb "Fotis P" <fo...@gmail.com>:
>
>> Hello everyone,
>>
>> I am getting this weird exception while running some simple counting
>> jobs in Flink.
>>
>> Exception in thread "main"
>> org.apache.flink.runtime.client.JobTimeoutException: Lost connection to
>> JobManager
>> at
>> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
>> at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
>> at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
>> at
>> org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
>> at
>> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
>> at
>> trackers.preprocessing.ExtractInfoFromLogs.main(ExtractInfoFromLogs.java:133)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:606)
>> at
>> com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>> [100000 milliseconds]
>> at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>> at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>> at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>> at scala.concurrent.Await$.result(package.scala:107)
>> at scala.concurrent.Await.result(package.scala)
>> at
>> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
>> ... 10 more
>>
>>
>> The only call above which comes from my code is
>> ExtractInfoFromLogs.java:133 which is the environment.execute() method.
>>
>> This exception comes when dealing with largish files (>10GB). No
>> exception is thrown when I am working with a smaller subset of my data.
>> Also I would swear that it was working fine until a few days ago, and
>> the code has not been changed :S Only change was a re-import of maven
>> dependencies.
>>
>> I am unsure what other information I could provide that would help you
>> help me :)
>>
>> I am running everything locally through the intelij IDE. Maven
>> dependency is set to 0.9-SNAPSHOT.
>> I have an 8-core Ubuntu 14.04 machine.
>>
>> Thanks in advance :D
>>
>
> --
> Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候, تحياتي.
> Mohamed Nadjib Mami
> PhD Student - EIS Department - Bonn University, Germany.
> About me! <http://www.strikingly.com/mohamed-nadjib-mami>
> LinkedIn
>
Re: JobTimeoutException: Lost connection to JobManager
Posted by Mohamed Nadjib MAMI <ma...@iai.uni-bonn.de>.
Hello,
Once I got the message, few seconds, I received your email. Well, this
just to cast a need for a fix.
Happy to feel the dynamism of the work. Great work.
On 14.04.2015 21:50, Stephan Ewen wrote:
>
> You are on the latest snapshot version? I think there is an
> inconsistency in there. Will try to fix that toning.
>
> Can you actually use the milestone1 version? That one should be good.
>
> Greetings,
> Stephan
>
> Am 14.04.2015 20:31 schrieb "Fotis P" <fotisp87@gmail.com
> <ma...@gmail.com>>:
>
> Hello everyone,
>
> I am getting this weird exception while running some simple
> counting jobs in Flink.
>
> Exception in thread "main"
> org.apache.flink.runtime.client.JobTimeoutException: Lost
> connection to JobManager
> at
> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
> at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
> at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
> at
> org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
> at
> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
> at
> trackers.preprocessing.ExtractInfoFromLogs.main(ExtractInfoFromLogs.java:133)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at
> com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
> Caused by: java.util.concurrent.TimeoutException: Futures timed
> out after [100000 milliseconds]
> at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
> at
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
> at
> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
> at
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
> at scala.concurrent.Await$.result(package.scala:107)
> at scala.concurrent.Await.result(package.scala)
> at
> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
> ... 10 more
>
>
> The only call above which comes from my code is
> ExtractInfoFromLogs.java:133 which is the environment.execute()
> method.
>
> This exception comes when dealing with largish files (>10GB). No
> exception is thrown when I am working with a smaller subset of my
> data.
> Also I would swear that it was working fine until a few days ago,
> and the code has not been changed :S Only change was a re-import
> of maven dependencies.
>
> I am unsure what other information I could provide that would help
> you help me :)
>
> I am running everything locally through the intelij IDE. Maven
> dependency is set to 0.9-SNAPSHOT.
> I have an 8-core Ubuntu 14.04 machine.
>
> Thanks in advance :D
>
--
Regards, Grüße, Cordialement, Recuerdos, Saluti, προσρήσεις, 问候,
تحياتي. Mohamed Nadjib Mami
PhD Student - EIS Department - Bonn University, Germany.
About me! <http://www.strikingly.com/mohamed-nadjib-mami>
LinkedIn
Re: JobTimeoutException: Lost connection to JobManager
Posted by Stephan Ewen <se...@apache.org>.
You are on the latest snapshot version? I think there is an inconsistency
in there. Will try to fix that toning.
Can you actually use the milestone1 version? That one should be good.
Greetings,
Stephan
Am 14.04.2015 20:31 schrieb "Fotis P" <fo...@gmail.com>:
> Hello everyone,
>
> I am getting this weird exception while running some simple counting jobs
> in Flink.
>
> Exception in thread "main"
> org.apache.flink.runtime.client.JobTimeoutException: Lost connection to
> JobManager
> at
> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:164)
> at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:198)
> at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.submitJobAndWait(FlinkMiniCluster.scala:188)
> at
> org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:179)
> at
> org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:54)
> at
> trackers.preprocessing.ExtractInfoFromLogs.main(ExtractInfoFromLogs.java:133)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
> [100000 milliseconds]
> at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
> at
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
> at
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
> at scala.concurrent.Await$.result(package.scala:107)
> at scala.concurrent.Await.result(package.scala)
> at
> org.apache.flink.runtime.client.JobClient.submitJobAndWait(JobClient.java:143)
> ... 10 more
>
>
> The only call above which comes from my code is
> ExtractInfoFromLogs.java:133 which is the environment.execute() method.
>
> This exception comes when dealing with largish files (>10GB). No exception
> is thrown when I am working with a smaller subset of my data.
> Also I would swear that it was working fine until a few days ago, and the
> code has not been changed :S Only change was a re-import of maven
> dependencies.
>
> I am unsure what other information I could provide that would help you
> help me :)
>
> I am running everything locally through the intelij IDE. Maven dependency
> is set to 0.9-SNAPSHOT.
> I have an 8-core Ubuntu 14.04 machine.
>
> Thanks in advance :D
>