You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Ufuk Celebi (JIRA)" <ji...@apache.org> on 2014/07/03 11:52:25 UTC

[jira] [Commented] (FLINK-941) Possible deadlock after increasing my data set size

    [ https://issues.apache.org/jira/browse/FLINK-941?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14051273#comment-14051273 ] 

Ufuk Celebi commented on FLINK-941:
-----------------------------------

Is there an update? I am curious whether you tried it with a version after [f13ad5b415a57e7d1c97319935a04f076cc1776b|https://github.com/apache/incubator-flink/commit/f13ad5b415a57e7d1c97319935a04f076cc1776b]?

> Possible deadlock after increasing my data set size
> ---------------------------------------------------
>
>                 Key: FLINK-941
>                 URL: https://issues.apache.org/jira/browse/FLINK-941
>             Project: Flink
>          Issue Type: Bug
>    Affects Versions: pre-apache-0.5.1
>            Reporter: Bastian Köcher
>            Assignee: Stephan Ewen
>             Fix For: 0.6-incubating
>
>         Attachments: IMPRO-3.SS14.G03.zip
>
>
> If I increase my data set, my algorithm stops at some point and doesn't continue anymore. I already waited a quite amount of time, but nothing happens. The linux processor explorer also displays that the process is sleeping and waiting for something to happen, could maybe be a deadlock.
> I attached the source of my program, the class HAC_2 is the actual algorithm.
> Changing the line 271 from "if(Integer.parseInt(tokens[0]) > 282)" to "if(Integer.parseInt(tokens[0]) > 283)" at my PC "enables" the bug. The numbers 282, 283 are the numbers of the documents in my test data and this line skips all documents with an id greater than that.



--
This message was sent by Atlassian JIRA
(v6.2#6252)