You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Julius Michaelis (Jira)" <ji...@apache.org> on 2020/10/16 09:59:00 UTC

[jira] [Comment Edited] (FLINK-18712) Flink RocksDB statebackend memory leak issue

    [ https://issues.apache.org/jira/browse/FLINK-18712?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17215118#comment-17215118 ] 

Julius Michaelis edited comment on FLINK-18712 at 10/16/20, 9:58 AM:
---------------------------------------------------------------------

[~yunta] , may I ask: are you still working on this?
If so, can I ask you to check that this bug indeed doesn't appear with {{state.backend.rocksdb.memory.managed: false}}? I'd like to make sure that the bug I'm seeing is really the same bug.

(MALLOC_ARENA_MAX does seem to take care of it, so it's likely the same bug.)


was (Author: caesar):
[~yunta] , may I ask: are you still working on this?
If so, can I ask you to check that this bug indeed doesn't appear with {{state.backend.rocksdb.memory.managed: false}}? I'd like to make sure that the bug I'm seeing is really the same bug.

> Flink RocksDB statebackend memory leak issue 
> ---------------------------------------------
>
>                 Key: FLINK-18712
>                 URL: https://issues.apache.org/jira/browse/FLINK-18712
>             Project: Flink
>          Issue Type: Bug
>          Components: Runtime / State Backends
>    Affects Versions: 1.10.0
>            Reporter: Farnight
>            Assignee: Yun Tang
>            Priority: Critical
>             Fix For: 1.12.0
>
>
> When using RocksDB as our statebackend, we found it will lead to memory leak when restarting job (manually or in recovery case).
>  
> How to reproduce:
>  # increase RocksDB blockcache size(e.g. 1G), it is easier to monitor and reproduce.
>  # start a job using RocksDB statebackend.
>  # when the RocksDB blockcache reachs maximum size, restart the job. and monitor the memory usage (k8s pod working set) of the TM.
>  # go through step 2-3 few more times. and memory will keep raising.
>  
> Any solution or suggestion for this? Thanks!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)