You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by "Jay Kreps (JIRA)" <ji...@apache.org> on 2015/02/07 23:15:35 UTC

[jira] [Updated] (KAFKA-1404) Close unused log file

     [ https://issues.apache.org/jira/browse/KAFKA-1404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Jay Kreps updated KAFKA-1404:
-----------------------------
    Priority: Major  (was: Critical)

> Close unused log file
> ---------------------
>
>                 Key: KAFKA-1404
>                 URL: https://issues.apache.org/jira/browse/KAFKA-1404
>             Project: Kafka
>          Issue Type: Improvement
>          Components: core
>    Affects Versions: 0.8.1
>            Reporter: Xinyao Hu
>   Original Estimate: 336h
>  Remaining Estimate: 336h
>
> This is somewhat related to KAFKA-1403. 
> One way to hack KAFKA-1403 is to roll a new file in a short period of time. However, this will result in many file descriptors open. Take our application for example, each server hosts about 5k topic-partition, if we roll a new file per hour, we will add ~100k file descriptors per day (I checked only .log is open but not .index which might be pinned in memory). We will run out of 1M file descriptor in about a week. However our disk can host much longer.  
> In reality very few of these file descriptors will be used. The most recent fd will be used to append data and the old file descriptor will be used for query. We should provide a parameter like max.num.fds and do LRU to decide which fds should be open. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)