You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@stanbol.apache.org by "Rupert Westenthaler (JIRA)" <ji...@apache.org> on 2015/06/12 10:39:00 UTC

[jira] [Resolved] (STANBOL-1424) The commons.opennlp module can load the same model twice in parallel

     [ https://issues.apache.org/jira/browse/STANBOL-1424?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Rupert Westenthaler resolved STANBOL-1424.
------------------------------------------
    Resolution: Fixed

fixed with http://svn.apache.org/r1685046 for trunk

> The commons.opennlp module can load the same model twice in parallel
> --------------------------------------------------------------------
>
>                 Key: STANBOL-1424
>                 URL: https://issues.apache.org/jira/browse/STANBOL-1424
>             Project: Stanbol
>          Issue Type: Bug
>    Affects Versions: 0.12.0
>            Reporter: Rupert Westenthaler
>            Assignee: Rupert Westenthaler
>            Priority: Minor
>             Fix For: 1.0.0, 0.12.1
>
>
> The commons.opennlp model allows to load models by their names via the DataFileProvider infrastructure. Loaded models are cached in memory. 
> If two components do request the same model in a short time. Especially when the 2md request for a model comes before the first was completed the same model is loaded twice in parallel. This will result that two instances of the model are loaded.
> While the 2nd request will override the cached model of the first the first component requesting the model might still hold a reference. In this case two instances of the model are holded in-memory.
> To solve those situations the OpenNLP service needs to use lock while loading models.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)