You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@lucene.apache.org by "Adrien Grand (Jira)" <ji...@apache.org> on 2021/06/07 16:41:00 UTC

[jira] [Created] (LUCENE-9994) Can IndexingChain better protect against large documents?

Adrien Grand created LUCENE-9994:
------------------------------------

             Summary: Can IndexingChain better protect against large documents?
                 Key: LUCENE-9994
                 URL: https://issues.apache.org/jira/browse/LUCENE-9994
             Project: Lucene - Core
          Issue Type: Improvement
            Reporter: Adrien Grand


It's easy for a single document to use several times the amount of RAM that is configured on IndexWriter by having many fields or many terms on a single field. Could we improve IndexingChain to reject such documents before they may cause an out-of-memory error? We could make such documents born deleted in the new segment like we already do when consuming a TokenStream raises an exception.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@lucene.apache.org
For additional commands, e-mail: issues-help@lucene.apache.org