You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@daffodil.apache.org by "Dave Thompson (Jira)" <ji...@apache.org> on 2021/02/03 13:53:00 UTC

[jira] [Created] (DAFFODIL-2468) Uparsing an infoset for an 800mb csv file runs out of memory

Dave Thompson created DAFFODIL-2468:
---------------------------------------

             Summary: Uparsing an infoset for an 800mb csv file runs out of memory 
                 Key: DAFFODIL-2468
                 URL: https://issues.apache.org/jira/browse/DAFFODIL-2468
             Project: Daffodil
          Issue Type: Bug
    Affects Versions: 3.1.0
            Reporter: Dave Thompson
         Attachments: csv_data800m.csv.gz

While verifying DAFFODIL-2455 - - Large CSV file causes "Attempting to backtrack too far" exception, found that unparsing the successfully parsed 800mb CSV files infoset ran out of memory.

Increased the DAFFODIL_JAVA_OPTS memory setting several time up to 32gb and tried unparsing the infoset, each time running out of memory. Ran on test platform which has 90+GB of memory. 

Parsed and unparsed using the shema from dfdl-shemas/dfdl-csv repo.

The 800gb csv file (csv_data800m.csv) gzipped.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)