You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@nifi.apache.org by "Mark Payne (JIRA)" <ji...@apache.org> on 2015/04/28 15:32:06 UTC
[jira] [Resolved] (NIFI-527) Persistent Prov Repo should compress
write-ahead-log files in chunks
[ https://issues.apache.org/jira/browse/NIFI-527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Mark Payne resolved NIFI-527.
-----------------------------
Resolution: Fixed
> Persistent Prov Repo should compress write-ahead-log files in chunks
> --------------------------------------------------------------------
>
> Key: NIFI-527
> URL: https://issues.apache.org/jira/browse/NIFI-527
> Project: Apache NiFi
> Issue Type: Bug
> Components: Extensions
> Reporter: Mark Payne
> Assignee: Mark Payne
> Fix For: 0.1.0
>
> Attachments: 0001-NIFI-527-Refactored-the-serialization-format-of-the-.patch, 0002-NIFI-527-More-performance-improvements-including-reu.patch, 0003-NIFI-527-Cleaned-up-log-messages.patch, 0004-NIFI-527-Added-unit-test-to-verify-backpressure.patch
>
>
> Currently when we rollover a prov log, we compress the entire thing. This means that when we want to jump to a particular offset we have to open a GZIPInputStream and read through all of the data. If we instead compress the logs in chunks, we can actually jump to a particular chunk using FileInputStream.skip and then open a GZIPInputStream from there. Currently, this is by far the biggest bottleneck in the prov repo when doing queries.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)