You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@parquet.apache.org by "Wes McKinney (JIRA)" <ji...@apache.org> on 2016/02/02 21:25:40 UTC
[jira] [Created] (PARQUET-505) Column reader: automatically handle
large data pages
Wes McKinney created PARQUET-505:
------------------------------------
Summary: Column reader: automatically handle large data pages
Key: PARQUET-505
URL: https://issues.apache.org/jira/browse/PARQUET-505
Project: Parquet
Issue Type: Bug
Components: parquet-cpp
Reporter: Wes McKinney
Currently, we are only supporting data pages whose headers are 64K or less (see {{parquet/column/serialized-page.cc}}. Since page headers can essentially be arbitrarily large (in pathological cases) because of the page statistics, if deserializing the page header fails, we should attempt to read a progressively larger amount of file data in effort to find the end of the page header.
As part of this (and to make testing easier!), the maximum data page header size should be configurable. We can write test cases by defining appropriate Statistics structs to yield serialized page headers of whatever desired size.
On malformed files, we may run past the end of the file, in such cases we should raise a reasonable exception.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)