You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@parquet.apache.org by "Gabor Szadovszky (JIRA)" <ji...@apache.org> on 2019/02/06 09:22:00 UTC

[jira] [Resolved] (PARQUET-1514) ParquetFileWriter Records Compressed Bytes instead of Uncompressed Bytes

     [ https://issues.apache.org/jira/browse/PARQUET-1514?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Gabor Szadovszky resolved PARQUET-1514.
---------------------------------------
    Resolution: Fixed

> ParquetFileWriter Records Compressed Bytes instead of Uncompressed Bytes
> ------------------------------------------------------------------------
>
>                 Key: PARQUET-1514
>                 URL: https://issues.apache.org/jira/browse/PARQUET-1514
>             Project: Parquet
>          Issue Type: Bug
>            Reporter: BELUGA BEHR
>            Assignee: BELUGA BEHR
>            Priority: Minor
>              Labels: pull-request-available
>
> As I understand it, the class {{BlockMetaData}} has a {{getTotalByteSize}} and a {{getCompressedSize}}.  However, in the case of {{ParquetFileWriter}}, there is one place where the compressed size is being written into {{totalByteSize}} instead of the uncompressed size.  This looks incorrect.
> {code:java|title=ParquetFileWriter}
> private BlockMetaData currentBlock;
> ...
> public void appendRowGroup(SeekableInputStream from, BlockMetaData rowGroup,
>     boolean dropColumns) throws IOException {
>     ...
>     currentBlock.setTotalByteSize(blockCompressedSize);
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)