You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@parquet.apache.org by "Gabor Szadovszky (JIRA)" <ji...@apache.org> on 2019/02/06 09:22:00 UTC
[jira] [Resolved] (PARQUET-1514) ParquetFileWriter Records
Compressed Bytes instead of Uncompressed Bytes
[ https://issues.apache.org/jira/browse/PARQUET-1514?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Gabor Szadovszky resolved PARQUET-1514.
---------------------------------------
Resolution: Fixed
> ParquetFileWriter Records Compressed Bytes instead of Uncompressed Bytes
> ------------------------------------------------------------------------
>
> Key: PARQUET-1514
> URL: https://issues.apache.org/jira/browse/PARQUET-1514
> Project: Parquet
> Issue Type: Bug
> Reporter: BELUGA BEHR
> Assignee: BELUGA BEHR
> Priority: Minor
> Labels: pull-request-available
>
> As I understand it, the class {{BlockMetaData}} has a {{getTotalByteSize}} and a {{getCompressedSize}}. However, in the case of {{ParquetFileWriter}}, there is one place where the compressed size is being written into {{totalByteSize}} instead of the uncompressed size. This looks incorrect.
> {code:java|title=ParquetFileWriter}
> private BlockMetaData currentBlock;
> ...
> public void appendRowGroup(SeekableInputStream from, BlockMetaData rowGroup,
> boolean dropColumns) throws IOException {
> ...
> currentBlock.setTotalByteSize(blockCompressedSize);
> }
> {code}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)