You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "Rahul Challapalli (JIRA)" <ji...@apache.org> on 2014/07/21 22:30:42 UTC

[jira] [Updated] (DRILL-1161) Drill Parquet writer fails with an Out of Memory issue when the data is large enough

     [ https://issues.apache.org/jira/browse/DRILL-1161?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Rahul Challapalli updated DRILL-1161:
-------------------------------------

    Attachment: error.log

> Drill Parquet writer fails with an Out of Memory issue when the data is large enough
> ------------------------------------------------------------------------------------
>
>                 Key: DRILL-1161
>                 URL: https://issues.apache.org/jira/browse/DRILL-1161
>             Project: Apache Drill
>          Issue Type: Bug
>          Components: Storage - Parquet, Storage - Writer
>            Reporter: Rahul Challapalli
>         Attachments: error.log
>
>
> git.commit.id.abbrev=e5c2da0
> The below query fails with an out of memory issue :
> create table `wide-columns-100000` as select columns[0] col0, cast(columns[1] as int) col1 from `wide-columns-100000.tbl`;
> The source file contains 100000 records. Each record has 2 columns. The first column is a string with 100000 characters in it and the second is an integer. Adding a limit to the above query succeeds. I attached the error messages from drillbit.log and drillbit.out
> Let me know if you need anything more



--
This message was sent by Atlassian JIRA
(v6.2#6252)