You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@drill.apache.org by "Abhishek Girish (JIRA)" <ji...@apache.org> on 2015/10/08 01:43:26 UTC
[jira] [Updated] (DRILL-3913) Possible memory leak during CTAS
using 30 TB TPC-H dataset
[ https://issues.apache.org/jira/browse/DRILL-3913?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Abhishek Girish updated DRILL-3913:
-----------------------------------
Attachment: query_profile.json
create_table_sf30000.txt
> Possible memory leak during CTAS using 30 TB TPC-H dataset
> ----------------------------------------------------------
>
> Key: DRILL-3913
> URL: https://issues.apache.org/jira/browse/DRILL-3913
> Project: Apache Drill
> Issue Type: Bug
> Components: Execution - Flow
> Affects Versions: 1.2.0
> Environment: 47 nodes configured with 32 GB Drill Direct memory
> Reporter: Abhishek Girish
> Attachments: create_table_sf30000.txt, query_profile.json
>
>
> 8 CTAS queries were executed sequentially to write TPC-H text data into Parquet. After successfully writing a few tables, CTAS failed with OOM.
> Restarting Drillbits fixed the problem and re-run of pending CTAS queries completed. This process was done twice in-order to complete all 8 tables to be written. Overall source was 30TB in size.
> Queries attached. Query profile for one of the CTAS which failed is attached. Logs indicated that the Drillbit was out of Direct Memory.
> Can share more details as required.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)