You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "Wes McKinney (JIRA)" <ji...@apache.org> on 2017/08/28 22:02:00 UTC

[jira] [Assigned] (ARROW-1400) [Python] Ability to create partitions when writing to Parquet

     [ https://issues.apache.org/jira/browse/ARROW-1400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Wes McKinney reassigned ARROW-1400:
-----------------------------------

    Assignee: Safyre Anderson

> [Python] Ability to create partitions when writing to Parquet
> -------------------------------------------------------------
>
>                 Key: ARROW-1400
>                 URL: https://issues.apache.org/jira/browse/ARROW-1400
>             Project: Apache Arrow
>          Issue Type: New Feature
>          Components: Python
>    Affects Versions: 0.6.0
>         Environment: Mac OS Sierra 10.12.6
>            Reporter: Safyre Anderson
>            Assignee: Safyre Anderson
>            Priority: Minor
>             Fix For: 0.7.0
>
>
> I'm fairly new to pyarrow so I apologize if this is already a feature, but I couldn't find a solution in the documentation nor an existing issue.  Basically I'm trying to export pandas dataframes to .parquet files with partitions. I can see that pyarrow.parquet has a way of reading .parquet files with partitions, but there's no indication that it can write with partitions. E.g., it would be nice if there was a parameter in pyarrow.Table.write_table() that took a list of columns to partition the table similar to the pyspark implementation: spark.write.parquet's "partitionBy" parameter.
> Referenced links:
> https://arrow.apache.org/docs/python/parquet.html
> https://arrow.apache.org/docs/python/parquet.html?highlight=pyarrow%20parquet%20partition



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)