You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "George Sakkis (JIRA)" <ji...@apache.org> on 2018/12/19 08:34:00 UTC
[jira] [Updated] (ARROW-4076) [Python] schema validation and
filters
[ https://issues.apache.org/jira/browse/ARROW-4076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
George Sakkis updated ARROW-4076:
---------------------------------
Description:
Currently [schema validation|https://github.com/apache/arrow/blob/758bd557584107cb336cbc3422744dacd93978af/python/pyarrow/parquet.py#L900] of {{ParquetDataset}} takes place before filtering. This may raise a {{ValueError}} if the schema is different in some dataset pieces, even if these pieces would be subsequently filtered out. I think validation should happen after filtering to prevent such spurious errors:
{noformat}
--- a/pyarrow/parquet.py
+++ b/pyarrow/parquet.py
@@ -878,13 +878,13 @@
if split_row_groups:
raise NotImplementedError("split_row_groups not yet implemented")
- if validate_schema:
- self.validate_schemas()
-
if filters is not None:
filters = _check_filters(filters)
self._filter(filters)
+ if validate_schema:
+ self.validate_schemas()
+
def validate_schemas(self):
open_file = self._get_open_file_func()
{noformat}
was:
Currently [schema validation|https://github.com/apache/arrow/blob/758bd557584107cb336cbc3422744dacd93978af/python/pyarrow/parquet.py#L900] of {{ParquetDataset}} takes place before filtering. This may raise a {{ValueError}}if the schema is different in some dataset pieces, even if these pieces would be subsequently filtered out. I think validation should happen after filtering to prevent such spurious errors:
{noformat}
--- a/pyarrow/parquet.py
+++ b/pyarrow/parquet.py
@@ -878,13 +878,13 @@
if split_row_groups:
raise NotImplementedError("split_row_groups not yet implemented")
- if validate_schema:
- self.validate_schemas()
-
if filters is not None:
filters = _check_filters(filters)
self._filter(filters)
+ if validate_schema:
+ self.validate_schemas()
+
def validate_schemas(self):
open_file = self._get_open_file_func()
{noformat}
> [Python] schema validation and filters
> --------------------------------------
>
> Key: ARROW-4076
> URL: https://issues.apache.org/jira/browse/ARROW-4076
> Project: Apache Arrow
> Issue Type: Bug
> Components: Python
> Reporter: George Sakkis
> Priority: Minor
>
> Currently [schema validation|https://github.com/apache/arrow/blob/758bd557584107cb336cbc3422744dacd93978af/python/pyarrow/parquet.py#L900] of {{ParquetDataset}} takes place before filtering. This may raise a {{ValueError}} if the schema is different in some dataset pieces, even if these pieces would be subsequently filtered out. I think validation should happen after filtering to prevent such spurious errors:
> {noformat}
> --- a/pyarrow/parquet.py
> +++ b/pyarrow/parquet.py
> @@ -878,13 +878,13 @@
> if split_row_groups:
> raise NotImplementedError("split_row_groups not yet implemented")
>
> - if validate_schema:
> - self.validate_schemas()
> -
> if filters is not None:
> filters = _check_filters(filters)
> self._filter(filters)
>
> + if validate_schema:
> + self.validate_schemas()
> +
> def validate_schemas(self):
> open_file = self._get_open_file_func()
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)