You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Wenchen Fan (JIRA)" <ji...@apache.org> on 2016/09/29 15:31:21 UTC
[jira] [Created] (SPARK-17725) Spark should not write out parquet
files with schema containing non-nullable fields
Wenchen Fan created SPARK-17725:
-----------------------------------
Summary: Spark should not write out parquet files with schema containing non-nullable fields
Key: SPARK-17725
URL: https://issues.apache.org/jira/browse/SPARK-17725
Project: Spark
Issue Type: Bug
Components: SQL
Affects Versions: 2.0.0
Reporter: Wenchen Fan
Since Spark 1.3, after PR https://github.com/apache/spark/pull/4826 , Spark SQL will always set all schema fields to nullable before writing out parquet files, to make the data pipeline more robust.
However, this behaviour has been changed in 2.0 accidently by PR https://github.com/apache/spark/pull/11509
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org