You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by Liz Bai <li...@icloud.com> on 2016/09/02 05:51:36 UTC

Dynamic Partitions When Writing Parquet

Hi there,

I have a question about writing Parquet using SparkSQL. Spark 1.4 has already supported writing DataFrames as Parquet files with “partitionBy(colNames: String*)”, as Spark-6561 fixed.
Is there any method or plan to write Parquet with dynamic partitions? For example, instead of partitioning on the column Year(range:1900-2016) directly, do partition on the *decade* of the Year(range:190-201).
Thanks.

Best,
Ran
---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscribe@spark.apache.org