You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Nikita Poberezkin (JIRA)" <ji...@apache.org> on 2018/08/28 10:44:00 UTC

[jira] [Comment Edited] (SPARK-25102) Write Spark version information to Parquet file footers

    [ https://issues.apache.org/jira/browse/SPARK-25102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594809#comment-16594809 ] 

Nikita Poberezkin edited comment on SPARK-25102 at 8/28/18 10:43 AM:
---------------------------------------------------------------------

Hi, [~zi]. I've tried to override getName method in org.apache.spark.sql.execution.datasources.parquet.ParquetWriteSupport. The problem is that the only way (that i know about) to find Spark version programmatically on an executor node is SparkSession.builder().getOrCreate().version. But, when I ran tests i received the following error: 
 Caused by: java.lang.IllegalStateException: SparkSession should only be created and accessed on the driver.
 Do you know any other way to find Spark version in an executor?


was (Author: npoberezkin):
Hi, Zoltan. I've tried to override getName method in org.apache.spark.sql.execution.datasources.parquet.ParquetWriteSupport. The problem is that the only way (that i know about) to find Spark version programmatically on an executor node is SparkSession.builder().getOrCreate().version. But, when I ran tests i received the following error: 
Caused by: java.lang.IllegalStateException: SparkSession should only be created and accessed on the driver.
Do you know any other way to find Spark version in an executor?

> Write Spark version information to Parquet file footers
> -------------------------------------------------------
>
>                 Key: SPARK-25102
>                 URL: https://issues.apache.org/jira/browse/SPARK-25102
>             Project: Spark
>          Issue Type: Improvement
>          Components: Spark Core
>    Affects Versions: 2.3.1
>            Reporter: Zoltan Ivanfi
>            Priority: Major
>
> -PARQUET-352- added support for the "writer.model.name" property in the Parquet metadata to identify the object model (application) that wrote the file.
> The easiest way to write this property is by overriding getName() of org.apache.parquet.hadoop.api.WriteSupport. In Spark, this would mean adding getName() to the org.apache.spark.sql.execution.datasources.parquet.ParquetWriteSupport class.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org