You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Nikita Poberezkin (JIRA)" <ji...@apache.org> on 2018/08/28 10:44:00 UTC
[jira] [Comment Edited] (SPARK-25102) Write Spark version
information to Parquet file footers
[ https://issues.apache.org/jira/browse/SPARK-25102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594809#comment-16594809 ]
Nikita Poberezkin edited comment on SPARK-25102 at 8/28/18 10:43 AM:
---------------------------------------------------------------------
Hi, [~zi]. I've tried to override getName method in org.apache.spark.sql.execution.datasources.parquet.ParquetWriteSupport. The problem is that the only way (that i know about) to find Spark version programmatically on an executor node is SparkSession.builder().getOrCreate().version. But, when I ran tests i received the following error:
Caused by: java.lang.IllegalStateException: SparkSession should only be created and accessed on the driver.
Do you know any other way to find Spark version in an executor?
was (Author: npoberezkin):
Hi, Zoltan. I've tried to override getName method in org.apache.spark.sql.execution.datasources.parquet.ParquetWriteSupport. The problem is that the only way (that i know about) to find Spark version programmatically on an executor node is SparkSession.builder().getOrCreate().version. But, when I ran tests i received the following error:
Caused by: java.lang.IllegalStateException: SparkSession should only be created and accessed on the driver.
Do you know any other way to find Spark version in an executor?
> Write Spark version information to Parquet file footers
> -------------------------------------------------------
>
> Key: SPARK-25102
> URL: https://issues.apache.org/jira/browse/SPARK-25102
> Project: Spark
> Issue Type: Improvement
> Components: Spark Core
> Affects Versions: 2.3.1
> Reporter: Zoltan Ivanfi
> Priority: Major
>
> -PARQUET-352- added support for the "writer.model.name" property in the Parquet metadata to identify the object model (application) that wrote the file.
> The easiest way to write this property is by overriding getName() of org.apache.parquet.hadoop.api.WriteSupport. In Spark, this would mean adding getName() to the org.apache.spark.sql.execution.datasources.parquet.ParquetWriteSupport class.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org