You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hyukjin Kwon (Jira)" <ji...@apache.org> on 2020/02/25 07:56:00 UTC

[jira] [Commented] (SPARK-30833) Extract o.a.s.sql.catalyst.util.ParseMode to external project

    [ https://issues.apache.org/jira/browse/SPARK-30833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17044201#comment-17044201 ] 

Hyukjin Kwon commented on SPARK-30833:
--------------------------------------

It just needs one line change in User's application as you pointed out. If this is exposed as APIs, we should think about how to expose it in PySpark and SparkR sides too. I think it's not worthwhile for now.

> Extract o.a.s.sql.catalyst.util.ParseMode to external project
> -------------------------------------------------------------
>
>                 Key: SPARK-30833
>                 URL: https://issues.apache.org/jira/browse/SPARK-30833
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL
>    Affects Versions: 2.4.5
>            Reporter: Michael Heuer
>            Priority: Minor
>
> In multiple projects downstream of Spark, we use an enum
> {code:java}
> enum ValidationStringency { STRICT, LENIENT, SILENT }
> {code}
> This is similar to but not the same as [ParseMode|https://github.com/apache/spark/blob/master/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/ParseMode.scala] in the SQL Catalyst utility package
> {code:java}
> enum ParseMode { FAILFAST, DROPMALFORMED, PERMISSIVE }
> {code}
> I wonder if there might be interest in extracting this enum to a project external to Spark, along with some related utility code, and refactoring Spark to use that external reference?
> It would also be more convenient to use a Java enum instead of Scala one, if that does not cause trouble for Spark.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org