You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "melin (Jira)" <ji...@apache.org> on 2022/08/05 13:09:00 UTC

[jira] [Created] (SPARK-39990) Restrict special characters in field name, which can be controlled by switches

melin created SPARK-39990:
-----------------------------

             Summary:  Restrict special characters in field name, which can be controlled by switches
                 Key: SPARK-39990
                 URL: https://issues.apache.org/jira/browse/SPARK-39990
             Project: Spark
          Issue Type: Improvement
          Components: SQL
    Affects Versions: 3.4.0
            Reporter: melin


The hive metastore restricts field name to only contain alphanumerics and underscores. If the custom catalog does not use hms, these restrictions may not exist, such as reading excel data, writing paruqet, and column names are prone to special characters such as spaces, parentheses, etc

hack way forbidden:
{code:java}
@Around("execution(public * org.apache.spark.sql.execution.datasources.DataSourceUtils.checkFieldNames(..))")
public void checkFieldNames_1(ProceedingJoinPoint pjp) throws Throwable {
LOG.info("skip checkFieldNames 1");
}

@Around("execution(public * org.apache.spark.sql.execution.datasources.parquet.ParquetSchemaConverter$.checkFieldNames(..))")
public void checkFieldNames_2(ProceedingJoinPoint pjp) throws Throwable {
LOG.info("skip checkFieldNames 2");
}{code}
CREATE OR REPLACE TABLE huaixin_rp.bigdata.parquet_orders_rp5 USING PARQUET  select 12 as id, 'ceity' as `address(地  址)`



@Around("execution(public * org.apache.spark.sql.execution.datasources.parquet.ParquetSchemaConverter$.checkFieldName(..))")
public void checkFieldNames_3(ProceedingJoinPoint pjp) throws Throwable {
LOG.info("skip checkFieldNames 3");
}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org