You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Steve Loughran (Jira)" <ji...@apache.org> on 2021/10/08 12:31:00 UTC

[jira] [Commented] (SPARK-36766) Spark SQL DDL does not recognize fs.s3.impl implied filesystem in LOCATION tag

    [ https://issues.apache.org/jira/browse/SPARK-36766?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17426152#comment-17426152 ] 

Steve Loughran commented on SPARK-36766:
----------------------------------------

I can see why you'd want to do this (consistent URLs on EMR and ASF spark builds), but be aware that there's a risk the s3a connector might assume its got an s3a URL, and sometimes it'll leak.

> Spark SQL DDL does not recognize fs.s3.impl implied filesystem in LOCATION tag
> ------------------------------------------------------------------------------
>
>                 Key: SPARK-36766
>                 URL: https://issues.apache.org/jira/browse/SPARK-36766
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 3.0.2
>            Reporter: Sungwon
>            Priority: Major
>
> CREATE TABLE with the LOCATION tag via SPARK SQL DDL does not seem to recognize the fileSystem implied by setting fs.s3.impl.
> The environment setting is set to:
> |spark.hadoop.fs.s3.impl|org.apache.hadoop.fs.s3a.S3AFileSystem|
> The following spark.sql ddl raises an exception:
>  spark.sql("""CREATE TABLE local.db.table (
>  col1 string,
>  col2 string,
>  )
>  LOCATION 's3://bucket/prefix'
>  """)
>  Exception: 
> Py4JJavaError: An error occurred while calling o86.sql. : org.apache.iceberg.exceptions.CommitStateUnknownException: Got exception: org.apache.hadoop.fs.UnsupportedFileSystemException No FileSystem for scheme "s3"
> This is inconsistent with other spark functions such as spark.read.parquet("s3://bucket/prefix") which succeeds with the same configuration



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org