You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Yun Tang (Jira)" <ji...@apache.org> on 2020/10/03 10:03:00 UTC

[jira] [Commented] (FLINK-19481) Add support for a flink native GCS FileSystem

    [ https://issues.apache.org/jira/browse/FLINK-19481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17206666#comment-17206666 ] 

Yun Tang commented on FLINK-19481:
----------------------------------

[~baugarten], from my point of view, we could still checkpoint on Google cloud storage via [hadoop file system|https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/filesystems/#hadoop-file-system-hdfs-and-its-other-implementations] without this improvement, is that right?

> Add support for a flink native GCS FileSystem
> ---------------------------------------------
>
>                 Key: FLINK-19481
>                 URL: https://issues.apache.org/jira/browse/FLINK-19481
>             Project: Flink
>          Issue Type: Improvement
>          Components: Connectors / FileSystem, FileSystems
>    Affects Versions: 1.12.0
>            Reporter: Ben Augarten
>            Priority: Major
>             Fix For: 1.12.0
>
>
> Currently, GCS is supported but only by using the hadoop connector[1]
>  
> The objective of this improvement is to add support for checkpointing to Google Cloud Storage with the Flink File System,
>  
> This would allow the `gs://` scheme to be used for savepointing and checkpointing. Long term, it would be nice if we could use the GCS FileSystem as a source and sink in flink jobs as well. 
>  
> Long term, I hope that implementing a flink native GCS FileSystem will simplify usage of GCS because the hadoop FileSystem ends up bringing in many unshaded dependencies.
>  
> [1] [https://github.com/GoogleCloudDataproc/hadoop-connectors|https://github.com/GoogleCloudDataproc/hadoop-connectors)]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)