You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Yun Tang (Jira)" <ji...@apache.org> on 2020/10/03 10:03:00 UTC
[jira] [Commented] (FLINK-19481) Add support for a flink native GCS
FileSystem
[ https://issues.apache.org/jira/browse/FLINK-19481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17206666#comment-17206666 ]
Yun Tang commented on FLINK-19481:
----------------------------------
[~baugarten], from my point of view, we could still checkpoint on Google cloud storage via [hadoop file system|https://ci.apache.org/projects/flink/flink-docs-release-1.11/ops/filesystems/#hadoop-file-system-hdfs-and-its-other-implementations] without this improvement, is that right?
> Add support for a flink native GCS FileSystem
> ---------------------------------------------
>
> Key: FLINK-19481
> URL: https://issues.apache.org/jira/browse/FLINK-19481
> Project: Flink
> Issue Type: Improvement
> Components: Connectors / FileSystem, FileSystems
> Affects Versions: 1.12.0
> Reporter: Ben Augarten
> Priority: Major
> Fix For: 1.12.0
>
>
> Currently, GCS is supported but only by using the hadoop connector[1]
>
> The objective of this improvement is to add support for checkpointing to Google Cloud Storage with the Flink File System,
>
> This would allow the `gs://` scheme to be used for savepointing and checkpointing. Long term, it would be nice if we could use the GCS FileSystem as a source and sink in flink jobs as well.
>
> Long term, I hope that implementing a flink native GCS FileSystem will simplify usage of GCS because the hadoop FileSystem ends up bringing in many unshaded dependencies.
>
> [1] [https://github.com/GoogleCloudDataproc/hadoop-connectors|https://github.com/GoogleCloudDataproc/hadoop-connectors)]
--
This message was sent by Atlassian Jira
(v8.3.4#803005)