You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "ouyangwulin (Jira)" <ji...@apache.org> on 2020/01/02 03:26:00 UTC

[jira] [Comment Edited] (FLINK-15378) StreamFileSystemSink supported mutil hdfs plugins.

    [ https://issues.apache.org/jira/browse/FLINK-15378?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17006557#comment-17006557 ] 

ouyangwulin edited comment on FLINK-15378 at 1/2/20 3:25 AM:
-------------------------------------------------------------

{code:java}
 do I understand your problem correctly, that you are trying to use the same plugin, but with different configs?{code}
   yes , but the same plugin how to use different configs when it only create one FilesystemFactory.
{code:java}
Can not you create a separate plugin but just with a different schema, instead of adding different identity? {code}
   I want to sink to different hdfs cluster. Schema is native same in different cluster. So I want add a identify to get the different.
{code:java}
where are the "conf A", "conf B"  and hdfs-site.xml files located? Are they bundled inside the plugin's fat jar? {code}
yes. It can from the fat jar. or From flink config


was (Author: ouyangwuli):
{code:java}
 do I understand your problem correctly, that you are trying to use the same plugin, but with different configs?{code}
   yes , but the same plugin how to use different configs when it only create one FilesystemFactory.
{code:java}
Can not you create a separate plugin but just with a different schema, instead of adding different identity? {code}
   I want to sink to different hdfs cluster. Schema is native same in different cluster. So I want add a identify to get the different.
{code:java}
where are the "conf A", "conf B"  and hdfs-site.xml files located? Are they bundled inside the plugin's fat jar? {code}
yes.

> StreamFileSystemSink supported mutil hdfs plugins.
> --------------------------------------------------
>
>                 Key: FLINK-15378
>                 URL: https://issues.apache.org/jira/browse/FLINK-15378
>             Project: Flink
>          Issue Type: Improvement
>          Components: Connectors / FileSystem, FileSystems
>    Affects Versions: 1.9.2, 1.10.0
>            Reporter: ouyangwulin
>            Priority: Major
>              Labels: pull-request-available
>             Fix For: 1.11.0
>
>         Attachments: jobmananger.log
>
>          Time Spent: 10m
>  Remaining Estimate: 0h
>
> [As report from maillist|[https://lists.apache.org/thread.html/7a6b1e341bde0ef632a82f8d46c9c93da358244b6bac0d8d544d11cb%40%3Cuser.flink.apache.org%3E]]
> Request 1:  FileSystem plugins not effect the default yarn dependecies.
> Request 2:  StreamFileSystemSink supported mutil hdfs plugins under the same schema
> As Problem describe :
>     when I put a ' filesystem plugin to FLINK_HOME/pulgins in flink', and the clas{color:#172b4d}s '*com.filesystem.plugin.FileSystemFactoryEnhance*' implements '*FileSystemFactory*', when jm start, It will call FileSystem.initialize(configuration, PluginUtils.createPluginManagerFromRootFolder(configuration)) to load factories to map  FileSystem#**{color}FS_FACTORIES, and the key is only schema. When tm/jm use local hadoop conf A ,   the user code use hadoop conf Bin 'filesystem plugin',  Conf A and Conf B is used to different hadoop cluster. and The Jm will start failed, beacuse of the blodserver in JM will load Conf B to get filesystem. the full log add appendix.
>  
> AS reslove method:
>     use  schema and spec identify as key for ' FileSystem#**FS_FACTORIES '
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)