You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by "sivabalan narayanan (Jira)" <ji...@apache.org> on 2021/01/26 19:57:00 UTC
[jira] [Updated] (HUDI-914) support different target data clusters
[ https://issues.apache.org/jira/browse/HUDI-914?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
sivabalan narayanan updated HUDI-914:
-------------------------------------
Labels: user-support-issues (was: )
> support different target data clusters
> --------------------------------------
>
> Key: HUDI-914
> URL: https://issues.apache.org/jira/browse/HUDI-914
> Project: Apache Hudi
> Issue Type: New Feature
> Components: DeltaStreamer
> Reporter: liujinhui
> Assignee: liujinhui
> Priority: Major
> Labels: user-support-issues
> Original Estimate: 168h
> Remaining Estimate: 168h
>
> Currently hudi-DeltaStreamer does not support writing to different target clusters. The specific scenarios are as follows: Generally, Hudi tasks run on an independent cluster. If you want to write data to the target data cluster, you generally rely on core-site.xml and hdfs-site.xml; sometimes you will encounter different targets. The data cluster writes data, but the cluster running the hudi task does not have the core-site.xml and hdfs-site.xml of the target cluster. Although specifying the namenode IP address of the target cluster can be written, this loses HDFS high availability, so I plan to Use the contents of the core-site.xml and hdfs-site.xml files of the target cluster as configuration items and configure them in the dfs-source.properties or kafka-source.properties file of Hudi.
> Is there a better way to solve this problem?
--
This message was sent by Atlassian Jira
(v8.3.4#803005)