You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@mnemonic.apache.org by "Yanhui Zhao (Jira)" <ji...@apache.org> on 2021/08/08 05:02:00 UTC

[jira] [Commented] (MNEMONIC-511) Apache Flink integration

    [ https://issues.apache.org/jira/browse/MNEMONIC-511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17395460#comment-17395460 ] 

Yanhui Zhao commented on MNEMONIC-511:
--------------------------------------

One of the possible integration point is at https://github.com/apache/flink/blob/779070274cf00cb475cb7bab1332317ba39ac16e/flink-streaming-scala/src/main/scala/org/apache/flink/streaming/api/scala/StreamExecutionEnvironment.scala

def readTextFile(filePath: String): DataStream[String] = asScalaStream(javaEnv.readTextFile(filePath))

which traces to the JavaEnv at org.apache.flink.api.java.{CollectionEnvironment, ExecutionEnvironment => JavaEnv}, need to further dig the JavaEnv class to see the return type of readTextFile function call.

> Apache Flink integration
> ------------------------
>
>                 Key: MNEMONIC-511
>                 URL: https://issues.apache.org/jira/browse/MNEMONIC-511
>             Project: Mnemonic
>          Issue Type: New Feature
>    Affects Versions: 0.16.0
>            Reporter: Kevin Ratnasekera
>            Assignee: Yanhui Zhao
>            Priority: Major
>             Fix For: 0.16.0
>
>
> I noticed existing Hadoop and Spark integration to Mnemonic project. I would like to propose further integration Flink Engine to the project. We might be able to reuse existing Hadoop input/output formats or write these input formats from the scratch feed/sink data in/out for Flink jobs. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)