You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@falcon.apache.org by "Jean-Baptiste Onofré (JIRA)" <ji...@apache.org> on 2014/09/23 13:46:33 UTC
[jira] [Commented] (FALCON-349) Support "MapReduce" workflow in
process entity
[ https://issues.apache.org/jira/browse/FALCON-349?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14144698#comment-14144698 ]
Jean-Baptiste Onofré commented on FALCON-349:
---------------------------------------------
What I have in mind as a first step: add some specific attributes/elements in the process entity XML (like mapper/reducer class, the location of the job jar). Falcon can create the oozie workflow.xml for the user using these metadata.
> Support "MapReduce" workflow in process entity
> ----------------------------------------------
>
> Key: FALCON-349
> URL: https://issues.apache.org/jira/browse/FALCON-349
> Project: Falcon
> Issue Type: Wish
> Components: process
> Reporter: Jean-Baptiste Onofré
> Assignee: Jean-Baptiste Onofré
>
> Currently, a process entity supports the following workflow:
> - oozie
> - pig
> - hive
> If an user has a "pure" MapReduce job, the only way to use it in Falcon is via the oozie workflow. It means he has to create the workflow.xml describing the nodes (mapred.mapper.class, etc elements).
> So, it may look like a overhead for the user who just wants to "schedule" the job/process.
> I would propose to create a "mapreduce" workflow, taken directly a MapReduce jar from the filesystem, and behind the hood, create a simple workflow.xml and schedule it in oozie.
> Thoughts ?
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)