You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Joydeep Sen Sarma (JIRA)" <ji...@apache.org> on 2009/05/15 06:07:45 UTC

[jira] Updated: (HADOOP-5839) fixes to ec2 scripts to allow remote job submission

     [ https://issues.apache.org/jira/browse/HADOOP-5839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Joydeep Sen Sarma updated HADOOP-5839:
--------------------------------------

    Attachment: 5839.1.patch

> fixes to ec2 scripts to allow remote job submission
> ---------------------------------------------------
>
>                 Key: HADOOP-5839
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5839
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: contrib/ec2
>            Reporter: Joydeep Sen Sarma
>         Attachments: 5839.1.patch
>
>
> i would very much like the option of submitting jobs from a workstation outside ec2 to a hadoop cluster in ec2. This has been explored here:
> http://www.nabble.com/public-IP-for-datanode-on-EC2-tt19336240.html
> the net result of this is that we can make this work (along with using a socks proxy) with a couple of changes in the ec2 scripts:
> a) use public 'hostname' for fs.default.name setting (instead of the private hostname being used currently)
> b) mark hadoop.rpc.socket.factory.class.default as final variable in the generated hadoop-site.xml (that applies to server side)
> #a has no downside as far as i can tell since public hostnames resolve to internal/private IP addresses within ec2 (so traffic is optimally routed).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.