You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@pig.apache.org by "Olga Natkovich (JIRA)" <ji...@apache.org> on 2010/07/15 01:44:52 UTC

[jira] Assigned: (PIG-103) Shared Job /tmp location should be configurable

     [ https://issues.apache.org/jira/browse/PIG-103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Olga Natkovich reassigned PIG-103:
----------------------------------

    Assignee: niraj rai

> Shared Job /tmp location should be configurable
> -----------------------------------------------
>
>                 Key: PIG-103
>                 URL: https://issues.apache.org/jira/browse/PIG-103
>             Project: Pig
>          Issue Type: Improvement
>          Components: impl
>         Environment: Partially shared file:// filesystem (eg NFS)
>            Reporter: Craig Macdonald
>            Assignee: niraj rai
>             Fix For: 0.8.0
>
>
> Hello,
> I'm investigating running pig in an environment where various parts of the file:// filesystem are available on all nodes. I can tell hadoop to use a file:// file system location for it's default, by seting fs.default.name=file://path/to/shared/folder
> However, this creates issues for Pig, as Pig writes it's job information in a folder that it assumes is a shared FS (eg DFS). However, in this scenario /tmp is not shared on each machine.
> So /tmp should either be configurable, or Hadoop should tell you the actual full location set in fs.default.name?
> Straightforward solution is to make "/tmp/" a property in src/org/apache/pig/impl/io/FileLocalizer.java init(PigContext)
> Any suggestions of property names?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.