You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@sqoop.apache.org by "Jarek Jarcec Cecho (Updated) (JIRA)" <ji...@apache.org> on 2012/02/18 22:57:59 UTC

[jira] [Updated] (SQOOP-443) Calling sqoop with hive import is not working multiple times due to kept output directory

     [ https://issues.apache.org/jira/browse/SQOOP-443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Jarek Jarcec Cecho updated SQOOP-443:
-------------------------------------

    Status: Patch Available  (was: Open)
    
> Calling sqoop with hive import is not working multiple times due to  kept output directory
> ------------------------------------------------------------------------------------------
>
>                 Key: SQOOP-443
>                 URL: https://issues.apache.org/jira/browse/SQOOP-443
>             Project: Sqoop
>          Issue Type: Improvement
>    Affects Versions: 1.4.0-incubating, 1.4.1-incubating
>            Reporter: Jarek Jarcec Cecho
>            Assignee: Jarek Jarcec Cecho
>            Priority: Minor
>         Attachments: SQOOP-443.patch
>
>
> Hive is not removing input directory when doing "LOAD DATA" command in all cases. This input directory is actually sqoop's export directory. Because this directory is kept, calling same sqoop command twice is failing on exception "org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory $table already exists".
> This issue might be easily overcome by manual directory removal, however it's putting unnecessary burden on users. It's also complicating executing saved jobs as there is additional script execution needed.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira