You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@sqoop.apache.org by "Sqoop QA bot (JIRA)" <ji...@apache.org> on 2015/03/17 03:55:40 UTC

[jira] [Commented] (SQOOP-2201) Sqoop2: Add possibility to read Hadoop configuration files to HFDS connector

    [ https://issues.apache.org/jira/browse/SQOOP-2201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14364470#comment-14364470 ] 

Sqoop QA bot commented on SQOOP-2201:
-------------------------------------

Testing file [SQOOP-2201.patch|https://issues.apache.org/jira/secure/attachment/12704976/SQOOP-2201.patch] against branch sqoop2 took 0:44:50.548976.

{color:red}Overall:{color} -1 due to 5 errors

{color:red}ERROR:{color} Some of fast integration tests failed ([report|https://builds.apache.org/job/PreCommit-SQOOP-Build/1168/artifact/patch-process/test_integration_fast.txt], executed 23 tests)
{color:red}ERROR:{color} Failed fast integration test: {{org.apache.sqoop.integration.repository.derby.upgrade.Derby1_99_4UpgradeTest}}
{color:red}ERROR:{color} Failed fast integration test: {{org.apache.sqoop.integration.server.SubmissionWithDisabledModelObjectsTest}}
{color:red}ERROR:{color} Failed fast integration test: {{org.apache.sqoop.integration.repository.derby.upgrade.Derby1_99_3UpgradeTest}}
{color:red}ERROR:{color} Failed fast integration test: {{org.apache.sqoop.integration.connector.jdbc.generic.IncrementalReadTest}}
{color:green}SUCCESS:{color} Clean was successful
{color:green}SUCCESS:{color} Patch applied correctly
{color:green}SUCCESS:{color} Patch add/modify test case
{color:green}SUCCESS:{color} License check passed
{color:green}SUCCESS:{color} Patch compiled
{color:green}SUCCESS:{color} All fast unit tests passed (executed 272 tests)
{color:green}SUCCESS:{color} All slow unit tests passed (executed 1 tests)
{color:green}SUCCESS:{color} All slow integration tests passed (executed 0 tests)

Console output is available [here|https://builds.apache.org/job/PreCommit-SQOOP-Build/1168/console].

This message is automatically generated.

> Sqoop2: Add possibility to read Hadoop configuration files to HFDS connector
> ----------------------------------------------------------------------------
>
>                 Key: SQOOP-2201
>                 URL: https://issues.apache.org/jira/browse/SQOOP-2201
>             Project: Sqoop
>          Issue Type: Bug
>    Affects Versions: 1.99.5
>            Reporter: Jarek Jarcec Cecho
>            Assignee: Jarek Jarcec Cecho
>             Fix For: 1.99.6
>
>         Attachments: SQOOP-2201.patch
>
>
> Currently the HDFS connector is not explicitly reading Hadoop configuration files. During [Initialization|https://github.com/apache/sqoop/blob/sqoop2/connector/connector-hdfs/src/main/java/org/apache/sqoop/connector/hdfs/HdfsToInitializer.java] phase it doesn't do anything, so the configuration files are not needed. During other parts of the workflow, we're [explicitly casting|https://github.com/apache/sqoop/blob/sqoop2/connector/connector-hdfs/src/main/java/org/apache/sqoop/connector/hdfs/HdfsExtractor.java#L61] the general {{Context}} object to Hadoop {{Configuration}}.
> This is unfortunate because:
> * It couples HDFS connector to Mapreduce execution engine. It will break with adding non mapreduce based execution engine.
> * We can't do any HDFS specific checks in {{Initializer}} as the Hadoop {{Configuration}} object is not available there.
> As a result I would like to propose breaking this coupling between HDFS connector and Mapreduce execution engine and add configuration option to HDFS Link to specify directory from which we should read the appropriate Hadoop configuration files (with reasonable defaults such as {{/etc/conf/hadoop}}).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)