You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Amr Awadallah (JIRA)" <ji...@apache.org> on 2009/05/22 00:55:45 UTC
[jira] Commented: (HADOOP-5887) Sqoop should create tables in Hive
metastore after importing to HDFS
[ https://issues.apache.org/jira/browse/HADOOP-5887?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12711851#action_12711851 ]
Amr Awadallah commented on HADOOP-5887:
---------------------------------------
alternatively you can also just point to the original HDFS files as an external table and avoid the copy.
e.g.
CREATE EXTERNAL TABLE mytable( schema )
ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' LINES TERMINATED BY '\n'
STORED AS TEXTFILE
LOCATION '<hdfs_path>';
-- amr
> Sqoop should create tables in Hive metastore after importing to HDFS
> --------------------------------------------------------------------
>
> Key: HADOOP-5887
> URL: https://issues.apache.org/jira/browse/HADOOP-5887
> Project: Hadoop Core
> Issue Type: New Feature
> Reporter: Aaron Kimball
> Assignee: Aaron Kimball
> Attachments: HADOOP-5887.patch
>
>
> Sqoop (HADOOP-5815) imports tables into HDFS; it is a straightforward enhancement to then generate a Hive DDL statement to recreate the table definition in the Hive metastore and move the imported table into the Hive warehouse directory from its upload target.
> This feature enhancement makes this process automatic. An import is performed with sqoop in the usual way; providing the argument "--hive-import" will cause it to then issue a CREATE TABLE .. LOAD DATA INTO statement to a Hive shell. It generates a script file and then attempts to run "$HIVE_HOME/bin/hive" on it, or failing that, any "hive" on the $PATH; $HIVE_HOME can be overridden with --hive-home. As a result, no direct linking against Hive is necessary.
> The unit tests provided with this enhancement use a mock implementation of 'bin/hive' that compares the script it's fed with one from a directory full of "expected" scripts. The exact script file referenced is controlled via an environment variable. It doesn't actually load into a proper Hive metastore, but manual testing has shown that this process works in practice, so the mock implementation is a reasonable unit testing tool.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.