You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@sqoop.apache.org by "Markus Kemper (JIRA)" <ji...@apache.org> on 2016/10/26 13:41:58 UTC
[jira] [Commented] (SQOOP-3038) Sqoop export using --hcatalog with
RDBMS reserved word column name results in "null" value
[ https://issues.apache.org/jira/browse/SQOOP-3038?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608476#comment-15608476 ]
Markus Kemper commented on SQOOP-3038:
--------------------------------------
Workaround:
1. Use (export --export-dir) instead of (export --hcatalog)
2. Use non-reserved word as column name
> Sqoop export using --hcatalog with RDBMS reserved word column name results in "null" value
> ------------------------------------------------------------------------------------------
>
> Key: SQOOP-3038
> URL: https://issues.apache.org/jira/browse/SQOOP-3038
> Project: Sqoop
> Issue Type: Bug
> Components: hive-integration
> Reporter: Markus Kemper
>
> -----------------------
> | c1 | value | <=== RDBMS reserved word "value"
> -----------------------
> | 1 | (null) | <=== null?
> -----------------------
> TEST CASE:
> STEP 01 - Create MySQL Tables
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "drop table t1"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "create table t1 (c1 int, c2 varchar(5))"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "insert into t1 values (1, 'one')"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "show create table t1"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "select * from t1"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "drop table t1_value"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "create table t1_value (c1 int, \`value\` varchar(5))"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "insert into t1_value select * from t1"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "show create table t1_value"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "select * from t1_value"
> Output:
> -----------------------------------------------
> | Table | Create Table |
> -----------------------------------------------
> | t1 | CREATE TABLE `t1` (
> `c1` int(11) DEFAULT NULL,
> `c2` varchar(5) DEFAULT NULL
> ) ENGINE=InnoDB DEFAULT CHARSET=utf8 |
> -----------------------------------------------
> -----------------------
> | c1 | c2 |
> -----------------------
> | 1 | one |
> -----------------------
> -----------------------------------------------
> | Table | Create Table |
> -----------------------------------------------
> | t1_value | CREATE TABLE `t1_value` (
> `c1` int(11) DEFAULT NULL,
> `value` varchar(5) DEFAULT NULL
> ) ENGINE=InnoDB DEFAULT CHARSET=utf8 |
> -----------------------------------------------
> -----------------------
> | c1 | value |
> -----------------------
> | 1 | one |
> -----------------------
> STEP 02 - Verify Import/Export Using --target-dir and --export-dir
> sqoop import --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1 --target-dir /user/root/t1 --delete-target-dir --num-mappers 1
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "delete from t1"
> sqoop export --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1 --export-dir /user/root/t1 --num-mappers 1
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "select * from t1"
> sqoop import --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1_value --target-dir /user/root/t1 --delete-target-dir --num-mappers 1
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "delete from t1_value"
> sqoop export --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1_value --export-dir /user/root/t1 --num-mappers 1
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "select * from t1_value"
> Output:
> -----------------------
> | c1 | c2 |
> -----------------------
> | 1 | one |
> -----------------------
> -----------------------
> | c1 | value |
> -----------------------
> | 1 | one |
> -----------------------
> STEP 03 - Verify import with --hive-import and export with --hcatalog
> beeline -u jdbc:hive2:// -e "use default; drop table t1;"
> sqoop import --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1 --target-dir /user/root/t1 --delete-target-dir --hive-import --hive-database default --hive-table t1 --num-mappers 1
> beeline -u jdbc:hive2:// -e "use default; select * from t1;"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "delete from t1"
> sqoop export --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1 --hcatalog-database default --hcatalog-table t1 --num-mappers 1
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "select * from t1"
> Output:
> +--------+--------+--+
> | t1.c1 | t1.c2 |
> +--------+--------+--+
> | 1 | one |
> +--------+--------+--+
> -----------------------
> | c1 | c2 |
> -----------------------
> | 1 | one |
> -----------------------
> beeline -u jdbc:hive2:// -e "use default; drop table t1_value;"
> sqoop import --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1_value --target-dir /user/root/t1 --delete-target-dir --hive-import --hive-database default --hive-table t1_value --num-mappers 1
> beeline -u jdbc:hive2:// -e "use default; select * from t1_value;"
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "delete from t1_value"
> sqoop export --connect $MYCONN --username $MYUSER --password $MYPSWD --table t1_value --hcatalog-database default --hcatalog-table t1_value --num-mappers 1
> sqoop eval --connect $MYCONN --username $MYUSER --password $MYPSWD --query "select * from t1_value"
> Output:
> +--------------+-----------------+--+
> | t1_value.c1 | t1_value.value |
> +--------------+-----------------+--+
> | 1 | one |
> +--------------+-----------------+--+
> -----------------------
> | c1 | value |
> -----------------------
> | 1 | (null) | <========== null?
> -----------------------
> beeline -u jdbc:hive2:// -e "use default; show create table t1;"
> beeline -u jdbc:hive2:// -e "use default; show create table t1_value;"
> Output:
> +-------------------------------------------------------------------------------+--+
> | createtab_stmt |
> +-------------------------------------------------------------------------------+--+
> | CREATE TABLE `t1`( |
> | `c1` int, |
> | `c2` string) |
> | COMMENT 'Imported by sqoop on 2016/10/25 17:16:20' |
> | ROW FORMAT SERDE |
> | 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' |
> | WITH SERDEPROPERTIES ( |
> | 'field.delim'='\u0001', |
> | 'line.delim'='\n', |
> | 'serialization.format'='\u0001') |
> | STORED AS INPUTFORMAT |
> | 'org.apache.hadoop.mapred.TextInputFormat' |
> | OUTPUTFORMAT |
> | 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
> | LOCATION |
> | 'hdfs://<hostname>:8020/user/hive/warehouse/t1' |
> | TBLPROPERTIES ( |
> | 'COLUMN_STATS_ACCURATE'='true', |
> | 'numFiles'='1', |
> | 'totalSize'='6', |
> | 'transient_lastDdlTime'='1477440983') |
> +-------------------------------------------------------------------------------+--+
> +-------------------------------------------------------------------------------------+--+
> | createtab_stmt |
> +-------------------------------------------------------------------------------------+--+
> | CREATE TABLE `t1_value`( |
> | `c1` int, |
> | `value` string) |
> | COMMENT 'Imported by sqoop on 2016/10/25 17:23:02' |
> | ROW FORMAT SERDE |
> | 'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' |
> | WITH SERDEPROPERTIES ( |
> | 'field.delim'='\u0001', |
> | 'line.delim'='\n', |
> | 'serialization.format'='\u0001') |
> | STORED AS INPUTFORMAT |
> | 'org.apache.hadoop.mapred.TextInputFormat' |
> | OUTPUTFORMAT |
> | 'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' |
> | LOCATION |
> | 'hdfs://<hostname>:8020/user/hive/warehouse/t1_value' |
> | TBLPROPERTIES ( |
> | 'COLUMN_STATS_ACCURATE'='true', |
> | 'numFiles'='1', |
> | 'totalSize'='6', |
> | 'transient_lastDdlTime'='1477441386') |
> +-------------------------------------------------------------------------------------+--+
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)