You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@phoenix.apache.org by "ASF GitHub Bot (JIRA)" <ji...@apache.org> on 2015/09/09 12:13:46 UTC

[jira] [Commented] (PHOENIX-2238) Bulk dataload Hive data with \001 delimiter

    [ https://issues.apache.org/jira/browse/PHOENIX-2238?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14736607#comment-14736607 ] 

ASF GitHub Bot commented on PHOENIX-2238:
-----------------------------------------

GitHub user 979969786 opened a pull request:

    https://github.com/apache/phoenix/pull/116

    PHOENIX-2238 Support more delimiter when using CsvBulkImport

    Add a function to support  more delimiter when using CsvBulkImport, such as /001, /t, /n.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/979969786/phoenix master

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/phoenix/pull/116.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #116
    
----
commit 53992542040bde33a8f3f93560adaac007b353c6
Author: yuming.wang <q7...@gmail.com>
Date:   2015-09-09T10:05:45Z

    support  more delimiter when using CsvBulkImport
    
    Add a function to support  more delimiter when using CsvBulkImport, such as /001, /t.

----


> Bulk dataload Hive data with \001 delimiter
> -------------------------------------------
>
>                 Key: PHOENIX-2238
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-2238
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.3.0
>            Reporter: q79969786
>
> Bulk dataload :
> hadoop jar /opt/cloudera/parcels/CLABS_PHOENIX-4.3.0-1.clabs_phoenix1.0.0.p0.78/lib/phoenix/phoenix-4.3.0-clabs-phoenix-1.0.0-client.jar org.apache.phoenix.mapreduce.CsvBulkLoadTool --table test_phoenix_import  --zookeeper n1,n2,n3 --delimiter \001 --input  /user/wym/test_phoenix_import
> The path ‘/user/wym/test_phoenix_import’ is a directory contains Hive data, so the default delimiter is \001.
> Exception:
> 15/09/08 15:22:17 INFO zookeeper.ClientCnxn: EventThread shut down
> 15/09/08 15:22:17 INFO zookeeper.ZooKeeper: Session: 0x34f1c9be8ba5667 closed
> Exception in thread "main" java.lang.IllegalArgumentException: Illegal delimiter character: 001
>         at org.apache.phoenix.mapreduce.CsvBulkLoadTool.configureOptions(CsvBulkLoadTool.java:327)
>         at org.apache.phoenix.mapreduce.CsvBulkLoadTool.loadData(CsvBulkLoadTool.java:201)
>         at org.apache.phoenix.mapreduce.CsvBulkLoadTool.run(CsvBulkLoadTool.java:186)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>         at org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoadTool.java:97)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>         at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> 15/09/08 15:22:18 INFO client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x34f1c9be8ba5665
> 15/09/08 15:22:18 INFO zookeeper.ClientCnxn: EventThread shut down
> 15/09/08 15:22:18 INFO zookeeper.ZooKeeper: Session: 0x34f1c9be8ba5665 closed



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)