You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Arun Patel <ar...@gmail.com> on 2015/08/29 16:31:55 UTC

HBase Import Error

I was able to successfully export data to S3 using below command.

hbase org.apache.hadoop.hbase.mapreduce.Export docs
s3n://KEY:ACCESSKEY@fdocshbase/data/bkp1 1440760612 1440848237

and I was able to import data to a new table (after creation) with command

hbase org.apache.hadoop.hbase.mapreduce.Import docsnew
s3n://KEY:ACCESSKEY@fdocshbase/data/bkp1

I exported with different time ranges to other directories like bkp2 bkp3.

When I was trying to import all directories into Hbase, getting a
filenotfound exception.

[hdfs@ip-172-31-59-10 ~]$ hbase org.apache.hadoop.hbase.mapreduce.Import
docsnew s3n://KEY:ACCESSKEY@fdocshbase

SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/usr/hdp/2.3.0.0-2557/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/hdp/2.3.0.0-2557/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
2015-08-29 09:21:02,058 INFO  [main] impl.TimelineClientImpl: Timeline
service address: http://ip-XXXXXXXXX.ec2.internal:8188/ws/v1/timeline/
2015-08-29 09:21:02,214 INFO  [main] client.RMProxy: Connecting to
ResourceManager at ip-XXXXXXXXX.ec2.internal/XXXXXXXXX:8050
2015-08-29 09:21:03,961 INFO  [main] input.FileInputFormat: Total input
paths to process : 1
2015-08-29 09:21:04,021 INFO  [main] mapreduce.JobSubmitter: Cleaning up
the staging area /user/hdfs/.staging/job_1440846401761_0026
Exception in thread "main" java.io.FileNotFoundException: No such file or
directory 's3n://XXXXXXX:YYYYYYYYYYY@fdocshbase/data/data'
        at
org.apache.hadoop.fs.s3native.NativeS3FileSystem.getFileStatus(NativeS3FileSystem.java:507)
        at
org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat.listStatus(SequenceFileInputFormat.java:67)
        at
org.apache.hadoop.mapreduce.lib.input.FileInputFormat.getSplits(FileInputFormat.java:387)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:301)
        at
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:318)
        at
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at org.apache.hadoop.hbase.mapreduce.Import.main(Import.java:544)


Its looking for data directory always?   I did export data to HDFS as
well.  Had a similar issue in HDFS as well.  I created bakcup directory
like /fdocshbase/bkp1, /fdocshbase/bkp2 and so on....
This created similar problem above while importing.  But, When I created
directories like /fdocshbase/data/bkp1, /fdocshbase/data/bkp2......and
provide /fdocshbase/data/ as input to import, it works.

But, this is not working on S3. I tried creating directory like
fdocshbase/data/data
 and placed all bkp* directories in it.

Any thoughts?  Any quick resolutions?

Regards,
Arun