You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Ryan LeCompte <le...@gmail.com> on 2008/09/22 17:08:31 UTC
NotYetReplicated exceptions when pushing large files into HDFS
Hello all,
I'd love to be able to upload into HDFS very large files (e.g., 8 or
10GB), but it seems like my only option is to chop up the file into
smaller pieces. Otherwise, after a while I get NotYetReplication
exceptions while the transfer is in progress. I'm using 0.18.1. Is
there any way I can do this? Perhaps use something else besides
bin/hadoop -put input output?
Thanks,
Ryan
Re: NotYetReplicated exceptions when pushing large files into HDFS
Posted by Ryan LeCompte <le...@gmail.com>.
I've noticed that although I get a few of these exceptions, the file
is ultimately uploaded to the HDFS cluster. Does this mean that my
file ended up getting there in 1 piece? The exceptions are just logged
at the WARN level and indicate retry attempts.
Thanks,
Ryan
On Mon, Sep 22, 2008 at 11:08 AM, Ryan LeCompte <le...@gmail.com> wrote:
> Hello all,
>
> I'd love to be able to upload into HDFS very large files (e.g., 8 or
> 10GB), but it seems like my only option is to chop up the file into
> smaller pieces. Otherwise, after a while I get NotYetReplication
> exceptions while the transfer is in progress. I'm using 0.18.1. Is
> there any way I can do this? Perhaps use something else besides
> bin/hadoop -put input output?
>
> Thanks,
> Ryan
>