You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Rita <rm...@gmail.com> on 2011/07/12 12:06:55 UTC

is mapreduce needed?

Hello,


I have a dataset which is several terabytes in size. I would like to query
this data using hbase (sql). Would I need to setup mapreduce to use hbase?
Currently the data is stored in hdfs and I am using `hdfs -cat ` to get the
data and pipe it into stdin.



-- 
--- Get your facts first, then you can distort them as you please.--

Re: is mapreduce needed?

Posted by Doug Meil <do...@explorysmedical.com>.
Hi there-

I think you probably want to start with these links...

http://hadoop.apache.org/hdfs/
http://hadoop.apache.org/mapreduce/





On 7/12/11 8:27 AM, "Rita" <rm...@gmail.com> wrote:

>Thanks for the response.
>
>Can you give me some use cases of hbase and hdfs only? I am
>really hesitant to implement MR because for batch jobs we already use
>torque.
>
>
>
>On Tue, Jul 12, 2011 at 8:11 AM, Doug Meil
><do...@explorysmedical.com>wrote:
>
>>
>> Hi there-
>>
>> I think you might want to start with the Hbase book, and specifically
>>this
>> entry...
>>
>> http://hbase.apache.org/book.html#faq.sql
>>
>> ... and then this one...
>>
>> http://hbase.apache.org/book.html#datamodel
>>
>> .. then this one.
>>
>> http://hbase.apache.org/book.html#mapreduce
>>
>> MapReduce is not required with Hbase, but it is extremely useful.
>>
>>
>>
>> On 7/12/11 6:06 AM, "Rita" <rm...@gmail.com> wrote:
>>
>> >Hello,
>> >
>> >
>> >I have a dataset which is several terabytes in size. I would like to
>>query
>> >this data using hbase (sql). Would I need to setup mapreduce to use
>>hbase?
>> >Currently the data is stored in hdfs and I am using `hdfs -cat ` to get
>> >the
>> >data and pipe it into stdin.
>> >
>> >
>> >
>> >--
>> >--- Get your facts first, then you can distort them as you please.--
>>
>>
>
>
>-- 
>--- Get your facts first, then you can distort them as you please.--


Re: is mapreduce needed?

Posted by Rita <rm...@gmail.com>.
Thanks for the response.

Can you give me some use cases of hbase and hdfs only? I am
really hesitant to implement MR because for batch jobs we already use
torque.



On Tue, Jul 12, 2011 at 8:11 AM, Doug Meil <do...@explorysmedical.com>wrote:

>
> Hi there-
>
> I think you might want to start with the Hbase book, and specifically this
> entry...
>
> http://hbase.apache.org/book.html#faq.sql
>
> ... and then this one...
>
> http://hbase.apache.org/book.html#datamodel
>
> .. then this one.
>
> http://hbase.apache.org/book.html#mapreduce
>
> MapReduce is not required with Hbase, but it is extremely useful.
>
>
>
> On 7/12/11 6:06 AM, "Rita" <rm...@gmail.com> wrote:
>
> >Hello,
> >
> >
> >I have a dataset which is several terabytes in size. I would like to query
> >this data using hbase (sql). Would I need to setup mapreduce to use hbase?
> >Currently the data is stored in hdfs and I am using `hdfs -cat ` to get
> >the
> >data and pipe it into stdin.
> >
> >
> >
> >--
> >--- Get your facts first, then you can distort them as you please.--
>
>


-- 
--- Get your facts first, then you can distort them as you please.--

Re: is mapreduce needed?

Posted by Doug Meil <do...@explorysmedical.com>.
Hi there-

I think you might want to start with the Hbase book, and specifically this
entry...

http://hbase.apache.org/book.html#faq.sql

... and then this one...

http://hbase.apache.org/book.html#datamodel

.. then this one.

http://hbase.apache.org/book.html#mapreduce

MapReduce is not required with Hbase, but it is extremely useful.



On 7/12/11 6:06 AM, "Rita" <rm...@gmail.com> wrote:

>Hello,
>
>
>I have a dataset which is several terabytes in size. I would like to query
>this data using hbase (sql). Would I need to setup mapreduce to use hbase?
>Currently the data is stored in hdfs and I am using `hdfs -cat ` to get
>the
>data and pipe it into stdin.
>
>
>
>-- 
>--- Get your facts first, then you can distort them as you please.--