You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Jignesh Patel <ji...@gmail.com> on 2011/10/18 20:02:10 UTC

integrating hadoop and Hbase with eclipse

I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't have
hbase specific jars. By googling I found some articles about only setting
base but none of them talks about setting up hadoop and hbase with eclipse.

-jignesh

Re: Custom timestamps

Posted by Doug Meil <do...@explorysmedical.com>.
Stack, he might be referring to this...

http://hbase.apache.org/book.html#versions

... I updated this recently based on an exchange with JD and somebody else.





On 10/21/11 1:08 AM, "Stuti Awasthi" <st...@hcl.com> wrote:

>Hi St. Ack,
>
>I read something while browsing . Right now don't have link but if I come
>across something similar , I will let you know. Thanks for info. It is
>really a big relief :)
>
>-----Original Message-----
>From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of Stack
>Sent: Thursday, October 20, 2011 9:27 PM
>To: user@hbase.apache.org
>Subject: Re: Custom timestamps
>
>On Wed, Oct 19, 2011 at 9:59 PM, Stuti Awasthi <st...@hcl.com>
>wrote:
>> Hi St. Ack , Ben
>>
>> I also have a scenario that in my case I have to take periodical backup
>>of Hbase data. For that I have will be using export/import tool. I have
>>decided to take backup based on time range interval. I have read it in
>>some other posts also that it is not good idea for one to use timestamp
>>field of Hbase.
>
>What you are doing sounds fine to me.  Which posts say it bad so I can
>see what issues in particular they are referring to.
>Thanks,
>St.Ack
>
>::DISCLAIMER::
>--------------------------------------------------------------------------
>---------------------------------------------
>
>The contents of this e-mail and any attachment(s) are confidential and
>intended for the named recipient(s) only.
>It shall not attach any liability on the originator or HCL or its
>affiliates. Any views or opinions presented in
>this email are solely those of the author and may not necessarily reflect
>the opinions of HCL or its affiliates.
>Any form of reproduction, dissemination, copying, disclosure,
>modification, distribution and / or publication of
>this message without the prior written consent of the author of this
>e-mail is strictly prohibited. If you have
>received this email in error please delete it and notify the sender
>immediately. Before opening any mail and
>attachments please check them for viruses and defect.
>
>--------------------------------------------------------------------------
>---------------------------------------------


RE: Custom timestamps

Posted by Stuti Awasthi <st...@hcl.com>.
Hi St. Ack,

I read something while browsing . Right now don't have link but if I come across something similar , I will let you know. Thanks for info. It is really a big relief :)

-----Original Message-----
From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of Stack
Sent: Thursday, October 20, 2011 9:27 PM
To: user@hbase.apache.org
Subject: Re: Custom timestamps

On Wed, Oct 19, 2011 at 9:59 PM, Stuti Awasthi <st...@hcl.com> wrote:
> Hi St. Ack , Ben
>
> I also have a scenario that in my case I have to take periodical backup of Hbase data. For that I have will be using export/import tool. I have decided to take backup based on time range interval. I have read it in some other posts also that it is not good idea for one to use timestamp field of Hbase.

What you are doing sounds fine to me.  Which posts say it bad so I can see what issues in particular they are referring to.
Thanks,
St.Ack

::DISCLAIMER::
-----------------------------------------------------------------------------------------------------------------------

The contents of this e-mail and any attachment(s) are confidential and intended for the named recipient(s) only.
It shall not attach any liability on the originator or HCL or its affiliates. Any views or opinions presented in
this email are solely those of the author and may not necessarily reflect the opinions of HCL or its affiliates.
Any form of reproduction, dissemination, copying, disclosure, modification, distribution and / or publication of
this message without the prior written consent of the author of this e-mail is strictly prohibited. If you have
received this email in error please delete it and notify the sender immediately. Before opening any mail and
attachments please check them for viruses and defect.

-----------------------------------------------------------------------------------------------------------------------

Re: Custom timestamps

Posted by Stack <st...@duboce.net>.
On Wed, Oct 19, 2011 at 9:59 PM, Stuti Awasthi <st...@hcl.com> wrote:
> Hi St. Ack , Ben
>
> I also have a scenario that in my case I have to take periodical backup of Hbase data. For that I have will be using export/import tool. I have decided to take backup based on time range interval. I have read it in some other posts also that it is not good idea for one to use timestamp field of Hbase.

What you are doing sounds fine to me.  Which posts say it bad so I can
see what issues in particular they are referring to.
Thanks,
St.Ack

RE: Custom timestamps

Posted by Stuti Awasthi <st...@hcl.com>.
Hi St. Ack , Ben

I also have a scenario that in my case I have to take periodical backup of Hbase data. For that I have will be using export/import tool. I have decided to take backup based on time range interval. I have read it in some other posts also that it is not good idea for one to use timestamp field of Hbase.
Till now my POC works fine at my end but I want to know that if I put same scenario in production will there be any issues to worry ?

Please Suggest.

-----Original Message-----
From: saint.ack@gmail.com [mailto:saint.ack@gmail.com] On Behalf Of Stack
Sent: Thursday, October 20, 2011 3:54 AM
To: user@hbase.apache.org; Ben West
Subject: Re: Custom timestamps

On Wed, Oct 19, 2011 at 12:18 PM, Ben West <bw...@yahoo.com> wrote:
> We're storing timestamped data in HBase; from lurking on the mailing list it seems like the recommendation is usually to make the timestamp part of the row key. I'm curious why this is - is scanning over rows more efficient than scanning over timestamps within a cell?
>

I'd be surprised if a noticeable difference.

It depends on how you are to access the data.  In the tsdb case for instance, it wants to get all metrics within a particular time range.
If the timestamp it used were that of the hbase system, then you'd have to do a full table scan each time to find metrics that had been fired during a particular time period -- i.e. you'd check each row and see if any entries on the row for the time period you are interested in -- whereas if the timestamp part of the row key, you instead just have to start scanning at the opening of the time range you are querying about.


> The book says: "the version timestamp is internally by HBase for things like time-to-live calculations. It's usually best to avoid setting this timestamp yourself. Prefer using a separate timestamp attribute of the row, or have the timestamp a part of the rowkey, or both." I understand that TTL would be ruined (or saved, depending on your goal) by custom timestamps, and I also gather that the way HBase handles concurrency is through MVCC. But we are using application level locks, and HBase's TTL functionality applying is a bonus if anything.
>

The books advice errs on the side of being conservative I'd say.

The MVCC that we do internally does not use the cell timestamp but instead a different running sequence number that is associated
(internally) with cells (I've not heard of an application atop hbase using the hbase timestamps to do MVCC at the application level).

The locks you talk of, are these the locks provided in hbase HTable API?  If so, are you aware they are dangerous (see back in this mailing list for explaination)?

> So is there any reason why we shouldn't set the timestamps manually?
>

Generally, hbase works fine with user set timestamps; there can be issues ordering edits if clients have divergent clocks and the version being set is time-based but I'm probably not telling you something you don't already know.

St.Ack

::DISCLAIMER::
-----------------------------------------------------------------------------------------------------------------------

The contents of this e-mail and any attachment(s) are confidential and intended for the named recipient(s) only.
It shall not attach any liability on the originator or HCL or its affiliates. Any views or opinions presented in
this email are solely those of the author and may not necessarily reflect the opinions of HCL or its affiliates.
Any form of reproduction, dissemination, copying, disclosure, modification, distribution and / or publication of
this message without the prior written consent of the author of this e-mail is strictly prohibited. If you have
received this email in error please delete it and notify the sender immediately. Before opening any mail and
attachments please check them for viruses and defect.

-----------------------------------------------------------------------------------------------------------------------

Re: Custom timestamps

Posted by Stack <st...@duboce.net>.
On Thu, Oct 20, 2011 at 8:11 AM, Ben West <bw...@yahoo.com> wrote:
> Actually, another question: are there issues with multiple puts having the same timestamp? I.e. I write a value with timestamp = today 12:00. I later change my mind and want to rewrite a different value but with the same timestamp. Would that present problems?
>

When you multiput, all get same timestamp.

There is no notion of 'changing' values in hbase.  You just add a new
version.  When hbase returns you the values, it will return them
ordered by timestamp (if the new timestamp is ahead of 12:00, it will
come out first.... else afterward).

St.Ack

Re: Custom timestamps

Posted by Ben West <bw...@yahoo.com>.
Actually, another question: are there issues with multiple puts having the same timestamp? I.e. I write a value with timestamp = today 12:00. I later change my mind and want to rewrite a different value but with the same timestamp. Would that present problems?

Thanks!
-Ben


----- Original Message -----
From: Ben West <bw...@yahoo.com>
To: "user@hbase.apache.org" <us...@hbase.apache.org>
Cc: 
Sent: Thursday, October 20, 2011 9:13 AM
Subject: Re: Custom timestamps

Thanks Stack. We are indeed using locks outside of HBase, but I hadn't heard about the problems with HBase's locks. Good to know.

-Ben


----- Original Message -----
From: Stack <st...@duboce.net>
To: user@hbase.apache.org; Ben West <bw...@yahoo.com>
Cc: 
Sent: Wednesday, October 19, 2011 5:24 PM
Subject: Re: Custom timestamps

On Wed, Oct 19, 2011 at 12:18 PM, Ben West <bw...@yahoo.com> wrote:
> We're storing timestamped data in HBase; from lurking on the mailing list it seems like the recommendation is usually to make the timestamp part of the row key. I'm curious why this is - is scanning over rows more efficient than scanning over timestamps within a cell?
>

I'd be surprised if a noticeable difference.

It depends on how you are to access the data.  In the tsdb case for
instance, it wants to get all metrics within a particular time range.
If the timestamp it used were that of the hbase system, then you'd
have to do a full table scan each time to find metrics that had been
fired during a particular time period -- i.e. you'd check each row and
see if any entries on the row for the time period you are interested
in -- whereas if the timestamp part of the row key, you instead just
have to start scanning at the opening of the time range you are
querying about.


> The book says: "the version timestamp is internally by HBase for things like time-to-live calculations. It's usually best to avoid setting this timestamp yourself. Prefer using a separate timestamp attribute of the row, or have the timestamp a part of the rowkey, or both." I understand that TTL would be ruined (or saved, depending on your goal) by custom timestamps, and I also gather that the way HBase handles concurrency is through MVCC. But we are using application level locks, and HBase's TTL functionality applying is a bonus if anything.
>

The books advice errs on the side of being conservative I'd say.

The MVCC that we do internally does not use the cell timestamp but
instead a different running sequence number that is associated
(internally) with cells (I've not heard of an application atop hbase
using the hbase timestamps to do MVCC at the application level).

The locks you talk of, are these the locks provided in hbase HTable
API?  If so, are you aware they are dangerous (see back in this
mailing list for explaination)?

> So is there any reason why we shouldn't set the timestamps manually?
>

Generally, hbase works fine with user set timestamps; there can be
issues ordering edits if clients have divergent clocks and the version
being set is time-based but I'm probably not telling you something you
don't already know.

St.Ack

Re: Custom timestamps

Posted by Ben West <bw...@yahoo.com>.
Thanks Stack. We are indeed using locks outside of HBase, but I hadn't heard about the problems with HBase's locks. Good to know.

-Ben


----- Original Message -----
From: Stack <st...@duboce.net>
To: user@hbase.apache.org; Ben West <bw...@yahoo.com>
Cc: 
Sent: Wednesday, October 19, 2011 5:24 PM
Subject: Re: Custom timestamps

On Wed, Oct 19, 2011 at 12:18 PM, Ben West <bw...@yahoo.com> wrote:
> We're storing timestamped data in HBase; from lurking on the mailing list it seems like the recommendation is usually to make the timestamp part of the row key. I'm curious why this is - is scanning over rows more efficient than scanning over timestamps within a cell?
>

I'd be surprised if a noticeable difference.

It depends on how you are to access the data.  In the tsdb case for
instance, it wants to get all metrics within a particular time range.
If the timestamp it used were that of the hbase system, then you'd
have to do a full table scan each time to find metrics that had been
fired during a particular time period -- i.e. you'd check each row and
see if any entries on the row for the time period you are interested
in -- whereas if the timestamp part of the row key, you instead just
have to start scanning at the opening of the time range you are
querying about.


> The book says: "the version timestamp is internally by HBase for things like time-to-live calculations. It's usually best to avoid setting this timestamp yourself. Prefer using a separate timestamp attribute of the row, or have the timestamp a part of the rowkey, or both." I understand that TTL would be ruined (or saved, depending on your goal) by custom timestamps, and I also gather that the way HBase handles concurrency is through MVCC. But we are using application level locks, and HBase's TTL functionality applying is a bonus if anything.
>

The books advice errs on the side of being conservative I'd say.

The MVCC that we do internally does not use the cell timestamp but
instead a different running sequence number that is associated
(internally) with cells (I've not heard of an application atop hbase
using the hbase timestamps to do MVCC at the application level).

The locks you talk of, are these the locks provided in hbase HTable
API?  If so, are you aware they are dangerous (see back in this
mailing list for explaination)?

> So is there any reason why we shouldn't set the timestamps manually?
>

Generally, hbase works fine with user set timestamps; there can be
issues ordering edits if clients have divergent clocks and the version
being set is time-based but I'm probably not telling you something you
don't already know.

St.Ack


Re: Custom timestamps

Posted by Stack <st...@duboce.net>.
On Wed, Oct 19, 2011 at 12:18 PM, Ben West <bw...@yahoo.com> wrote:
> We're storing timestamped data in HBase; from lurking on the mailing list it seems like the recommendation is usually to make the timestamp part of the row key. I'm curious why this is - is scanning over rows more efficient than scanning over timestamps within a cell?
>

I'd be surprised if a noticeable difference.

It depends on how you are to access the data.  In the tsdb case for
instance, it wants to get all metrics within a particular time range.
If the timestamp it used were that of the hbase system, then you'd
have to do a full table scan each time to find metrics that had been
fired during a particular time period -- i.e. you'd check each row and
see if any entries on the row for the time period you are interested
in -- whereas if the timestamp part of the row key, you instead just
have to start scanning at the opening of the time range you are
querying about.


> The book says: "the version timestamp is internally by HBase for things like time-to-live calculations. It's usually best to avoid setting this timestamp yourself. Prefer using a separate timestamp attribute of the row, or have the timestamp a part of the rowkey, or both." I understand that TTL would be ruined (or saved, depending on your goal) by custom timestamps, and I also gather that the way HBase handles concurrency is through MVCC. But we are using application level locks, and HBase's TTL functionality applying is a bonus if anything.
>

The books advice errs on the side of being conservative I'd say.

The MVCC that we do internally does not use the cell timestamp but
instead a different running sequence number that is associated
(internally) with cells (I've not heard of an application atop hbase
using the hbase timestamps to do MVCC at the application level).

The locks you talk of, are these the locks provided in hbase HTable
API?  If so, are you aware they are dangerous (see back in this
mailing list for explaination)?

> So is there any reason why we shouldn't set the timestamps manually?
>

Generally, hbase works fine with user set timestamps; there can be
issues ordering edits if clients have divergent clocks and the version
being set is time-based but I'm probably not telling you something you
don't already know.

St.Ack

Custom timestamps

Posted by Ben West <bw...@yahoo.com>.
Hi all,

We're storing timestamped data in HBase; from lurking on the mailing list it seems like the recommendation is usually to make the timestamp part of the row key. I'm curious why this is - is scanning over rows more efficient than scanning over timestamps within a cell? 

The book says: "the version timestamp is internally by HBase for things like time-to-live calculations. It's usually best to avoid setting this timestamp yourself. Prefer using a separate timestamp attribute of the row, or have the timestamp a part of the rowkey, or both." I understand that TTL would be ruined (or saved, depending on your goal) by custom timestamps, and I also gather that the way HBase handles concurrency is through MVCC. But we are using application level locks, and HBase's TTL functionality applying is a bonus if anything.

So is there any reason why we shouldn't set the timestamps manually?

Thanks!
-Ben

Re: integrating hadoop and Hbase with eclipse

Posted by Stack <st...@duboce.net>.
On Wed, Oct 19, 2011 at 2:19 PM, Jignesh Patel <ji...@gmail.com> wrote:
> St.Ack,
>
> changin hadoop-version in pom.xml doesn't help as during runtime it still
> tries to load from m2./repository folder. Is there a way where I can say
> that don't go to .m2/repository folder but goto /lib directory.
>

You can rig maven to use explicit dir for a particular jar  -- set its
scope=system [1] -- rather than read from .m2.

Or just change the pom and do a command line build and it will pull
down 205 from maven repository into your .m2 repo (There is likely a
way to do this from eclipse maven plugin too but I've not used it).
Do your mvn eclipse:eclipse and you should be good to go then.

St.Ack

1. http://stackoverflow.com/questions/364114/can-i-add-jars-to-maven-2-build-classpath-without-installing-them

Re: integrating hadoop and Hbase with eclipse

Posted by Jignesh Patel <ji...@gmail.com>.
St.Ack,

changin hadoop-version in pom.xml doesn't help as during runtime it still
tries to load from m2./repository folder. Is there a way where I can say
that don't go to .m2/repository folder but goto /lib directory.

-jignesh

On Wed, Oct 19, 2011 at 4:42 PM, Stack <st...@duboce.net> wrote:

> On Wed, Oct 19, 2011 at 1:10 PM, Jignesh Patel <ji...@gmail.com>
> wrote:
> > just figured out that while running start-hbase.sh it is taking files
> from
> > ~/.m2/repository then usual hadoop-hbase/lib.
> >
> > What I need to change so that it will take the files only from
> > hadoop-hbase/lib. Because repository has hadoop-core append file which is
> > not matching with mine hadoop-core-2.205.0.jar
> >
>
> Change the hadoop version in your pom.
> St.Ack
>

Re: integrating hadoop and Hbase with eclipse

Posted by Stack <st...@duboce.net>.
On Wed, Oct 19, 2011 at 1:10 PM, Jignesh Patel <ji...@gmail.com> wrote:
> just figured out that while running start-hbase.sh it is taking files from
> ~/.m2/repository then usual hadoop-hbase/lib.
>
> What I need to change so that it will take the files only from
> hadoop-hbase/lib. Because repository has hadoop-core append file which is
> not matching with mine hadoop-core-2.205.0.jar
>

Change the hadoop version in your pom.
St.Ack

Re: integrating hadoop and Hbase with eclipse

Posted by Jignesh Patel <ji...@gmail.com>.
just figured out that while running start-hbase.sh it is taking files from
~/.m2/repository then usual hadoop-hbase/lib.

What I need to change so that it will take the files only from
hadoop-hbase/lib. Because repository has hadoop-core append file which is
not matching with mine hadoop-core-2.205.0.jar

-Jignesh

On Wed, Oct 19, 2011 at 2:37 PM, Jignesh Patel <ji...@gmail.com>wrote:

> After compiling with with maven, I started getting back the following
> error.
>
> 2011-10-19 14:02:38,089 FATAL org.apache.hadoop.hbase.master.HMaster:
> Unhandled exception. Starting shutdown.
> java.io.IOException: Call to localhost/127.0.0.1:9000 failed on local
> exception: java.io.EOFException
>  at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
> at org.apache.hadoop.ipc.Client.call(Client.java:743)
>  at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
> at $Proxy6.getProtocolVersion(Unknown Source)
>  at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
> at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:113)
>  at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:215)
> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:177)
>  at
> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82)
> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
>  at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
>  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175)
>  at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:364)
> at
> org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:81)
>  at
> org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:346)
> at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:282)
>  at
> org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster.run(HMasterCommandLine.java:193)
> at java.lang.Thread.run(Thread.java:680)
> Caused by: java.io.EOFException
>
>
> However, the hadoop-hbase still shows that I have
> required hadoop-core-0.20.205.0.jar in the path.
>
> The only change I see in the folder are creation of build and target folder
> as well as .m2 folder at different location. Is there anything changed
> because I used mvn install command.
>
>
> On Wed, Oct 19, 2011 at 11:25 AM, Jignesh Patel <ji...@gmail.com>wrote:
>
>> Just thinking it may be more easier to refer the hbase lib/jar by creating
>> map reduce project.
>> But I don't know in which jar following package exist.
>>
>> org.apache.hadoop.hbase.
>>
>>
>> -Jignesh
>>
>> On Wed, Oct 19, 2011 at 10:14 AM, Jignesh Patel <ji...@gmail.com>wrote:
>>
>>> Dough and Jonathan,
>>> Yes I did follow the steps and with the help of Maven, I have  created
>>> target folder.
>>> But that folder doesn't have anything like eclipse plugin. Something
>>> readily available in hadoop/contrib folder.
>>>
>>> What I am looking at this time, when I create a project through hadoop it
>>> shouldn't cry for the missing hbase classes. The way hadoop plugin adds all
>>> the required jars when I create map reduce project, similarly I should have
>>> hbase files available.
>>>
>>> So I am looking support for the code level. The link you provided for the
>>> Maven, is able to help till the compiling all the source code. But then I
>>> have to add appropriate jar so that whenever I create project for
>>> hadoop(0.20.205.0) and base(0.90.4) I have all the required jars available.
>>>
>>> For the database creation there is tool from Toad but I have not tried
>>> yet.
>>>
>>> -Jignesh
>>>
>>> On Wed, Oct 19, 2011 at 9:43 AM, Doug Meil <
>>> doug.meil@explorysmedical.com> wrote:
>>>
>>>>
>>>> In addition to what Jonathan just said, see
>>>>
>>>> http://hbase.apache.org/book.html#ides
>>>>
>>>>
>>>>
>>>> On 10/19/11 3:05 AM, "Jonathan Gray" <jg...@fb.com> wrote:
>>>>
>>>> >Not sure what kind of integration you're talking about, but if just
>>>> want
>>>> >to create a project with the HBase source then just grab an SVN
>>>> checkout
>>>> >of an HBase repo and just do:
>>>> >
>>>> >mvn eclipse:eclipse
>>>> >
>>>> >This creates all the necessary project files.  Then just add new
>>>> project
>>>> >from existing source.
>>>> >
>>>> >That's the extent to which there exists any HBase integration with
>>>> >eclipse that I know of.
>>>> >
>>>> >> -----Original Message-----
>>>> >> From: Jignesh Patel [mailto:jigneshmpatel@gmail.com]
>>>> >> Sent: Tuesday, October 18, 2011 11:02 AM
>>>> >> To: user@hbase.apache.org
>>>> >> Subject: integrating hadoop and Hbase with eclipse
>>>> >>
>>>> >> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't
>>>> >>have
>>>> >> hbase specific jars. By googling I found some articles about only
>>>> >>setting base
>>>> >> but none of them talks about setting up hadoop and hbase with
>>>> eclipse.
>>>> >>
>>>> >> -jignesh
>>>>
>>>>
>>>
>>
>

Re: integrating hadoop and Hbase with eclipse

Posted by Jignesh Patel <ji...@gmail.com>.
After compiling with with maven, I started getting back the following error.

2011-10-19 14:02:38,089 FATAL org.apache.hadoop.hbase.master.HMaster:
Unhandled exception. Starting shutdown.
java.io.IOException: Call to localhost/127.0.0.1:9000 failed on local
exception: java.io.EOFException
at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)
at org.apache.hadoop.ipc.Client.call(Client.java:743)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)
at $Proxy6.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)
at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:113)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:215)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:177)
at
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175)
at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:364)
at
org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:81)
at
org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:346)
at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:282)
at
org.apache.hadoop.hbase.master.HMasterCommandLine$LocalHMaster.run(HMasterCommandLine.java:193)
at java.lang.Thread.run(Thread.java:680)
Caused by: java.io.EOFException


However, the hadoop-hbase still shows that I have
required hadoop-core-0.20.205.0.jar in the path.

The only change I see in the folder are creation of build and target folder
as well as .m2 folder at different location. Is there anything changed
because I used mvn install command.

On Wed, Oct 19, 2011 at 11:25 AM, Jignesh Patel <ji...@gmail.com>wrote:

> Just thinking it may be more easier to refer the hbase lib/jar by creating
> map reduce project.
> But I don't know in which jar following package exist.
>
> org.apache.hadoop.hbase.
>
>
> -Jignesh
>
> On Wed, Oct 19, 2011 at 10:14 AM, Jignesh Patel <ji...@gmail.com>wrote:
>
>> Dough and Jonathan,
>> Yes I did follow the steps and with the help of Maven, I have  created
>> target folder.
>> But that folder doesn't have anything like eclipse plugin. Something
>> readily available in hadoop/contrib folder.
>>
>> What I am looking at this time, when I create a project through hadoop it
>> shouldn't cry for the missing hbase classes. The way hadoop plugin adds all
>> the required jars when I create map reduce project, similarly I should have
>> hbase files available.
>>
>> So I am looking support for the code level. The link you provided for the
>> Maven, is able to help till the compiling all the source code. But then I
>> have to add appropriate jar so that whenever I create project for
>> hadoop(0.20.205.0) and base(0.90.4) I have all the required jars available.
>>
>> For the database creation there is tool from Toad but I have not tried
>> yet.
>>
>> -Jignesh
>>
>> On Wed, Oct 19, 2011 at 9:43 AM, Doug Meil <doug.meil@explorysmedical.com
>> > wrote:
>>
>>>
>>> In addition to what Jonathan just said, see
>>>
>>> http://hbase.apache.org/book.html#ides
>>>
>>>
>>>
>>> On 10/19/11 3:05 AM, "Jonathan Gray" <jg...@fb.com> wrote:
>>>
>>> >Not sure what kind of integration you're talking about, but if just want
>>> >to create a project with the HBase source then just grab an SVN checkout
>>> >of an HBase repo and just do:
>>> >
>>> >mvn eclipse:eclipse
>>> >
>>> >This creates all the necessary project files.  Then just add new project
>>> >from existing source.
>>> >
>>> >That's the extent to which there exists any HBase integration with
>>> >eclipse that I know of.
>>> >
>>> >> -----Original Message-----
>>> >> From: Jignesh Patel [mailto:jigneshmpatel@gmail.com]
>>> >> Sent: Tuesday, October 18, 2011 11:02 AM
>>> >> To: user@hbase.apache.org
>>> >> Subject: integrating hadoop and Hbase with eclipse
>>> >>
>>> >> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't
>>> >>have
>>> >> hbase specific jars. By googling I found some articles about only
>>> >>setting base
>>> >> but none of them talks about setting up hadoop and hbase with eclipse.
>>> >>
>>> >> -jignesh
>>>
>>>
>>
>

Re: integrating hadoop and Hbase with eclipse

Posted by Jignesh Patel <ji...@gmail.com>.
Just thinking it may be more easier to refer the hbase lib/jar by creating
map reduce project.
But I don't know in which jar following package exist.

org.apache.hadoop.hbase.


-Jignesh

On Wed, Oct 19, 2011 at 10:14 AM, Jignesh Patel <ji...@gmail.com>wrote:

> Dough and Jonathan,
> Yes I did follow the steps and with the help of Maven, I have  created
> target folder.
> But that folder doesn't have anything like eclipse plugin. Something
> readily available in hadoop/contrib folder.
>
> What I am looking at this time, when I create a project through hadoop it
> shouldn't cry for the missing hbase classes. The way hadoop plugin adds all
> the required jars when I create map reduce project, similarly I should have
> hbase files available.
>
> So I am looking support for the code level. The link you provided for the
> Maven, is able to help till the compiling all the source code. But then I
> have to add appropriate jar so that whenever I create project for
> hadoop(0.20.205.0) and base(0.90.4) I have all the required jars available.
>
> For the database creation there is tool from Toad but I have not tried yet.
>
> -Jignesh
>
> On Wed, Oct 19, 2011 at 9:43 AM, Doug Meil <do...@explorysmedical.com>wrote:
>
>>
>> In addition to what Jonathan just said, see
>>
>> http://hbase.apache.org/book.html#ides
>>
>>
>>
>> On 10/19/11 3:05 AM, "Jonathan Gray" <jg...@fb.com> wrote:
>>
>> >Not sure what kind of integration you're talking about, but if just want
>> >to create a project with the HBase source then just grab an SVN checkout
>> >of an HBase repo and just do:
>> >
>> >mvn eclipse:eclipse
>> >
>> >This creates all the necessary project files.  Then just add new project
>> >from existing source.
>> >
>> >That's the extent to which there exists any HBase integration with
>> >eclipse that I know of.
>> >
>> >> -----Original Message-----
>> >> From: Jignesh Patel [mailto:jigneshmpatel@gmail.com]
>> >> Sent: Tuesday, October 18, 2011 11:02 AM
>> >> To: user@hbase.apache.org
>> >> Subject: integrating hadoop and Hbase with eclipse
>> >>
>> >> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't
>> >>have
>> >> hbase specific jars. By googling I found some articles about only
>> >>setting base
>> >> but none of them talks about setting up hadoop and hbase with eclipse.
>> >>
>> >> -jignesh
>>
>>
>

Re: integrating hadoop and Hbase with eclipse

Posted by Jignesh Patel <ji...@gmail.com>.
Dough and Jonathan,
Yes I did follow the steps and with the help of Maven, I have  created
target folder.
But that folder doesn't have anything like eclipse plugin. Something readily
available in hadoop/contrib folder.

What I am looking at this time, when I create a project through hadoop it
shouldn't cry for the missing hbase classes. The way hadoop plugin adds all
the required jars when I create map reduce project, similarly I should have
hbase files available.

So I am looking support for the code level. The link you provided for the
Maven, is able to help till the compiling all the source code. But then I
have to add appropriate jar so that whenever I create project for
hadoop(0.20.205.0) and base(0.90.4) I have all the required jars available.

For the database creation there is tool from Toad but I have not tried yet.

-Jignesh

On Wed, Oct 19, 2011 at 9:43 AM, Doug Meil <do...@explorysmedical.com>wrote:

>
> In addition to what Jonathan just said, see
>
> http://hbase.apache.org/book.html#ides
>
>
>
> On 10/19/11 3:05 AM, "Jonathan Gray" <jg...@fb.com> wrote:
>
> >Not sure what kind of integration you're talking about, but if just want
> >to create a project with the HBase source then just grab an SVN checkout
> >of an HBase repo and just do:
> >
> >mvn eclipse:eclipse
> >
> >This creates all the necessary project files.  Then just add new project
> >from existing source.
> >
> >That's the extent to which there exists any HBase integration with
> >eclipse that I know of.
> >
> >> -----Original Message-----
> >> From: Jignesh Patel [mailto:jigneshmpatel@gmail.com]
> >> Sent: Tuesday, October 18, 2011 11:02 AM
> >> To: user@hbase.apache.org
> >> Subject: integrating hadoop and Hbase with eclipse
> >>
> >> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't
> >>have
> >> hbase specific jars. By googling I found some articles about only
> >>setting base
> >> but none of them talks about setting up hadoop and hbase with eclipse.
> >>
> >> -jignesh
>
>

Re: integrating hadoop and Hbase with eclipse

Posted by Doug Meil <do...@explorysmedical.com>.
In addition to what Jonathan just said, see

http://hbase.apache.org/book.html#ides



On 10/19/11 3:05 AM, "Jonathan Gray" <jg...@fb.com> wrote:

>Not sure what kind of integration you're talking about, but if just want
>to create a project with the HBase source then just grab an SVN checkout
>of an HBase repo and just do:
>
>mvn eclipse:eclipse
>
>This creates all the necessary project files.  Then just add new project
>from existing source.
>
>That's the extent to which there exists any HBase integration with
>eclipse that I know of.
>
>> -----Original Message-----
>> From: Jignesh Patel [mailto:jigneshmpatel@gmail.com]
>> Sent: Tuesday, October 18, 2011 11:02 AM
>> To: user@hbase.apache.org
>> Subject: integrating hadoop and Hbase with eclipse
>> 
>> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't
>>have
>> hbase specific jars. By googling I found some articles about only
>>setting base
>> but none of them talks about setting up hadoop and hbase with eclipse.
>> 
>> -jignesh


RE: integrating hadoop and Hbase with eclipse

Posted by Jonathan Gray <jg...@fb.com>.
Not sure what kind of integration you're talking about, but if just want to create a project with the HBase source then just grab an SVN checkout of an HBase repo and just do:

mvn eclipse:eclipse

This creates all the necessary project files.  Then just add new project from existing source.

That's the extent to which there exists any HBase integration with eclipse that I know of.

> -----Original Message-----
> From: Jignesh Patel [mailto:jigneshmpatel@gmail.com]
> Sent: Tuesday, October 18, 2011 11:02 AM
> To: user@hbase.apache.org
> Subject: integrating hadoop and Hbase with eclipse
> 
> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't have
> hbase specific jars. By googling I found some articles about only setting base
> but none of them talks about setting up hadoop and hbase with eclipse.
> 
> -jignesh

Re: integrating hadoop and Hbase with eclipse

Posted by Harsh J <ha...@cloudera.com>.
Jignesh,

AFAIK, there isn't a HBase eclipse plugin integration available yet, to browse tables and such (was that your intention?).

I don't think Karmasphere integration tools have HBase in it yet, but there is work ongoing I guess, from [0].

The only visual browser for HBase I've seen and used so far is Lars George's https://github.com/larsgeorge/hbase-explorer, which is a HUE application [1].

I'm interested in hearing about others if they've popped up somewhere.

Otherwise, you need to make do with HBase's web UI, and the shell.

[0] - http://old.nabble.com/karmasphere-hbase-td27316495.html
[1] - http://cloudera.github.com/hue/

On 18-Oct-2011, at 11:32 PM, Jignesh Patel wrote:

> I am able to integrate hadoop 0.20.205.0 with eclipse. But it doesn't have
> hbase specific jars. By googling I found some articles about only setting
> base but none of them talks about setting up hadoop and hbase with eclipse.
> 
> -jignesh