You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@mahout.apache.org by Wojciech Indyk <wo...@gmail.com> on 2011/05/24 22:57:12 UTC

AdaBoost

Hi!
I want implement AdaBoost in Mahout. Could it be useful in Mahout? I
think so, because it's strong algorithm and very powerful, but Mahout
is specific, so who knows :)
I thought about training data and I know, that I must parallelize by
data, rather than by algorithms, so it will be not so easy - I must
run all mapers of chosed algorithms in my training maper, but I have
no idea how could i choose algorithms to adaboost (in architecture
way) like a parameter.

Regards

Re: AdaBoost

Posted by Hector Yee <he...@gmail.com>.
It is if you use the grad boost variant. I'll work on it next week on vacation...

Sent from my iPad

On May 24, 2011, at 4:48 PM, Ted Dunning <te...@gmail.com> wrote:

> Is AdaBoost a scalable algorithm?
> 
> It seems to me that it is inherently very sequential.
> 
> On Tue, May 24, 2011 at 1:57 PM, Wojciech Indyk <wo...@gmail.com>wrote:
> 
>> Hi!
>> I want implement AdaBoost in Mahout. Could it be useful in Mahout? I
>> think so, because it's strong algorithm and very powerful, but Mahout
>> is specific, so who knows :)
>> I thought about training data and I know, that I must parallelize by
>> data, rather than by algorithms, so it will be not so easy - I must
>> run all mapers of chosed algorithms in my training maper, but I have
>> no idea how could i choose algorithms to adaboost (in architecture
>> way) like a parameter.
>> 
>> Regards
>> 

Re: AdaBoost

Posted by Ted Dunning <te...@gmail.com>.
Is AdaBoost a scalable algorithm?

It seems to me that it is inherently very sequential.

On Tue, May 24, 2011 at 1:57 PM, Wojciech Indyk <wo...@gmail.com>wrote:

> Hi!
> I want implement AdaBoost in Mahout. Could it be useful in Mahout? I
> think so, because it's strong algorithm and very powerful, but Mahout
> is specific, so who knows :)
> I thought about training data and I know, that I must parallelize by
> data, rather than by algorithms, so it will be not so easy - I must
> run all mapers of chosed algorithms in my training maper, but I have
> no idea how could i choose algorithms to adaboost (in architecture
> way) like a parameter.
>
> Regards
>

Re: AdaBoost

Posted by Hector Yee <he...@gmail.com>.
Wojciech, I've opened a ticked you can watch

 https://issues.apache.org/jira/browse/MAHOUT-716

I should have the in core code ready in ~3 days. The gradient portion is
easily parallelizable if you want to implement it as mapreduce.

On Tue, May 24, 2011 at 1:57 PM, Wojciech Indyk <wo...@gmail.com>wrote:

> Hi!
> I want implement AdaBoost in Mahout. Could it be useful in Mahout? I
> think so, because it's strong algorithm and very powerful, but Mahout
> is specific, so who knows :)
> I thought about training data and I know, that I must parallelize by
> data, rather than by algorithms, so it will be not so easy - I must
> run all mapers of chosed algorithms in my training maper, but I have
> no idea how could i choose algorithms to adaboost (in architecture
> way) like a parameter.
>
> Regards
>



-- 
Yee Yang Li Hector
http://hectorgon.blogspot.com/ (tech + travel)
http://hectorgon.com (book reviews)

Re: AdaBoost

Posted by Hector Yee <he...@gmail.com>.
Patch uploaded

https://issues.apache.org/jira/browse/MAHOUT-716

On Tue, May 24, 2011 at 1:57 PM, Wojciech Indyk <wo...@gmail.com>wrote:

> Hi!
> I want implement AdaBoost in Mahout. Could it be useful in Mahout? I
> think so, because it's strong algorithm and very powerful, but Mahout
> is specific, so who knows :)
> I thought about training data and I know, that I must parallelize by
> data, rather than by algorithms, so it will be not so easy - I must
> run all mapers of chosed algorithms in my training maper, but I have
> no idea how could i choose algorithms to adaboost (in architecture
> way) like a parameter.
>
> Regards
>



-- 
Yee Yang Li Hector
http://hectorgon.blogspot.com/ (tech + travel)
http://hectorgon.com (book reviews)