You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hbase.apache.org by Ferdy Galema <fe...@kalooga.com> on 2011/03/16 18:28:51 UTC

after upgrade, fatal error in regionserver compacter, LzoCompressor, "AbstractMethodError"

We upgraded to Hadoop 0.20.1 and Hbase 0.90.1 (both CDH3B4). We are 
using 64bit machines.

Starting goes great, only right after the first compaction we get this 
error:
Uncaught exception in service thread regionserver60020.compactor
java.lang.AbstractMethodError: 
com.hadoop.compression.lzo.LzoCompressor.reinit(Lorg/apache/hadoop/conf/Configuration;)V
         at 
org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:105)
         at 
org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
         at 
org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:200)
         at 
org.apache.hadoop.hbase.io.hfile.HFile$Writer.getCompressingStream(HFile.java:397)
         at 
org.apache.hadoop.hbase.io.hfile.HFile$Writer.newBlock(HFile.java:383)
         at 
org.apache.hadoop.hbase.io.hfile.HFile$Writer.checkBlockBoundary(HFile.java:354)
         at 
org.apache.hadoop.hbase.io.hfile.HFile$Writer.append(HFile.java:536)
         at 
org.apache.hadoop.hbase.io.hfile.HFile$Writer.append(HFile.java:501)
         at 
org.apache.hadoop.hbase.regionserver.StoreFile$Writer.append(StoreFile.java:836)
         at 
org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:935)
         at 
org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:733)
         at 
org.apache.hadoop.hbase.regionserver.HRegion.compactStores(HRegion.java:769)
         at 
org.apache.hadoop.hbase.regionserver.HRegion.compactStores(HRegion.java:714)
         at 
org.apache.hadoop.hbase.regionserver.CompactSplitThread.run(CompactSplitThread.java:81)

Lzo worked fine. This is how I believe we used it.
# LZO compression in Hbase will pass through three layers:
# 1) hadoop-gpl-compression-*.jar in the hbase/lib directory; the entry 
point
# 2) libgplcompression.* in the hbase native lib directory; the native 
connectors
# 3) liblzo2.so.2 in the hbase native lib directory; the base native library

Anyway, it would be great if somebody could help us out.

Re: after upgrade, fatal error in regionserver compacter, LzoCompressor, "AbstractMethodError"

Posted by Jiajun Chen <cj...@gmail.com>.
use this version http://code.google.com/a/apache-extras.org/p/hadoop-gpl-
compression/?redir=1

add codes in LzoDecompressor

@Override
    public int getRemaining()
    {
        return uncompressedDirectBuf.remaining();
    }

change return type in LzopCodec

protected int getCompressedData()
{
...
return len;
}

and then ant build successful .

On 10 April 2012 18:10, Jiajun Chen <cj...@gmail.com> wrote:

> Why can't i use this version:
> http://code.google.com/a/apache-extras.org/p/hadoop-gpl-compression/?redir=1
>
>
> On 17 March 2011 17:29, Ferdy Galema <fe...@kalooga.com> wrote:
>
>> Updating the lzo libraries resolved the problem. Thanks for pointing it
>> out and thanks to Todd Lipcon for his hadoop-lzo-packager.
>>
>>
>> On 03/16/2011 06:35 PM, Stack wrote:
>>
>>> Poking in our mail archives, does it help?  For example:
>>> http://search-hadoop.com/m/**QMDV41Sh1GI/lzo+compression&**
>>> subj=LZO+Compression<http://search-hadoop.com/m/QMDV41Sh1GI/lzo+compression&subj=LZO+Compression>
>>> St.Ack
>>>
>>> On Wed, Mar 16, 2011 at 10:28 AM, Ferdy Galema<ferdy.galema@kalooga.**
>>> com <fe...@kalooga.com>>  wrote:
>>>
>>>> We upgraded to Hadoop 0.20.1 and Hbase 0.90.1 (both CDH3B4). We are
>>>> using
>>>> 64bit machines.
>>>>
>>>> Starting goes great, only right after the first compaction we get this
>>>> error:
>>>> Uncaught exception in service thread regionserver60020.compactor
>>>> java.lang.AbstractMethodError:
>>>> com.hadoop.compression.lzo.**LzoCompressor.reinit(Lorg/**
>>>> apache/hadoop/conf/**Configuration;)V
>>>>        at
>>>> org.apache.hadoop.io.compress.**CodecPool.getCompressor(**
>>>> CodecPool.java:105)
>>>>        at
>>>> org.apache.hadoop.io.compress.**CodecPool.getCompressor(**
>>>> CodecPool.java:112)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.Compression$Algorithm.**
>>>> getCompressor(Compression.**java:200)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.**
>>>> getCompressingStream(HFile.**java:397)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.newBlock(**
>>>> HFile.java:383)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.**
>>>> checkBlockBoundary(HFile.java:**354)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.append(**
>>>> HFile.java:536)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.append(**
>>>> HFile.java:501)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.StoreFile$Writer.**
>>>> append(StoreFile.java:836)
>>>>        at org.apache.hadoop.hbase.**regionserver.Store.compact(**
>>>> Store.java:935)
>>>>        at org.apache.hadoop.hbase.**regionserver.Store.compact(**
>>>> Store.java:733)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.HRegion.**
>>>> compactStores(HRegion.java:**769)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.HRegion.**
>>>> compactStores(HRegion.java:**714)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.**CompactSplitThread.run(**
>>>> CompactSplitThread.java:81)
>>>>
>>>> Lzo worked fine. This is how I believe we used it.
>>>> # LZO compression in Hbase will pass through three layers:
>>>> # 1) hadoop-gpl-compression-*.jar in the hbase/lib directory; the entry
>>>> point
>>>> # 2) libgplcompression.* in the hbase native lib directory; the native
>>>> connectors
>>>> # 3) liblzo2.so.2 in the hbase native lib directory; the base native
>>>> library
>>>>
>>>> Anyway, it would be great if somebody could help us out.
>>>>
>>>>
>
>
> --
>
> 陈加俊  项目经理
> 优讯时代(北京)网络技术有限公司
> 优讯网 www.uuwatch.com
>
> 地址:北京市海淀区上地五街7号昊海大厦207
>
> 电话:010-82895510
> 传真:010-82896636
> 手机:15110038983
> 电邮:*cjjvictory@gmail.com*
>
>


-- 

陈加俊  项目经理
优讯时代(北京)网络技术有限公司
优讯网 www.uuwatch.com

地址:北京市海淀区上地五街7号昊海大厦207

电话:010-82895510
传真:010-82896636
手机:15110038983
电邮:*cjjvictory@gmail.com*

Re: after upgrade, fatal error in regionserver compacter, LzoCompressor, "AbstractMethodError"

Posted by Jiajun Chen <cj...@gmail.com>.
I replaced the hadoop-core-0.20.2-cdh3u1.jar to hadoop-core-1.0.1.jar ,but
build failed .

compile-java:
    [javac] Compiling 24 source files to
/app/setup/cloud/toddlipcon-hadoop-lzo-c7d54ff/build/classes
    [javac]
/app/setup/cloud/toddlipcon-hadoop-lzo-c7d54ff/src/java/com/hadoop/compression/lzo/LzoDecompressor.java:34:
com.hadoop.compression.lzo.LzoDecompressor is not abstract and does not
override abstract method getRemaining() in
org.apache.hadoop.io.compress.Decompressor
    [javac] class LzoDecompressor implements Decompressor {
    [javac] ^
    [javac]
/app/setup/cloud/toddlipcon-hadoop-lzo-c7d54ff/src/java/com/hadoop/compression/lzo/LzopInputStream.java:277:
getCompressedData() in com.hadoop.compression.lzo.LzopInputStream cannot
override getCompressedData() in
org.apache.hadoop.io.compress.BlockDecompressorStream; attempting to use
incompatible return type
    [javac] found   : void
    [javac] required: int
    [javac]   protected void getCompressedData() throws IOException {
    [javac]                  ^
    [javac]
/app/setup/cloud/toddlipcon-hadoop-lzo-c7d54ff/src/java/com/hadoop/compression/lzo/LzopInputStream.java:276:
method does not override or implement a method from a supertype
    [javac]   @Override
    [javac]   ^
    [javac]
/app/setup/cloud/toddlipcon-hadoop-lzo-c7d54ff/src/java/com/hadoop/mapreduce/LzoIndexOutputFormat.java:31:
warning: [deprecation] cleanupJob(org.apache.hadoop.mapreduce.JobContext)
in org.apache.hadoop.mapreduce.OutputCommitter has been deprecated
    [javac]       @Override public void cleanupJob(JobContext jobContext)
throws IOException {}
    [javac]                             ^
    [javac] 3 errors
    [javac] 1 warning

BUILD FAILED
/app/setup/cloud/toddlipcon-hadoop-lzo-c7d54ff/build.xml:216: Compile
failed; see the compiler error output for details.


On 10 April 2012 18:10, Jiajun Chen <cj...@gmail.com> wrote:

> Why can't i use this version:
> http://code.google.com/a/apache-extras.org/p/hadoop-gpl-compression/?redir=1
>
>
> On 17 March 2011 17:29, Ferdy Galema <fe...@kalooga.com> wrote:
>
>> Updating the lzo libraries resolved the problem. Thanks for pointing it
>> out and thanks to Todd Lipcon for his hadoop-lzo-packager.
>>
>>
>> On 03/16/2011 06:35 PM, Stack wrote:
>>
>>> Poking in our mail archives, does it help?  For example:
>>> http://search-hadoop.com/m/**QMDV41Sh1GI/lzo+compression&**
>>> subj=LZO+Compression<http://search-hadoop.com/m/QMDV41Sh1GI/lzo+compression&subj=LZO+Compression>
>>> St.Ack
>>>
>>> On Wed, Mar 16, 2011 at 10:28 AM, Ferdy Galema<ferdy.galema@kalooga.**
>>> com <fe...@kalooga.com>>  wrote:
>>>
>>>> We upgraded to Hadoop 0.20.1 and Hbase 0.90.1 (both CDH3B4). We are
>>>> using
>>>> 64bit machines.
>>>>
>>>> Starting goes great, only right after the first compaction we get this
>>>> error:
>>>> Uncaught exception in service thread regionserver60020.compactor
>>>> java.lang.AbstractMethodError:
>>>> com.hadoop.compression.lzo.**LzoCompressor.reinit(Lorg/**
>>>> apache/hadoop/conf/**Configuration;)V
>>>>        at
>>>> org.apache.hadoop.io.compress.**CodecPool.getCompressor(**
>>>> CodecPool.java:105)
>>>>        at
>>>> org.apache.hadoop.io.compress.**CodecPool.getCompressor(**
>>>> CodecPool.java:112)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.Compression$Algorithm.**
>>>> getCompressor(Compression.**java:200)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.**
>>>> getCompressingStream(HFile.**java:397)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.newBlock(**
>>>> HFile.java:383)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.**
>>>> checkBlockBoundary(HFile.java:**354)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.append(**
>>>> HFile.java:536)
>>>>        at
>>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.append(**
>>>> HFile.java:501)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.StoreFile$Writer.**
>>>> append(StoreFile.java:836)
>>>>        at org.apache.hadoop.hbase.**regionserver.Store.compact(**
>>>> Store.java:935)
>>>>        at org.apache.hadoop.hbase.**regionserver.Store.compact(**
>>>> Store.java:733)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.HRegion.**
>>>> compactStores(HRegion.java:**769)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.HRegion.**
>>>> compactStores(HRegion.java:**714)
>>>>        at
>>>> org.apache.hadoop.hbase.**regionserver.**CompactSplitThread.run(**
>>>> CompactSplitThread.java:81)
>>>>
>>>> Lzo worked fine. This is how I believe we used it.
>>>> # LZO compression in Hbase will pass through three layers:
>>>> # 1) hadoop-gpl-compression-*.jar in the hbase/lib directory; the entry
>>>> point
>>>> # 2) libgplcompression.* in the hbase native lib directory; the native
>>>> connectors
>>>> # 3) liblzo2.so.2 in the hbase native lib directory; the base native
>>>> library
>>>>
>>>> Anyway, it would be great if somebody could help us out.
>>>>
>>>>
>
>
> --
>
> 陈加俊  项目经理
> 优讯时代(北京)网络技术有限公司
> 优讯网 www.uuwatch.com
>
> 地址:北京市海淀区上地五街7号昊海大厦207
>
> 电话:010-82895510
> 传真:010-82896636
> 手机:15110038983
> 电邮:*cjjvictory@gmail.com*
>
>


-- 

陈加俊  项目经理
优讯时代(北京)网络技术有限公司
优讯网 www.uuwatch.com

地址:北京市海淀区上地五街7号昊海大厦207

电话:010-82895510
传真:010-82896636
手机:15110038983
电邮:*cjjvictory@gmail.com*

Re: after upgrade, fatal error in regionserver compacter, LzoCompressor, "AbstractMethodError"

Posted by Jiajun Chen <cj...@gmail.com>.
Why can't i use this version:
http://code.google.com/a/apache-extras.org/p/hadoop-gpl-compression/?redir=1


On 17 March 2011 17:29, Ferdy Galema <fe...@kalooga.com> wrote:

> Updating the lzo libraries resolved the problem. Thanks for pointing it
> out and thanks to Todd Lipcon for his hadoop-lzo-packager.
>
>
> On 03/16/2011 06:35 PM, Stack wrote:
>
>> Poking in our mail archives, does it help?  For example:
>> http://search-hadoop.com/m/**QMDV41Sh1GI/lzo+compression&**
>> subj=LZO+Compression<http://search-hadoop.com/m/QMDV41Sh1GI/lzo+compression&subj=LZO+Compression>
>> St.Ack
>>
>> On Wed, Mar 16, 2011 at 10:28 AM, Ferdy Galema<fe...@kalooga.com>>
>>  wrote:
>>
>>> We upgraded to Hadoop 0.20.1 and Hbase 0.90.1 (both CDH3B4). We are using
>>> 64bit machines.
>>>
>>> Starting goes great, only right after the first compaction we get this
>>> error:
>>> Uncaught exception in service thread regionserver60020.compactor
>>> java.lang.AbstractMethodError:
>>> com.hadoop.compression.lzo.**LzoCompressor.reinit(Lorg/**
>>> apache/hadoop/conf/**Configuration;)V
>>>        at
>>> org.apache.hadoop.io.compress.**CodecPool.getCompressor(**
>>> CodecPool.java:105)
>>>        at
>>> org.apache.hadoop.io.compress.**CodecPool.getCompressor(**
>>> CodecPool.java:112)
>>>        at
>>> org.apache.hadoop.hbase.io.**hfile.Compression$Algorithm.**
>>> getCompressor(Compression.**java:200)
>>>        at
>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.**
>>> getCompressingStream(HFile.**java:397)
>>>        at
>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.newBlock(**
>>> HFile.java:383)
>>>        at
>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.**
>>> checkBlockBoundary(HFile.java:**354)
>>>        at
>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.append(**HFile.java:536)
>>>        at
>>> org.apache.hadoop.hbase.io.**hfile.HFile$Writer.append(**HFile.java:501)
>>>        at
>>> org.apache.hadoop.hbase.**regionserver.StoreFile$Writer.**
>>> append(StoreFile.java:836)
>>>        at org.apache.hadoop.hbase.**regionserver.Store.compact(**
>>> Store.java:935)
>>>        at org.apache.hadoop.hbase.**regionserver.Store.compact(**
>>> Store.java:733)
>>>        at
>>> org.apache.hadoop.hbase.**regionserver.HRegion.**
>>> compactStores(HRegion.java:**769)
>>>        at
>>> org.apache.hadoop.hbase.**regionserver.HRegion.**
>>> compactStores(HRegion.java:**714)
>>>        at
>>> org.apache.hadoop.hbase.**regionserver.**CompactSplitThread.run(**
>>> CompactSplitThread.java:81)
>>>
>>> Lzo worked fine. This is how I believe we used it.
>>> # LZO compression in Hbase will pass through three layers:
>>> # 1) hadoop-gpl-compression-*.jar in the hbase/lib directory; the entry
>>> point
>>> # 2) libgplcompression.* in the hbase native lib directory; the native
>>> connectors
>>> # 3) liblzo2.so.2 in the hbase native lib directory; the base native
>>> library
>>>
>>> Anyway, it would be great if somebody could help us out.
>>>
>>>


-- 

陈加俊  项目经理
优讯时代(北京)网络技术有限公司
优讯网 www.uuwatch.com

地址:北京市海淀区上地五街7号昊海大厦207

电话:010-82895510
传真:010-82896636
手机:15110038983
电邮:*cjjvictory@gmail.com*

Re: after upgrade, fatal error in regionserver compacter, LzoCompressor, "AbstractMethodError"

Posted by Ferdy Galema <fe...@kalooga.com>.
Updating the lzo libraries resolved the problem. Thanks for pointing it 
out and thanks to Todd Lipcon for his hadoop-lzo-packager.

On 03/16/2011 06:35 PM, Stack wrote:
> Poking in our mail archives, does it help?  For example:
> http://search-hadoop.com/m/QMDV41Sh1GI/lzo+compression&subj=LZO+Compression
> St.Ack
>
> On Wed, Mar 16, 2011 at 10:28 AM, Ferdy Galema<fe...@kalooga.com>  wrote:
>> We upgraded to Hadoop 0.20.1 and Hbase 0.90.1 (both CDH3B4). We are using
>> 64bit machines.
>>
>> Starting goes great, only right after the first compaction we get this
>> error:
>> Uncaught exception in service thread regionserver60020.compactor
>> java.lang.AbstractMethodError:
>> com.hadoop.compression.lzo.LzoCompressor.reinit(Lorg/apache/hadoop/conf/Configuration;)V
>>         at
>> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:105)
>>         at
>> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
>>         at
>> org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:200)
>>         at
>> org.apache.hadoop.hbase.io.hfile.HFile$Writer.getCompressingStream(HFile.java:397)
>>         at
>> org.apache.hadoop.hbase.io.hfile.HFile$Writer.newBlock(HFile.java:383)
>>         at
>> org.apache.hadoop.hbase.io.hfile.HFile$Writer.checkBlockBoundary(HFile.java:354)
>>         at
>> org.apache.hadoop.hbase.io.hfile.HFile$Writer.append(HFile.java:536)
>>         at
>> org.apache.hadoop.hbase.io.hfile.HFile$Writer.append(HFile.java:501)
>>         at
>> org.apache.hadoop.hbase.regionserver.StoreFile$Writer.append(StoreFile.java:836)
>>         at org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:935)
>>         at org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:733)
>>         at
>> org.apache.hadoop.hbase.regionserver.HRegion.compactStores(HRegion.java:769)
>>         at
>> org.apache.hadoop.hbase.regionserver.HRegion.compactStores(HRegion.java:714)
>>         at
>> org.apache.hadoop.hbase.regionserver.CompactSplitThread.run(CompactSplitThread.java:81)
>>
>> Lzo worked fine. This is how I believe we used it.
>> # LZO compression in Hbase will pass through three layers:
>> # 1) hadoop-gpl-compression-*.jar in the hbase/lib directory; the entry
>> point
>> # 2) libgplcompression.* in the hbase native lib directory; the native
>> connectors
>> # 3) liblzo2.so.2 in the hbase native lib directory; the base native library
>>
>> Anyway, it would be great if somebody could help us out.
>>

Re: after upgrade, fatal error in regionserver compacter, LzoCompressor, "AbstractMethodError"

Posted by Stack <st...@duboce.net>.
Poking in our mail archives, does it help?  For example:
http://search-hadoop.com/m/QMDV41Sh1GI/lzo+compression&subj=LZO+Compression
St.Ack

On Wed, Mar 16, 2011 at 10:28 AM, Ferdy Galema <fe...@kalooga.com> wrote:
> We upgraded to Hadoop 0.20.1 and Hbase 0.90.1 (both CDH3B4). We are using
> 64bit machines.
>
> Starting goes great, only right after the first compaction we get this
> error:
> Uncaught exception in service thread regionserver60020.compactor
> java.lang.AbstractMethodError:
> com.hadoop.compression.lzo.LzoCompressor.reinit(Lorg/apache/hadoop/conf/Configuration;)V
>        at
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:105)
>        at
> org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
>        at
> org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:200)
>        at
> org.apache.hadoop.hbase.io.hfile.HFile$Writer.getCompressingStream(HFile.java:397)
>        at
> org.apache.hadoop.hbase.io.hfile.HFile$Writer.newBlock(HFile.java:383)
>        at
> org.apache.hadoop.hbase.io.hfile.HFile$Writer.checkBlockBoundary(HFile.java:354)
>        at
> org.apache.hadoop.hbase.io.hfile.HFile$Writer.append(HFile.java:536)
>        at
> org.apache.hadoop.hbase.io.hfile.HFile$Writer.append(HFile.java:501)
>        at
> org.apache.hadoop.hbase.regionserver.StoreFile$Writer.append(StoreFile.java:836)
>        at org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:935)
>        at org.apache.hadoop.hbase.regionserver.Store.compact(Store.java:733)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.compactStores(HRegion.java:769)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.compactStores(HRegion.java:714)
>        at
> org.apache.hadoop.hbase.regionserver.CompactSplitThread.run(CompactSplitThread.java:81)
>
> Lzo worked fine. This is how I believe we used it.
> # LZO compression in Hbase will pass through three layers:
> # 1) hadoop-gpl-compression-*.jar in the hbase/lib directory; the entry
> point
> # 2) libgplcompression.* in the hbase native lib directory; the native
> connectors
> # 3) liblzo2.so.2 in the hbase native lib directory; the base native library
>
> Anyway, it would be great if somebody could help us out.
>