You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@giraph.apache.org by Toshio ITO <to...@toshiba.co.jp> on 2014/06/26 12:34:25 UTC

Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Hi all,

I recently tried Giraph.

It compiled fine, but when I did the unit test (i.e. `mvn test`) it
failed.  I did the test several times with different settings. See
below for details.

My question is:

- Is it normal for Giraph? Is the unit test supposed to fail for now?
- If it's not normal, what do you think is wrong?


My environment:

- Ubuntu Server 14.04 64-bit
- Dynamic address by DHCP
- Followed the "Quick Start" guide
  https://giraph.apache.org/quick_start.html
  except I used "localhost" for the node hostname.
-- Installed openjdk-7-jdk, git, maven by apt-get
-- Installed hadoop-0.20.203.0rc1 as instructed


Test settings:

(1) 1.1.0-SNAPSHOT (Commit ID: b218d72cedc52467e691c6002e596e482d8583e4)
    with LocalJobRunner
(2) 1.1.0-SNAPSHOT with the running Hadoop instance
(3) Tag "release-1.0.0-RC3" with LocalJobRunner
(4) Tag "release-1.0.0-RC3" with the running Hadoop instance

Cases (1) and (3) are run by the following commands.

    $ mvn clean
    $ mvn package -Phadoop_0.20.203 -DskipTests
    $ mvn test -Phadoop_0.20.203

Cases (2) and (4) are run by the following commands.

    $ mvn clean
    $ mvn package -Phadoop_0.20.203 -DskipTests
    $ mvn test -Phadoop_0.20.203 -Dprop.mapred.job.tracker=localhost:54311


Test results:

(1) Failed at "Rexster I/O Format". At some point, the test
endlessly tried to connect to ZooKeeper, repeating the following log
messages. I had to terminate the test by Ctrl-C.

    14/06/26 18:10:27 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:22182. Will not attempt to authenticate using SASL (unknown error)
    14/06/26 18:10:27 WARN zookeeper.ClientCnxn: Session 0x146d76e16750002 for server null, unexpected error, closing socket connection and attempting reconnect
    java.net.ConnectException: Connection refused
            at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
            at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
            at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
            at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)


(2) Failed at "Examples". Hadoop was stuck at "testBspFail"
job. Logs like the following line were printed every 5 seconds for
10 minutes.  After that, the test endlessly tried to connect to
ZooKeeper, just like the case (1). I killed the test process and the
Hadoop job.

    14/06/26 18:40:30 INFO job.JobProgressTracker: Data from 2 workers - Compute superstep 20: 10 out of 10 vertices computed; 6 out of 6 partitions computed; min free memory on worker 1 - 137.42MB, average 149.12MB


(3) Failed at "HBase I/O" in testHBaseInputOutput
(org.apache.giraph.io.hbase.TestHBaseRootMarkerVertextFormat). Before
it reported failure, it blocked for 10 minutes.


(4) Failed at "Core" in testContinue
(org.apache.giraph.io.TestJsonBase64Format) . Hadoop was stuck at
the second "testContinue" job. After 10 minutes, the test went on
and reported the failure. The "testContinue" map task was aborted
with the following error.

    java.io.FileNotFoundException: /tmp/_giraphTests/testContinue/_logs (Is a directory)



Thanks in advance.

------------------------------------
Toshio Ito

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Toshio ITO <to...@toshiba.co.jp>.
Hi Akila,

Thank you for the pointer.

I think this is it. I got the following error messages.


14/07/08 13:42:36 INFO http.HttpServer: Jetty bound to port 60030
14/07/08 13:42:36 INFO mortbay.log: jetty-6.1.26
14/07/08 13:42:36 INFO http.HttpServer: Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
14/07/08 13:42:36 INFO http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener on 60030
14/07/08 13:42:36 INFO regionserver.HRegionServer: STOPPED: Failed initialization
14/07/08 13:42:36 ERROR regionserver.HRegionServer: Failed init
java.net.BindException: Address already in use
        at sun.nio.ch.Net.bind0(Native Method)
        at sun.nio.ch.Net.bind(Net.java:444)
        at sun.nio.ch.Net.bind(Net.java:436)
        at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
        at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
        at org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
        at org.apache.hadoop.http.HttpServer.start(HttpServer.java:562)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.startServiceThreads(HRegionServer.java:1292)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.handleReportForDutyResponse(HRegionServer.java:890)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.handleReportForDutyResponse(MiniHBaseCluster.java:187)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.tryReportForDuty(HRegionServer.java:1509)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:568)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(MiniHBaseCluster.java:213)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.access$000(MiniHBaseCluster.java:163)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer$1.run(MiniHBaseCluster.java:197)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:356)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1042)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.hadoop.hbase.util.Methods.call(Methods.java:37)
        at org.apache.hadoop.hbase.security.User.call(User.java:457)
        at org.apache.hadoop.hbase.security.User.access$500(User.java:49)
        at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:346)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.run(MiniHBaseCluster.java:195)
        at java.lang.Thread.run(Thread.java:744)

(snip)

14/07/08 13:42:36 FATAL regionserver.HRegionServer: ABORTING region server serverName=localhost,34825,1404794553051, load=(requests=0, regions=0, usedHeap=110, maxHeap=21465): Unhandled exception: Address already in use
java.net.BindException: Address already in use

(snip)

14/07/08 13:42:39 WARN master.AssignmentManager: Failed assignment of -ROOT-,,0.70236052 to serverName=localhost,34825,1404794553051, load=(requests=0, regions=0, usedHeap=0, maxHeap=0), trying to assign elsewhere instead; retry=0
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed setting up proxy interface org.apache.hadoop.hbase.ipc.HRegionInterface to /127.0.0.1:34825 after attempts=1


The last message repeated 10 times.



> 
> [1  <text/plain; ISO-8859-1 (7bit)>]
> Hi Toshio, Roman,
> 
> The HBase I/O test failure (no 3) happens may be due to this issue
> 
> https://issues.apache.org/jira/browse/GIRAPH-926
> 
> Toshio, Can you check whether you get an error similar to this?
> 
> 14/07/07 15:53:58 INFO hbase.metrics: new MBeanInfo
> 14/07/07 15:53:58 INFO metrics.RegionServerMetrics: Initialized
> 14/07/07 15:53:58 INFO http.HttpServer: Added global filtersafety
> (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
> 14/07/07 15:53:58 INFO http.HttpServer: Port returned by
> webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening
> the listener on 60030
> 14/07/07 15:53:58 WARN regionserver.HRegionServer: Exception in region
> server :
> java.net.BindException: Address already in use
>         at sun.nio.ch.Net.bind0(Native Method)
>         at sun.nio.ch.Net.bind(Net.java:444)
>         at sun.nio.ch.Net.bind(Net.java:436)
>         at
> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
>         at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
>         at
> org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
>         at org.apache.hadoop.http.HttpServer.start(HttpServer.java:602)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.putUpWebUI(HRegionServer.java:1760)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.startServiceThreads(HRegionServer.java:1715)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.handleReportForDutyResponse(HRegionServer.java:1108)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.handleReportForDutyResponse(MiniHBaseCluster.java:122)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:752)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(MiniHBaseCluster.java:148)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.access$000(MiniHBaseCluster.java:101)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer$1.run(MiniHBaseCluster.java:132)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:356)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1172)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at org.apache.hadoop.hbase.util.Methods.call(Methods.java:37)
>         at org.apache.hadoop.hbase.security.User.call(User.java:624)
>         at org.apache.hadoop.hbase.security.User.access$600(User.java:52)
>         at
> org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:464)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.run(MiniHBaseCluster.java:130)
>         at java.lang.Thread.run(Thread.java:724)
> 14/07/07 15:53:58 INFO regionserver.HRegionServer: STOPPED: Failed
> initialization
> 14/07/07 15:53:58 ERROR regionserver.HRegionServer: Failed init
> java.net.BindException: Address already in use
> 
> Thanks,
> 
> Akila
> 
> 
> On Wed, Jul 2, 2014 at 12:10 AM, Roman Shaposhnik <ro...@shaposhnik.org>
> wrote:
> 
> > Yes, the failures around Accumulo in hadoop_2 profile are expected and
> > nothing
> > to worry about. I should've probably mentioned it in my RC announcement
> > email.
> > Sorry about that.
> >
> > Any failures in hadoop_1 profile would be a reason to reconsider RC0.
> >
> > Thanks,
> > Roman.
> >
> > P.S. This is one of the reasons we're still running with hadoop_1 as a
> > default
> > profile.
> >
> > On Mon, Jun 30, 2014 at 3:09 AM, Akila Wajirasena
> > <ak...@gmail.com> wrote:
> > > Hi Roman,
> > >
> > > I got the same error when running hadoop_2 profile.
> > > According to this [1] the Accumulo version we use in giraph (1.4) is not
> > > compatible with Hadoop 2.
> > > I think this is the issue.
> > >
> > > [1]
> > >
> > http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html
> > >
> > > Thanks
> > >
> > > Akila
> > >
> > >
> > > On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
> > > wrote:
> > >>
> > >> Hi Roman.
> > >>
> > >> I checked out release-1.1.0-RC0 and succeeded to build it.
> > >>
> > >>     $ git checkout release-1.1.0-RC0
> > >>     $ mvn clean
> > >>     $ mvn package -Phadoop_2 -DskipTests
> > >>     ## SUCCESS
> > >>
> > >> However, when I ran the tests with LocalJobRunner, it failed.
> > >>
> > >>     $ mvn clean
> > >>     $ mvn package -Phadoop_2
> > >>
> > >> It passed tests from "Core" and "Examples", but it failed at
> > >> "Accumulo I/O".
> > >>
> > >>
> > >>
> > testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
> > >>
> > >> The error log contained the following exception
> > >>
> > >>     java.lang.IncompatibleClassChangeError: Found interface
> > >> org.apache.hadoop.mapreduce.JobContext, but class was expected
> > >>
> > >>
> > >> Next I wanted to run the tests with a running Hadoop2 instance, but
> > >> I'm having trouble to set it up (I'm quite new to Hadoop).
> > >>
> > >> Could you show me some example configuration (etc/hadoop/* files) of
> > >> Hadoop 2.2.0 single-node cluster? That would be very helpful.
> > >>
> > >>
> > >>
> > >>
> > >> >
> > >> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <
> > toshio9.ito@toshiba.co.jp>
> > >> > wrote:
> > >> > > Hi Roman.
> > >> > >
> > >> > > Thanks for the reply.
> > >> > >
> > >> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
> > >> > > release-1.1.0-RC0 and report the result.
> > >> >
> > >> > That would be extremely helpful!
> > >> >
> > >> > And speaking of which -- I'd like to remind folks
> > >> > that taking RC0 for a spin would really help
> > >> > at this point. If we ever want to have 1.1.0 out
> > >> > we need the required PMC votes.
> > >> >
> > >> > Thanks,
> > >> > Roman.
> > >> ------------------------------------
> > >> Toshio Ito
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> >
> 
> 
> 
> -- 
> Regards
> Akila Wajirasena
> [2  <text/html; ISO-8859-1 (quoted-printable)>]
> 
------------------------------------
Toshio Ito

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Toshio ITO <to...@toshiba.co.jp>.
Hi Akila,

Thank you for the pointer.

I think this is it. I got the following error messages.


14/07/08 13:42:36 INFO http.HttpServer: Jetty bound to port 60030
14/07/08 13:42:36 INFO mortbay.log: jetty-6.1.26
14/07/08 13:42:36 INFO http.HttpServer: Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
14/07/08 13:42:36 INFO http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener on 60030
14/07/08 13:42:36 INFO regionserver.HRegionServer: STOPPED: Failed initialization
14/07/08 13:42:36 ERROR regionserver.HRegionServer: Failed init
java.net.BindException: Address already in use
        at sun.nio.ch.Net.bind0(Native Method)
        at sun.nio.ch.Net.bind(Net.java:444)
        at sun.nio.ch.Net.bind(Net.java:436)
        at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
        at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
        at org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
        at org.apache.hadoop.http.HttpServer.start(HttpServer.java:562)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.startServiceThreads(HRegionServer.java:1292)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.handleReportForDutyResponse(HRegionServer.java:890)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.handleReportForDutyResponse(MiniHBaseCluster.java:187)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.tryReportForDuty(HRegionServer.java:1509)
        at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:568)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(MiniHBaseCluster.java:213)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.access$000(MiniHBaseCluster.java:163)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer$1.run(MiniHBaseCluster.java:197)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:356)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1042)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.hadoop.hbase.util.Methods.call(Methods.java:37)
        at org.apache.hadoop.hbase.security.User.call(User.java:457)
        at org.apache.hadoop.hbase.security.User.access$500(User.java:49)
        at org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:346)
        at org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.run(MiniHBaseCluster.java:195)
        at java.lang.Thread.run(Thread.java:744)

(snip)

14/07/08 13:42:36 FATAL regionserver.HRegionServer: ABORTING region server serverName=localhost,34825,1404794553051, load=(requests=0, regions=0, usedHeap=110, maxHeap=21465): Unhandled exception: Address already in use
java.net.BindException: Address already in use

(snip)

14/07/08 13:42:39 WARN master.AssignmentManager: Failed assignment of -ROOT-,,0.70236052 to serverName=localhost,34825,1404794553051, load=(requests=0, regions=0, usedHeap=0, maxHeap=0), trying to assign elsewhere instead; retry=0
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed setting up proxy interface org.apache.hadoop.hbase.ipc.HRegionInterface to /127.0.0.1:34825 after attempts=1


The last message repeated 10 times.



> 
> [1  <text/plain; ISO-8859-1 (7bit)>]
> Hi Toshio, Roman,
> 
> The HBase I/O test failure (no 3) happens may be due to this issue
> 
> https://issues.apache.org/jira/browse/GIRAPH-926
> 
> Toshio, Can you check whether you get an error similar to this?
> 
> 14/07/07 15:53:58 INFO hbase.metrics: new MBeanInfo
> 14/07/07 15:53:58 INFO metrics.RegionServerMetrics: Initialized
> 14/07/07 15:53:58 INFO http.HttpServer: Added global filtersafety
> (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
> 14/07/07 15:53:58 INFO http.HttpServer: Port returned by
> webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening
> the listener on 60030
> 14/07/07 15:53:58 WARN regionserver.HRegionServer: Exception in region
> server :
> java.net.BindException: Address already in use
>         at sun.nio.ch.Net.bind0(Native Method)
>         at sun.nio.ch.Net.bind(Net.java:444)
>         at sun.nio.ch.Net.bind(Net.java:436)
>         at
> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
>         at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
>         at
> org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
>         at org.apache.hadoop.http.HttpServer.start(HttpServer.java:602)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.putUpWebUI(HRegionServer.java:1760)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.startServiceThreads(HRegionServer.java:1715)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.handleReportForDutyResponse(HRegionServer.java:1108)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.handleReportForDutyResponse(MiniHBaseCluster.java:122)
>         at
> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:752)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(MiniHBaseCluster.java:148)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.access$000(MiniHBaseCluster.java:101)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer$1.run(MiniHBaseCluster.java:132)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:356)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1172)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at org.apache.hadoop.hbase.util.Methods.call(Methods.java:37)
>         at org.apache.hadoop.hbase.security.User.call(User.java:624)
>         at org.apache.hadoop.hbase.security.User.access$600(User.java:52)
>         at
> org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:464)
>         at
> org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.run(MiniHBaseCluster.java:130)
>         at java.lang.Thread.run(Thread.java:724)
> 14/07/07 15:53:58 INFO regionserver.HRegionServer: STOPPED: Failed
> initialization
> 14/07/07 15:53:58 ERROR regionserver.HRegionServer: Failed init
> java.net.BindException: Address already in use
> 
> Thanks,
> 
> Akila
> 
> 
> On Wed, Jul 2, 2014 at 12:10 AM, Roman Shaposhnik <ro...@shaposhnik.org>
> wrote:
> 
> > Yes, the failures around Accumulo in hadoop_2 profile are expected and
> > nothing
> > to worry about. I should've probably mentioned it in my RC announcement
> > email.
> > Sorry about that.
> >
> > Any failures in hadoop_1 profile would be a reason to reconsider RC0.
> >
> > Thanks,
> > Roman.
> >
> > P.S. This is one of the reasons we're still running with hadoop_1 as a
> > default
> > profile.
> >
> > On Mon, Jun 30, 2014 at 3:09 AM, Akila Wajirasena
> > <ak...@gmail.com> wrote:
> > > Hi Roman,
> > >
> > > I got the same error when running hadoop_2 profile.
> > > According to this [1] the Accumulo version we use in giraph (1.4) is not
> > > compatible with Hadoop 2.
> > > I think this is the issue.
> > >
> > > [1]
> > >
> > http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html
> > >
> > > Thanks
> > >
> > > Akila
> > >
> > >
> > > On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
> > > wrote:
> > >>
> > >> Hi Roman.
> > >>
> > >> I checked out release-1.1.0-RC0 and succeeded to build it.
> > >>
> > >>     $ git checkout release-1.1.0-RC0
> > >>     $ mvn clean
> > >>     $ mvn package -Phadoop_2 -DskipTests
> > >>     ## SUCCESS
> > >>
> > >> However, when I ran the tests with LocalJobRunner, it failed.
> > >>
> > >>     $ mvn clean
> > >>     $ mvn package -Phadoop_2
> > >>
> > >> It passed tests from "Core" and "Examples", but it failed at
> > >> "Accumulo I/O".
> > >>
> > >>
> > >>
> > testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
> > >>
> > >> The error log contained the following exception
> > >>
> > >>     java.lang.IncompatibleClassChangeError: Found interface
> > >> org.apache.hadoop.mapreduce.JobContext, but class was expected
> > >>
> > >>
> > >> Next I wanted to run the tests with a running Hadoop2 instance, but
> > >> I'm having trouble to set it up (I'm quite new to Hadoop).
> > >>
> > >> Could you show me some example configuration (etc/hadoop/* files) of
> > >> Hadoop 2.2.0 single-node cluster? That would be very helpful.
> > >>
> > >>
> > >>
> > >>
> > >> >
> > >> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <
> > toshio9.ito@toshiba.co.jp>
> > >> > wrote:
> > >> > > Hi Roman.
> > >> > >
> > >> > > Thanks for the reply.
> > >> > >
> > >> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
> > >> > > release-1.1.0-RC0 and report the result.
> > >> >
> > >> > That would be extremely helpful!
> > >> >
> > >> > And speaking of which -- I'd like to remind folks
> > >> > that taking RC0 for a spin would really help
> > >> > at this point. If we ever want to have 1.1.0 out
> > >> > we need the required PMC votes.
> > >> >
> > >> > Thanks,
> > >> > Roman.
> > >> ------------------------------------
> > >> Toshio Ito
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> >
> 
> 
> 
> -- 
> Regards
> Akila Wajirasena
> [2  <text/html; ISO-8859-1 (quoted-printable)>]
> 
------------------------------------
Toshio Ito

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Akila Wajirasena <ak...@gmail.com>.
Hi Toshio, Roman,

The HBase I/O test failure (no 3) happens may be due to this issue

https://issues.apache.org/jira/browse/GIRAPH-926

Toshio, Can you check whether you get an error similar to this?

14/07/07 15:53:58 INFO hbase.metrics: new MBeanInfo
14/07/07 15:53:58 INFO metrics.RegionServerMetrics: Initialized
14/07/07 15:53:58 INFO http.HttpServer: Added global filtersafety
(class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
14/07/07 15:53:58 INFO http.HttpServer: Port returned by
webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening
the listener on 60030
14/07/07 15:53:58 WARN regionserver.HRegionServer: Exception in region
server :
java.net.BindException: Address already in use
        at sun.nio.ch.Net.bind0(Native Method)
        at sun.nio.ch.Net.bind(Net.java:444)
        at sun.nio.ch.Net.bind(Net.java:436)
        at
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
        at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
        at
org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
        at org.apache.hadoop.http.HttpServer.start(HttpServer.java:602)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.putUpWebUI(HRegionServer.java:1760)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.startServiceThreads(HRegionServer.java:1715)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.handleReportForDutyResponse(HRegionServer.java:1108)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.handleReportForDutyResponse(MiniHBaseCluster.java:122)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:752)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(MiniHBaseCluster.java:148)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.access$000(MiniHBaseCluster.java:101)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer$1.run(MiniHBaseCluster.java:132)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:356)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1172)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.hadoop.hbase.util.Methods.call(Methods.java:37)
        at org.apache.hadoop.hbase.security.User.call(User.java:624)
        at org.apache.hadoop.hbase.security.User.access$600(User.java:52)
        at
org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:464)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.run(MiniHBaseCluster.java:130)
        at java.lang.Thread.run(Thread.java:724)
14/07/07 15:53:58 INFO regionserver.HRegionServer: STOPPED: Failed
initialization
14/07/07 15:53:58 ERROR regionserver.HRegionServer: Failed init
java.net.BindException: Address already in use

Thanks,

Akila


On Wed, Jul 2, 2014 at 12:10 AM, Roman Shaposhnik <ro...@shaposhnik.org>
wrote:

> Yes, the failures around Accumulo in hadoop_2 profile are expected and
> nothing
> to worry about. I should've probably mentioned it in my RC announcement
> email.
> Sorry about that.
>
> Any failures in hadoop_1 profile would be a reason to reconsider RC0.
>
> Thanks,
> Roman.
>
> P.S. This is one of the reasons we're still running with hadoop_1 as a
> default
> profile.
>
> On Mon, Jun 30, 2014 at 3:09 AM, Akila Wajirasena
> <ak...@gmail.com> wrote:
> > Hi Roman,
> >
> > I got the same error when running hadoop_2 profile.
> > According to this [1] the Accumulo version we use in giraph (1.4) is not
> > compatible with Hadoop 2.
> > I think this is the issue.
> >
> > [1]
> >
> http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html
> >
> > Thanks
> >
> > Akila
> >
> >
> > On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
> > wrote:
> >>
> >> Hi Roman.
> >>
> >> I checked out release-1.1.0-RC0 and succeeded to build it.
> >>
> >>     $ git checkout release-1.1.0-RC0
> >>     $ mvn clean
> >>     $ mvn package -Phadoop_2 -DskipTests
> >>     ## SUCCESS
> >>
> >> However, when I ran the tests with LocalJobRunner, it failed.
> >>
> >>     $ mvn clean
> >>     $ mvn package -Phadoop_2
> >>
> >> It passed tests from "Core" and "Examples", but it failed at
> >> "Accumulo I/O".
> >>
> >>
> >>
> testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
> >>
> >> The error log contained the following exception
> >>
> >>     java.lang.IncompatibleClassChangeError: Found interface
> >> org.apache.hadoop.mapreduce.JobContext, but class was expected
> >>
> >>
> >> Next I wanted to run the tests with a running Hadoop2 instance, but
> >> I'm having trouble to set it up (I'm quite new to Hadoop).
> >>
> >> Could you show me some example configuration (etc/hadoop/* files) of
> >> Hadoop 2.2.0 single-node cluster? That would be very helpful.
> >>
> >>
> >>
> >>
> >> >
> >> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <
> toshio9.ito@toshiba.co.jp>
> >> > wrote:
> >> > > Hi Roman.
> >> > >
> >> > > Thanks for the reply.
> >> > >
> >> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
> >> > > release-1.1.0-RC0 and report the result.
> >> >
> >> > That would be extremely helpful!
> >> >
> >> > And speaking of which -- I'd like to remind folks
> >> > that taking RC0 for a spin would really help
> >> > at this point. If we ever want to have 1.1.0 out
> >> > we need the required PMC votes.
> >> >
> >> > Thanks,
> >> > Roman.
> >> ------------------------------------
> >> Toshio Ito
> >
> >
> >
> >
> >
> >
> >
>



-- 
Regards
Akila Wajirasena

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Akila Wajirasena <ak...@gmail.com>.
Hi Toshio, Roman,

The HBase I/O test failure (no 3) happens may be due to this issue

https://issues.apache.org/jira/browse/GIRAPH-926

Toshio, Can you check whether you get an error similar to this?

14/07/07 15:53:58 INFO hbase.metrics: new MBeanInfo
14/07/07 15:53:58 INFO metrics.RegionServerMetrics: Initialized
14/07/07 15:53:58 INFO http.HttpServer: Added global filtersafety
(class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
14/07/07 15:53:58 INFO http.HttpServer: Port returned by
webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening
the listener on 60030
14/07/07 15:53:58 WARN regionserver.HRegionServer: Exception in region
server :
java.net.BindException: Address already in use
        at sun.nio.ch.Net.bind0(Native Method)
        at sun.nio.ch.Net.bind(Net.java:444)
        at sun.nio.ch.Net.bind(Net.java:436)
        at
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
        at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
        at
org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
        at org.apache.hadoop.http.HttpServer.start(HttpServer.java:602)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.putUpWebUI(HRegionServer.java:1760)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.startServiceThreads(HRegionServer.java:1715)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.handleReportForDutyResponse(HRegionServer.java:1108)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.handleReportForDutyResponse(MiniHBaseCluster.java:122)
        at
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:752)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.runRegionServer(MiniHBaseCluster.java:148)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.access$000(MiniHBaseCluster.java:101)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer$1.run(MiniHBaseCluster.java:132)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:356)
        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1172)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.hadoop.hbase.util.Methods.call(Methods.java:37)
        at org.apache.hadoop.hbase.security.User.call(User.java:624)
        at org.apache.hadoop.hbase.security.User.access$600(User.java:52)
        at
org.apache.hadoop.hbase.security.User$SecureHadoopUser.runAs(User.java:464)
        at
org.apache.hadoop.hbase.MiniHBaseCluster$MiniHBaseClusterRegionServer.run(MiniHBaseCluster.java:130)
        at java.lang.Thread.run(Thread.java:724)
14/07/07 15:53:58 INFO regionserver.HRegionServer: STOPPED: Failed
initialization
14/07/07 15:53:58 ERROR regionserver.HRegionServer: Failed init
java.net.BindException: Address already in use

Thanks,

Akila


On Wed, Jul 2, 2014 at 12:10 AM, Roman Shaposhnik <ro...@shaposhnik.org>
wrote:

> Yes, the failures around Accumulo in hadoop_2 profile are expected and
> nothing
> to worry about. I should've probably mentioned it in my RC announcement
> email.
> Sorry about that.
>
> Any failures in hadoop_1 profile would be a reason to reconsider RC0.
>
> Thanks,
> Roman.
>
> P.S. This is one of the reasons we're still running with hadoop_1 as a
> default
> profile.
>
> On Mon, Jun 30, 2014 at 3:09 AM, Akila Wajirasena
> <ak...@gmail.com> wrote:
> > Hi Roman,
> >
> > I got the same error when running hadoop_2 profile.
> > According to this [1] the Accumulo version we use in giraph (1.4) is not
> > compatible with Hadoop 2.
> > I think this is the issue.
> >
> > [1]
> >
> http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html
> >
> > Thanks
> >
> > Akila
> >
> >
> > On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
> > wrote:
> >>
> >> Hi Roman.
> >>
> >> I checked out release-1.1.0-RC0 and succeeded to build it.
> >>
> >>     $ git checkout release-1.1.0-RC0
> >>     $ mvn clean
> >>     $ mvn package -Phadoop_2 -DskipTests
> >>     ## SUCCESS
> >>
> >> However, when I ran the tests with LocalJobRunner, it failed.
> >>
> >>     $ mvn clean
> >>     $ mvn package -Phadoop_2
> >>
> >> It passed tests from "Core" and "Examples", but it failed at
> >> "Accumulo I/O".
> >>
> >>
> >>
> testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
> >>
> >> The error log contained the following exception
> >>
> >>     java.lang.IncompatibleClassChangeError: Found interface
> >> org.apache.hadoop.mapreduce.JobContext, but class was expected
> >>
> >>
> >> Next I wanted to run the tests with a running Hadoop2 instance, but
> >> I'm having trouble to set it up (I'm quite new to Hadoop).
> >>
> >> Could you show me some example configuration (etc/hadoop/* files) of
> >> Hadoop 2.2.0 single-node cluster? That would be very helpful.
> >>
> >>
> >>
> >>
> >> >
> >> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <
> toshio9.ito@toshiba.co.jp>
> >> > wrote:
> >> > > Hi Roman.
> >> > >
> >> > > Thanks for the reply.
> >> > >
> >> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
> >> > > release-1.1.0-RC0 and report the result.
> >> >
> >> > That would be extremely helpful!
> >> >
> >> > And speaking of which -- I'd like to remind folks
> >> > that taking RC0 for a spin would really help
> >> > at this point. If we ever want to have 1.1.0 out
> >> > we need the required PMC votes.
> >> >
> >> > Thanks,
> >> > Roman.
> >> ------------------------------------
> >> Toshio Ito
> >
> >
> >
> >
> >
> >
> >
>



-- 
Regards
Akila Wajirasena

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Roman Shaposhnik <ro...@shaposhnik.org>.
Yes, the failures around Accumulo in hadoop_2 profile are expected and nothing
to worry about. I should've probably mentioned it in my RC announcement email.
Sorry about that.

Any failures in hadoop_1 profile would be a reason to reconsider RC0.

Thanks,
Roman.

P.S. This is one of the reasons we're still running with hadoop_1 as a default
profile.

On Mon, Jun 30, 2014 at 3:09 AM, Akila Wajirasena
<ak...@gmail.com> wrote:
> Hi Roman,
>
> I got the same error when running hadoop_2 profile.
> According to this [1] the Accumulo version we use in giraph (1.4) is not
> compatible with Hadoop 2.
> I think this is the issue.
>
> [1]
> http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html
>
> Thanks
>
> Akila
>
>
> On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
> wrote:
>>
>> Hi Roman.
>>
>> I checked out release-1.1.0-RC0 and succeeded to build it.
>>
>>     $ git checkout release-1.1.0-RC0
>>     $ mvn clean
>>     $ mvn package -Phadoop_2 -DskipTests
>>     ## SUCCESS
>>
>> However, when I ran the tests with LocalJobRunner, it failed.
>>
>>     $ mvn clean
>>     $ mvn package -Phadoop_2
>>
>> It passed tests from "Core" and "Examples", but it failed at
>> "Accumulo I/O".
>>
>>
>> testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
>>
>> The error log contained the following exception
>>
>>     java.lang.IncompatibleClassChangeError: Found interface
>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>
>>
>> Next I wanted to run the tests with a running Hadoop2 instance, but
>> I'm having trouble to set it up (I'm quite new to Hadoop).
>>
>> Could you show me some example configuration (etc/hadoop/* files) of
>> Hadoop 2.2.0 single-node cluster? That would be very helpful.
>>
>>
>>
>>
>> >
>> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp>
>> > wrote:
>> > > Hi Roman.
>> > >
>> > > Thanks for the reply.
>> > >
>> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
>> > > release-1.1.0-RC0 and report the result.
>> >
>> > That would be extremely helpful!
>> >
>> > And speaking of which -- I'd like to remind folks
>> > that taking RC0 for a spin would really help
>> > at this point. If we ever want to have 1.1.0 out
>> > we need the required PMC votes.
>> >
>> > Thanks,
>> > Roman.
>> ------------------------------------
>> Toshio Ito
>
>
>
>
>
>
>

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Roman Shaposhnik <ro...@shaposhnik.org>.
Yes, the failures around Accumulo in hadoop_2 profile are expected and nothing
to worry about. I should've probably mentioned it in my RC announcement email.
Sorry about that.

Any failures in hadoop_1 profile would be a reason to reconsider RC0.

Thanks,
Roman.

P.S. This is one of the reasons we're still running with hadoop_1 as a default
profile.

On Mon, Jun 30, 2014 at 3:09 AM, Akila Wajirasena
<ak...@gmail.com> wrote:
> Hi Roman,
>
> I got the same error when running hadoop_2 profile.
> According to this [1] the Accumulo version we use in giraph (1.4) is not
> compatible with Hadoop 2.
> I think this is the issue.
>
> [1]
> http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html
>
> Thanks
>
> Akila
>
>
> On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
> wrote:
>>
>> Hi Roman.
>>
>> I checked out release-1.1.0-RC0 and succeeded to build it.
>>
>>     $ git checkout release-1.1.0-RC0
>>     $ mvn clean
>>     $ mvn package -Phadoop_2 -DskipTests
>>     ## SUCCESS
>>
>> However, when I ran the tests with LocalJobRunner, it failed.
>>
>>     $ mvn clean
>>     $ mvn package -Phadoop_2
>>
>> It passed tests from "Core" and "Examples", but it failed at
>> "Accumulo I/O".
>>
>>
>> testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
>>
>> The error log contained the following exception
>>
>>     java.lang.IncompatibleClassChangeError: Found interface
>> org.apache.hadoop.mapreduce.JobContext, but class was expected
>>
>>
>> Next I wanted to run the tests with a running Hadoop2 instance, but
>> I'm having trouble to set it up (I'm quite new to Hadoop).
>>
>> Could you show me some example configuration (etc/hadoop/* files) of
>> Hadoop 2.2.0 single-node cluster? That would be very helpful.
>>
>>
>>
>>
>> >
>> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp>
>> > wrote:
>> > > Hi Roman.
>> > >
>> > > Thanks for the reply.
>> > >
>> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
>> > > release-1.1.0-RC0 and report the result.
>> >
>> > That would be extremely helpful!
>> >
>> > And speaking of which -- I'd like to remind folks
>> > that taking RC0 for a spin would really help
>> > at this point. If we ever want to have 1.1.0 out
>> > we need the required PMC votes.
>> >
>> > Thanks,
>> > Roman.
>> ------------------------------------
>> Toshio Ito
>
>
>
>
>
>
>

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Akila Wajirasena <ak...@gmail.com>.
Hi Roman,

I got the same error when running hadoop_2 profile.
According to this [1] the Accumulo version we use in giraph (1.4) is not
compatible with Hadoop 2.
I think this is the issue.

[1]
http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html

Thanks

Akila

On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
wrote:

> Hi Roman.
>
> I checked out release-1.1.0-RC0 and succeeded to build it.
>
>     $ git checkout release-1.1.0-RC0
>     $ mvn clean
>     $ mvn package -Phadoop_2 -DskipTests
>     ## SUCCESS
>
> However, when I ran the tests with LocalJobRunner, it failed.
>
>     $ mvn clean
>     $ mvn package -Phadoop_2
>
> It passed tests from "Core" and "Examples", but it failed at
> "Accumulo I/O".
>
>
> testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
>
> The error log contained the following exception
>
>     java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>
>
> Next I wanted to run the tests with a running Hadoop2 instance, but
> I'm having trouble to set it up (I'm quite new to Hadoop).
>
> Could you show me some example configuration (etc/hadoop/* files) of
> Hadoop 2.2.0 single-node cluster? That would be very helpful.
>
>
>
>
> >
> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp>
> wrote:
> > > Hi Roman.
> > >
> > > Thanks for the reply.
> > >
> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
> > > release-1.1.0-RC0 and report the result.
> >
> > That would be extremely helpful!
> >
> > And speaking of which -- I'd like to remind folks
> > that taking RC0 for a spin would really help
> > at this point. If we ever want to have 1.1.0 out
> > we need the required PMC votes.
> >
> > Thanks,
> > Roman.
> ------------------------------------
> Toshio Ito
>

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Akila Wajirasena <ak...@gmail.com>.
Hi Roman,

I got the same error when running hadoop_2 profile.
According to this [1] the Accumulo version we use in giraph (1.4) is not
compatible with Hadoop 2.
I think this is the issue.

[1]
http://apache-accumulo.1065345.n5.nabble.com/Accumulo-Hadoop-version-compatibility-matrix-tp3893p3894.html

Thanks

Akila

On Mon, Jun 30, 2014 at 2:21 PM, Toshio ITO <to...@toshiba.co.jp>
wrote:

> Hi Roman.
>
> I checked out release-1.1.0-RC0 and succeeded to build it.
>
>     $ git checkout release-1.1.0-RC0
>     $ mvn clean
>     $ mvn package -Phadoop_2 -DskipTests
>     ## SUCCESS
>
> However, when I ran the tests with LocalJobRunner, it failed.
>
>     $ mvn clean
>     $ mvn package -Phadoop_2
>
> It passed tests from "Core" and "Examples", but it failed at
> "Accumulo I/O".
>
>
> testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)
>
> The error log contained the following exception
>
>     java.lang.IncompatibleClassChangeError: Found interface
> org.apache.hadoop.mapreduce.JobContext, but class was expected
>
>
> Next I wanted to run the tests with a running Hadoop2 instance, but
> I'm having trouble to set it up (I'm quite new to Hadoop).
>
> Could you show me some example configuration (etc/hadoop/* files) of
> Hadoop 2.2.0 single-node cluster? That would be very helpful.
>
>
>
>
> >
> > On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp>
> wrote:
> > > Hi Roman.
> > >
> > > Thanks for the reply.
> > >
> > > OK, I'll try hadoop_1 and hadoop_2 with the latest
> > > release-1.1.0-RC0 and report the result.
> >
> > That would be extremely helpful!
> >
> > And speaking of which -- I'd like to remind folks
> > that taking RC0 for a spin would really help
> > at this point. If we ever want to have 1.1.0 out
> > we need the required PMC votes.
> >
> > Thanks,
> > Roman.
> ------------------------------------
> Toshio Ito
>

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Toshio ITO <to...@toshiba.co.jp>.
Hi Roman.

I checked out release-1.1.0-RC0 and succeeded to build it.

    $ git checkout release-1.1.0-RC0
    $ mvn clean
    $ mvn package -Phadoop_2 -DskipTests
    ## SUCCESS

However, when I ran the tests with LocalJobRunner, it failed.

    $ mvn clean
    $ mvn package -Phadoop_2

It passed tests from "Core" and "Examples", but it failed at
"Accumulo I/O".

    testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)

The error log contained the following exception

    java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected


Next I wanted to run the tests with a running Hadoop2 instance, but
I'm having trouble to set it up (I'm quite new to Hadoop).

Could you show me some example configuration (etc/hadoop/* files) of
Hadoop 2.2.0 single-node cluster? That would be very helpful.




> 
> On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp> wrote:
> > Hi Roman.
> >
> > Thanks for the reply.
> >
> > OK, I'll try hadoop_1 and hadoop_2 with the latest
> > release-1.1.0-RC0 and report the result.
> 
> That would be extremely helpful!
> 
> And speaking of which -- I'd like to remind folks
> that taking RC0 for a spin would really help
> at this point. If we ever want to have 1.1.0 out
> we need the required PMC votes.
> 
> Thanks,
> Roman.
------------------------------------
Toshio Ito

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Toshio ITO <to...@toshiba.co.jp>.
Hi Roman.

I checked out release-1.1.0-RC0 and succeeded to build it.

    $ git checkout release-1.1.0-RC0
    $ mvn clean
    $ mvn package -Phadoop_2 -DskipTests
    ## SUCCESS

However, when I ran the tests with LocalJobRunner, it failed.

    $ mvn clean
    $ mvn package -Phadoop_2

It passed tests from "Core" and "Examples", but it failed at
"Accumulo I/O".

    testAccumuloInputOutput(org.apache.giraph.io.accumulo.TestAccumuloVertexFormat)

The error log contained the following exception

    java.lang.IncompatibleClassChangeError: Found interface org.apache.hadoop.mapreduce.JobContext, but class was expected


Next I wanted to run the tests with a running Hadoop2 instance, but
I'm having trouble to set it up (I'm quite new to Hadoop).

Could you show me some example configuration (etc/hadoop/* files) of
Hadoop 2.2.0 single-node cluster? That would be very helpful.




> 
> On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp> wrote:
> > Hi Roman.
> >
> > Thanks for the reply.
> >
> > OK, I'll try hadoop_1 and hadoop_2 with the latest
> > release-1.1.0-RC0 and report the result.
> 
> That would be extremely helpful!
> 
> And speaking of which -- I'd like to remind folks
> that taking RC0 for a spin would really help
> at this point. If we ever want to have 1.1.0 out
> we need the required PMC votes.
> 
> Thanks,
> Roman.
------------------------------------
Toshio Ito

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Roman Shaposhnik <ro...@shaposhnik.org>.
On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp> wrote:
> Hi Roman.
>
> Thanks for the reply.
>
> OK, I'll try hadoop_1 and hadoop_2 with the latest
> release-1.1.0-RC0 and report the result.

That would be extremely helpful!

And speaking of which -- I'd like to remind folks
that taking RC0 for a spin would really help
at this point. If we ever want to have 1.1.0 out
we need the required PMC votes.

Thanks,
Roman.

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Roman Shaposhnik <ro...@shaposhnik.org>.
On Sun, Jun 29, 2014 at 5:06 PM, Toshio ITO <to...@toshiba.co.jp> wrote:
> Hi Roman.
>
> Thanks for the reply.
>
> OK, I'll try hadoop_1 and hadoop_2 with the latest
> release-1.1.0-RC0 and report the result.

That would be extremely helpful!

And speaking of which -- I'd like to remind folks
that taking RC0 for a spin would really help
at this point. If we ever want to have 1.1.0 out
we need the required PMC votes.

Thanks,
Roman.

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Toshio ITO <to...@toshiba.co.jp>.
Hi Roman.

Thanks for the reply.

OK, I'll try hadoop_1 and hadoop_2 with the latest
release-1.1.0-RC0 and report the result.

I still wonder why release-1.0.0-RC3 failed, though...



> 
> Please try profiles hadoop_1 and hadoop_2. These
> are the two profiles that we're targeting for the upcoming
> 1.1.0 release. The rest of profiles are optional and
> may or may NOT work.
> 
> Thanks,
> Roman.
> 
> On Thu, Jun 26, 2014 at 3:34 AM, Toshio ITO <to...@toshiba.co.jp> wrote:
> > Hi all,
> >
> > I recently tried Giraph.
> >
> > It compiled fine, but when I did the unit test (i.e. `mvn test`) it
> > failed.  I did the test several times with different settings. See
> > below for details.
> >
> > My question is:
> >
> > - Is it normal for Giraph? Is the unit test supposed to fail for now?
> > - If it's not normal, what do you think is wrong?
> >
> >
> > My environment:
> >
> > - Ubuntu Server 14.04 64-bit
> > - Dynamic address by DHCP
> > - Followed the "Quick Start" guide
> >   https://giraph.apache.org/quick_start.html
> >   except I used "localhost" for the node hostname.
> > -- Installed openjdk-7-jdk, git, maven by apt-get
> > -- Installed hadoop-0.20.203.0rc1 as instructed
> >
> >
> > Test settings:
> >
> > (1) 1.1.0-SNAPSHOT (Commit ID: b218d72cedc52467e691c6002e596e482d8583e4)
> >     with LocalJobRunner
> > (2) 1.1.0-SNAPSHOT with the running Hadoop instance
> > (3) Tag "release-1.0.0-RC3" with LocalJobRunner
> > (4) Tag "release-1.0.0-RC3" with the running Hadoop instance
> >
> > Cases (1) and (3) are run by the following commands.
> >
> >     $ mvn clean
> >     $ mvn package -Phadoop_0.20.203 -DskipTests
> >     $ mvn test -Phadoop_0.20.203
> >
> > Cases (2) and (4) are run by the following commands.
> >
> >     $ mvn clean
> >     $ mvn package -Phadoop_0.20.203 -DskipTests
> >     $ mvn test -Phadoop_0.20.203 -Dprop.mapred.job.tracker=localhost:54311
> >
> >
> > Test results:
> >
> > (1) Failed at "Rexster I/O Format". At some point, the test
> > endlessly tried to connect to ZooKeeper, repeating the following log
> > messages. I had to terminate the test by Ctrl-C.
> >
> >     14/06/26 18:10:27 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:22182. Will not attempt to authenticate using SASL (unknown error)
> >     14/06/26 18:10:27 WARN zookeeper.ClientCnxn: Session 0x146d76e16750002 for server null, unexpected error, closing socket connection and attempting reconnect
> >     java.net.ConnectException: Connection refused
> >             at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
> >             at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
> >             at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
> >             at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
> >
> >
> > (2) Failed at "Examples". Hadoop was stuck at "testBspFail"
> > job. Logs like the following line were printed every 5 seconds for
> > 10 minutes.  After that, the test endlessly tried to connect to
> > ZooKeeper, just like the case (1). I killed the test process and the
> > Hadoop job.
> >
> >     14/06/26 18:40:30 INFO job.JobProgressTracker: Data from 2 workers - Compute superstep 20: 10 out of 10 vertices computed; 6 out of 6 partitions computed; min free memory on worker 1 - 137.42MB, average 149.12MB
> >
> >
> > (3) Failed at "HBase I/O" in testHBaseInputOutput
> > (org.apache.giraph.io.hbase.TestHBaseRootMarkerVertextFormat). Before
> > it reported failure, it blocked for 10 minutes.
> >
> >
> > (4) Failed at "Core" in testContinue
> > (org.apache.giraph.io.TestJsonBase64Format) . Hadoop was stuck at
> > the second "testContinue" job. After 10 minutes, the test went on
> > and reported the failure. The "testContinue" map task was aborted
> > with the following error.
> >
> >     java.io.FileNotFoundException: /tmp/_giraphTests/testContinue/_logs (Is a directory)
> >
> >
> >
> > Thanks in advance.
> >
> > ------------------------------------
> > Toshio Ito
------------------------------------
Toshio Ito

Re: Giraph (1.1.0-SNAPSHOT and 1.0.0-RC3) unit tests fail

Posted by Roman Shaposhnik <ro...@shaposhnik.org>.
Please try profiles hadoop_1 and hadoop_2. These
are the two profiles that we're targeting for the upcoming
1.1.0 release. The rest of profiles are optional and
may or may NOT work.

Thanks,
Roman.

On Thu, Jun 26, 2014 at 3:34 AM, Toshio ITO <to...@toshiba.co.jp> wrote:
> Hi all,
>
> I recently tried Giraph.
>
> It compiled fine, but when I did the unit test (i.e. `mvn test`) it
> failed.  I did the test several times with different settings. See
> below for details.
>
> My question is:
>
> - Is it normal for Giraph? Is the unit test supposed to fail for now?
> - If it's not normal, what do you think is wrong?
>
>
> My environment:
>
> - Ubuntu Server 14.04 64-bit
> - Dynamic address by DHCP
> - Followed the "Quick Start" guide
>   https://giraph.apache.org/quick_start.html
>   except I used "localhost" for the node hostname.
> -- Installed openjdk-7-jdk, git, maven by apt-get
> -- Installed hadoop-0.20.203.0rc1 as instructed
>
>
> Test settings:
>
> (1) 1.1.0-SNAPSHOT (Commit ID: b218d72cedc52467e691c6002e596e482d8583e4)
>     with LocalJobRunner
> (2) 1.1.0-SNAPSHOT with the running Hadoop instance
> (3) Tag "release-1.0.0-RC3" with LocalJobRunner
> (4) Tag "release-1.0.0-RC3" with the running Hadoop instance
>
> Cases (1) and (3) are run by the following commands.
>
>     $ mvn clean
>     $ mvn package -Phadoop_0.20.203 -DskipTests
>     $ mvn test -Phadoop_0.20.203
>
> Cases (2) and (4) are run by the following commands.
>
>     $ mvn clean
>     $ mvn package -Phadoop_0.20.203 -DskipTests
>     $ mvn test -Phadoop_0.20.203 -Dprop.mapred.job.tracker=localhost:54311
>
>
> Test results:
>
> (1) Failed at "Rexster I/O Format". At some point, the test
> endlessly tried to connect to ZooKeeper, repeating the following log
> messages. I had to terminate the test by Ctrl-C.
>
>     14/06/26 18:10:27 INFO zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:22182. Will not attempt to authenticate using SASL (unknown error)
>     14/06/26 18:10:27 WARN zookeeper.ClientCnxn: Session 0x146d76e16750002 for server null, unexpected error, closing socket connection and attempting reconnect
>     java.net.ConnectException: Connection refused
>             at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>             at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
>             at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
>             at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
>
>
> (2) Failed at "Examples". Hadoop was stuck at "testBspFail"
> job. Logs like the following line were printed every 5 seconds for
> 10 minutes.  After that, the test endlessly tried to connect to
> ZooKeeper, just like the case (1). I killed the test process and the
> Hadoop job.
>
>     14/06/26 18:40:30 INFO job.JobProgressTracker: Data from 2 workers - Compute superstep 20: 10 out of 10 vertices computed; 6 out of 6 partitions computed; min free memory on worker 1 - 137.42MB, average 149.12MB
>
>
> (3) Failed at "HBase I/O" in testHBaseInputOutput
> (org.apache.giraph.io.hbase.TestHBaseRootMarkerVertextFormat). Before
> it reported failure, it blocked for 10 minutes.
>
>
> (4) Failed at "Core" in testContinue
> (org.apache.giraph.io.TestJsonBase64Format) . Hadoop was stuck at
> the second "testContinue" job. After 10 minutes, the test went on
> and reported the failure. The "testContinue" map task was aborted
> with the following error.
>
>     java.io.FileNotFoundException: /tmp/_giraphTests/testContinue/_logs (Is a directory)
>
>
>
> Thanks in advance.
>
> ------------------------------------
> Toshio Ito