You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Nigel Daley (JIRA)" <ji...@apache.org> on 2007/07/20 07:57:06 UTC

[jira] Created: (HADOOP-1640) TestDecommission fails on Windows

TestDecommission fails on Windows
---------------------------------

                 Key: HADOOP-1640
                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
             Project: Hadoop
          Issue Type: Bug
          Components: dfs
    Affects Versions: 0.14.0
            Reporter: Nigel Daley
            Priority: Blocker
             Fix For: 0.14.0


In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:

    [junit] Created file decommission.dat with 2 replicas.
    [junit] Block[0] : xxx xxx 
    [junit] Block[1] : xxx xxx 
    [junit] Decommissioning node: 127.0.0.1:50013
    [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
    [junit] Name: 127.0.0.1:50013
    [junit] State          : Decommission in progress
    [junit] Total raw bytes: 80030941184 (74.53 GB)
    [junit] Used raw bytes: 33940945746 (31.60 GB)
    [junit] % used: 42.40%
    [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007

    [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
    [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
    [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
    [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
    [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
    [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
    [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
    [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
    [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
    [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
    [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
    [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
    [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
    [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
    [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
    [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
    [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Commented: (HADOOP-1640) TestDecommission fails on Windows

Posted by "Hadoop QA (JIRA)" <ji...@apache.org>.
    [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12514817 ] 

Hadoop QA commented on HADOOP-1640:
-----------------------------------

+1

http://issues.apache.org/jira/secure/attachment/12362274/testDecommission1640.patch applied and successfully tested against trunk revision r558897.

Test results:   http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/455/testReport/
Console output: http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/455/console

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows

Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

dhruba borthakur updated HADOOP-1640:
-------------------------------------

    Status: Patch Available  (was: Open)

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows

Posted by "Owen O'Malley (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Owen O'Malley updated HADOOP-1640:
----------------------------------

    Status: Open  (was: Patch Available)

I'd like to have the wait time bounded by 1 minute or so, so that if the test is broken that it doesn't hang the unit tests.

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Commented: (HADOOP-1640) TestDecommission fails on Windows

Posted by "Jim Kellerman (JIRA)" <ji...@apache.org>.
    [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12514653 ] 

Jim Kellerman commented on HADOOP-1640:
---------------------------------------

So long as the timeout applies just to this test I'd agree.

There are a couple of HBase tests that take about two minutes or a bit more, so applying a universal timeout would be unacceptable.

However, if we could specify test timeouts on a per test basis, that would be a ++1.

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows

Posted by "Doug Cutting (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Doug Cutting updated HADOOP-1640:
---------------------------------

    Resolution: Fixed
        Status: Resolved  (was: Patch Available)

I just committed this.  Thanks, Dhruba!

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Commented: (HADOOP-1640) TestDecommission fails on Windows

Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
    [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12514698 ] 

dhruba borthakur commented on HADOOP-1640:
------------------------------------------

In general, I do not like the idea of having timeouts for each test. Especially because the same test can take different times to complete on different hardware platforms and different operating systems. Currently, the junit testing framework enforces a timeout for each test. If the test takes exceptionally long, then the junit framework kills the test and declares it as "failed".


> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows

Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

dhruba borthakur updated HADOOP-1640:
-------------------------------------

    Attachment: testDecommission1640.patch

Make the test wait till all replicas check in with the namenode,.

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Assigned: (HADOOP-1640) TestDecommission fails on Windows

Posted by "Nigel Daley (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Nigel Daley reassigned HADOOP-1640:
-----------------------------------

    Assignee: dhruba borthakur

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows

Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
     [ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

dhruba borthakur updated HADOOP-1640:
-------------------------------------

    Status: Patch Available  (was: Open)

> TestDecommission fails on Windows
> ---------------------------------
>
>                 Key: HADOOP-1640
>                 URL: https://issues.apache.org/jira/browse/HADOOP-1640
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.14.0
>            Reporter: Nigel Daley
>            Assignee: dhruba borthakur
>            Priority: Blocker
>             Fix For: 0.14.0
>
>         Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
>     [junit] Created file decommission.dat with 2 replicas.
>     [junit] Block[0] : xxx xxx 
>     [junit] Block[1] : xxx xxx 
>     [junit] Decommissioning node: 127.0.0.1:50013
>     [junit] 2007-07-19 19:12:45,059 INFO  fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
>     [junit] Name: 127.0.0.1:50013
>     [junit] State          : Decommission in progress
>     [junit] Total raw bytes: 80030941184 (74.53 GB)
>     [junit] Used raw bytes: 33940945746 (31.60 GB)
>     [junit] % used: 42.40%
>     [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
>     [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
>     [junit] 2007-07-19 19:12:45,199 INFO  http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
>     [junit] 2007-07-19 19:12:45,199 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
>     [junit] 2007-07-19 19:12:45,293 INFO  util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
>     [junit] 2007-07-19 19:12:45,402 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
>     [junit] 2007-07-19 19:12:45,481 INFO  util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
>     [junit] 2007-07-19 19:12:45,496 INFO  dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
>     [junit] 2007-07-19 19:12:45,496 WARN  dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
>     [junit] 	at org.apache.hadoop.fs.DF.doDF(DF.java:71)
>     [junit] 	at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
>     [junit] 	at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
>     [junit] 	at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
>     [junit] 	at java.lang.Thread.run(Thread.java:595)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.