You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-dev@hadoop.apache.org by "Nigel Daley (JIRA)" <ji...@apache.org> on 2007/07/20 07:57:06 UTC
[jira] Created: (HADOOP-1640) TestDecommission fails on Windows
TestDecommission fails on Windows
---------------------------------
Key: HADOOP-1640
URL: https://issues.apache.org/jira/browse/HADOOP-1640
Project: Hadoop
Issue Type: Bug
Components: dfs
Affects Versions: 0.14.0
Reporter: Nigel Daley
Priority: Blocker
Fix For: 0.14.0
In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
[junit] Created file decommission.dat with 2 replicas.
[junit] Block[0] : xxx xxx
[junit] Block[1] : xxx xxx
[junit] Decommissioning node: 127.0.0.1:50013
[junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
[junit] Name: 127.0.0.1:50013
[junit] State : Decommission in progress
[junit] Total raw bytes: 80030941184 (74.53 GB)
[junit] Used raw bytes: 33940945746 (31.60 GB)
[junit] % used: 42.40%
[junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
[junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
[junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
[junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
[junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
[junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
[junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
[junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
[junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
[junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
[junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
[junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
[junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
[junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
[junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
[junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
[junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
[junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Commented: (HADOOP-1640) TestDecommission fails on Windows
Posted by "Hadoop QA (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12514817 ]
Hadoop QA commented on HADOOP-1640:
-----------------------------------
+1
http://issues.apache.org/jira/secure/attachment/12362274/testDecommission1640.patch applied and successfully tested against trunk revision r558897.
Test results: http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/455/testReport/
Console output: http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/455/console
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows
Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
dhruba borthakur updated HADOOP-1640:
-------------------------------------
Status: Patch Available (was: Open)
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows
Posted by "Owen O'Malley (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Owen O'Malley updated HADOOP-1640:
----------------------------------
Status: Open (was: Patch Available)
I'd like to have the wait time bounded by 1 minute or so, so that if the test is broken that it doesn't hang the unit tests.
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Commented: (HADOOP-1640) TestDecommission fails on Windows
Posted by "Jim Kellerman (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12514653 ]
Jim Kellerman commented on HADOOP-1640:
---------------------------------------
So long as the timeout applies just to this test I'd agree.
There are a couple of HBase tests that take about two minutes or a bit more, so applying a universal timeout would be unacceptable.
However, if we could specify test timeouts on a per test basis, that would be a ++1.
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows
Posted by "Doug Cutting (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Doug Cutting updated HADOOP-1640:
---------------------------------
Resolution: Fixed
Status: Resolved (was: Patch Available)
I just committed this. Thanks, Dhruba!
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Commented: (HADOOP-1640) TestDecommission fails on Windows
Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12514698 ]
dhruba borthakur commented on HADOOP-1640:
------------------------------------------
In general, I do not like the idea of having timeouts for each test. Especially because the same test can take different times to complete on different hardware platforms and different operating systems. Currently, the junit testing framework enforces a timeout for each test. If the test takes exceptionally long, then the junit framework kills the test and declares it as "failed".
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows
Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
dhruba borthakur updated HADOOP-1640:
-------------------------------------
Attachment: testDecommission1640.patch
Make the test wait till all replicas check in with the namenode,.
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Assigned: (HADOOP-1640) TestDecommission fails on Windows
Posted by "Nigel Daley (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Nigel Daley reassigned HADOOP-1640:
-----------------------------------
Assignee: dhruba borthakur
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.
[jira] Updated: (HADOOP-1640) TestDecommission fails on Windows
Posted by "dhruba borthakur (JIRA)" <ji...@apache.org>.
[ https://issues.apache.org/jira/browse/HADOOP-1640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
dhruba borthakur updated HADOOP-1640:
-------------------------------------
Status: Patch Available (was: Open)
> TestDecommission fails on Windows
> ---------------------------------
>
> Key: HADOOP-1640
> URL: https://issues.apache.org/jira/browse/HADOOP-1640
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.14.0
> Reporter: Nigel Daley
> Assignee: dhruba borthakur
> Priority: Blocker
> Fix For: 0.14.0
>
> Attachments: testDecommission1640.patch
>
>
> In the snippet of test log below, the exception happens every ~15 milliseconds for 15 minutes until the test is timed out:
> [junit] Created file decommission.dat with 2 replicas.
> [junit] Block[0] : xxx xxx
> [junit] Block[1] : xxx xxx
> [junit] Decommissioning node: 127.0.0.1:50013
> [junit] 2007-07-19 19:12:45,059 INFO fs.FSNamesystem (FSNamesystem.java:startDecommission(2572)) - Start Decommissioning node 127.0.0.1:50013
> [junit] Name: 127.0.0.1:50013
> [junit] State : Decommission in progress
> [junit] Total raw bytes: 80030941184 (74.53 GB)
> [junit] Used raw bytes: 33940945746 (31.60 GB)
> [junit] % used: 42.40%
> [junit] Last contact: Thu Jul 19 19:12:44 PDT 2007
> [junit] Waiting for node 127.0.0.1:50013 to change state to DECOMMISSIONED
> [junit] 2007-07-19 19:12:45,199 INFO http.SocketListener (SocketListener.java:stop(212)) - Stopped SocketListener on 0.0.0.0:3147
> [junit] 2007-07-19 19:12:45,199 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.servlet.WebApplicationHandler@1d98a
> [junit] 2007-07-19 19:12:45,293 INFO util.Container (Container.java:stop(156)) - Stopped WebApplicationContext[/,/]
> [junit] 2007-07-19 19:12:45,402 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/logs,/logs]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped HttpContext[/static,/static]
> [junit] 2007-07-19 19:12:45,481 INFO util.Container (Container.java:stop(156)) - Stopped org.mortbay.jetty.Server@f1916f
> [junit] 2007-07-19 19:12:45,496 INFO dfs.DataNode (DataNode.java:run(692)) - Exiting DataXceiveServer due to java.net.SocketException: socket closed
> [junit] 2007-07-19 19:12:45,496 WARN dfs.DataNode (DataNode.java:offerService(568)) - java.io.IOException: java.lang.InterruptedException
> [junit] at org.apache.hadoop.fs.DF.doDF(DF.java:71)
> [junit] at org.apache.hadoop.fs.DF.getCapacity(DF.java:89)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolume.getCapacity(FSDataset.java:292)
> [junit] at org.apache.hadoop.dfs.FSDataset$FSVolumeSet.getCapacity(FSDataset.java:379)
> [junit] at org.apache.hadoop.dfs.FSDataset.getCapacity(FSDataset.java:466)
> [junit] at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:493)
> [junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1306)
> [junit] at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.