You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cn...@apache.org on 2015/01/13 18:50:11 UTC

[10/50] [abbrv] hadoop git commit: MAPREDUCE-6206. TestAggregatedTransferRate fails on non-US systems. Contributed by Jens Rabe

MAPREDUCE-6206. TestAggregatedTransferRate fails on non-US systems. Contributed by Jens Rabe


Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/a0aeed10
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/a0aeed10
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/a0aeed10

Branch: refs/heads/HDFS-6994
Commit: a0aeed100b839d5c122974bd0ea12f9168ee36a5
Parents: 60103fc
Author: Jason Lowe <jl...@apache.org>
Authored: Tue Jan 6 23:39:43 2015 +0000
Committer: Jason Lowe <jl...@apache.org>
Committed: Tue Jan 6 23:39:43 2015 +0000

----------------------------------------------------------------------
 hadoop-mapreduce-project/CHANGES.txt            |  3 +
 .../task/reduce/TestShuffleScheduler.java       | 68 +++++++++-----------
 2 files changed, 35 insertions(+), 36 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hadoop/blob/a0aeed10/hadoop-mapreduce-project/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt
index 82295de..3bf34d8 100644
--- a/hadoop-mapreduce-project/CHANGES.txt
+++ b/hadoop-mapreduce-project/CHANGES.txt
@@ -294,6 +294,9 @@ Release 2.7.0 - UNRELEASED
     MAPREDUCE-6199. AbstractCounters are not reset completely on
     deserialization (adhoot via rkanter)
 
+    MAPREDUCE-6206. TestAggregatedTransferRate fails on non-US systems (Jens
+    Rabe via jlowe)
+
 Release 2.6.0 - 2014-11-18
 
   INCOMPATIBLE CHANGES

http://git-wip-us.apache.org/repos/asf/hadoop/blob/a0aeed10/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestShuffleScheduler.java
----------------------------------------------------------------------
diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestShuffleScheduler.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestShuffleScheduler.java
index 905fd44..6ac2320 100644
--- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestShuffleScheduler.java
+++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/task/reduce/TestShuffleScheduler.java
@@ -77,7 +77,7 @@ public class TestShuffleScheduler {
         0.0f);
     Assert.assertTrue(scheduler.waitUntilDone(1));
   }
-  
+
   @SuppressWarnings("rawtypes")
   @Test
   public <K, V> void TestAggregatedTransferRate() throws Exception {
@@ -102,7 +102,7 @@ public class TestShuffleScheduler {
     Task mockTask = mock(Task.class);
     @SuppressWarnings("unchecked")
     MapOutput<K, V> output = mock(MapOutput.class);
-     
+
     ShuffleConsumerPlugin.Context<K, V> context =
         new ShuffleConsumerPlugin.Context<K, V>(mockTaskAttemptID, job, mockFileSystem,
                                                      mockUmbilical, mockLocalDirAllocator,
@@ -125,56 +125,51 @@ public class TestShuffleScheduler {
     ShuffleSchedulerImpl<K, V> scheduler = new ShuffleSchedulerImpl<K, V>(job, status, null,
         null, progress, context.getShuffledMapsCounter(),
         context.getReduceShuffleBytes(), context.getFailedShuffleCounter());
-    TaskAttemptID attemptID0 = new TaskAttemptID( 
+    TaskAttemptID attemptID0 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
         new JobID("test",0), TaskType.MAP, 0), 0);
-     
+
     //adding the 1st interval, 40MB from 60s to 100s
     long bytes = (long)40 * 1024 * 1024;
     scheduler.copySucceeded(attemptID0, new MapHost(null, null), bytes, 60000, 100000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000000_0 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(1 of 10 at 1.00 MB/s)", progress.toString());
-     
-    TaskAttemptID attemptID1 = new TaskAttemptID( 
+    Assert.assertEquals(copyMessage(1, 1, 1), progress.toString());
+
+    TaskAttemptID attemptID1 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
         new JobID("test",0), TaskType.MAP, 1), 1);
-     
+
     //adding the 2nd interval before the 1st interval, 50MB from 0s to 50s
     bytes = (long)50 * 1024 * 1024;
     scheduler.copySucceeded(attemptID1, new MapHost(null, null), bytes, 0, 50000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000001_1 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(2 of 10 at 1.00 MB/s)", progress.toString());
-     
-    TaskAttemptID attemptID2 = new TaskAttemptID( 
+    Assert.assertEquals(copyMessage(2, 1, 1), progress.toString());
+
+    TaskAttemptID attemptID2 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
         new JobID("test",0), TaskType.MAP, 2), 2);
-         
+
     //adding the 3rd interval overlapping with the 1st and the 2nd interval
     //110MB from 25s to 80s
     bytes = (long)110 * 1024 * 1024;
     scheduler.copySucceeded(attemptID2, new MapHost(null, null), bytes, 25000, 80000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000002_2 succeeded at 2.00 MB/s)"
-        + " Aggregated copy rate(3 of 10 at 2.00 MB/s)", progress.toString());
-     
-    TaskAttemptID attemptID3 = new TaskAttemptID( 
+    Assert.assertEquals(copyMessage(3, 2, 2), progress.toString());
+
+    TaskAttemptID attemptID3 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
         new JobID("test",0), TaskType.MAP, 3), 3);
-         
+
     //adding the 4th interval just after the 2nd interval, 100MB from 100s to 300s
     bytes = (long)100 * 1024 * 1024;
     scheduler.copySucceeded(attemptID3, new MapHost(null, null), bytes, 100000, 300000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000003_3 succeeded at 0.50 MB/s)"
-        + " Aggregated copy rate(4 of 10 at 1.00 MB/s)", progress.toString());
-     
-    TaskAttemptID attemptID4 = new TaskAttemptID( 
+    Assert.assertEquals(copyMessage(4, 0.5, 1), progress.toString());
+
+    TaskAttemptID attemptID4 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
         new JobID("test",0), TaskType.MAP, 4), 4);
-         
+
     //adding the 5th interval between after 4th, 50MB from 350s to 400s
     bytes = (long)50 * 1024 * 1024;
     scheduler.copySucceeded(attemptID4, new MapHost(null, null), bytes, 350000, 400000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000004_4 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(5 of 10 at 1.00 MB/s)", progress.toString());
+    Assert.assertEquals(copyMessage(5, 1, 1), progress.toString());
 
 
     TaskAttemptID attemptID5 = new TaskAttemptID(
@@ -183,8 +178,7 @@ public class TestShuffleScheduler {
     //adding the 6th interval between after 5th, 50MB from 450s to 500s
     bytes = (long)50 * 1024 * 1024;
     scheduler.copySucceeded(attemptID5, new MapHost(null, null), bytes, 450000, 500000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000005_5 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(6 of 10 at 1.00 MB/s)", progress.toString());
+    Assert.assertEquals(copyMessage(6, 1, 1), progress.toString());
 
     TaskAttemptID attemptID6 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
@@ -192,8 +186,7 @@ public class TestShuffleScheduler {
     //adding the 7th interval between after 5th and 6th interval, 20MB from 320s to 340s
     bytes = (long)20 * 1024 * 1024;
     scheduler.copySucceeded(attemptID6, new MapHost(null, null), bytes, 320000, 340000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000006_6 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(7 of 10 at 1.00 MB/s)", progress.toString());
+    Assert.assertEquals(copyMessage(7, 1, 1), progress.toString());
 
     TaskAttemptID attemptID7 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
@@ -201,8 +194,7 @@ public class TestShuffleScheduler {
     //adding the 8th interval overlapping with 4th, 5th, and 7th 30MB from 290s to 350s
     bytes = (long)30 * 1024 * 1024;
     scheduler.copySucceeded(attemptID7, new MapHost(null, null), bytes, 290000, 350000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000007_7 succeeded at 0.50 MB/s)"
-        + " Aggregated copy rate(8 of 10 at 1.00 MB/s)", progress.toString());
+    Assert.assertEquals(copyMessage(8, 0.5, 1), progress.toString());
 
     TaskAttemptID attemptID8 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
@@ -210,8 +202,7 @@ public class TestShuffleScheduler {
     //adding the 9th interval overlapping with 5th and 6th, 50MB from 400s to 450s
     bytes = (long)50 * 1024 * 1024;
     scheduler.copySucceeded(attemptID8, new MapHost(null, null), bytes, 400000, 450000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000008_8 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(9 of 10 at 1.00 MB/s)", progress.toString());
+    Assert.assertEquals(copyMessage(9, 1, 1), progress.toString());
 
     TaskAttemptID attemptID9 = new TaskAttemptID(
         new org.apache.hadoop.mapred.TaskID(
@@ -219,8 +210,13 @@ public class TestShuffleScheduler {
     //adding the 10th interval overlapping with all intervals, 500MB from 0s to 500s
     bytes = (long)500 * 1024 * 1024;
     scheduler.copySucceeded(attemptID9, new MapHost(null, null), bytes, 0, 500000, output);
-    Assert.assertEquals("copy task(attempt_test_0000_m_000009_9 succeeded at 1.00 MB/s)"
-        + " Aggregated copy rate(10 of 10 at 2.00 MB/s)", progress.toString());
+    Assert.assertEquals(copyMessage(10, 1, 2), progress.toString());
+  }
 
+  private static String copyMessage(int attemptNo, double rate1, double rate2) {
+    int attemptZero = attemptNo - 1;
+    return String.format("copy task(attempt_test_0000_m_%06d_%d succeeded at %1.2f MB/s)"
+            + " Aggregated copy rate(%d of 10 at %1.2f MB/s)", attemptZero
+            , attemptZero, rate1, attemptNo, rate2);
   }
 }