You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by om...@apache.org on 2008/06/06 22:10:52 UTC

svn commit: r664111 - /hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobClient.java

Author: omalley
Date: Fri Jun  6 13:10:52 2008
New Revision: 664111

URL: http://svn.apache.org/viewvc?rev=664111&view=rev
Log:
HADOOP-3230. Missed test file. Contributed by Tom White.

Added:
    hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobClient.java

Added: hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobClient.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobClient.java?rev=664111&view=auto
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobClient.java (added)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestJobClient.java Fri Jun  6 13:10:52 2008
@@ -0,0 +1,85 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.mapred;
+
+import java.io.ByteArrayOutputStream;
+import java.io.OutputStream;
+import java.io.OutputStreamWriter;
+import java.io.PrintStream;
+import java.io.Writer;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.io.LongWritable;
+import org.apache.hadoop.io.Text;
+import org.apache.hadoop.util.Tool;
+import org.apache.hadoop.util.ToolRunner;
+
+public class TestJobClient extends ClusterMapReduceTestCase {
+  private String runJob() throws Exception {
+    OutputStream os = getFileSystem().create(new Path(getInputDir(), "text.txt"));
+    Writer wr = new OutputStreamWriter(os);
+    wr.write("hello1\n");
+    wr.write("hello2\n");
+    wr.write("hello3\n");
+    wr.close();
+
+    JobConf conf = createJobConf();
+    conf.setJobName("mr");
+
+    conf.setInputFormat(TextInputFormat.class);
+
+    conf.setMapOutputKeyClass(LongWritable.class);
+    conf.setMapOutputValueClass(Text.class);
+
+    conf.setOutputFormat(TextOutputFormat.class);
+    conf.setOutputKeyClass(LongWritable.class);
+    conf.setOutputValueClass(Text.class);
+
+    conf.setMapperClass(org.apache.hadoop.mapred.lib.IdentityMapper.class);
+    conf.setReducerClass(org.apache.hadoop.mapred.lib.IdentityReducer.class);
+
+    FileInputFormat.setInputPaths(conf, getInputDir());
+    FileOutputFormat.setOutputPath(conf, getOutputDir());
+
+    return JobClient.runJob(conf).getID().toString();
+  }
+  
+  private int runTool(Configuration conf, Tool tool, String[] args, OutputStream out) throws Exception {
+    PrintStream oldOut = System.out;
+    PrintStream newOut = new PrintStream(out, true);
+    try {
+      System.setOut(newOut);
+      return ToolRunner.run(conf, tool, args);
+    } finally {
+      System.setOut(oldOut);
+    }
+  }
+
+  public void testGetCounter() throws Exception {
+    String jobId = runJob();
+    ByteArrayOutputStream out = new ByteArrayOutputStream();
+    int exitCode = runTool(createJobConf(), new JobClient(),
+        new String[] { "-counter", jobId,
+        "org.apache.hadoop.mapred.Task$Counter", "MAP_INPUT_RECORDS" },
+        out);
+    assertEquals("Exit code", 0, exitCode);
+    assertEquals("Counter", "3", out.toString().trim());
+  }
+
+}