You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pig.apache.org by da...@apache.org on 2014/09/27 01:23:25 UTC

svn commit: r1627909 - in /pig/trunk: CHANGES.txt shims/src/hadoop23/org/apache/pig/backend/hadoop/executionengine/shims/HadoopShims.java

Author: daijy
Date: Fri Sep 26 23:23:25 2014
New Revision: 1627909

URL: http://svn.apache.org/r1627909
Log:
Revert PIG-4164: After Pig job finish, Pig client spend too much time retry to connect to AM

Modified:
    pig/trunk/CHANGES.txt
    pig/trunk/shims/src/hadoop23/org/apache/pig/backend/hadoop/executionengine/shims/HadoopShims.java

Modified: pig/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/pig/trunk/CHANGES.txt?rev=1627909&r1=1627908&r2=1627909&view=diff
==============================================================================
--- pig/trunk/CHANGES.txt (original)
+++ pig/trunk/CHANGES.txt Fri Sep 26 23:23:25 2014
@@ -86,8 +86,6 @@ BUG FIXES
 
 PIG-4176: Fix tez e2e test Bloom_[1-3] (daijy)
 
-PIG-4164: After Pig job finish, Pig client spend too much time retry to connect to AM (daijy)
-
 PIG-4195: Support loading char/varchar data in OrcStorage (daijy)
 
 PIG-4201: Native e2e tests fail when run against old version of pig (rohini)

Modified: pig/trunk/shims/src/hadoop23/org/apache/pig/backend/hadoop/executionengine/shims/HadoopShims.java
URL: http://svn.apache.org/viewvc/pig/trunk/shims/src/hadoop23/org/apache/pig/backend/hadoop/executionengine/shims/HadoopShims.java?rev=1627909&r1=1627908&r2=1627909&view=diff
==============================================================================
--- pig/trunk/shims/src/hadoop23/org/apache/pig/backend/hadoop/executionengine/shims/HadoopShims.java (original)
+++ pig/trunk/shims/src/hadoop23/org/apache/pig/backend/hadoop/executionengine/shims/HadoopShims.java Fri Sep 26 23:23:25 2014
@@ -32,7 +32,6 @@ import org.apache.hadoop.mapred.TIPStatu
 import org.apache.hadoop.mapred.TaskReport;
 import org.apache.hadoop.mapred.jobcontrol.Job;
 import org.apache.hadoop.mapred.jobcontrol.JobControl;
-import org.apache.hadoop.mapreduce.Cluster;
 import org.apache.hadoop.mapreduce.ContextFactory;
 import org.apache.hadoop.mapreduce.JobContext;
 import org.apache.hadoop.mapreduce.JobID;
@@ -121,8 +120,7 @@ public class HadoopShims {
 
     public static Counters getCounters(Job job) throws IOException {
         try {
-            Cluster cluster = new Cluster(job.getJobConf());
-            return new Counters(cluster.getJob(job.getAssignedJobID()).getCounters());
+            return new Counters(job.getJob().getCounters());
         } catch (Exception ir) {
             throw new IOException(ir);
         }
@@ -221,9 +219,8 @@ public class HadoopShims {
             LOG.info("TaskReports are disabled for job: " + job.getAssignedJobID());
             return null;
         }
-        Cluster cluster = new Cluster(job.getJobConf());
+        org.apache.hadoop.mapreduce.Job mrJob = job.getJob();
         try {
-            org.apache.hadoop.mapreduce.Job mrJob = cluster.getJob(job.getAssignedJobID());
             org.apache.hadoop.mapreduce.TaskReport[] reports = mrJob.getTaskReports(type);
             return DowngradeHelper.downgradeTaskReports(reports);
         } catch (InterruptedException ir) {