You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pig.apache.org by dv...@apache.org on 2012/04/15 17:27:42 UTC
svn commit: r1326364 - in /pig/branches/branch-0.10: CHANGES.txt
src/org/apache/pig/backend/hadoop/executionengine/mapReduceLayer/MRCompiler.java
Author: dvryaboy
Date: Sun Apr 15 15:27:41 2012
New Revision: 1326364
URL: http://svn.apache.org/viewvc?rev=1326364&view=rev
Log:
Rolling back PIG-2652
Modified:
pig/branches/branch-0.10/CHANGES.txt
pig/branches/branch-0.10/src/org/apache/pig/backend/hadoop/executionengine/mapReduceLayer/MRCompiler.java
Modified: pig/branches/branch-0.10/CHANGES.txt
URL: http://svn.apache.org/viewvc/pig/branches/branch-0.10/CHANGES.txt?rev=1326364&r1=1326363&r2=1326364&view=diff
==============================================================================
--- pig/branches/branch-0.10/CHANGES.txt (original)
+++ pig/branches/branch-0.10/CHANGES.txt Sun Apr 15 15:27:41 2012
@@ -192,8 +192,6 @@ BUG FIXES
PIG-2578: Multiple Store-commands mess up mapred.output.dir. (daijy)
-PIG-2652: Skew join and order by don't trigger reducer estimation (billgraham via dvryaboy)
-
PIG-2623: Support S3 paths for registering UDFs (nshkrob via daijy)
PIG-2540: [piggybank] AvroStorage can't read schema on amazon s3 in elastic mapreduce (rjurney via jcoveney)
Modified: pig/branches/branch-0.10/src/org/apache/pig/backend/hadoop/executionengine/mapReduceLayer/MRCompiler.java
URL: http://svn.apache.org/viewvc/pig/branches/branch-0.10/src/org/apache/pig/backend/hadoop/executionengine/mapReduceLayer/MRCompiler.java?rev=1326364&r1=1326363&r2=1326364&view=diff
==============================================================================
--- pig/branches/branch-0.10/src/org/apache/pig/backend/hadoop/executionengine/mapReduceLayer/MRCompiler.java (original)
+++ pig/branches/branch-0.10/src/org/apache/pig/backend/hadoop/executionengine/mapReduceLayer/MRCompiler.java Sun Apr 15 15:27:41 2012
@@ -2521,8 +2521,8 @@ public class MRCompiler extends PhyPlanV
val = pigContext.defaultParallel;
if (val<=0)
val = eng.getJobConf().getNumReduceTasks();
- if (val<=1)
- val = -1;
+ if (val<=0)
+ val = 1;
} catch (Exception e) {
int errCode = 6015;
String msg = "Problem getting the default number of reduces from the Job Client.";