You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pig.apache.org by zl...@apache.org on 2017/04/07 01:41:05 UTC
svn commit: r1790492 -
/pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/optimizer/CombinerOptimizer.java
Author: zly
Date: Fri Apr 7 01:41:05 2017
New Revision: 1790492
URL: http://svn.apache.org/viewvc?rev=1790492&view=rev
Log:
PIG-5203:Partitioner E2E test fails on spark(Adam via Liyun)
Modified:
pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/optimizer/CombinerOptimizer.java
Modified: pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/optimizer/CombinerOptimizer.java
URL: http://svn.apache.org/viewvc/pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/optimizer/CombinerOptimizer.java?rev=1790492&r1=1790491&r2=1790492&view=diff
==============================================================================
--- pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/optimizer/CombinerOptimizer.java (original)
+++ pig/branches/spark/src/org/apache/pig/backend/hadoop/executionengine/spark/optimizer/CombinerOptimizer.java Fri Apr 7 01:41:05 2017
@@ -248,9 +248,9 @@ public class CombinerOptimizer extends S
phyPlan.replace(rearrange, combinerLocalRearrange);
// Create a reduceBy operator.
- POReduceBySpark reduceOperator = new POReduceBySpark(cfe.getOperatorKey(), cfe
- .getRequestedParallelism(),
- cfe.getInputPlans(), cfe.getToBeFlattened(), combinePack, newRearrange);
+ POReduceBySpark reduceOperator = new POReduceBySpark(cfe.getOperatorKey(), combinerLocalRearrange
+ .getRequestedParallelism(), cfe.getInputPlans(), cfe.getToBeFlattened(), combinePack,
+ newRearrange);
reduceOperator.setCustomPartitioner(glr.getCustomPartitioner());
fixReduceSideFE(postReduceFE, algebraicOps);
CombinerOptimizerUtil.changeFunc(reduceOperator, POUserFunc.INTERMEDIATE);