You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by GitBox <gi...@apache.org> on 2018/12/11 02:50:38 UTC

[GitHub] lcqzte10192193 opened a new pull request #23281: [MINOR][DOC]update the condition description of BypassMergeSortShuffl…

lcqzte10192193 opened a new pull request #23281: [MINOR][DOC]update the condition description of BypassMergeSortShuffl…
URL: https://github.com/apache/spark/pull/23281
 
 
   …eWriter
   
   ## What changes were proposed in this pull request?
   These three condition description should update, follow #23228  :
   <li>no Ordering is specified,</li>
   <li>no Aggregator is specified, and</li>
   <li>the number of partitions is less than
   <code>spark.shuffle.sort.bypassMergeThreshold</code>.
   </li>
   1、If the shuffle dependency specifies aggregation, but it only aggregates at the reduce-side, serialized shuffle can still be used.
   2、If the number of output partitions is 16777216 , we can use serialized shuffle.
   
   
   ## How was this patch tested?
   N/A
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org