You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by squito <gi...@git.apache.org> on 2018/11/26 20:47:18 UTC
[GitHub] spark pull request #23106: [SPARK-26141] Enable custom metrics implementatio...
Github user squito commented on a diff in the pull request:
https://github.com/apache/spark/pull/23106#discussion_r236418843
--- Diff: core/src/main/java/org/apache/spark/shuffle/sort/ShuffleExternalSorter.java ---
@@ -242,8 +243,13 @@ private void writeSortedFile(boolean isLastFile) {
// Note that we intentionally ignore the value of `writeMetricsToUse.shuffleWriteTime()`.
// Consistent with ExternalSorter, we do not count this IO towards shuffle write time.
// This means that this IO time is not accounted for anywhere; SPARK-3577 will fix this.
- writeMetrics.incRecordsWritten(writeMetricsToUse.recordsWritten());
- taskContext.taskMetrics().incDiskBytesSpilled(writeMetricsToUse.bytesWritten());
+
+ // This is guaranteed to be a ShuffleWriteMetrics based on the if check in the beginning
+ // of this file.
--- End diff --
I found "beginning of this file" to be confusing, I thought you meant the beginning of `ShuffleExternalSorter.java`, not the spill file. maybe "beginning of this method"
also is the comment above this about SPARK-3577 out of date now that has been fixed?
---
---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org