You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Noritaka Sekiyama (Jira)" <ji...@apache.org> on 2020/10/28 02:35:00 UTC
[jira] [Created] (SPARK-33266) Add total duration, read duration,
and write duration as task level metrics
Noritaka Sekiyama created SPARK-33266:
-----------------------------------------
Summary: Add total duration, read duration, and write duration as task level metrics
Key: SPARK-33266
URL: https://issues.apache.org/jira/browse/SPARK-33266
Project: Spark
Issue Type: Improvement
Components: Spark Core
Affects Versions: 3.0.1
Reporter: Noritaka Sekiyama
Sometimes we need to identify performance bottlenecks, for example, how long it took to read from data store, how long it took to write into another data store.
It would be great if we can have total duration, read duration, and write duration as task level metrics.
Currently it seems that both `InputMetrics` and `OutputMetrics` do not have duration related metrics.
[https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/executor/InputMetrics.scala#L42-L58]
[https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/executor/OutputMetrics.scala#L41-L56
]
On the other hand, other metrics such as `ShuffleWriteMetrics` has write time. We might need similar metrics for input/output.
[https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/executor/ShuffleWriteMetrics.scala]
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org