You are viewing a plain text version of this content. The canonical link for it is here.
Posted to mapreduce-issues@hadoop.apache.org by "ASF GitHub Bot (JIRA)" <ji...@apache.org> on 2017/08/03 00:44:00 UTC

[jira] [Commented] (MAPREDUCE-6931) Fix TestDFSIO "Total Throughput" calculation

    [ https://issues.apache.org/jira/browse/MAPREDUCE-6931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16111987#comment-16111987 ] 

ASF GitHub Bot commented on MAPREDUCE-6931:
-------------------------------------------

GitHub user dennishuo opened a pull request:

    https://github.com/apache/hadoop/pull/259

    MAPREDUCE-6931. Fix TestDFSIO "Total Throughput" calculation.

    Previously it failed to convert ms to seconds and thus reports aggregate
    throughput as 1/1000x actual numbers. Also, make all the bytes-to-mb
    and milliseconds-to-seconds conversions consistent in the reporting
    messages to help avoid this type of error in the future.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/dennishuo/hadoop trunk

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/hadoop/pull/259.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #259
    
----
commit 23a374d6e818004daa0b1cf16c4fedb439ab4943
Author: Dennis Huo <dh...@google.com>
Date:   2017-08-02T21:30:51Z

    MAPREDUCE-6931. Fix TestDFSIO "Total Throughput" calculation.
    
    Previously it failed to convert ms to seconds and thus reports aggregate
    throughput as 1/1000x actual numbers. Also, make all the bytes-to-mb
    and milliseconds-to-seconds conversions consistent in the reporting
    messages to help avoid this type of error in the future.

----


> Fix TestDFSIO "Total Throughput" calculation
> --------------------------------------------
>
>                 Key: MAPREDUCE-6931
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6931
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>          Components: benchmarks, test
>    Affects Versions: 2.8.0
>            Reporter: Dennis Huo
>            Priority: Trivial
>
> The new "Total Throughput" line added in https://issues.apache.org/jira/browse/HDFS-9153 is currently calculated as {{toMB(size) / ((float)execTime)}} and claims to be in units of "MB/s", but {{execTime}} is in milliseconds; thus, the reported number is 1/1000x the actual value:
> {code:java}
>     String resultLines[] = {
>         "----- TestDFSIO ----- : " + testType,
>         "            Date & time: " + new Date(System.currentTimeMillis()),
>         "        Number of files: " + tasks,
>         " Total MBytes processed: " + df.format(toMB(size)),
>         "      Throughput mb/sec: " + df.format(size * 1000.0 / (time * MEGA)),
>         "Total Throughput mb/sec: " + df.format(toMB(size) / ((float)execTime)),
>         " Average IO rate mb/sec: " + df.format(med),
>         "  IO rate std deviation: " + df.format(stdDev),
>         "     Test exec time sec: " + df.format((float)execTime / 1000),
>         "" };
> {code}
> The different calculated fields can also use toMB and a shared milliseconds-to-seconds conversion to make it easier to keep units consistent.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: mapreduce-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: mapreduce-issues-help@hadoop.apache.org