You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Alex Amato (JIRA)" <ji...@apache.org> on 2019/04/12 20:33:00 UTC

[jira] [Commented] (BEAM-7058) Python SDK metric process_bundle_msecs reported as zero

    [ https://issues.apache.org/jira/browse/BEAM-7058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16816622#comment-16816622 ] 

Alex Amato commented on BEAM-7058:
----------------------------------

We have an end to end integration test successfully collecting these metrics in Dataflow python and several unit tests, These tests do force some sleep times though.

So the the python SDK is emitting the metrics in that case. One theory is that small bundles may not trigger the state sampler code properly. Or the particular test is too small, and executes too fast to exercise the sampling code at all (and it really isn't running much). We should test this on a test with a high element count.

If that's the case, the state sampler should be setup to trigger intervals periodically, and not reset the interval on new bundles.


It could be related to the behaviour of this specific test. We would need someone to debug this test, is there a repro?

 

 

> Python SDK metric process_bundle_msecs reported as zero
> -------------------------------------------------------
>
>                 Key: BEAM-7058
>                 URL: https://issues.apache.org/jira/browse/BEAM-7058
>             Project: Beam
>          Issue Type: Bug
>          Components: runner-flink, sdk-py-harness
>            Reporter: Thomas Weise
>            Assignee: Alex Amato
>            Priority: Major
>              Labels: portability-flink
>
> With the portable Flink runner, the metric is reported as 0, while the count metric works as expected.
> [https://lists.apache.org/thread.html/25eec8104bda6e4c71cc6c5e9864c335833c3ae2afe225d372479f30@%3Cdev.beam.apache.org%3E]
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)