You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Anand Inguva (Jira)" <ji...@apache.org> on 2022/03/07 22:00:00 UTC

[jira] [Created] (BEAM-14068) RunInference Benchmarking tests

Anand Inguva created BEAM-14068:
-----------------------------------

             Summary: RunInference Benchmarking tests
                 Key: BEAM-14068
                 URL: https://issues.apache.org/jira/browse/BEAM-14068
             Project: Beam
          Issue Type: Sub-task
          Components: sdk-py-core
            Reporter: Anand Inguva
            Assignee: Anand Inguva


RunInference benchmarks will evaluate performance of Pipelines, which represent common use cases of Beam + Dataflow in Pytorch, sklearn and possibly TFX. These benchmarks would be the integration tests that exercise several software components using Beam, PyTorch, Scikit learn and TensorFlow extended.

we would use the datasets that's available publicly (Eg; Kaggle). 

Size: small / 10 GB / 1 TB etc

The default execution runner would be Dataflow unless specified otherwise.

These tests would be run very less frequently(every release cycle).  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)