You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Anand Inguva (Jira)" <ji...@apache.org> on 2022/03/07 22:00:00 UTC
[jira] [Created] (BEAM-14068) RunInference Benchmarking tests
Anand Inguva created BEAM-14068:
-----------------------------------
Summary: RunInference Benchmarking tests
Key: BEAM-14068
URL: https://issues.apache.org/jira/browse/BEAM-14068
Project: Beam
Issue Type: Sub-task
Components: sdk-py-core
Reporter: Anand Inguva
Assignee: Anand Inguva
RunInference benchmarks will evaluate performance of Pipelines, which represent common use cases of Beam + Dataflow in Pytorch, sklearn and possibly TFX. These benchmarks would be the integration tests that exercise several software components using Beam, PyTorch, Scikit learn and TensorFlow extended.
we would use the datasets that's available publicly (Eg; Kaggle).
Size: small / 10 GB / 1 TB etc
The default execution runner would be Dataflow unless specified otherwise.
These tests would be run very less frequently(every release cycle).
--
This message was sent by Atlassian Jira
(v8.20.1#820001)