You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@beam.apache.org by "Beam JIRA Bot (Jira)" <ji...@apache.org> on 2022/04/02 16:59:00 UTC
[jira] [Updated] (BEAM-13970) RunInference V1
[ https://issues.apache.org/jira/browse/BEAM-13970?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Beam JIRA Bot updated BEAM-13970:
---------------------------------
Labels: run-inference stale-assigned (was: run-inference)
> RunInference V1
> ---------------
>
> Key: BEAM-13970
> URL: https://issues.apache.org/jira/browse/BEAM-13970
> Project: Beam
> Issue Type: New Feature
> Components: sdk-py-core
> Reporter: Andy Ye
> Assignee: Andy Ye
> Priority: P2
> Labels: run-inference, stale-assigned
>
> Users of machine learning frameworks must currently implement their own transforms for running ML inferences. The exception is the TensorFlow [RunInference transform|https://github.com/tensorflow/tfx-bsl/blob/master/tfx_bsl/beam/run_inference.py]. However, this is hosted in its own [repo|https://github.com/tensorflow/tfx-bsl], and has an [API|https://www.tensorflow.org/tfx/tfx_bsl/api_docs/python/tfx_bsl/public/beam/RunInference] that is exclusively geared towards the TensorFlow TFX library. Our goal is to add new implementations of RunInference for the two other popular machine learning frameworks: scikit-learn and Pytorch.
> Please see main design document [here|https://s.apache.org/inference-sklearn-pytorch].
--
This message was sent by Atlassian Jira
(v8.20.1#820001)