You are viewing a plain text version of this content. The canonical link for it is here.
Posted to github@beam.apache.org by GitBox <gi...@apache.org> on 2022/09/01 21:44:03 UTC

[GitHub] [beam] TheNeuralBit commented on issue #22813: [Feature Request]: Add a WriteBatchedParquet option for python parquet-io

TheNeuralBit commented on issue #22813:
URL: https://github.com/apache/beam/issues/22813#issuecomment-1234813284

   Hi @esadler-hbo, this would be a very welcome contribution :) Please let me know if you need any advice on how to proceed.
   
   I'm actually currently working on integrating ParquetIO with Batched DoFns (https://s.apache.org/batched-dofns) which will hopefully unify things a bit (https://github.com/apache/beam/pull/22947). But in the meantime adding a specifc "Batched" version would be great.
   
   CC: @rezarokni (since you mentioned model inference, Reza may be interested in learning more about your use-case and whether the new RunInference API will work for you)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: github-unsubscribe@beam.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org