You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "amoeba (via GitHub)" <gi...@apache.org> on 2023/05/16 23:21:37 UTC

[GitHub] [arrow] amoeba opened a new issue, #35630: [R] Expose AWS SDK retry strategies to s3_bucket/S3FileSystem

amoeba opened a new issue, #35630:
URL: https://github.com/apache/arrow/issues/35630

   ### Describe the enhancement requested
   
   In PyArrow, we expose a `retry_strategy` argument that lets the user pick from the pre-packaged retry strategies offered by the AWS C++ SDK, see https://arrow.apache.org/docs/python/generated/pyarrow.fs.S3FileSystem.html. These aren't exposed from R. We might match up the default behavior in R with that of PyArrow which uses the standard retry strategy (exponential backoff with jitter) and a maximum of three attempts. I'm not quite sure what the best API would be but the PyArrow API looks like this:
   
   ```python
   import pyarrow.parquet as pq
   from pyarrow import fs
   
   fs.S3FileSystem(retry_strategy=fs.AwsStandardS3RetryStrategy(max_attempts=3))
   ```
   
   ### Component(s)
   
   R


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscribe@arrow.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org