You are viewing a plain text version of this content. The canonical link for it is here.
Posted to github@arrow.apache.org by "wjones127 (via GitHub)" <gi...@apache.org> on 2023/05/28 23:08:36 UTC

[GitHub] [arrow] wjones127 commented on a diff in pull request #35808: GH-34363: [C++] Use equal size parts in S3 upload for R2 compatibility

wjones127 commented on code in PR #35808:
URL: https://github.com/apache/arrow/pull/35808#discussion_r1208683874


##########
cpp/src/arrow/filesystem/s3fs.cc:
##########
@@ -1154,7 +1154,7 @@ class ObjectInputFile final : public io::RandomAccessFile {
 // AWS doc says "5 MB" but it's not clear whether those are MB or MiB,
 // so I chose the safer value.
 // (see https://docs.aws.amazon.com/AmazonS3/latest/API/mpUploadUploadPart.html)
-static constexpr int64_t kMinimumPartUpload = 5 * 1024 * 1024;
+static constexpr int64_t kMinimumPartUpload = 10 * 1024 * 1024;

Review Comment:
   Previously our limit was ~2.4TB. In the new design, it is 100GB. It would be only 50GB if it was left at 5MB per part. I feel like this is an okay balance, but willing to hear other's opinions.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: github-unsubscribe@arrow.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org