You are viewing a plain text version of this content. The canonical link for it is here.
Posted to notifications@jclouds.apache.org by "Andrew Gaul (JIRA)" <ji...@apache.org> on 2018/01/05 08:18:00 UTC
[jira] [Commented] (JCLOUDS-1366) OutOfMemory when InputStream
referencing to big file is used as payload
[ https://issues.apache.org/jira/browse/JCLOUDS-1366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16312682#comment-16312682 ]
Andrew Gaul commented on JCLOUDS-1366:
--------------------------------------
42079e1392fb5b2b792f518812689854c375445f introduced this regression with the parallel upload feature. Previously {{BaseBlobStore.putMultipartBlob}} prepared a single MPU part and uploaded it, looping until complete. Now it prepares all MPU parts simultaneously and submits them to an {{ExecutorService}}. Combined with JCLOUDS-814, this buffers the entire blob in-memory and results in {{OutOfMemoryError}}. Instead we should limit the number of simultaneous uploads with {{InputStream}} payloads. [~zack-s] [~dgyurdzhekliev] Could you investigate this?
> OutOfMemory when InputStream referencing to big file is used as payload
> -----------------------------------------------------------------------
>
> Key: JCLOUDS-1366
> URL: https://issues.apache.org/jira/browse/JCLOUDS-1366
> Project: jclouds
> Issue Type: Bug
> Components: jclouds-blobstore
> Affects Versions: 2.0.3
> Environment: Linux and Windows
> Reporter: Deyan
> Priority: Critical
>
> If I use InputStream which source is large file (lets say 3GB) I am getting OOE. This is with default java VM options.
> Here is the code I am using to construct the blob:
> {code:java}
> File bigFile = new File(file);
> try (InputStream inputStream = new FileInputStream(f)) {
> Blob b = blobStore.blobBuilder(blobName)
> .payload(inputStream).contentLength(f.length())
> .contentDisposition(blobName)
> .contentType(
> MediaType.OCTET_STREAM)
> .userMetadata(ImmutableMap.of("a", "b", "test", "beta"))
> .build();
> blobStore.putBlob("test", bbbbb, multipart());
> }
> {code}
> Stacktrace:
> {code:java}
> java.lang.OutOfMemoryError: Java heap space
> at org.jclouds.io.internal.BasePayloadSlicer$InputStreamPayloadIterator.getNextPayload(BasePayloadSlicer.java:101)
> at org.jclouds.io.internal.BasePayloadSlicer$InputStreamPayloadIterator.next(BasePayloadSlicer.java:90)
> at org.jclouds.io.internal.BasePayloadSlicer$InputStreamPayloadIterator.next(BasePayloadSlicer.java:63)
> at org.jclouds.blobstore.internal.BaseBlobStore.putMultipartBlob(BaseBlobStore.java:363)
> at org.jclouds.blobstore.internal.BaseBlobStore.putMultipartBlob(BaseBlobStore.java:349)
> at org.jclouds.s3.blobstore.S3BlobStore.putBlob(S3BlobStore.java:262)
> {code}
> If 'bigFile' is used as payload the bug is not reproducible.
>
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)