You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@camel.apache.org by ac...@apache.org on 2021/07/22 06:47:52 UTC

[camel-kafka-connector] 02/05: Fix InputStream

This is an automated email from the ASF dual-hosted git repository.

acosentino pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/camel-kafka-connector.git

commit d25529dc92f03b1fb2afd3b9883ce2a8f5a86a9c
Author: Mathieu <ma...@gmail.com>
AuthorDate: Wed Jun 23 23:02:50 2021 +0200

    Fix InputStream
---
 .../kafkaconnector/aws2s3/transformers/RecordToJSONTransforms.java     | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/connectors/camel-aws2-s3-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2s3/transformers/RecordToJSONTransforms.java b/connectors/camel-aws2-s3-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2s3/transformers/RecordToJSONTransforms.java
index fdb2e9f..203ab14 100644
--- a/connectors/camel-aws2-s3-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2s3/transformers/RecordToJSONTransforms.java
+++ b/connectors/camel-aws2-s3-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2s3/transformers/RecordToJSONTransforms.java
@@ -29,6 +29,7 @@ import org.apache.kafka.connect.transforms.Transformation;
 import org.apache.camel.kafkaconnector.aws2s3.models.StorageHeader;
 import org.apache.camel.kafkaconnector.aws2s3.models.StorageRecord;
 import java.io.ByteArrayInputStream;
+import java.io.InputStream;
 
 public class RecordToJSONTransforms<R extends ConnectRecord<R>> implements Transformation<R> {
   public static final String FIELD_KEY_CONFIG = "key";
@@ -61,7 +62,7 @@ public class RecordToJSONTransforms<R extends ConnectRecord<R>> implements Trans
     GsonBuilder gsonBuilder = new GsonBuilder();
     Gson gson = gsonBuilder.create();
     String storageRecordJSON = gson.toJson(storageRecord, StorageRecord.class);
-    InputStream storageRecordStream = new ByteArrayInputStream(storageRecordJSON.getBytes())
+    InputStream storageRecordStream = new ByteArrayInputStream(storageRecordJSON.getBytes());
     return record.newRecord(
         record.topic(),
         record.kafkaPartition(),