You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pekko.apache.org by md...@apache.org on 2023/01/06 08:57:35 UTC
[incubator-pekko-connectors] branch main updated (ec5347d6 -> 433daa5c)
This is an automated email from the ASF dual-hosted git repository.
mdedetrich pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-pekko-connectors.git
from ec5347d6 Artifact rename (#10)
new 54c516e8 Update scalafmt
new 0e9e6bd5 Add .gitattributes to enforce unix line endings
new 433daa5c format source with scalafmt, #12
The 3 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails. The revisions
listed as "add" were already present in the repository and have only
been added to this reference.
Summary of changes:
.gitattributes | 5 +
.scalafmt.conf | 90 +++-
.../alpakka/amqp/AmqpConnectionProvider.scala | 59 ++-
.../alpakka/amqp/AmqpConnectorSettings.scala | 58 +--
.../impl/AbstractAmqpAsyncFlowStageLogic.scala | 41 +-
.../amqp/impl/AbstractAmqpFlowStageLogic.scala | 19 +-
.../alpakka/amqp/impl/AmqpAsyncFlowStage.scala | 61 ++-
.../amqp/impl/AmqpAsyncUnorderedFlowStage.scala | 36 +-
.../alpakka/amqp/impl/AmqpConnectorLogic.scala | 13 +-
.../alpakka/amqp/impl/AmqpReplyToSinkStage.scala | 104 ++--
.../alpakka/amqp/impl/AmqpRpcFlowStage.scala | 326 ++++++-------
.../alpakka/amqp/impl/AmqpSimpleFlowStage.scala | 35 +-
.../stream/alpakka/amqp/impl/AmqpSourceStage.scala | 29 +-
.../stream/alpakka/amqp/javadsl/AmqpFlow.scala | 16 +-
.../alpakka/amqp/javadsl/AmqpFlowWithContext.scala | 8 +-
.../stream/alpakka/amqp/javadsl/AmqpRpcFlow.scala | 11 +-
.../stream/alpakka/amqp/javadsl/AmqpSink.scala | 3 +-
.../stream/alpakka/amqp/javadsl/AmqpSource.scala | 2 +-
.../scala/akka/stream/alpakka/amqp/model.scala | 12 +-
.../stream/alpakka/amqp/scaladsl/AmqpFlow.scala | 25 +-
.../amqp/scaladsl/AmqpFlowWithContext.scala | 14 +-
.../stream/alpakka/amqp/scaladsl/AmqpRpcFlow.scala | 10 +-
.../stream/alpakka/amqp/scaladsl/AmqpSink.scala | 2 +-
.../stream/alpakka/amqp/scaladsl/AmqpSource.scala | 2 +-
.../stream/alpakka/amqp/AmqpProxyConnection.scala | 2 +-
.../scaladsl/AmqpConnectionProvidersSpec.scala | 8 +-
.../alpakka/amqp/scaladsl/AmqpConnectorsSpec.scala | 74 +--
.../alpakka/amqp/scaladsl/AmqpFlowSpec.scala | 22 +-
...AmqpGraphStageLogicConnectionShutdownSpec.scala | 5 +-
.../stream/alpakka/amqp/scaladsl/AmqpMocking.scala | 2 +-
.../test/scala/docs/scaladsl/AmqpDocsSpec.scala | 90 ++--
.../alpakka/avroparquet/impl/AvroParquetFlow.scala | 16 +-
.../avroparquet/impl/AvroParquetSource.scala | 7 +-
.../avroparquet/javadsl/AvroParquetSink.scala | 4 +-
.../avroparquet/scaladsl/AvroParquetSink.scala | 2 +-
.../scala/docs/scaladsl/AbstractAvroParquet.scala | 11 +-
.../scala/docs/scaladsl/AvroParquetFlowSpec.scala | 14 +-
.../scala/docs/scaladsl/AvroParquetSinkSpec.scala | 16 +-
.../docs/scaladsl/AvroParquetSourceSpec.scala | 16 +-
.../eventbridge/EventBridgePublishSettings.scala | 2 +-
.../eventbridge/javadsl/EventBridgePublisher.scala | 18 +-
.../scaladsl/EventBridgePublisher.scala | 25 +-
.../aws/eventbridge/DefaultTestContext.scala | 2 +-
.../eventbridge/EventBridgePublishMockSpec.scala | 7 +-
.../aws/eventbridge/IntegrationTestContext.scala | 15 +-
.../docs/scaladsl/EventBridgePublisherSpec.scala | 4 +-
.../alpakka/awslambda/javadsl/AwsLambdaFlow.scala | 2 +-
.../alpakka/awslambda/scaladsl/AwsLambdaFlow.scala | 5 +-
.../scala/docs/scaladsl/AwsLambdaFlowSpec.scala | 16 +-
.../src/test/scala/docs/scaladsl/Examples.scala | 16 +-
.../impl/AzureQueueSinkFunctions.scala | 16 +-
.../storagequeue/impl/AzureQueueSourceStage.scala | 11 +-
.../storagequeue/javadsl/AzureQueueSink.scala | 19 +-
.../storagequeue/javadsl/AzureQueueSource.scala | 2 +-
.../storagequeue/scaladsl/AzureQueueSink.scala | 18 +-
.../storagequeue/scaladsl/AzureQueueSource.scala | 5 +-
.../alpakka/azure/storagequeue/settings.scala | 11 +-
.../test/scala/docs/scaladsl/AzureQueueSpec.scala | 26 +-
build.sbt | 241 ++++-----
.../cassandra/AkkaDiscoverySessionProvider.scala | 15 +-
.../cassandra/CassandraMetricsRegistry.scala | 2 +-
.../cassandra/CassandraSessionSettings.scala | 8 +-
.../alpakka/cassandra/CassandraWriteSettings.scala | 14 +-
.../alpakka/cassandra/CqlSessionProvider.scala | 10 +-
.../alpakka/cassandra/javadsl/CassandraFlow.scala | 34 +-
.../cassandra/javadsl/CassandraSession.scala | 47 +-
.../javadsl/CassandraSessionRegistry.scala | 4 +-
.../cassandra/javadsl/CassandraSource.scala | 2 +-
.../alpakka/cassandra/scaladsl/CassandraFlow.scala | 18 +-
.../cassandra/scaladsl/CassandraSession.scala | 26 +-
.../scaladsl/CassandraSessionRegistry.scala | 22 +-
.../cassandra/scaladsl/CassandraSource.scala | 2 +-
.../cassandra/scaladsl/CassandraLifecycle.scala | 14 +-
.../scaladsl/CassandraSessionPerformanceSpec.scala | 7 +-
.../cassandra/scaladsl/CassandraSpecBase.scala | 4 +-
.../scala/docs/javadsl/CassandraSessionSpec.scala | 14 +-
.../scala/docs/scaladsl/AkkaDiscoverySpec.scala | 2 +-
.../scala/docs/scaladsl/CassandraFlowSpec.scala | 36 +-
.../scala/docs/scaladsl/CassandraSourceSpec.scala | 8 +-
.../couchbase/CouchbaseResponseException.scala | 3 +-
.../couchbase/CouchbaseSessionRegistry.scala | 11 +-
.../couchbase/impl/CouchbaseClusterRegistry.scala | 7 +-
.../couchbase/impl/CouchbaseSessionImpl.scala | 80 ++-
.../impl/CouchbaseSessionJavaAdapter.scala | 19 +-
.../alpakka/couchbase/impl/RxUtilities.scala | 6 +-
.../alpakka/couchbase/javadsl/CouchbaseFlow.scala | 38 +-
.../couchbase/javadsl/CouchbaseSession.scala | 28 +-
.../alpakka/couchbase/javadsl/CouchbaseSink.scala | 26 +-
.../couchbase/javadsl/CouchbaseSource.scala | 12 +-
.../akka/stream/alpakka/couchbase/model.scala | 53 +-
.../alpakka/couchbase/scaladsl/CouchbaseFlow.scala | 132 +++--
.../couchbase/scaladsl/CouchbaseSession.scala | 22 +-
.../alpakka/couchbase/scaladsl/CouchbaseSink.scala | 16 +-
.../couchbase/scaladsl/CouchbaseSource.scala | 12 +-
.../couchbase/scaladsl/DiscoverySupport.scala | 18 +-
.../couchbase/testing/CouchbaseSupport.scala | 18 +-
.../scala/docs/scaladsl/CouchbaseFlowSpec.scala | 167 +++----
.../scaladsl/CouchbaseSessionExamplesSpec.scala | 6 +-
.../scala/docs/scaladsl/CouchbaseSourceSpec.scala | 10 +-
.../test/scala/docs/scaladsl/DiscoverySpec.scala | 10 +-
.../stream/alpakka/csv/scaladsl/CsvBench.scala | 6 +-
.../stream/alpakka/csv/impl/CsvFormatter.scala | 12 +-
.../akka/stream/alpakka/csv/impl/CsvParser.scala | 27 +-
.../stream/alpakka/csv/impl/CsvParsingStage.scala | 25 +-
.../alpakka/csv/impl/CsvToMapJavaStage.scala | 54 +--
.../stream/alpakka/csv/impl/CsvToMapStage.scala | 43 +-
.../alpakka/csv/scaladsl/CsvFormatting.scala | 10 +-
.../stream/alpakka/csv/scaladsl/CsvParsing.scala | 9 +-
.../alpakka/csv/scaladsl/CsvQuotingStyle.scala | 2 +-
.../stream/alpakka/csv/scaladsl/CsvToMap.scala | 67 ++-
.../akka/stream/alpakka/csv/CsvParserSpec.scala | 20 +-
.../scala/docs/scaladsl/CsvFormattingSpec.scala | 9 +-
.../test/scala/docs/scaladsl/CsvParsingSpec.scala | 37 +-
csv/src/test/scala/docs/scaladsl/CsvSpec.scala | 2 +-
.../test/scala/docs/scaladsl/CsvToMapSpec.scala | 107 ++--
.../alpakka/eip/scaladsl/PassThroughExamples.scala | 24 +-
.../akka/stream/alpakka/dynamodb/DynamoDbOp.scala | 6 +-
.../stream/alpakka/dynamodb/javadsl/DynamoDb.scala | 29 +-
.../alpakka/dynamodb/scaladsl/DynamoDb.scala | 34 +-
.../akka/stream/alpakka/dynamodb/ItemSpec.scala | 4 +-
.../akka/stream/alpakka/dynamodb/TableSpec.scala | 4 +-
.../akka/stream/alpakka/dynamodb/TestOps.scala | 48 +-
.../src/test/scala/docs/scaladsl/ExampleSpec.scala | 24 +-
.../src/test/scala/docs/scaladsl/RetrySpec.scala | 8 +-
.../ElasticsearchConnectionSettings.scala | 30 +-
.../ElasticsearchSourceSettings.scala | 49 +-
.../elasticsearch/ElasticsearchWriteSettings.scala | 38 +-
.../elasticsearch/OpensearchSourceSettings.scala | 49 +-
.../elasticsearch/OpensearchWriteSettings.scala | 30 +-
.../stream/alpakka/elasticsearch/ReadResult.scala | 4 +-
.../alpakka/elasticsearch/SourceSettingsBase.scala | 22 +-
.../alpakka/elasticsearch/WriteMessage.scala | 52 +-
.../alpakka/elasticsearch/WriteSettingsBase.scala | 13 +-
.../elasticsearch/impl/ElasticsearchApi.scala | 10 +-
.../impl/ElasticsearchSimpleFlowStage.scala | 35 +-
.../impl/ElasticsearchSourceStage.scala | 61 +--
.../elasticsearch/impl/NDJsonProtocol.scala | 2 +-
.../alpakka/elasticsearch/impl/RestBulkApi.scala | 16 +-
.../alpakka/elasticsearch/impl/RestBulkApiV5.scala | 21 +-
.../alpakka/elasticsearch/impl/RestBulkApiV7.scala | 18 +-
.../elasticsearch/javadsl/ElasticsearchFlow.scala | 35 +-
.../elasticsearch/javadsl/ElasticsearchSink.scala | 5 +-
.../javadsl/ElasticsearchSource.scala | 64 ++-
.../elasticsearch/scaladsl/ElasticsearchFlow.scala | 67 ++-
.../elasticsearch/scaladsl/ElasticsearchSink.scala | 7 +-
.../scaladsl/ElasticsearchSource.scala | 40 +-
.../elasticsearch/testkit/MessageFactory.scala | 26 +-
.../impl/ElasticsearchSimpleFlowStageTest.scala | 23 +-
.../impl/ElasticsearchSourcStageTest.scala | 8 +-
.../scaladsl/ElasticsearchConnectorBehaviour.scala | 228 ++++-----
.../scala/docs/scaladsl/ElasticsearchSpec.scala | 10 +-
.../docs/scaladsl/ElasticsearchSpecBase.scala | 10 +-
.../docs/scaladsl/ElasticsearchSpecUtils.scala | 23 +-
.../scala/docs/scaladsl/ElasticsearchV5Spec.scala | 179 +++----
.../scala/docs/scaladsl/ElasticsearchV7Spec.scala | 135 ++----
.../scaladsl/OpensearchConnectorBehaviour.scala | 228 ++++-----
.../scala/docs/scaladsl/OpensearchV1Spec.scala | 169 +++----
.../file/impl/archive/EnsureByteStreamSize.scala | 18 +-
.../file/impl/archive/TarArchiveEntry.scala | 27 +-
.../file/impl/archive/TarArchiveManager.scala | 2 +-
.../alpakka/file/impl/archive/TarReaderStage.scala | 29 +-
.../alpakka/file/impl/archive/ZipArchiveFlow.scala | 17 +-
.../file/impl/archive/ZipArchiveManager.scala | 2 +-
.../file/impl/archive/ZipReaderSource.scala | 16 +-
.../akka/stream/alpakka/file/javadsl/Archive.scala | 12 +-
.../alpakka/file/javadsl/LogRotatorSink.scala | 27 +-
.../scala/akka/stream/alpakka/file/model.scala | 63 ++-
.../stream/alpakka/file/scaladsl/Archive.scala | 14 +-
.../stream/alpakka/file/scaladsl/Directory.scala | 8 +-
.../file/scaladsl/DirectoryChangesSource.scala | 7 +-
.../alpakka/file/scaladsl/FileTailSource.scala | 19 +-
.../alpakka/file/scaladsl/LogRotatorSink.scala | 49 +-
.../file/impl/archive/TarArchiveEntrySpec.scala | 2 +-
.../file/impl/archive/ZipArchiveFlowTest.scala | 2 +-
.../src/test/scala/docs/scaladsl/ArchiveSpec.scala | 19 +-
.../test/scala/docs/scaladsl/DirectorySpec.scala | 10 +-
.../test/scala/docs/scaladsl/ExecutableUtils.scala | 8 +-
.../docs/scaladsl/FileTailSourceExtrasSpec.scala | 20 +-
.../scala/docs/scaladsl/FileTailSourceSpec.scala | 3 +-
.../scala/docs/scaladsl/LogRotatorSinkSpec.scala | 59 +--
.../test/scala/docs/scaladsl/TarArchiveSpec.scala | 43 +-
.../alpakka/ftp/impl/CommonFtpOperations.scala | 11 +-
.../alpakka/ftp/impl/FtpBrowserGraphStage.scala | 5 +-
.../impl/FtpDirectoryOperationsGraphStage.scala | 5 +-
.../stream/alpakka/ftp/impl/FtpGraphStage.scala | 2 +-
.../alpakka/ftp/impl/FtpGraphStageLogic.scala | 5 +-
.../stream/alpakka/ftp/impl/FtpIOGraphStage.scala | 41 +-
.../akka/stream/alpakka/ftp/impl/FtpLike.scala | 4 +-
.../stream/alpakka/ftp/impl/FtpOperations.scala | 13 +-
.../stream/alpakka/ftp/impl/FtpSourceFactory.scala | 58 +--
.../stream/alpakka/ftp/impl/FtpsOperations.scala | 10 +-
.../stream/alpakka/ftp/impl/SftpOperations.scala | 25 +-
.../akka/stream/alpakka/ftp/javadsl/FtpApi.scala | 164 +++----
.../main/scala/akka/stream/alpakka/ftp/model.scala | 91 ++--
.../akka/stream/alpakka/ftp/scaladsl/FtpApi.scala | 92 ++--
.../akka/stream/alpakka/ftp/BaseFtpSpec.scala | 16 +-
.../akka/stream/alpakka/ftp/BaseFtpsSpec.scala | 16 +-
.../akka/stream/alpakka/ftp/BaseSftpSpec.scala | 16 +-
.../scala/akka/stream/alpakka/ftp/BaseSpec.scala | 17 +-
.../stream/alpakka/ftp/CommonFtpStageSpec.scala | 56 +--
.../test/scala/docs/scaladsl/FtpExamplesSpec.scala | 16 +-
.../test/scala/docs/scaladsl/scalaExamples.scala | 44 +-
.../akka/stream/alpakka/geode/GeodeSettings.scala | 14 +-
.../stream/alpakka/geode/impl/GeodeCache.scala | 4 +-
.../alpakka/geode/impl/GeodeCapabilities.scala | 2 +-
.../geode/impl/pdx/DelegatingPdxSerializer.scala | 5 +-
.../stream/alpakka/geode/impl/pdx/PdxDecoder.scala | 12 +-
.../stream/alpakka/geode/impl/pdx/PdxEncoder.scala | 8 +-
.../geode/impl/pdx/ShapelessPdxSerializer.scala | 7 +-
.../geode/impl/stage/GeodeCQueryGraphLogic.scala | 14 +-
.../impl/stage/GeodeContinuousSourceStage.scala | 61 ++-
.../geode/impl/stage/GeodeFiniteSourceStage.scala | 31 +-
.../alpakka/geode/impl/stage/GeodeFlowStage.scala | 24 +-
.../geode/impl/stage/GeodeQueryGraphLogic.scala | 4 +-
.../geode/impl/stage/GeodeSourceStageLogic.scala | 4 +-
.../akka/stream/alpakka/geode/scaladsl/Geode.scala | 30 +-
.../alpakka/geode/impl/pdx/PDXDecoderSpec.scala | 2 +-
.../alpakka/geode/impl/pdx/PDXEncodeSpec.scala | 2 +-
.../alpakka/geode/impl/pdx/PdxWriterMock.scala | 8 +-
.../test/scala/docs/scaladsl/GeodeBaseSpec.scala | 8 +-
.../docs/scaladsl/GeodeContinuousSourceSpec.scala | 14 +-
.../docs/scaladsl/GeodeFiniteSourceSpec.scala | 4 +-
.../test/scala/docs/scaladsl/GeodeFlowSpec.scala | 14 +-
.../test/scala/docs/scaladsl/GeodeSinkSpec.scala | 8 +-
geode/src/test/scala/docs/scaladsl/Model.scala | 2 +-
.../scala/docs/scaladsl/PersonPdxSerializer.scala | 2 +-
.../bigquery/storage/BigQueryRecordMapImpl.scala | 4 +-
.../bigquery/storage/BigQueryStorageSettings.scala | 5 +-
.../bigquery/storage/ProtobufConverters.scala | 6 +-
.../bigquery/storage/impl/AkkaGrpcSettings.scala | 8 +-
.../bigquery/storage/impl/ArrowSource.scala | 21 +-
.../bigquery/storage/impl/AvroDecoder.scala | 2 +-
.../bigquery/storage/impl/AvroSource.scala | 18 +-
.../bigquery/storage/impl/SDKClientSource.scala | 2 +-
.../storage/javadsl/BigQueryArrowStorage.scala | 82 ++--
.../storage/javadsl/BigQueryAvroStorage.scala | 86 ++--
.../bigquery/storage/javadsl/BigQueryStorage.scala | 65 ++-
.../storage/scaladsl/BigQueryArrowStorage.scala | 88 ++--
.../storage/scaladsl/BigQueryAvroStorage.scala | 88 ++--
.../storage/scaladsl/BigQueryStorage.scala | 36 +-
.../scaladsl/BigQueryStorageAttributes.scala | 3 +-
.../scaladsl/GrpcBigQueryStorageReader.scala | 2 +-
.../storage/javadsl/AvroByteStringDecoder.scala | 6 +-
.../bigquery/storage/BigQueryStorageSpecBase.scala | 7 +-
.../bigquery/storage/mock/BigQueryMockData.scala | 10 +-
.../bigquery/storage/mock/BigQueryMockServer.scala | 23 +-
.../storage/scaladsl/ArrowByteStringDecoder.scala | 19 +-
.../storage/scaladsl/AvroByteStringDecoder.scala | 6 +-
.../scaladsl/BigQueryArrowStorageSpec.scala | 7 +-
.../storage/scaladsl/BigQueryStorageSpec.scala | 30 +-
.../test/scala/docs/scaladsl/ExampleReader.scala | 37 +-
.../googlecloud/bigquery/BigQueryAttributes.scala | 6 +-
.../googlecloud/bigquery/BigQueryException.scala | 9 +-
.../alpakka/googlecloud/bigquery/BigQueryExt.scala | 8 +-
.../googlecloud/bigquery/javadsl/BigQuery.scala | 118 +++--
.../javadsl/jackson/BigQueryMarshallers.scala | 11 +-
.../bigquery/model/DatasetJsonProtocol.scala | 14 +-
.../bigquery/model/ErrorProtoJsonProtocol.scala | 6 +-
.../bigquery/model/JobJsonProtocol.scala | 35 +-
.../bigquery/model/QueryJsonProtocol.scala | 117 +++--
.../googlecloud/bigquery/model/StringEnum.scala | 4 +-
.../bigquery/model/TableDataJsonProtocol.scala | 31 +-
.../bigquery/model/TableJsonProtocol.scala | 61 ++-
.../bigquery/scaladsl/BigQueryDatasets.scala | 22 +-
.../bigquery/scaladsl/BigQueryJobs.scala | 34 +-
.../bigquery/scaladsl/BigQueryQueries.scala | 41 +-
.../bigquery/scaladsl/BigQueryRest.scala | 4 +-
.../bigquery/scaladsl/BigQueryTableData.scala | 34 +-
.../bigquery/scaladsl/BigQueryTables.scala | 19 +-
.../bigquery/scaladsl/schema/BasicSchemas.scala | 2 +-
.../bigquery/scaladsl/schema/JavaTimeSchemas.scala | 4 +-
.../scaladsl/schema/PrimitiveSchemaWriter.scala | 2 +-
.../bigquery/scaladsl/schema/ProductSchemas.scala | 4 +-
.../bigquery/scaladsl/schema/SchemaWriter.scala | 2 +-
.../bigquery/scaladsl/schema/StandardSchemas.scala | 2 +-
.../scaladsl/spray/BigQueryBasicFormats.scala | 46 +-
.../scaladsl/spray/BigQueryCollectionFormats.scala | 8 +-
.../scaladsl/spray/BigQueryJavaTimeFormats.scala | 14 +-
.../scaladsl/spray/BigQueryJsonFormat.scala | 2 +-
.../scaladsl/spray/BigQueryJsonProtocol.scala | 2 +-
.../scaladsl/spray/BigQueryProductFormats.scala | 5 +-
.../scaladsl/spray/BigQueryRestBasicFormats.scala | 12 +-
.../scaladsl/spray/BigQueryRestJsonProtocol.scala | 2 +-
.../scaladsl/spray/BigQueryStandardFormats.scala | 2 +-
.../googlecloud/bigquery/HoverflySupport.scala | 7 +-
.../alpakka/googlecloud/bigquery/e2e/A.scala | 13 +-
.../e2e/scaladsl/BigQueryEndToEndSpec.scala | 13 +-
.../bigquery/e2e/scaladsl/EndToEndHelper.scala | 17 +-
.../bigquery/scaladsl/BigQueryQueriesSpec.scala | 43 +-
.../scaladsl/schema/BigQuerySchemasSpec.scala | 11 +-
.../scaladsl/spray/BigQueryJsonProtocolSpec.scala | 6 +-
.../src/test/scala/docs/scaladsl/BigQueryDoc.scala | 50 +-
.../googlecloud/pubsub/grpc/PubSubSettings.scala | 24 +-
.../pubsub/grpc/impl/AkkaGrpcSettings.scala | 8 +-
.../pubsub/grpc/impl/DeprecatedCredentials.scala | 4 +-
.../pubsub/grpc/javadsl/GooglePubSub.scala | 21 +-
.../pubsub/grpc/javadsl/GrpcPublisher.scala | 2 +-
.../pubsub/grpc/javadsl/GrpcSubscriber.scala | 6 +-
.../pubsub/grpc/scaladsl/GooglePubSub.scala | 28 +-
.../pubsub/grpc/scaladsl/GrpcPublisher.scala | 4 +-
.../pubsub/grpc/scaladsl/GrpcSubscriber.scala | 4 +-
.../src/test/scala/docs/scaladsl/ExampleApp.scala | 14 +-
.../test/scala/docs/scaladsl/IntegrationSpec.scala | 42 +-
.../googlecloud/pubsub/impl/PubSubApi.scala | 44 +-
.../googlecloud/pubsub/javadsl/GooglePubSub.scala | 28 +-
.../stream/alpakka/googlecloud/pubsub/model.scala | 116 +++--
.../googlecloud/pubsub/scaladsl/GooglePubSub.scala | 39 +-
.../googlecloud/pubsub/GooglePubSubSpec.scala | 27 +-
.../googlecloud/pubsub/impl/PubSubApiSpec.scala | 133 ++---
.../test/scala/docs/scaladsl/ExampleUsage.scala | 54 +--
.../test/scala/docs/scaladsl/IntegrationSpec.scala | 6 +-
.../alpakka/googlecloud/storage/Bucket.scala | 21 +-
.../googlecloud/storage/CustomerEncryption.scala | 2 +-
.../alpakka/googlecloud/storage/FailedUpload.scala | 3 +-
.../alpakka/googlecloud/storage/GCSExt.scala | 2 +-
.../alpakka/googlecloud/storage/GCStorageExt.scala | 2 +-
.../googlecloud/storage/GCStorageSettings.scala | 26 +-
.../googlecloud/storage/ObjectAccessControls.scala | 147 +++---
.../googlecloud/storage/StorageObject.scala | 24 +-
.../googlecloud/storage/StorageSettings.scala | 4 +-
.../storage/impl/BucketListResult.scala | 12 +-
.../alpakka/googlecloud/storage/impl/Formats.scala | 102 ++--
.../googlecloud/storage/impl/GCStorageStream.scala | 81 ++--
.../googlecloud/storage/impl/RewriteResponse.scala | 3 +-
.../storage/impl/UploadPartResponse.scala | 4 +-
.../googlecloud/storage/javadsl/GCStorage.scala | 77 ++-
.../googlecloud/storage/scaladsl/GCStorage.scala | 57 ++-
.../alpakka/googlecloud/storage/settings.scala | 2 +-
.../alpakka/googlecloud/storage/GCSExtSpec.scala | 4 +-
.../googlecloud/storage/GCSSettingsSpec.scala | 4 +-
.../googlecloud/storage/GCStorageExtSpec.scala | 4 +-
.../storage/GCStorageSettingsSpec.scala | 4 +-
.../storage/WithMaterializerGlobal.scala | 4 +-
.../impl/GCStorageStreamIntegrationSpec.scala | 49 +-
.../storage/scaladsl/GCStorageWiremockBase.scala | 212 +++-----
.../scala/docs/scaladsl/GCStorageSinkSpec.scala | 21 +-
.../scala/docs/scaladsl/GCStorageSourceSpec.scala | 127 ++---
.../stream/alpakka/google/GoogleAttributes.scala | 4 +-
.../akka/stream/alpakka/google/GoogleExt.scala | 8 +-
.../stream/alpakka/google/GoogleSettings.scala | 66 ++-
.../stream/alpakka/google/PaginatedRequest.scala | 5 +-
.../stream/alpakka/google/ResumableUpload.scala | 34 +-
.../stream/alpakka/google/auth/AccessToken.scala | 2 +-
.../google/auth/ComputeEngineCredentials.scala | 4 +-
.../stream/alpakka/google/auth/Credentials.scala | 18 +-
.../google/auth/GoogleComputeMetadata.scala | 6 +-
.../stream/alpakka/google/auth/GoogleOAuth2.scala | 19 +-
.../google/auth/GoogleOAuth2Credentials.scala | 11 +-
.../google/auth/GoogleOAuth2Exception.scala | 11 +-
.../stream/alpakka/google/auth/NoCredentials.scala | 4 +-
.../alpakka/google/auth/OAuth2Credentials.scala | 20 +-
.../google/auth/ServiceAccountCredentials.scala | 24 +-
.../google/auth/UserAccessCredentials.scala | 36 +-
.../alpakka/google/auth/UserAccessMetadata.scala | 8 +-
.../google/http/ForwardProxyHttpsContext.scala | 6 +-
.../google/http/ForwardProxyPoolSettings.scala | 26 +-
.../stream/alpakka/google/http/GoogleHttp.scala | 119 +++--
.../akka/stream/alpakka/google/implicits.scala | 7 +-
.../stream/alpakka/google/javadsl/Google.scala | 19 +-
.../stream/alpakka/google/scaladsl/Google.scala | 7 +-
.../google/scaladsl/`X-Upload-Content-Type`.scala | 11 +-
.../stream/alpakka/google/util/AnnotateLast.scala | 2 +-
.../stream/alpakka/google/util/EitherFlow.scala | 14 +-
.../akka/stream/alpakka/google/util/Retry.scala | 18 +-
.../alpakka/google/GoogleHttpException.scala | 2 +-
.../stream/alpakka/google/GoogleSettingsSpec.scala | 6 +-
.../stream/alpakka/google/HoverflySupport.scala | 7 +-
.../alpakka/google/PaginatedRequestSpec.scala | 22 +-
.../alpakka/google/ResumableUploadSpec.scala | 16 +-
.../alpakka/google/auth/GoogleOAuth2Spec.scala | 8 +-
.../google/auth/OAuth2CredentialsSpec.scala | 8 +-
.../alpakka/google/http/GoogleHttpSpec.scala | 62 +--
.../test/scala/docs/scaladsl/GoogleCommonDoc.scala | 10 +-
.../firebase/fcm/FcmNotificationModels.scala | 15 +-
.../alpakka/google/firebase/fcm/FcmSettings.scala | 79 ++-
.../google/firebase/fcm/impl/FcmFlows.scala | 12 +-
.../google/firebase/fcm/impl/FcmJsonSupport.scala | 21 +-
.../google/firebase/fcm/impl/FcmSender.scala | 9 +-
.../google/firebase/fcm/javadsl/GoogleFcm.scala | 6 +-
.../google/firebase/fcm/scaladsl/GoogleFcm.scala | 6 +-
.../google/firebase/fcm/v1/impl/FcmFlows.scala | 2 +-
.../firebase/fcm/v1/impl/FcmJsonSupport.scala | 61 ++-
.../google/firebase/fcm/v1/impl/FcmSender.scala | 9 +-
.../google/firebase/fcm/v1/javadsl/GoogleFcm.scala | 4 +-
.../firebase/fcm/v1/models/AndroidConfig.scala | 9 +-
.../google/firebase/fcm/v1/models/ApnsConfig.scala | 3 +-
.../firebase/fcm/v1/models/FcmNotification.scala | 7 +-
.../firebase/fcm/v1/models/WebPushConfig.scala | 6 +-
.../firebase/fcm/v1/scaladsl/GoogleFcm.scala | 6 +-
.../firebase/fcm/v1/FcmNotificationSpec.scala | 2 +-
.../firebase/fcm/v1/impl/FcmSenderSpec.scala | 48 +-
.../src/test/scala/docs/scaladsl/FcmExamples.scala | 32 +-
.../akka/stream/alpakka/hbase/HTableSettings.scala | 24 +-
.../alpakka/hbase/impl/HBaseCapabilities.scala | 11 +-
.../stream/alpakka/hbase/impl/HBaseFlowStage.scala | 22 +-
.../alpakka/hbase/impl/HBaseSourceStage.scala | 12 +-
.../stream/alpakka/hbase/javadsl/HTableStage.scala | 8 +-
.../alpakka/hbase/scaladsl/HTableStage.scala | 8 +-
.../test/scala/docs/scaladsl/HBaseStageSpec.scala | 40 +-
.../stream/alpakka/hdfs/impl/HdfsFlowStage.scala | 20 +-
.../impl/strategy/DefaultRotationStrategy.scala | 6 +-
.../hdfs/impl/strategy/DefaultSyncStrategy.scala | 3 +-
.../hdfs/impl/writer/CompressedDataWriter.scala | 10 +-
.../alpakka/hdfs/impl/writer/DataWriter.scala | 5 +-
.../alpakka/hdfs/impl/writer/HdfsWriter.scala | 2 +-
.../alpakka/hdfs/impl/writer/SequenceWriter.scala | 33 +-
.../stream/alpakka/hdfs/javadsl/HdfsFlow.scala | 43 +-
.../stream/alpakka/hdfs/javadsl/HdfsSource.scala | 21 +-
.../scala/akka/stream/alpakka/hdfs/model.scala | 12 +-
.../stream/alpakka/hdfs/scaladsl/HdfsFlow.scala | 60 +--
.../stream/alpakka/hdfs/scaladsl/HdfsSource.scala | 17 +-
.../akka/stream/alpakka/hdfs/util/TestUtils.scala | 33 +-
.../test/scala/docs/scaladsl/HdfsReaderSpec.scala | 40 +-
.../test/scala/docs/scaladsl/HdfsWriterSpec.scala | 116 ++---
.../huawei/pushkit/ForwardProxyHttpsContext.scala | 8 +-
.../huawei/pushkit/ForwardProxyPoolSettings.scala | 10 +-
.../alpakka/huawei/pushkit/HmsSettingExt.scala | 8 +-
.../alpakka/huawei/pushkit/HmsSettings.scala | 49 +-
.../alpakka/huawei/pushkit/impl/HmsTokenApi.scala | 17 +-
.../alpakka/huawei/pushkit/impl/PushKitFlows.scala | 12 +-
.../huawei/pushkit/impl/PushKitJsonSupport.scala | 33 +-
.../huawei/pushkit/impl/PushKitSender.scala | 19 +-
.../huawei/pushkit/javadsl/HmsPushKit.scala | 4 +-
.../huawei/pushkit/models/AndroidConfig.scala | 104 ++--
.../alpakka/huawei/pushkit/models/ApnsConfig.scala | 4 +-
.../pushkit/models/PushKitNotification.scala | 18 +-
.../alpakka/huawei/pushkit/models/WebConfig.scala | 30 +-
.../huawei/pushkit/scaladsl/HmsPushKit.scala | 6 +-
.../huawei/pushkit/ConditionBuilderSpec.scala | 2 +-
.../huawei/pushkit/impl/HmsTokenApiSpec.scala | 41 +-
.../huawei/pushkit/impl/PushKitSenderSpec.scala | 54 +--
.../test/scala/docs/scaladsl/PushKitExamples.scala | 22 +-
.../alpakka/influxdb/InfluxDbReadSettings.scala | 6 +-
.../alpakka/influxdb/InfluxDbWriteMessage.scala | 18 +-
.../influxdb/impl/AlpakkaResultMapperHelper.scala | 36 +-
.../alpakka/influxdb/impl/InfluxDbFlowStage.scala | 44 +-
.../influxdb/impl/InfluxDbSourceStage.scala | 38 +-
.../alpakka/influxdb/javadsl/InfluxDbFlow.scala | 20 +-
.../alpakka/influxdb/javadsl/InfluxDbSink.scala | 16 +-
.../alpakka/influxdb/javadsl/InfluxDbSource.scala | 8 +-
.../alpakka/influxdb/scaladsl/InfluxDbFlow.scala | 20 +-
.../alpakka/influxdb/scaladsl/InfluxDbSink.scala | 8 +-
.../alpakka/influxdb/scaladsl/InfluxDbSource.scala | 8 +-
.../src/test/scala/docs/scaladsl/FlowSpec.scala | 35 +-
.../scala/docs/scaladsl/InfluxDbSourceSpec.scala | 10 +-
.../test/scala/docs/scaladsl/InfluxDbSpec.scala | 32 +-
.../stream/alpakka/ironmq/IronMqSettings.scala | 41 +-
.../scala/akka/stream/alpakka/ironmq/domain.scala | 2 +-
.../akka/stream/alpakka/ironmq/impl/Codec.scala | 2 +-
.../stream/alpakka/ironmq/impl/IronMqClient.scala | 62 ++-
.../alpakka/ironmq/impl/IronMqPullStage.scala | 10 +-
.../alpakka/ironmq/impl/IronMqPushStage.scala | 16 +-
.../alpakka/ironmq/javadsl/IronMqConsumer.scala | 2 +-
.../alpakka/ironmq/javadsl/IronMqProducer.scala | 26 +-
.../alpakka/ironmq/scaladsl/IronMqProducer.scala | 15 +-
.../akka/stream/alpakka/ironmq/IronMqSpec.scala | 4 +-
.../alpakka/ironmq/impl/IronMqClientForTests.scala | 2 +-
.../alpakka/ironmq/impl/IronMqClientSpec.scala | 2 +-
.../alpakka/ironmq/impl/IronMqPullStageSpec.scala | 4 +-
.../alpakka/ironmq/impl/IronMqPushStageSpec.scala | 2 +-
.../ironmq/scaladsl/IronMqConsumerSpec.scala | 6 +-
.../ironmq/scaladsl/IronMqProducerSpec.scala | 14 +-
.../test/scala/docs/scaladsl/IronMqDocsSpec.scala | 10 +-
.../akka/stream/alpakka/jms/AcknowledgeMode.scala | 15 +-
.../alpakka/jms/ConnectionRetrySettings.scala | 14 +-
.../akka/stream/alpakka/jms/Credentials.scala | 24 +-
.../akka/stream/alpakka/jms/Destinations.scala | 2 +-
.../scala/akka/stream/alpakka/jms/Envelopes.scala | 4 +-
.../stream/alpakka/jms/JmsBrowseSettings.scala | 20 +-
.../stream/alpakka/jms/JmsConsumerSettings.scala | 21 +-
.../akka/stream/alpakka/jms/JmsExceptions.scala | 15 +-
.../akka/stream/alpakka/jms/JmsMessages.scala | 279 +++++------
.../stream/alpakka/jms/JmsProducerSettings.scala | 23 +-
.../stream/alpakka/jms/SendRetrySettings.scala | 17 +-
.../alpakka/jms/impl/InternalConnectionState.scala | 6 +-
.../alpakka/jms/impl/JmsAckSourceStage.scala | 12 +-
.../stream/alpakka/jms/impl/JmsBrowseStage.scala | 6 +-
.../stream/alpakka/jms/impl/JmsConnector.scala | 56 +--
.../stream/alpakka/jms/impl/JmsConsumerStage.scala | 5 +-
.../alpakka/jms/impl/JmsMessageProducer.scala | 48 +-
.../stream/alpakka/jms/impl/JmsMessageReader.scala | 12 +-
.../stream/alpakka/jms/impl/JmsProducerStage.scala | 37 +-
.../stream/alpakka/jms/impl/JmsTxSourceStage.scala | 16 +-
.../akka/stream/alpakka/jms/impl/Sessions.scala | 27 +-
.../stream/alpakka/jms/impl/SourceStageLogic.scala | 23 +-
.../stream/alpakka/jms/javadsl/JmsConsumer.scala | 6 +-
.../stream/alpakka/jms/javadsl/JmsProducer.scala | 22 +-
.../alpakka/jms/scaladsl/JmsConnectorState.scala | 12 +-
.../stream/alpakka/jms/scaladsl/JmsConsumer.scala | 2 +-
.../stream/alpakka/jms/scaladsl/JmsProducer.scala | 22 +-
.../akka/stream/alpakka/jms/scaladsl/package.scala | 18 +-
.../alpakka/jms/JmsConnectionStatusSpec.scala | 36 +-
.../stream/alpakka/jms/JmsProducerRetrySpec.scala | 40 +-
.../scala/akka/stream/alpakka/jms/JmsSpec.scala | 22 +-
.../alpakka/jms/impl/JmsMessageProducerSpec.scala | 20 +-
.../alpakka/jms/impl/SoftReferenceCacheSpec.scala | 4 +-
.../jms/scaladsl/CachedConnectionFactory.scala | 2 +-
.../jms/scaladsl/JmsAckConnectorsSpec.scala | 101 ++--
.../scaladsl/JmsBufferedAckConnectorsSpec.scala | 96 ++--
.../scala/docs/scaladsl/JmsConnectorsSpec.scala | 359 ++++++--------
.../docs/scaladsl/JmsIbmmqConnectorsSpec.scala | 60 ++-
.../test/scala/docs/scaladsl/JmsSettingsSpec.scala | 22 +-
.../scala/docs/scaladsl/JmsTxConnectorsSpec.scala | 147 +++---
.../alpakka/json/impl/JsonStreamReader.scala | 13 +-
.../test/scala/docs/scaladsl/JsonReaderTest.scala | 2 +-
.../stream/alpakka/kinesis/CommittableRecord.scala | 11 +-
.../stream/alpakka/kinesis/KinesisErrors.scala | 2 +-
.../alpakka/kinesis/KinesisFlowSettings.scala | 18 +-
.../alpakka/kinesis/KinesisSchedulerSettings.scala | 6 +-
.../stream/alpakka/kinesis/ShardSettings.scala | 21 +-
.../kinesis/impl/KinesisSchedulerSourceStage.scala | 16 +-
.../alpakka/kinesis/impl/KinesisSourceStage.scala | 20 +-
.../alpakka/kinesis/impl/ShardProcessor.scala | 21 +-
.../alpakka/kinesis/javadsl/KinesisFlow.scala | 18 +-
.../kinesis/javadsl/KinesisSchedulerSource.scala | 19 +-
.../alpakka/kinesis/javadsl/KinesisSink.scala | 6 +-
.../alpakka/kinesis/javadsl/KinesisSource.scala | 4 +-
.../alpakka/kinesis/scaladsl/KinesisFlow.scala | 66 +--
.../kinesis/scaladsl/KinesisSchedulerSource.scala | 18 +-
.../alpakka/kinesis/scaladsl/KinesisSink.scala | 15 +-
.../alpakka/kinesis/scaladsl/KinesisSource.scala | 8 +-
.../KinesisFirehoseFlowSettings.scala | 18 +-
.../javadsl/KinesisFirehoseFlow.scala | 10 +-
.../javadsl/KinesisFirehoseSink.scala | 8 +-
.../scaladsl/KinesisFirehoseFlow.scala | 28 +-
.../scaladsl/KinesisFirehoseSink.scala | 3 +-
.../alpakka/kinesis/DefaultTestContext.scala | 9 +-
.../stream/alpakka/kinesis/KinesisFlowSpec.scala | 22 +-
.../akka/stream/alpakka/kinesis/KinesisMock.scala | 2 +-
.../kinesis/KinesisSchedulerSourceSpec.scala | 76 ++-
.../stream/alpakka/kinesis/KinesisSourceSpec.scala | 15 +-
.../scala/akka/stream/alpakka/kinesis/Valve.scala | 13 +-
.../kinesisfirehose/KinesisFirehoseFlowSpec.scala | 2 +-
.../kinesisfirehose/KinesisFirehoseMock.scala | 2 +-
.../src/test/scala/docs/scaladsl/KclSnippets.scala | 34 +-
.../docs/scaladsl/KinesisFirehoseSnippets.scala | 22 +-
.../test/scala/docs/scaladsl/KinesisSnippets.scala | 39 +-
.../akka/stream/alpakka/kudu/KuduClientExt.scala | 2 +-
.../stream/alpakka/kudu/KuduTableSettings.scala | 33 +-
.../alpakka/kudu/impl/KuduCapabilities.scala | 6 +-
.../stream/alpakka/kudu/impl/KuduFlowStage.scala | 16 +-
.../stream/alpakka/kudu/javadsl/KuduTable.scala | 4 +-
.../stream/alpakka/kudu/scaladsl/KuduTable.scala | 8 +-
.../test/scala/docs/scaladsl/KuduTableSpec.scala | 24 +-
.../stream/alpakka/mongodb/DocumentReplace.scala | 1 -
.../stream/alpakka/mongodb/DocumentUpdate.scala | 1 -
.../stream/alpakka/mongodb/javadsl/MongoFlow.scala | 29 +-
.../stream/alpakka/mongodb/javadsl/MongoSink.scala | 17 +-
.../alpakka/mongodb/scaladsl/MongoFlow.scala | 56 +--
.../alpakka/mongodb/scaladsl/MongoSink.scala | 22 +-
.../test/scala/docs/scaladsl/MongoSinkSpec.scala | 43 +-
.../test/scala/docs/scaladsl/MongoSourceSpec.scala | 6 +-
.../scala/akka/stream/alpakka/mqtt/MqttPerf.scala | 17 +-
.../stream/alpakka/mqtt/streaming/MqttPerf.scala | 18 +-
.../mqtt/streaming/MqttSessionSettings.scala | 64 ++-
.../mqtt/streaming/impl/BehaviorRunner.scala | 4 +-
.../alpakka/mqtt/streaming/impl/ClientState.scala | 228 ++++-----
.../mqtt/streaming/impl/MqttFrameStage.scala | 19 +-
.../mqtt/streaming/impl/QueueOfferState.scala | 8 +-
.../alpakka/mqtt/streaming/impl/RequestState.scala | 138 +++---
.../alpakka/mqtt/streaming/impl/ServerState.scala | 360 ++++++--------
.../alpakka/mqtt/streaming/javadsl/Mqtt.scala | 11 +-
.../akka/stream/alpakka/mqtt/streaming/model.scala | 175 ++++---
.../alpakka/mqtt/streaming/scaladsl/Mqtt.scala | 20 +-
.../mqtt/streaming/scaladsl/MqttSession.scala | 178 ++++---
.../mqtt/streaming/impl/MqttFrameStageSpec.scala | 2 +-
.../mqtt/streaming/impl/QueueOfferStateSpec.scala | 13 +-
.../mqtt/streaming/impl/RequestStateSpec.scala | 12 +-
.../scala/docs/scaladsl/MqttActorSystemsSpec.scala | 2 +-
.../test/scala/docs/scaladsl/MqttCodecSpec.scala | 43 +-
.../test/scala/docs/scaladsl/MqttFlowSpec.scala | 58 ++-
.../test/scala/docs/scaladsl/MqttSessionSpec.scala | 150 +++---
.../stream/alpakka/mqtt/impl/MqttFlowStage.scala | 88 ++--
.../alpakka/mqtt/impl/MqttFlowStageWithAck.scala | 58 +--
.../stream/alpakka/mqtt/javadsl/MqttFlow.scala | 12 +-
.../stream/alpakka/mqtt/javadsl/MqttSink.scala | 4 +-
.../stream/alpakka/mqtt/javadsl/MqttSource.scala | 8 +-
.../scala/akka/stream/alpakka/mqtt/model.scala | 24 +-
.../stream/alpakka/mqtt/scaladsl/MqttFlow.scala | 40 +-
.../stream/alpakka/mqtt/scaladsl/MqttSink.scala | 2 +-
.../stream/alpakka/mqtt/scaladsl/MqttSource.scala | 16 +-
.../scala/akka/stream/alpakka/mqtt/settings.scala | 84 ++--
.../test/scala/docs/scaladsl/MqttFlowSpec.scala | 28 +-
.../test/scala/docs/scaladsl/MqttSinkSpec.scala | 22 +-
.../test/scala/docs/scaladsl/MqttSourceSpec.scala | 118 ++---
.../test/scala/docs/scaladsl/MqttSpecBase.scala | 5 +-
.../alpakka/orientdb/OrientDbSourceSettings.scala | 15 +-
.../alpakka/orientdb/OrientDbWriteSettings.scala | 17 +-
.../alpakka/orientdb/impl/OrientDbFlowStage.scala | 9 +-
.../orientdb/impl/OrientDbSourceStage.scala | 19 +-
.../alpakka/orientdb/javadsl/OrientDbFlow.scala | 20 +-
.../alpakka/orientdb/javadsl/OrientDbSink.scala | 14 +-
.../alpakka/orientdb/javadsl/OrientDbSource.scala | 22 +-
.../alpakka/orientdb/scaladsl/OrientDbFlow.scala | 36 +-
.../alpakka/orientdb/scaladsl/OrientDbSink.scala | 10 +-
.../alpakka/orientdb/scaladsl/OrientDbSource.scala | 18 +-
.../test/scala/docs/scaladsl/OrientDbSpec.scala | 87 ++--
.../akka/stream/alpakka/pravega/PravegaEvent.scala | 2 +-
.../pravega/PravegaReaderGroupManager.scala | 12 +-
.../stream/alpakka/pravega/PravegaSettings.scala | 327 ++++++-------
.../alpakka/pravega/impl/PravegaCapabilities.scala | 2 +-
.../stream/alpakka/pravega/impl/PravegaFlow.scala | 27 +-
.../alpakka/pravega/impl/PravegaSource.scala | 30 +-
.../pravega/impl/PravegaTableReadFlow.scala | 20 +-
.../alpakka/pravega/impl/PravegaTableSource.scala | 26 +-
.../pravega/impl/PravegaTableWriteFlow.scala | 31 +-
.../alpakka/pravega/impl/PravegaWriter.scala | 3 +-
.../stream/alpakka/pravega/scaladsl/Pravega.scala | 8 +-
.../alpakka/pravega/scaladsl/PravegaTable.scala | 31 +-
.../stream/alpakka/pravega/PravegaBaseSpec.scala | 2 +-
.../stream/alpakka/pravega/PravegaGraphSpec.scala | 13 +-
.../alpakka/pravega/PravegaKVTableSpec.scala | 2 +-
.../pravega/PravegaStreamAndTableSpec.scala | 16 +-
.../scala/docs/scaladsl/PravegaReadWriteDocs.scala | 4 +-
.../scala/docs/scaladsl/PravegaSettingsSpec.scala | 23 +-
project/AutomaticModuleName.scala | 5 +-
project/Common.scala | 169 ++++---
project/Dependencies.scala | 534 +++++++++-----------
project/TestChanged.scala | 15 +-
.../reference/impl/ReferenceFlowStage.scala | 19 +-
.../reference/impl/ReferenceSourceStage.scala | 21 +-
.../impl/ReferenceWithResourceFlowStage.scala | 19 +-
.../alpakka/reference/javadsl/Reference.scala | 13 +-
.../akka/stream/alpakka/reference/model.scala | 14 +-
.../alpakka/reference/scaladsl/Reference.scala | 8 +-
.../akka/stream/alpakka/reference/settings.scala | 14 +-
.../alpakka/reference/testkit/MessageFactory.scala | 8 +-
.../test/scala/docs/scaladsl/ReferenceSpec.scala | 17 +-
.../scala/akka/stream/alpakka/s3/S3Exception.scala | 13 +-
.../main/scala/akka/stream/alpakka/s3/S3Ext.scala | 2 +-
.../scala/akka/stream/alpakka/s3/S3Headers.scala | 19 +-
.../main/scala/akka/stream/alpakka/s3/Utils.scala | 3 +-
.../alpakka/s3/headers/ServerSideEncryption.scala | 34 +-
.../scala/akka/stream/alpakka/s3/impl/Chunk.scala | 2 +-
.../akka/stream/alpakka/s3/impl/DiskBuffer.scala | 2 +-
.../akka/stream/alpakka/s3/impl/HttpRequests.scala | 122 ++---
.../akka/stream/alpakka/s3/impl/Marshalling.scala | 81 ++--
.../akka/stream/alpakka/s3/impl/MemoryBuffer.scala | 4 +-
.../stream/alpakka/s3/impl/MemoryWithContext.scala | 4 +-
.../akka/stream/alpakka/s3/impl/S3Stream.scala | 538 +++++++++------------
.../stream/alpakka/s3/impl/SplitAfterSize.scala | 2 +-
.../s3/impl/SplitAfterSizeWithContext.scala | 10 +-
.../alpakka/s3/impl/auth/CanonicalRequest.scala | 19 +-
.../akka/stream/alpakka/s3/impl/auth/Signer.scala | 26 +-
.../stream/alpakka/s3/impl/auth/SigningKey.scala | 10 +-
.../scala/akka/stream/alpakka/s3/javadsl/S3.scala | 434 ++++++++---------
.../main/scala/akka/stream/alpakka/s3/model.scala | 270 +++++------
.../scala/akka/stream/alpakka/s3/scaladsl/S3.scala | 219 ++++-----
.../scala/akka/stream/alpakka/s3/settings.scala | 100 ++--
.../akka/stream/alpakka/s3/MinioContainer.scala | 4 +-
.../scala/akka/stream/alpakka/s3/MinioS3Test.scala | 5 +-
.../akka/stream/alpakka/s3/S3SettingsSpec.scala | 37 +-
.../stream/alpakka/s3/impl/DiskBufferSpec.scala | 18 +-
.../stream/alpakka/s3/impl/HttpRequestsSpec.scala | 57 +--
.../stream/alpakka/s3/impl/MarshallingSpec.scala | 93 ++--
.../stream/alpakka/s3/impl/MemoryBufferSpec.scala | 9 +-
.../stream/alpakka/s3/impl/S3HeadersSpec.scala | 15 +-
.../akka/stream/alpakka/s3/impl/S3StreamSpec.scala | 56 +--
.../alpakka/s3/impl/SplitAfterSizeSpec.scala | 40 +-
.../s3/impl/auth/CanonicalRequestSpec.scala | 58 +--
.../stream/alpakka/s3/impl/auth/SignerSpec.scala | 34 +-
.../alpakka/s3/impl/auth/SigningKeySpec.scala | 10 +-
.../impl/auth/SplitAfterSizeWithContextSpec.scala | 19 +-
.../alpakka/s3/impl/auth/StreamUtilsSpec.scala | 10 +-
.../s3/scaladsl/S3ClientIntegrationSpec.scala | 2 +-
.../stream/alpakka/s3/scaladsl/S3ExtSpec.scala | 4 +-
.../alpakka/s3/scaladsl/S3IntegrationSpec.scala | 246 ++++------
.../s3/scaladsl/S3SlowMinioIntegrationSpec.scala | 9 +-
.../alpakka/s3/scaladsl/S3WireMockBase.scala | 222 +++------
s3/src/test/scala/docs/scaladsl/S3SinkSpec.scala | 120 ++---
s3/src/test/scala/docs/scaladsl/S3SourceSpec.scala | 58 ++-
.../recordio/impl/RecordIOFramingStage.scala | 10 +-
.../scala/docs/scaladsl/RecordIOFramingSpec.scala | 71 ++-
.../akka/stream/alpakka/slick/javadsl/Slick.scala | 67 ++-
.../stream/alpakka/slick/javadsl/package.scala | 3 +-
.../akka/stream/alpakka/slick/scaladsl/Slick.scala | 22 +-
.../stream/alpakka/slick/scaladsl/package.scala | 2 +-
.../src/test/scala/docs/scaladsl/DocSnippets.scala | 31 +-
slick/src/test/scala/docs/scaladsl/SlickSpec.scala | 56 +--
.../stream/alpakka/sns/javadsl/SnsPublisher.scala | 26 +-
.../stream/alpakka/sns/scaladsl/SnsPublisher.scala | 25 +-
.../stream/alpakka/sns/DefaultTestContext.scala | 2 +-
.../alpakka/sns/IntegrationTestContext.scala | 14 +-
.../stream/alpakka/sns/SnsPublishMockingSpec.scala | 6 +-
.../scala/docs/scaladsl/SnsPublisherSpec.scala | 26 +-
.../akka/stream/alpakka/solr/SolrMessages.scala | 42 +-
.../stream/alpakka/solr/SolrUpdateSettings.scala | 9 +-
.../stream/alpakka/solr/impl/SolrFlowStage.scala | 44 +-
.../stream/alpakka/solr/impl/SolrSourceStage.scala | 4 +-
.../stream/alpakka/solr/javadsl/SolrFlow.scala | 61 +--
.../stream/alpakka/solr/javadsl/SolrSink.scala | 21 +-
.../stream/alpakka/solr/scaladsl/SolrFlow.scala | 65 +--
.../stream/alpakka/solr/scaladsl/SolrSink.scala | 16 +-
solr/src/test/scala/docs/scaladsl/SolrSpec.scala | 165 +++----
.../stream/alpakka/sqs/SqsAckBatchSettings.scala | 3 +-
.../stream/alpakka/sqs/SqsAckGroupedSettings.scala | 33 +-
.../akka/stream/alpakka/sqs/SqsAckSettings.scala | 12 +-
.../scala/akka/stream/alpakka/sqs/SqsModel.scala | 39 +-
.../alpakka/sqs/SqsPublishBatchSettings.scala | 3 +-
.../alpakka/sqs/SqsPublishGroupedSettings.scala | 21 +-
.../stream/alpakka/sqs/SqsSourceSettings.scala | 22 +-
.../alpakka/sqs/impl/BalancingMapAsync.scala | 41 +-
.../stream/alpakka/sqs/javadsl/SqsAckFlow.scala | 8 +-
.../stream/alpakka/sqs/javadsl/SqsAckSink.scala | 10 +-
.../alpakka/sqs/javadsl/SqsPublishFlow.scala | 14 +-
.../alpakka/sqs/javadsl/SqsPublishSink.scala | 27 +-
.../stream/alpakka/sqs/scaladsl/SqsAckFlow.scala | 41 +-
.../stream/alpakka/sqs/scaladsl/SqsAckSink.scala | 10 +-
.../alpakka/sqs/scaladsl/SqsPublishFlow.scala | 39 +-
.../alpakka/sqs/scaladsl/SqsPublishSink.scala | 36 +-
.../stream/alpakka/sqs/scaladsl/SqsSource.scala | 11 +-
.../alpakka/sqs/testkit/MessageFactory.scala | 17 +-
.../alpakka/sqs/scaladsl/DefaultTestContext.scala | 20 +-
.../sqs/scaladsl/MessageAttributeNameSpec.scala | 3 +-
.../alpakka/sqs/scaladsl/SqsPublishSinkSpec.scala | 62 +--
.../alpakka/sqs/scaladsl/SqsSourceMockSpec.scala | 27 +-
sqs/src/test/scala/docs/scaladsl/SqsAckSpec.scala | 69 ++-
.../test/scala/docs/scaladsl/SqsPublishSpec.scala | 57 ++-
.../test/scala/docs/scaladsl/SqsSourceSpec.scala | 30 +-
.../stream/alpakka/sse/javadsl/EventSource.scala | 28 +-
.../stream/alpakka/sse/scaladsl/EventSource.scala | 23 +-
.../test/scala/docs/scaladsl/EventSourceSpec.scala | 28 +-
.../stream/alpakka/testkit/CapturingAppender.scala | 7 +-
.../akka/stream/alpakka/testkit/LogbackUtil.scala | 9 +-
.../testkit/javadsl/LogCapturingJunit4.scala | 9 +-
.../alpakka/testkit/scaladsl/LogCapturing.scala | 6 +-
.../stream/alpakka/testkit/scaladsl/Repeated.scala | 2 +-
.../alpakka/text/impl/CharsetDecodingFlow.scala | 4 +-
.../stream/alpakka/text/impl/CharsetLogic.scala | 28 +-
.../alpakka/text/impl/CharsetTranscodingFlow.scala | 4 +-
.../stream/alpakka/text/javadsl/TextFlow.scala | 2 +-
.../stream/alpakka/text/scaladsl/TextFlow.scala | 2 +-
.../text/scaladsl/CharsetCodingFlowsSpec.scala | 10 +-
.../docs/scaladsl/CharsetCodingFlowsDoc.scala | 4 +-
.../akka/stream/alpakka/udp/impl/UdpBind.scala | 27 +-
.../akka/stream/alpakka/udp/impl/UdpSend.scala | 20 +-
.../akka/stream/alpakka/udp/javadsl/Udp.scala | 20 +-
.../akka/stream/alpakka/udp/scaladsl/Udp.scala | 24 +-
udp/src/test/scala/docs/scaladsl/UdpSpec.scala | 4 +-
.../impl/UnixDomainSocketImpl.scala | 102 ++--
.../javadsl/UnixDomainSocket.scala | 22 +-
.../scaladsl/UnixDomainSocket.scala | 21 +-
.../scala/docs/scaladsl/UnixDomainSocketSpec.scala | 27 +-
.../akka/stream/alpakka/xml/impl/Coalesce.scala | 10 +-
.../alpakka/xml/impl/StreamingXmlParser.scala | 27 +-
.../alpakka/xml/impl/StreamingXmlWriter.scala | 6 +-
.../akka/stream/alpakka/xml/impl/Subslice.scala | 6 +-
.../akka/stream/alpakka/xml/impl/Subtree.scala | 6 +-
.../stream/alpakka/xml/javadsl/XmlParsing.scala | 6 +-
.../stream/alpakka/xml/javadsl/XmlWriting.scala | 4 +-
.../main/scala/akka/stream/alpakka/xml/model.scala | 46 +-
.../stream/alpakka/xml/scaladsl/XmlParsing.scala | 2 +-
.../stream/alpakka/xml/scaladsl/XmlWriting.scala | 2 +-
.../test/scala/docs/scaladsl/XmlCoalesceSpec.scala | 14 +-
.../scala/docs/scaladsl/XmlProcessingSpec.scala | 92 ++--
.../test/scala/docs/scaladsl/XmlSubsliceSpec.scala | 22 +-
.../test/scala/docs/scaladsl/XmlSubtreeSpec.scala | 22 +-
.../test/scala/docs/scaladsl/XmlWritingSpec.scala | 47 +-
758 files changed, 10276 insertions(+), 13309 deletions(-)
create mode 100644 .gitattributes
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@pekko.apache.org
For additional commands, e-mail: commits-help@pekko.apache.org
[incubator-pekko-connectors] 02/03: Add .gitattributes to enforce unix line endings
Posted by md...@apache.org.
This is an automated email from the ASF dual-hosted git repository.
mdedetrich pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-pekko-connectors.git
commit 0e9e6bd5491d8a66404fa109cb6e9df494372de3
Author: Matthew de Detrich <ma...@aiven.io>
AuthorDate: Wed Nov 9 13:55:00 2022 +0100
Add .gitattributes to enforce unix line endings
---
.gitattributes | 5 +++++
1 file changed, 5 insertions(+)
diff --git a/.gitattributes b/.gitattributes
new file mode 100644
index 00000000..9dde9b97
--- /dev/null
+++ b/.gitattributes
@@ -0,0 +1,5 @@
+# Activate line ending normalization, setting eol will make the behavior match core.autocrlf = input
+* text=auto eol=lf
+# Force batch scripts to always use CRLF line endings
+*.{cmd,[cC][mM][dD]} text eol=crlf
+*.{bat,[bB][aA][tT]} text eol=crlf
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@pekko.apache.org
For additional commands, e-mail: commits-help@pekko.apache.org
[incubator-pekko-connectors] 01/03: Update scalafmt
Posted by md...@apache.org.
This is an automated email from the ASF dual-hosted git repository.
mdedetrich pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-pekko-connectors.git
commit 54c516e80288b65d14c387eb11820c19ac819538
Author: Matthew de Detrich <ma...@aiven.io>
AuthorDate: Wed Nov 9 13:54:03 2022 +0100
Update scalafmt
---
.scalafmt.conf | 90 +++++++++++++++++++++++++++++++++++++++++++++++++---------
1 file changed, 77 insertions(+), 13 deletions(-)
diff --git a/.scalafmt.conf b/.scalafmt.conf
index 0f38b162..efa297ee 100644
--- a/.scalafmt.conf
+++ b/.scalafmt.conf
@@ -1,13 +1,77 @@
-version = 2.1.0
-
-style = defaultWithAlign
-
-align.tokens = [off]
-align.openParenDefnSite = true
-align.openParenCallSite = true
-danglingParentheses = true
-docstrings = JavaDoc
-indentOperator = spray
-maxColumn = 120
-rewrite.rules = [RedundantParens, SortImports]
-unindentTopLevelOperators = true
+version = 3.6.1
+runner.dialect = scala213
+project.git = true
+style = defaultWithAlign
+docstrings.style = Asterisk
+docstrings.wrap = false
+indentOperator.preset = spray
+maxColumn = 120
+lineEndings = preserve
+rewrite.rules = [RedundantParens, SortImports, AvoidInfix]
+indentOperator.exemptScope = all
+align.preset = some
+align.tokens."+" = [
+ {
+ code = "~>"
+ owners = [
+ { regex = "Term.ApplyInfix" }
+ ]
+ }
+]
+literals.hexDigits = upper
+literals.hexPrefix = lower
+binPack.unsafeCallSite = always
+binPack.unsafeDefnSite = always
+binPack.indentCallSiteSingleArg = false
+binPack.indentCallSiteOnce = true
+newlines.avoidForSimpleOverflow = [slc]
+newlines.source = keep
+newlines.beforeMultiline = keep
+align.openParenDefnSite = false
+align.openParenCallSite = false
+align.allowOverflow = true
+optIn.breakChainOnFirstMethodDot = false
+optIn.configStyleArguments = false
+danglingParentheses.preset = false
+spaces.inImportCurlyBraces = true
+rewrite.neverInfix.excludeFilters = [
+ and
+ min
+ max
+ until
+ to
+ by
+ eq
+ ne
+ "should.*"
+ "contain.*"
+ "must.*"
+ in
+ ignore
+ be
+ taggedAs
+ thrownBy
+ synchronized
+ have
+ when
+ size
+ only
+ noneOf
+ oneElementOf
+ noElementsOf
+ atLeastOneElementOf
+ atMostOneElementOf
+ allElementsOf
+ inOrderElementsOf
+ theSameElementsAs
+ theSameElementsInOrderAs
+]
+rewriteTokens = {
+ "⇒": "=>"
+ "→": "->"
+ "←": "<-"
+}
+project.excludeFilters = [
+ "scripts/authors.scala"
+]
+project.layout = StandardConvention
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@pekko.apache.org
For additional commands, e-mail: commits-help@pekko.apache.org
[incubator-pekko-connectors] 03/03: format source with scalafmt, #12
Posted by md...@apache.org.
This is an automated email from the ASF dual-hosted git repository.
mdedetrich pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-pekko-connectors.git
commit 433daa5c5b0b2c3f85c6f1c7e62a4f4171856468
Author: Auto Format <nobody>
AuthorDate: Fri Jan 6 09:48:29 2023 +0100
format source with scalafmt, #12
---
.../alpakka/amqp/AmqpConnectionProvider.scala | 59 ++-
.../alpakka/amqp/AmqpConnectorSettings.scala | 58 +--
.../impl/AbstractAmqpAsyncFlowStageLogic.scala | 41 +-
.../amqp/impl/AbstractAmqpFlowStageLogic.scala | 19 +-
.../alpakka/amqp/impl/AmqpAsyncFlowStage.scala | 61 ++-
.../amqp/impl/AmqpAsyncUnorderedFlowStage.scala | 36 +-
.../alpakka/amqp/impl/AmqpConnectorLogic.scala | 13 +-
.../alpakka/amqp/impl/AmqpReplyToSinkStage.scala | 104 ++--
.../alpakka/amqp/impl/AmqpRpcFlowStage.scala | 326 ++++++-------
.../alpakka/amqp/impl/AmqpSimpleFlowStage.scala | 35 +-
.../stream/alpakka/amqp/impl/AmqpSourceStage.scala | 29 +-
.../stream/alpakka/amqp/javadsl/AmqpFlow.scala | 16 +-
.../alpakka/amqp/javadsl/AmqpFlowWithContext.scala | 8 +-
.../stream/alpakka/amqp/javadsl/AmqpRpcFlow.scala | 11 +-
.../stream/alpakka/amqp/javadsl/AmqpSink.scala | 3 +-
.../stream/alpakka/amqp/javadsl/AmqpSource.scala | 2 +-
.../scala/akka/stream/alpakka/amqp/model.scala | 12 +-
.../stream/alpakka/amqp/scaladsl/AmqpFlow.scala | 25 +-
.../amqp/scaladsl/AmqpFlowWithContext.scala | 14 +-
.../stream/alpakka/amqp/scaladsl/AmqpRpcFlow.scala | 10 +-
.../stream/alpakka/amqp/scaladsl/AmqpSink.scala | 2 +-
.../stream/alpakka/amqp/scaladsl/AmqpSource.scala | 2 +-
.../stream/alpakka/amqp/AmqpProxyConnection.scala | 2 +-
.../scaladsl/AmqpConnectionProvidersSpec.scala | 8 +-
.../alpakka/amqp/scaladsl/AmqpConnectorsSpec.scala | 74 +--
.../alpakka/amqp/scaladsl/AmqpFlowSpec.scala | 22 +-
...AmqpGraphStageLogicConnectionShutdownSpec.scala | 5 +-
.../stream/alpakka/amqp/scaladsl/AmqpMocking.scala | 2 +-
.../test/scala/docs/scaladsl/AmqpDocsSpec.scala | 90 ++--
.../alpakka/avroparquet/impl/AvroParquetFlow.scala | 16 +-
.../avroparquet/impl/AvroParquetSource.scala | 7 +-
.../avroparquet/javadsl/AvroParquetSink.scala | 4 +-
.../avroparquet/scaladsl/AvroParquetSink.scala | 2 +-
.../scala/docs/scaladsl/AbstractAvroParquet.scala | 11 +-
.../scala/docs/scaladsl/AvroParquetFlowSpec.scala | 14 +-
.../scala/docs/scaladsl/AvroParquetSinkSpec.scala | 16 +-
.../docs/scaladsl/AvroParquetSourceSpec.scala | 16 +-
.../eventbridge/EventBridgePublishSettings.scala | 2 +-
.../eventbridge/javadsl/EventBridgePublisher.scala | 18 +-
.../scaladsl/EventBridgePublisher.scala | 25 +-
.../aws/eventbridge/DefaultTestContext.scala | 2 +-
.../eventbridge/EventBridgePublishMockSpec.scala | 7 +-
.../aws/eventbridge/IntegrationTestContext.scala | 15 +-
.../docs/scaladsl/EventBridgePublisherSpec.scala | 4 +-
.../alpakka/awslambda/javadsl/AwsLambdaFlow.scala | 2 +-
.../alpakka/awslambda/scaladsl/AwsLambdaFlow.scala | 5 +-
.../scala/docs/scaladsl/AwsLambdaFlowSpec.scala | 16 +-
.../src/test/scala/docs/scaladsl/Examples.scala | 16 +-
.../impl/AzureQueueSinkFunctions.scala | 16 +-
.../storagequeue/impl/AzureQueueSourceStage.scala | 11 +-
.../storagequeue/javadsl/AzureQueueSink.scala | 19 +-
.../storagequeue/javadsl/AzureQueueSource.scala | 2 +-
.../storagequeue/scaladsl/AzureQueueSink.scala | 18 +-
.../storagequeue/scaladsl/AzureQueueSource.scala | 5 +-
.../alpakka/azure/storagequeue/settings.scala | 11 +-
.../test/scala/docs/scaladsl/AzureQueueSpec.scala | 26 +-
build.sbt | 241 ++++-----
.../cassandra/AkkaDiscoverySessionProvider.scala | 15 +-
.../cassandra/CassandraMetricsRegistry.scala | 2 +-
.../cassandra/CassandraSessionSettings.scala | 8 +-
.../alpakka/cassandra/CassandraWriteSettings.scala | 14 +-
.../alpakka/cassandra/CqlSessionProvider.scala | 10 +-
.../alpakka/cassandra/javadsl/CassandraFlow.scala | 34 +-
.../cassandra/javadsl/CassandraSession.scala | 47 +-
.../javadsl/CassandraSessionRegistry.scala | 4 +-
.../cassandra/javadsl/CassandraSource.scala | 2 +-
.../alpakka/cassandra/scaladsl/CassandraFlow.scala | 18 +-
.../cassandra/scaladsl/CassandraSession.scala | 26 +-
.../scaladsl/CassandraSessionRegistry.scala | 22 +-
.../cassandra/scaladsl/CassandraSource.scala | 2 +-
.../cassandra/scaladsl/CassandraLifecycle.scala | 14 +-
.../scaladsl/CassandraSessionPerformanceSpec.scala | 7 +-
.../cassandra/scaladsl/CassandraSpecBase.scala | 4 +-
.../scala/docs/javadsl/CassandraSessionSpec.scala | 14 +-
.../scala/docs/scaladsl/AkkaDiscoverySpec.scala | 2 +-
.../scala/docs/scaladsl/CassandraFlowSpec.scala | 36 +-
.../scala/docs/scaladsl/CassandraSourceSpec.scala | 8 +-
.../couchbase/CouchbaseResponseException.scala | 3 +-
.../couchbase/CouchbaseSessionRegistry.scala | 11 +-
.../couchbase/impl/CouchbaseClusterRegistry.scala | 7 +-
.../couchbase/impl/CouchbaseSessionImpl.scala | 80 ++-
.../impl/CouchbaseSessionJavaAdapter.scala | 19 +-
.../alpakka/couchbase/impl/RxUtilities.scala | 6 +-
.../alpakka/couchbase/javadsl/CouchbaseFlow.scala | 38 +-
.../couchbase/javadsl/CouchbaseSession.scala | 28 +-
.../alpakka/couchbase/javadsl/CouchbaseSink.scala | 26 +-
.../couchbase/javadsl/CouchbaseSource.scala | 12 +-
.../akka/stream/alpakka/couchbase/model.scala | 53 +-
.../alpakka/couchbase/scaladsl/CouchbaseFlow.scala | 132 +++--
.../couchbase/scaladsl/CouchbaseSession.scala | 22 +-
.../alpakka/couchbase/scaladsl/CouchbaseSink.scala | 16 +-
.../couchbase/scaladsl/CouchbaseSource.scala | 12 +-
.../couchbase/scaladsl/DiscoverySupport.scala | 18 +-
.../couchbase/testing/CouchbaseSupport.scala | 18 +-
.../scala/docs/scaladsl/CouchbaseFlowSpec.scala | 167 +++----
.../scaladsl/CouchbaseSessionExamplesSpec.scala | 6 +-
.../scala/docs/scaladsl/CouchbaseSourceSpec.scala | 10 +-
.../test/scala/docs/scaladsl/DiscoverySpec.scala | 10 +-
.../stream/alpakka/csv/scaladsl/CsvBench.scala | 6 +-
.../stream/alpakka/csv/impl/CsvFormatter.scala | 12 +-
.../akka/stream/alpakka/csv/impl/CsvParser.scala | 27 +-
.../stream/alpakka/csv/impl/CsvParsingStage.scala | 25 +-
.../alpakka/csv/impl/CsvToMapJavaStage.scala | 54 +--
.../stream/alpakka/csv/impl/CsvToMapStage.scala | 43 +-
.../alpakka/csv/scaladsl/CsvFormatting.scala | 10 +-
.../stream/alpakka/csv/scaladsl/CsvParsing.scala | 9 +-
.../alpakka/csv/scaladsl/CsvQuotingStyle.scala | 2 +-
.../stream/alpakka/csv/scaladsl/CsvToMap.scala | 67 ++-
.../akka/stream/alpakka/csv/CsvParserSpec.scala | 20 +-
.../scala/docs/scaladsl/CsvFormattingSpec.scala | 9 +-
.../test/scala/docs/scaladsl/CsvParsingSpec.scala | 37 +-
csv/src/test/scala/docs/scaladsl/CsvSpec.scala | 2 +-
.../test/scala/docs/scaladsl/CsvToMapSpec.scala | 107 ++--
.../alpakka/eip/scaladsl/PassThroughExamples.scala | 24 +-
.../akka/stream/alpakka/dynamodb/DynamoDbOp.scala | 6 +-
.../stream/alpakka/dynamodb/javadsl/DynamoDb.scala | 29 +-
.../alpakka/dynamodb/scaladsl/DynamoDb.scala | 34 +-
.../akka/stream/alpakka/dynamodb/ItemSpec.scala | 4 +-
.../akka/stream/alpakka/dynamodb/TableSpec.scala | 4 +-
.../akka/stream/alpakka/dynamodb/TestOps.scala | 48 +-
.../src/test/scala/docs/scaladsl/ExampleSpec.scala | 24 +-
.../src/test/scala/docs/scaladsl/RetrySpec.scala | 8 +-
.../ElasticsearchConnectionSettings.scala | 30 +-
.../ElasticsearchSourceSettings.scala | 49 +-
.../elasticsearch/ElasticsearchWriteSettings.scala | 38 +-
.../elasticsearch/OpensearchSourceSettings.scala | 49 +-
.../elasticsearch/OpensearchWriteSettings.scala | 30 +-
.../stream/alpakka/elasticsearch/ReadResult.scala | 4 +-
.../alpakka/elasticsearch/SourceSettingsBase.scala | 22 +-
.../alpakka/elasticsearch/WriteMessage.scala | 52 +-
.../alpakka/elasticsearch/WriteSettingsBase.scala | 13 +-
.../elasticsearch/impl/ElasticsearchApi.scala | 10 +-
.../impl/ElasticsearchSimpleFlowStage.scala | 35 +-
.../impl/ElasticsearchSourceStage.scala | 61 +--
.../elasticsearch/impl/NDJsonProtocol.scala | 2 +-
.../alpakka/elasticsearch/impl/RestBulkApi.scala | 16 +-
.../alpakka/elasticsearch/impl/RestBulkApiV5.scala | 21 +-
.../alpakka/elasticsearch/impl/RestBulkApiV7.scala | 18 +-
.../elasticsearch/javadsl/ElasticsearchFlow.scala | 35 +-
.../elasticsearch/javadsl/ElasticsearchSink.scala | 5 +-
.../javadsl/ElasticsearchSource.scala | 64 ++-
.../elasticsearch/scaladsl/ElasticsearchFlow.scala | 67 ++-
.../elasticsearch/scaladsl/ElasticsearchSink.scala | 7 +-
.../scaladsl/ElasticsearchSource.scala | 40 +-
.../elasticsearch/testkit/MessageFactory.scala | 26 +-
.../impl/ElasticsearchSimpleFlowStageTest.scala | 23 +-
.../impl/ElasticsearchSourcStageTest.scala | 8 +-
.../scaladsl/ElasticsearchConnectorBehaviour.scala | 228 ++++-----
.../scala/docs/scaladsl/ElasticsearchSpec.scala | 10 +-
.../docs/scaladsl/ElasticsearchSpecBase.scala | 10 +-
.../docs/scaladsl/ElasticsearchSpecUtils.scala | 23 +-
.../scala/docs/scaladsl/ElasticsearchV5Spec.scala | 179 +++----
.../scala/docs/scaladsl/ElasticsearchV7Spec.scala | 135 ++----
.../scaladsl/OpensearchConnectorBehaviour.scala | 228 ++++-----
.../scala/docs/scaladsl/OpensearchV1Spec.scala | 169 +++----
.../file/impl/archive/EnsureByteStreamSize.scala | 18 +-
.../file/impl/archive/TarArchiveEntry.scala | 27 +-
.../file/impl/archive/TarArchiveManager.scala | 2 +-
.../alpakka/file/impl/archive/TarReaderStage.scala | 29 +-
.../alpakka/file/impl/archive/ZipArchiveFlow.scala | 17 +-
.../file/impl/archive/ZipArchiveManager.scala | 2 +-
.../file/impl/archive/ZipReaderSource.scala | 16 +-
.../akka/stream/alpakka/file/javadsl/Archive.scala | 12 +-
.../alpakka/file/javadsl/LogRotatorSink.scala | 27 +-
.../scala/akka/stream/alpakka/file/model.scala | 63 ++-
.../stream/alpakka/file/scaladsl/Archive.scala | 14 +-
.../stream/alpakka/file/scaladsl/Directory.scala | 8 +-
.../file/scaladsl/DirectoryChangesSource.scala | 7 +-
.../alpakka/file/scaladsl/FileTailSource.scala | 19 +-
.../alpakka/file/scaladsl/LogRotatorSink.scala | 49 +-
.../file/impl/archive/TarArchiveEntrySpec.scala | 2 +-
.../file/impl/archive/ZipArchiveFlowTest.scala | 2 +-
.../src/test/scala/docs/scaladsl/ArchiveSpec.scala | 19 +-
.../test/scala/docs/scaladsl/DirectorySpec.scala | 10 +-
.../test/scala/docs/scaladsl/ExecutableUtils.scala | 8 +-
.../docs/scaladsl/FileTailSourceExtrasSpec.scala | 20 +-
.../scala/docs/scaladsl/FileTailSourceSpec.scala | 3 +-
.../scala/docs/scaladsl/LogRotatorSinkSpec.scala | 59 +--
.../test/scala/docs/scaladsl/TarArchiveSpec.scala | 43 +-
.../alpakka/ftp/impl/CommonFtpOperations.scala | 11 +-
.../alpakka/ftp/impl/FtpBrowserGraphStage.scala | 5 +-
.../impl/FtpDirectoryOperationsGraphStage.scala | 5 +-
.../stream/alpakka/ftp/impl/FtpGraphStage.scala | 2 +-
.../alpakka/ftp/impl/FtpGraphStageLogic.scala | 5 +-
.../stream/alpakka/ftp/impl/FtpIOGraphStage.scala | 41 +-
.../akka/stream/alpakka/ftp/impl/FtpLike.scala | 4 +-
.../stream/alpakka/ftp/impl/FtpOperations.scala | 13 +-
.../stream/alpakka/ftp/impl/FtpSourceFactory.scala | 58 +--
.../stream/alpakka/ftp/impl/FtpsOperations.scala | 10 +-
.../stream/alpakka/ftp/impl/SftpOperations.scala | 25 +-
.../akka/stream/alpakka/ftp/javadsl/FtpApi.scala | 164 +++----
.../main/scala/akka/stream/alpakka/ftp/model.scala | 91 ++--
.../akka/stream/alpakka/ftp/scaladsl/FtpApi.scala | 92 ++--
.../akka/stream/alpakka/ftp/BaseFtpSpec.scala | 16 +-
.../akka/stream/alpakka/ftp/BaseFtpsSpec.scala | 16 +-
.../akka/stream/alpakka/ftp/BaseSftpSpec.scala | 16 +-
.../scala/akka/stream/alpakka/ftp/BaseSpec.scala | 17 +-
.../stream/alpakka/ftp/CommonFtpStageSpec.scala | 56 +--
.../test/scala/docs/scaladsl/FtpExamplesSpec.scala | 16 +-
.../test/scala/docs/scaladsl/scalaExamples.scala | 44 +-
.../akka/stream/alpakka/geode/GeodeSettings.scala | 14 +-
.../stream/alpakka/geode/impl/GeodeCache.scala | 4 +-
.../alpakka/geode/impl/GeodeCapabilities.scala | 2 +-
.../geode/impl/pdx/DelegatingPdxSerializer.scala | 5 +-
.../stream/alpakka/geode/impl/pdx/PdxDecoder.scala | 12 +-
.../stream/alpakka/geode/impl/pdx/PdxEncoder.scala | 8 +-
.../geode/impl/pdx/ShapelessPdxSerializer.scala | 7 +-
.../geode/impl/stage/GeodeCQueryGraphLogic.scala | 14 +-
.../impl/stage/GeodeContinuousSourceStage.scala | 61 ++-
.../geode/impl/stage/GeodeFiniteSourceStage.scala | 31 +-
.../alpakka/geode/impl/stage/GeodeFlowStage.scala | 24 +-
.../geode/impl/stage/GeodeQueryGraphLogic.scala | 4 +-
.../geode/impl/stage/GeodeSourceStageLogic.scala | 4 +-
.../akka/stream/alpakka/geode/scaladsl/Geode.scala | 30 +-
.../alpakka/geode/impl/pdx/PDXDecoderSpec.scala | 2 +-
.../alpakka/geode/impl/pdx/PDXEncodeSpec.scala | 2 +-
.../alpakka/geode/impl/pdx/PdxWriterMock.scala | 8 +-
.../test/scala/docs/scaladsl/GeodeBaseSpec.scala | 8 +-
.../docs/scaladsl/GeodeContinuousSourceSpec.scala | 14 +-
.../docs/scaladsl/GeodeFiniteSourceSpec.scala | 4 +-
.../test/scala/docs/scaladsl/GeodeFlowSpec.scala | 14 +-
.../test/scala/docs/scaladsl/GeodeSinkSpec.scala | 8 +-
geode/src/test/scala/docs/scaladsl/Model.scala | 2 +-
.../scala/docs/scaladsl/PersonPdxSerializer.scala | 2 +-
.../bigquery/storage/BigQueryRecordMapImpl.scala | 4 +-
.../bigquery/storage/BigQueryStorageSettings.scala | 5 +-
.../bigquery/storage/ProtobufConverters.scala | 6 +-
.../bigquery/storage/impl/AkkaGrpcSettings.scala | 8 +-
.../bigquery/storage/impl/ArrowSource.scala | 21 +-
.../bigquery/storage/impl/AvroDecoder.scala | 2 +-
.../bigquery/storage/impl/AvroSource.scala | 18 +-
.../bigquery/storage/impl/SDKClientSource.scala | 2 +-
.../storage/javadsl/BigQueryArrowStorage.scala | 82 ++--
.../storage/javadsl/BigQueryAvroStorage.scala | 86 ++--
.../bigquery/storage/javadsl/BigQueryStorage.scala | 65 ++-
.../storage/scaladsl/BigQueryArrowStorage.scala | 88 ++--
.../storage/scaladsl/BigQueryAvroStorage.scala | 88 ++--
.../storage/scaladsl/BigQueryStorage.scala | 36 +-
.../scaladsl/BigQueryStorageAttributes.scala | 3 +-
.../scaladsl/GrpcBigQueryStorageReader.scala | 2 +-
.../storage/javadsl/AvroByteStringDecoder.scala | 6 +-
.../bigquery/storage/BigQueryStorageSpecBase.scala | 7 +-
.../bigquery/storage/mock/BigQueryMockData.scala | 10 +-
.../bigquery/storage/mock/BigQueryMockServer.scala | 23 +-
.../storage/scaladsl/ArrowByteStringDecoder.scala | 19 +-
.../storage/scaladsl/AvroByteStringDecoder.scala | 6 +-
.../scaladsl/BigQueryArrowStorageSpec.scala | 7 +-
.../storage/scaladsl/BigQueryStorageSpec.scala | 30 +-
.../test/scala/docs/scaladsl/ExampleReader.scala | 37 +-
.../googlecloud/bigquery/BigQueryAttributes.scala | 6 +-
.../googlecloud/bigquery/BigQueryException.scala | 9 +-
.../alpakka/googlecloud/bigquery/BigQueryExt.scala | 8 +-
.../googlecloud/bigquery/javadsl/BigQuery.scala | 118 +++--
.../javadsl/jackson/BigQueryMarshallers.scala | 11 +-
.../bigquery/model/DatasetJsonProtocol.scala | 14 +-
.../bigquery/model/ErrorProtoJsonProtocol.scala | 6 +-
.../bigquery/model/JobJsonProtocol.scala | 35 +-
.../bigquery/model/QueryJsonProtocol.scala | 117 +++--
.../googlecloud/bigquery/model/StringEnum.scala | 4 +-
.../bigquery/model/TableDataJsonProtocol.scala | 31 +-
.../bigquery/model/TableJsonProtocol.scala | 61 ++-
.../bigquery/scaladsl/BigQueryDatasets.scala | 22 +-
.../bigquery/scaladsl/BigQueryJobs.scala | 34 +-
.../bigquery/scaladsl/BigQueryQueries.scala | 41 +-
.../bigquery/scaladsl/BigQueryRest.scala | 4 +-
.../bigquery/scaladsl/BigQueryTableData.scala | 34 +-
.../bigquery/scaladsl/BigQueryTables.scala | 19 +-
.../bigquery/scaladsl/schema/BasicSchemas.scala | 2 +-
.../bigquery/scaladsl/schema/JavaTimeSchemas.scala | 4 +-
.../scaladsl/schema/PrimitiveSchemaWriter.scala | 2 +-
.../bigquery/scaladsl/schema/ProductSchemas.scala | 4 +-
.../bigquery/scaladsl/schema/SchemaWriter.scala | 2 +-
.../bigquery/scaladsl/schema/StandardSchemas.scala | 2 +-
.../scaladsl/spray/BigQueryBasicFormats.scala | 46 +-
.../scaladsl/spray/BigQueryCollectionFormats.scala | 8 +-
.../scaladsl/spray/BigQueryJavaTimeFormats.scala | 14 +-
.../scaladsl/spray/BigQueryJsonFormat.scala | 2 +-
.../scaladsl/spray/BigQueryJsonProtocol.scala | 2 +-
.../scaladsl/spray/BigQueryProductFormats.scala | 5 +-
.../scaladsl/spray/BigQueryRestBasicFormats.scala | 12 +-
.../scaladsl/spray/BigQueryRestJsonProtocol.scala | 2 +-
.../scaladsl/spray/BigQueryStandardFormats.scala | 2 +-
.../googlecloud/bigquery/HoverflySupport.scala | 7 +-
.../alpakka/googlecloud/bigquery/e2e/A.scala | 13 +-
.../e2e/scaladsl/BigQueryEndToEndSpec.scala | 13 +-
.../bigquery/e2e/scaladsl/EndToEndHelper.scala | 17 +-
.../bigquery/scaladsl/BigQueryQueriesSpec.scala | 43 +-
.../scaladsl/schema/BigQuerySchemasSpec.scala | 11 +-
.../scaladsl/spray/BigQueryJsonProtocolSpec.scala | 6 +-
.../src/test/scala/docs/scaladsl/BigQueryDoc.scala | 50 +-
.../googlecloud/pubsub/grpc/PubSubSettings.scala | 24 +-
.../pubsub/grpc/impl/AkkaGrpcSettings.scala | 8 +-
.../pubsub/grpc/impl/DeprecatedCredentials.scala | 4 +-
.../pubsub/grpc/javadsl/GooglePubSub.scala | 21 +-
.../pubsub/grpc/javadsl/GrpcPublisher.scala | 2 +-
.../pubsub/grpc/javadsl/GrpcSubscriber.scala | 6 +-
.../pubsub/grpc/scaladsl/GooglePubSub.scala | 28 +-
.../pubsub/grpc/scaladsl/GrpcPublisher.scala | 4 +-
.../pubsub/grpc/scaladsl/GrpcSubscriber.scala | 4 +-
.../src/test/scala/docs/scaladsl/ExampleApp.scala | 14 +-
.../test/scala/docs/scaladsl/IntegrationSpec.scala | 42 +-
.../googlecloud/pubsub/impl/PubSubApi.scala | 44 +-
.../googlecloud/pubsub/javadsl/GooglePubSub.scala | 28 +-
.../stream/alpakka/googlecloud/pubsub/model.scala | 116 +++--
.../googlecloud/pubsub/scaladsl/GooglePubSub.scala | 39 +-
.../googlecloud/pubsub/GooglePubSubSpec.scala | 27 +-
.../googlecloud/pubsub/impl/PubSubApiSpec.scala | 133 ++---
.../test/scala/docs/scaladsl/ExampleUsage.scala | 54 +--
.../test/scala/docs/scaladsl/IntegrationSpec.scala | 6 +-
.../alpakka/googlecloud/storage/Bucket.scala | 21 +-
.../googlecloud/storage/CustomerEncryption.scala | 2 +-
.../alpakka/googlecloud/storage/FailedUpload.scala | 3 +-
.../alpakka/googlecloud/storage/GCSExt.scala | 2 +-
.../alpakka/googlecloud/storage/GCStorageExt.scala | 2 +-
.../googlecloud/storage/GCStorageSettings.scala | 26 +-
.../googlecloud/storage/ObjectAccessControls.scala | 147 +++---
.../googlecloud/storage/StorageObject.scala | 24 +-
.../googlecloud/storage/StorageSettings.scala | 4 +-
.../storage/impl/BucketListResult.scala | 12 +-
.../alpakka/googlecloud/storage/impl/Formats.scala | 102 ++--
.../googlecloud/storage/impl/GCStorageStream.scala | 81 ++--
.../googlecloud/storage/impl/RewriteResponse.scala | 3 +-
.../storage/impl/UploadPartResponse.scala | 4 +-
.../googlecloud/storage/javadsl/GCStorage.scala | 77 ++-
.../googlecloud/storage/scaladsl/GCStorage.scala | 57 ++-
.../alpakka/googlecloud/storage/settings.scala | 2 +-
.../alpakka/googlecloud/storage/GCSExtSpec.scala | 4 +-
.../googlecloud/storage/GCSSettingsSpec.scala | 4 +-
.../googlecloud/storage/GCStorageExtSpec.scala | 4 +-
.../storage/GCStorageSettingsSpec.scala | 4 +-
.../storage/WithMaterializerGlobal.scala | 4 +-
.../impl/GCStorageStreamIntegrationSpec.scala | 49 +-
.../storage/scaladsl/GCStorageWiremockBase.scala | 212 +++-----
.../scala/docs/scaladsl/GCStorageSinkSpec.scala | 21 +-
.../scala/docs/scaladsl/GCStorageSourceSpec.scala | 127 ++---
.../stream/alpakka/google/GoogleAttributes.scala | 4 +-
.../akka/stream/alpakka/google/GoogleExt.scala | 8 +-
.../stream/alpakka/google/GoogleSettings.scala | 66 ++-
.../stream/alpakka/google/PaginatedRequest.scala | 5 +-
.../stream/alpakka/google/ResumableUpload.scala | 34 +-
.../stream/alpakka/google/auth/AccessToken.scala | 2 +-
.../google/auth/ComputeEngineCredentials.scala | 4 +-
.../stream/alpakka/google/auth/Credentials.scala | 18 +-
.../google/auth/GoogleComputeMetadata.scala | 6 +-
.../stream/alpakka/google/auth/GoogleOAuth2.scala | 19 +-
.../google/auth/GoogleOAuth2Credentials.scala | 11 +-
.../google/auth/GoogleOAuth2Exception.scala | 11 +-
.../stream/alpakka/google/auth/NoCredentials.scala | 4 +-
.../alpakka/google/auth/OAuth2Credentials.scala | 20 +-
.../google/auth/ServiceAccountCredentials.scala | 24 +-
.../google/auth/UserAccessCredentials.scala | 36 +-
.../alpakka/google/auth/UserAccessMetadata.scala | 8 +-
.../google/http/ForwardProxyHttpsContext.scala | 6 +-
.../google/http/ForwardProxyPoolSettings.scala | 26 +-
.../stream/alpakka/google/http/GoogleHttp.scala | 119 +++--
.../akka/stream/alpakka/google/implicits.scala | 7 +-
.../stream/alpakka/google/javadsl/Google.scala | 19 +-
.../stream/alpakka/google/scaladsl/Google.scala | 7 +-
.../google/scaladsl/`X-Upload-Content-Type`.scala | 11 +-
.../stream/alpakka/google/util/AnnotateLast.scala | 2 +-
.../stream/alpakka/google/util/EitherFlow.scala | 14 +-
.../akka/stream/alpakka/google/util/Retry.scala | 18 +-
.../alpakka/google/GoogleHttpException.scala | 2 +-
.../stream/alpakka/google/GoogleSettingsSpec.scala | 6 +-
.../stream/alpakka/google/HoverflySupport.scala | 7 +-
.../alpakka/google/PaginatedRequestSpec.scala | 22 +-
.../alpakka/google/ResumableUploadSpec.scala | 16 +-
.../alpakka/google/auth/GoogleOAuth2Spec.scala | 8 +-
.../google/auth/OAuth2CredentialsSpec.scala | 8 +-
.../alpakka/google/http/GoogleHttpSpec.scala | 62 +--
.../test/scala/docs/scaladsl/GoogleCommonDoc.scala | 10 +-
.../firebase/fcm/FcmNotificationModels.scala | 15 +-
.../alpakka/google/firebase/fcm/FcmSettings.scala | 79 ++-
.../google/firebase/fcm/impl/FcmFlows.scala | 12 +-
.../google/firebase/fcm/impl/FcmJsonSupport.scala | 21 +-
.../google/firebase/fcm/impl/FcmSender.scala | 9 +-
.../google/firebase/fcm/javadsl/GoogleFcm.scala | 6 +-
.../google/firebase/fcm/scaladsl/GoogleFcm.scala | 6 +-
.../google/firebase/fcm/v1/impl/FcmFlows.scala | 2 +-
.../firebase/fcm/v1/impl/FcmJsonSupport.scala | 61 ++-
.../google/firebase/fcm/v1/impl/FcmSender.scala | 9 +-
.../google/firebase/fcm/v1/javadsl/GoogleFcm.scala | 4 +-
.../firebase/fcm/v1/models/AndroidConfig.scala | 9 +-
.../google/firebase/fcm/v1/models/ApnsConfig.scala | 3 +-
.../firebase/fcm/v1/models/FcmNotification.scala | 7 +-
.../firebase/fcm/v1/models/WebPushConfig.scala | 6 +-
.../firebase/fcm/v1/scaladsl/GoogleFcm.scala | 6 +-
.../firebase/fcm/v1/FcmNotificationSpec.scala | 2 +-
.../firebase/fcm/v1/impl/FcmSenderSpec.scala | 48 +-
.../src/test/scala/docs/scaladsl/FcmExamples.scala | 32 +-
.../akka/stream/alpakka/hbase/HTableSettings.scala | 24 +-
.../alpakka/hbase/impl/HBaseCapabilities.scala | 11 +-
.../stream/alpakka/hbase/impl/HBaseFlowStage.scala | 22 +-
.../alpakka/hbase/impl/HBaseSourceStage.scala | 12 +-
.../stream/alpakka/hbase/javadsl/HTableStage.scala | 8 +-
.../alpakka/hbase/scaladsl/HTableStage.scala | 8 +-
.../test/scala/docs/scaladsl/HBaseStageSpec.scala | 40 +-
.../stream/alpakka/hdfs/impl/HdfsFlowStage.scala | 20 +-
.../impl/strategy/DefaultRotationStrategy.scala | 6 +-
.../hdfs/impl/strategy/DefaultSyncStrategy.scala | 3 +-
.../hdfs/impl/writer/CompressedDataWriter.scala | 10 +-
.../alpakka/hdfs/impl/writer/DataWriter.scala | 5 +-
.../alpakka/hdfs/impl/writer/HdfsWriter.scala | 2 +-
.../alpakka/hdfs/impl/writer/SequenceWriter.scala | 33 +-
.../stream/alpakka/hdfs/javadsl/HdfsFlow.scala | 43 +-
.../stream/alpakka/hdfs/javadsl/HdfsSource.scala | 21 +-
.../scala/akka/stream/alpakka/hdfs/model.scala | 12 +-
.../stream/alpakka/hdfs/scaladsl/HdfsFlow.scala | 60 +--
.../stream/alpakka/hdfs/scaladsl/HdfsSource.scala | 17 +-
.../akka/stream/alpakka/hdfs/util/TestUtils.scala | 33 +-
.../test/scala/docs/scaladsl/HdfsReaderSpec.scala | 40 +-
.../test/scala/docs/scaladsl/HdfsWriterSpec.scala | 116 ++---
.../huawei/pushkit/ForwardProxyHttpsContext.scala | 8 +-
.../huawei/pushkit/ForwardProxyPoolSettings.scala | 10 +-
.../alpakka/huawei/pushkit/HmsSettingExt.scala | 8 +-
.../alpakka/huawei/pushkit/HmsSettings.scala | 49 +-
.../alpakka/huawei/pushkit/impl/HmsTokenApi.scala | 17 +-
.../alpakka/huawei/pushkit/impl/PushKitFlows.scala | 12 +-
.../huawei/pushkit/impl/PushKitJsonSupport.scala | 33 +-
.../huawei/pushkit/impl/PushKitSender.scala | 19 +-
.../huawei/pushkit/javadsl/HmsPushKit.scala | 4 +-
.../huawei/pushkit/models/AndroidConfig.scala | 104 ++--
.../alpakka/huawei/pushkit/models/ApnsConfig.scala | 4 +-
.../pushkit/models/PushKitNotification.scala | 18 +-
.../alpakka/huawei/pushkit/models/WebConfig.scala | 30 +-
.../huawei/pushkit/scaladsl/HmsPushKit.scala | 6 +-
.../huawei/pushkit/ConditionBuilderSpec.scala | 2 +-
.../huawei/pushkit/impl/HmsTokenApiSpec.scala | 41 +-
.../huawei/pushkit/impl/PushKitSenderSpec.scala | 54 +--
.../test/scala/docs/scaladsl/PushKitExamples.scala | 22 +-
.../alpakka/influxdb/InfluxDbReadSettings.scala | 6 +-
.../alpakka/influxdb/InfluxDbWriteMessage.scala | 18 +-
.../influxdb/impl/AlpakkaResultMapperHelper.scala | 36 +-
.../alpakka/influxdb/impl/InfluxDbFlowStage.scala | 44 +-
.../influxdb/impl/InfluxDbSourceStage.scala | 38 +-
.../alpakka/influxdb/javadsl/InfluxDbFlow.scala | 20 +-
.../alpakka/influxdb/javadsl/InfluxDbSink.scala | 16 +-
.../alpakka/influxdb/javadsl/InfluxDbSource.scala | 8 +-
.../alpakka/influxdb/scaladsl/InfluxDbFlow.scala | 20 +-
.../alpakka/influxdb/scaladsl/InfluxDbSink.scala | 8 +-
.../alpakka/influxdb/scaladsl/InfluxDbSource.scala | 8 +-
.../src/test/scala/docs/scaladsl/FlowSpec.scala | 35 +-
.../scala/docs/scaladsl/InfluxDbSourceSpec.scala | 10 +-
.../test/scala/docs/scaladsl/InfluxDbSpec.scala | 32 +-
.../stream/alpakka/ironmq/IronMqSettings.scala | 41 +-
.../scala/akka/stream/alpakka/ironmq/domain.scala | 2 +-
.../akka/stream/alpakka/ironmq/impl/Codec.scala | 2 +-
.../stream/alpakka/ironmq/impl/IronMqClient.scala | 62 ++-
.../alpakka/ironmq/impl/IronMqPullStage.scala | 10 +-
.../alpakka/ironmq/impl/IronMqPushStage.scala | 16 +-
.../alpakka/ironmq/javadsl/IronMqConsumer.scala | 2 +-
.../alpakka/ironmq/javadsl/IronMqProducer.scala | 26 +-
.../alpakka/ironmq/scaladsl/IronMqProducer.scala | 15 +-
.../akka/stream/alpakka/ironmq/IronMqSpec.scala | 4 +-
.../alpakka/ironmq/impl/IronMqClientForTests.scala | 2 +-
.../alpakka/ironmq/impl/IronMqClientSpec.scala | 2 +-
.../alpakka/ironmq/impl/IronMqPullStageSpec.scala | 4 +-
.../alpakka/ironmq/impl/IronMqPushStageSpec.scala | 2 +-
.../ironmq/scaladsl/IronMqConsumerSpec.scala | 6 +-
.../ironmq/scaladsl/IronMqProducerSpec.scala | 14 +-
.../test/scala/docs/scaladsl/IronMqDocsSpec.scala | 10 +-
.../akka/stream/alpakka/jms/AcknowledgeMode.scala | 15 +-
.../alpakka/jms/ConnectionRetrySettings.scala | 14 +-
.../akka/stream/alpakka/jms/Credentials.scala | 24 +-
.../akka/stream/alpakka/jms/Destinations.scala | 2 +-
.../scala/akka/stream/alpakka/jms/Envelopes.scala | 4 +-
.../stream/alpakka/jms/JmsBrowseSettings.scala | 20 +-
.../stream/alpakka/jms/JmsConsumerSettings.scala | 21 +-
.../akka/stream/alpakka/jms/JmsExceptions.scala | 15 +-
.../akka/stream/alpakka/jms/JmsMessages.scala | 279 +++++------
.../stream/alpakka/jms/JmsProducerSettings.scala | 23 +-
.../stream/alpakka/jms/SendRetrySettings.scala | 17 +-
.../alpakka/jms/impl/InternalConnectionState.scala | 6 +-
.../alpakka/jms/impl/JmsAckSourceStage.scala | 12 +-
.../stream/alpakka/jms/impl/JmsBrowseStage.scala | 6 +-
.../stream/alpakka/jms/impl/JmsConnector.scala | 56 +--
.../stream/alpakka/jms/impl/JmsConsumerStage.scala | 5 +-
.../alpakka/jms/impl/JmsMessageProducer.scala | 48 +-
.../stream/alpakka/jms/impl/JmsMessageReader.scala | 12 +-
.../stream/alpakka/jms/impl/JmsProducerStage.scala | 37 +-
.../stream/alpakka/jms/impl/JmsTxSourceStage.scala | 16 +-
.../akka/stream/alpakka/jms/impl/Sessions.scala | 27 +-
.../stream/alpakka/jms/impl/SourceStageLogic.scala | 23 +-
.../stream/alpakka/jms/javadsl/JmsConsumer.scala | 6 +-
.../stream/alpakka/jms/javadsl/JmsProducer.scala | 22 +-
.../alpakka/jms/scaladsl/JmsConnectorState.scala | 12 +-
.../stream/alpakka/jms/scaladsl/JmsConsumer.scala | 2 +-
.../stream/alpakka/jms/scaladsl/JmsProducer.scala | 22 +-
.../akka/stream/alpakka/jms/scaladsl/package.scala | 18 +-
.../alpakka/jms/JmsConnectionStatusSpec.scala | 36 +-
.../stream/alpakka/jms/JmsProducerRetrySpec.scala | 40 +-
.../scala/akka/stream/alpakka/jms/JmsSpec.scala | 22 +-
.../alpakka/jms/impl/JmsMessageProducerSpec.scala | 20 +-
.../alpakka/jms/impl/SoftReferenceCacheSpec.scala | 4 +-
.../jms/scaladsl/CachedConnectionFactory.scala | 2 +-
.../jms/scaladsl/JmsAckConnectorsSpec.scala | 101 ++--
.../scaladsl/JmsBufferedAckConnectorsSpec.scala | 96 ++--
.../scala/docs/scaladsl/JmsConnectorsSpec.scala | 359 ++++++--------
.../docs/scaladsl/JmsIbmmqConnectorsSpec.scala | 60 ++-
.../test/scala/docs/scaladsl/JmsSettingsSpec.scala | 22 +-
.../scala/docs/scaladsl/JmsTxConnectorsSpec.scala | 147 +++---
.../alpakka/json/impl/JsonStreamReader.scala | 13 +-
.../test/scala/docs/scaladsl/JsonReaderTest.scala | 2 +-
.../stream/alpakka/kinesis/CommittableRecord.scala | 11 +-
.../stream/alpakka/kinesis/KinesisErrors.scala | 2 +-
.../alpakka/kinesis/KinesisFlowSettings.scala | 18 +-
.../alpakka/kinesis/KinesisSchedulerSettings.scala | 6 +-
.../stream/alpakka/kinesis/ShardSettings.scala | 21 +-
.../kinesis/impl/KinesisSchedulerSourceStage.scala | 16 +-
.../alpakka/kinesis/impl/KinesisSourceStage.scala | 20 +-
.../alpakka/kinesis/impl/ShardProcessor.scala | 21 +-
.../alpakka/kinesis/javadsl/KinesisFlow.scala | 18 +-
.../kinesis/javadsl/KinesisSchedulerSource.scala | 19 +-
.../alpakka/kinesis/javadsl/KinesisSink.scala | 6 +-
.../alpakka/kinesis/javadsl/KinesisSource.scala | 4 +-
.../alpakka/kinesis/scaladsl/KinesisFlow.scala | 66 +--
.../kinesis/scaladsl/KinesisSchedulerSource.scala | 18 +-
.../alpakka/kinesis/scaladsl/KinesisSink.scala | 15 +-
.../alpakka/kinesis/scaladsl/KinesisSource.scala | 8 +-
.../KinesisFirehoseFlowSettings.scala | 18 +-
.../javadsl/KinesisFirehoseFlow.scala | 10 +-
.../javadsl/KinesisFirehoseSink.scala | 8 +-
.../scaladsl/KinesisFirehoseFlow.scala | 28 +-
.../scaladsl/KinesisFirehoseSink.scala | 3 +-
.../alpakka/kinesis/DefaultTestContext.scala | 9 +-
.../stream/alpakka/kinesis/KinesisFlowSpec.scala | 22 +-
.../akka/stream/alpakka/kinesis/KinesisMock.scala | 2 +-
.../kinesis/KinesisSchedulerSourceSpec.scala | 76 ++-
.../stream/alpakka/kinesis/KinesisSourceSpec.scala | 15 +-
.../scala/akka/stream/alpakka/kinesis/Valve.scala | 13 +-
.../kinesisfirehose/KinesisFirehoseFlowSpec.scala | 2 +-
.../kinesisfirehose/KinesisFirehoseMock.scala | 2 +-
.../src/test/scala/docs/scaladsl/KclSnippets.scala | 34 +-
.../docs/scaladsl/KinesisFirehoseSnippets.scala | 22 +-
.../test/scala/docs/scaladsl/KinesisSnippets.scala | 39 +-
.../akka/stream/alpakka/kudu/KuduClientExt.scala | 2 +-
.../stream/alpakka/kudu/KuduTableSettings.scala | 33 +-
.../alpakka/kudu/impl/KuduCapabilities.scala | 6 +-
.../stream/alpakka/kudu/impl/KuduFlowStage.scala | 16 +-
.../stream/alpakka/kudu/javadsl/KuduTable.scala | 4 +-
.../stream/alpakka/kudu/scaladsl/KuduTable.scala | 8 +-
.../test/scala/docs/scaladsl/KuduTableSpec.scala | 24 +-
.../stream/alpakka/mongodb/DocumentReplace.scala | 1 -
.../stream/alpakka/mongodb/DocumentUpdate.scala | 1 -
.../stream/alpakka/mongodb/javadsl/MongoFlow.scala | 29 +-
.../stream/alpakka/mongodb/javadsl/MongoSink.scala | 17 +-
.../alpakka/mongodb/scaladsl/MongoFlow.scala | 56 +--
.../alpakka/mongodb/scaladsl/MongoSink.scala | 22 +-
.../test/scala/docs/scaladsl/MongoSinkSpec.scala | 43 +-
.../test/scala/docs/scaladsl/MongoSourceSpec.scala | 6 +-
.../scala/akka/stream/alpakka/mqtt/MqttPerf.scala | 17 +-
.../stream/alpakka/mqtt/streaming/MqttPerf.scala | 18 +-
.../mqtt/streaming/MqttSessionSettings.scala | 64 ++-
.../mqtt/streaming/impl/BehaviorRunner.scala | 4 +-
.../alpakka/mqtt/streaming/impl/ClientState.scala | 228 ++++-----
.../mqtt/streaming/impl/MqttFrameStage.scala | 19 +-
.../mqtt/streaming/impl/QueueOfferState.scala | 8 +-
.../alpakka/mqtt/streaming/impl/RequestState.scala | 138 +++---
.../alpakka/mqtt/streaming/impl/ServerState.scala | 360 ++++++--------
.../alpakka/mqtt/streaming/javadsl/Mqtt.scala | 11 +-
.../akka/stream/alpakka/mqtt/streaming/model.scala | 175 ++++---
.../alpakka/mqtt/streaming/scaladsl/Mqtt.scala | 20 +-
.../mqtt/streaming/scaladsl/MqttSession.scala | 178 ++++---
.../mqtt/streaming/impl/MqttFrameStageSpec.scala | 2 +-
.../mqtt/streaming/impl/QueueOfferStateSpec.scala | 13 +-
.../mqtt/streaming/impl/RequestStateSpec.scala | 12 +-
.../scala/docs/scaladsl/MqttActorSystemsSpec.scala | 2 +-
.../test/scala/docs/scaladsl/MqttCodecSpec.scala | 43 +-
.../test/scala/docs/scaladsl/MqttFlowSpec.scala | 58 ++-
.../test/scala/docs/scaladsl/MqttSessionSpec.scala | 150 +++---
.../stream/alpakka/mqtt/impl/MqttFlowStage.scala | 88 ++--
.../alpakka/mqtt/impl/MqttFlowStageWithAck.scala | 58 +--
.../stream/alpakka/mqtt/javadsl/MqttFlow.scala | 12 +-
.../stream/alpakka/mqtt/javadsl/MqttSink.scala | 4 +-
.../stream/alpakka/mqtt/javadsl/MqttSource.scala | 8 +-
.../scala/akka/stream/alpakka/mqtt/model.scala | 24 +-
.../stream/alpakka/mqtt/scaladsl/MqttFlow.scala | 40 +-
.../stream/alpakka/mqtt/scaladsl/MqttSink.scala | 2 +-
.../stream/alpakka/mqtt/scaladsl/MqttSource.scala | 16 +-
.../scala/akka/stream/alpakka/mqtt/settings.scala | 84 ++--
.../test/scala/docs/scaladsl/MqttFlowSpec.scala | 28 +-
.../test/scala/docs/scaladsl/MqttSinkSpec.scala | 22 +-
.../test/scala/docs/scaladsl/MqttSourceSpec.scala | 118 ++---
.../test/scala/docs/scaladsl/MqttSpecBase.scala | 5 +-
.../alpakka/orientdb/OrientDbSourceSettings.scala | 15 +-
.../alpakka/orientdb/OrientDbWriteSettings.scala | 17 +-
.../alpakka/orientdb/impl/OrientDbFlowStage.scala | 9 +-
.../orientdb/impl/OrientDbSourceStage.scala | 19 +-
.../alpakka/orientdb/javadsl/OrientDbFlow.scala | 20 +-
.../alpakka/orientdb/javadsl/OrientDbSink.scala | 14 +-
.../alpakka/orientdb/javadsl/OrientDbSource.scala | 22 +-
.../alpakka/orientdb/scaladsl/OrientDbFlow.scala | 36 +-
.../alpakka/orientdb/scaladsl/OrientDbSink.scala | 10 +-
.../alpakka/orientdb/scaladsl/OrientDbSource.scala | 18 +-
.../test/scala/docs/scaladsl/OrientDbSpec.scala | 87 ++--
.../akka/stream/alpakka/pravega/PravegaEvent.scala | 2 +-
.../pravega/PravegaReaderGroupManager.scala | 12 +-
.../stream/alpakka/pravega/PravegaSettings.scala | 327 ++++++-------
.../alpakka/pravega/impl/PravegaCapabilities.scala | 2 +-
.../stream/alpakka/pravega/impl/PravegaFlow.scala | 27 +-
.../alpakka/pravega/impl/PravegaSource.scala | 30 +-
.../pravega/impl/PravegaTableReadFlow.scala | 20 +-
.../alpakka/pravega/impl/PravegaTableSource.scala | 26 +-
.../pravega/impl/PravegaTableWriteFlow.scala | 31 +-
.../alpakka/pravega/impl/PravegaWriter.scala | 3 +-
.../stream/alpakka/pravega/scaladsl/Pravega.scala | 8 +-
.../alpakka/pravega/scaladsl/PravegaTable.scala | 31 +-
.../stream/alpakka/pravega/PravegaBaseSpec.scala | 2 +-
.../stream/alpakka/pravega/PravegaGraphSpec.scala | 13 +-
.../alpakka/pravega/PravegaKVTableSpec.scala | 2 +-
.../pravega/PravegaStreamAndTableSpec.scala | 16 +-
.../scala/docs/scaladsl/PravegaReadWriteDocs.scala | 4 +-
.../scala/docs/scaladsl/PravegaSettingsSpec.scala | 23 +-
project/AutomaticModuleName.scala | 5 +-
project/Common.scala | 169 ++++---
project/Dependencies.scala | 534 +++++++++-----------
project/TestChanged.scala | 15 +-
.../reference/impl/ReferenceFlowStage.scala | 19 +-
.../reference/impl/ReferenceSourceStage.scala | 21 +-
.../impl/ReferenceWithResourceFlowStage.scala | 19 +-
.../alpakka/reference/javadsl/Reference.scala | 13 +-
.../akka/stream/alpakka/reference/model.scala | 14 +-
.../alpakka/reference/scaladsl/Reference.scala | 8 +-
.../akka/stream/alpakka/reference/settings.scala | 14 +-
.../alpakka/reference/testkit/MessageFactory.scala | 8 +-
.../test/scala/docs/scaladsl/ReferenceSpec.scala | 17 +-
.../scala/akka/stream/alpakka/s3/S3Exception.scala | 13 +-
.../main/scala/akka/stream/alpakka/s3/S3Ext.scala | 2 +-
.../scala/akka/stream/alpakka/s3/S3Headers.scala | 19 +-
.../main/scala/akka/stream/alpakka/s3/Utils.scala | 3 +-
.../alpakka/s3/headers/ServerSideEncryption.scala | 34 +-
.../scala/akka/stream/alpakka/s3/impl/Chunk.scala | 2 +-
.../akka/stream/alpakka/s3/impl/DiskBuffer.scala | 2 +-
.../akka/stream/alpakka/s3/impl/HttpRequests.scala | 122 ++---
.../akka/stream/alpakka/s3/impl/Marshalling.scala | 81 ++--
.../akka/stream/alpakka/s3/impl/MemoryBuffer.scala | 4 +-
.../stream/alpakka/s3/impl/MemoryWithContext.scala | 4 +-
.../akka/stream/alpakka/s3/impl/S3Stream.scala | 538 +++++++++------------
.../stream/alpakka/s3/impl/SplitAfterSize.scala | 2 +-
.../s3/impl/SplitAfterSizeWithContext.scala | 10 +-
.../alpakka/s3/impl/auth/CanonicalRequest.scala | 19 +-
.../akka/stream/alpakka/s3/impl/auth/Signer.scala | 26 +-
.../stream/alpakka/s3/impl/auth/SigningKey.scala | 10 +-
.../scala/akka/stream/alpakka/s3/javadsl/S3.scala | 434 ++++++++---------
.../main/scala/akka/stream/alpakka/s3/model.scala | 270 +++++------
.../scala/akka/stream/alpakka/s3/scaladsl/S3.scala | 219 ++++-----
.../scala/akka/stream/alpakka/s3/settings.scala | 100 ++--
.../akka/stream/alpakka/s3/MinioContainer.scala | 4 +-
.../scala/akka/stream/alpakka/s3/MinioS3Test.scala | 5 +-
.../akka/stream/alpakka/s3/S3SettingsSpec.scala | 37 +-
.../stream/alpakka/s3/impl/DiskBufferSpec.scala | 18 +-
.../stream/alpakka/s3/impl/HttpRequestsSpec.scala | 57 +--
.../stream/alpakka/s3/impl/MarshallingSpec.scala | 93 ++--
.../stream/alpakka/s3/impl/MemoryBufferSpec.scala | 9 +-
.../stream/alpakka/s3/impl/S3HeadersSpec.scala | 15 +-
.../akka/stream/alpakka/s3/impl/S3StreamSpec.scala | 56 +--
.../alpakka/s3/impl/SplitAfterSizeSpec.scala | 40 +-
.../s3/impl/auth/CanonicalRequestSpec.scala | 58 +--
.../stream/alpakka/s3/impl/auth/SignerSpec.scala | 34 +-
.../alpakka/s3/impl/auth/SigningKeySpec.scala | 10 +-
.../impl/auth/SplitAfterSizeWithContextSpec.scala | 19 +-
.../alpakka/s3/impl/auth/StreamUtilsSpec.scala | 10 +-
.../s3/scaladsl/S3ClientIntegrationSpec.scala | 2 +-
.../stream/alpakka/s3/scaladsl/S3ExtSpec.scala | 4 +-
.../alpakka/s3/scaladsl/S3IntegrationSpec.scala | 246 ++++------
.../s3/scaladsl/S3SlowMinioIntegrationSpec.scala | 9 +-
.../alpakka/s3/scaladsl/S3WireMockBase.scala | 222 +++------
s3/src/test/scala/docs/scaladsl/S3SinkSpec.scala | 120 ++---
s3/src/test/scala/docs/scaladsl/S3SourceSpec.scala | 58 ++-
.../recordio/impl/RecordIOFramingStage.scala | 10 +-
.../scala/docs/scaladsl/RecordIOFramingSpec.scala | 71 ++-
.../akka/stream/alpakka/slick/javadsl/Slick.scala | 67 ++-
.../stream/alpakka/slick/javadsl/package.scala | 3 +-
.../akka/stream/alpakka/slick/scaladsl/Slick.scala | 22 +-
.../stream/alpakka/slick/scaladsl/package.scala | 2 +-
.../src/test/scala/docs/scaladsl/DocSnippets.scala | 31 +-
slick/src/test/scala/docs/scaladsl/SlickSpec.scala | 56 +--
.../stream/alpakka/sns/javadsl/SnsPublisher.scala | 26 +-
.../stream/alpakka/sns/scaladsl/SnsPublisher.scala | 25 +-
.../stream/alpakka/sns/DefaultTestContext.scala | 2 +-
.../alpakka/sns/IntegrationTestContext.scala | 14 +-
.../stream/alpakka/sns/SnsPublishMockingSpec.scala | 6 +-
.../scala/docs/scaladsl/SnsPublisherSpec.scala | 26 +-
.../akka/stream/alpakka/solr/SolrMessages.scala | 42 +-
.../stream/alpakka/solr/SolrUpdateSettings.scala | 9 +-
.../stream/alpakka/solr/impl/SolrFlowStage.scala | 44 +-
.../stream/alpakka/solr/impl/SolrSourceStage.scala | 4 +-
.../stream/alpakka/solr/javadsl/SolrFlow.scala | 61 +--
.../stream/alpakka/solr/javadsl/SolrSink.scala | 21 +-
.../stream/alpakka/solr/scaladsl/SolrFlow.scala | 65 +--
.../stream/alpakka/solr/scaladsl/SolrSink.scala | 16 +-
solr/src/test/scala/docs/scaladsl/SolrSpec.scala | 165 +++----
.../stream/alpakka/sqs/SqsAckBatchSettings.scala | 3 +-
.../stream/alpakka/sqs/SqsAckGroupedSettings.scala | 33 +-
.../akka/stream/alpakka/sqs/SqsAckSettings.scala | 12 +-
.../scala/akka/stream/alpakka/sqs/SqsModel.scala | 39 +-
.../alpakka/sqs/SqsPublishBatchSettings.scala | 3 +-
.../alpakka/sqs/SqsPublishGroupedSettings.scala | 21 +-
.../stream/alpakka/sqs/SqsSourceSettings.scala | 22 +-
.../alpakka/sqs/impl/BalancingMapAsync.scala | 41 +-
.../stream/alpakka/sqs/javadsl/SqsAckFlow.scala | 8 +-
.../stream/alpakka/sqs/javadsl/SqsAckSink.scala | 10 +-
.../alpakka/sqs/javadsl/SqsPublishFlow.scala | 14 +-
.../alpakka/sqs/javadsl/SqsPublishSink.scala | 27 +-
.../stream/alpakka/sqs/scaladsl/SqsAckFlow.scala | 41 +-
.../stream/alpakka/sqs/scaladsl/SqsAckSink.scala | 10 +-
.../alpakka/sqs/scaladsl/SqsPublishFlow.scala | 39 +-
.../alpakka/sqs/scaladsl/SqsPublishSink.scala | 36 +-
.../stream/alpakka/sqs/scaladsl/SqsSource.scala | 11 +-
.../alpakka/sqs/testkit/MessageFactory.scala | 17 +-
.../alpakka/sqs/scaladsl/DefaultTestContext.scala | 20 +-
.../sqs/scaladsl/MessageAttributeNameSpec.scala | 3 +-
.../alpakka/sqs/scaladsl/SqsPublishSinkSpec.scala | 62 +--
.../alpakka/sqs/scaladsl/SqsSourceMockSpec.scala | 27 +-
sqs/src/test/scala/docs/scaladsl/SqsAckSpec.scala | 69 ++-
.../test/scala/docs/scaladsl/SqsPublishSpec.scala | 57 ++-
.../test/scala/docs/scaladsl/SqsSourceSpec.scala | 30 +-
.../stream/alpakka/sse/javadsl/EventSource.scala | 28 +-
.../stream/alpakka/sse/scaladsl/EventSource.scala | 23 +-
.../test/scala/docs/scaladsl/EventSourceSpec.scala | 28 +-
.../stream/alpakka/testkit/CapturingAppender.scala | 7 +-
.../akka/stream/alpakka/testkit/LogbackUtil.scala | 9 +-
.../testkit/javadsl/LogCapturingJunit4.scala | 9 +-
.../alpakka/testkit/scaladsl/LogCapturing.scala | 6 +-
.../stream/alpakka/testkit/scaladsl/Repeated.scala | 2 +-
.../alpakka/text/impl/CharsetDecodingFlow.scala | 4 +-
.../stream/alpakka/text/impl/CharsetLogic.scala | 28 +-
.../alpakka/text/impl/CharsetTranscodingFlow.scala | 4 +-
.../stream/alpakka/text/javadsl/TextFlow.scala | 2 +-
.../stream/alpakka/text/scaladsl/TextFlow.scala | 2 +-
.../text/scaladsl/CharsetCodingFlowsSpec.scala | 10 +-
.../docs/scaladsl/CharsetCodingFlowsDoc.scala | 4 +-
.../akka/stream/alpakka/udp/impl/UdpBind.scala | 27 +-
.../akka/stream/alpakka/udp/impl/UdpSend.scala | 20 +-
.../akka/stream/alpakka/udp/javadsl/Udp.scala | 20 +-
.../akka/stream/alpakka/udp/scaladsl/Udp.scala | 24 +-
udp/src/test/scala/docs/scaladsl/UdpSpec.scala | 4 +-
.../impl/UnixDomainSocketImpl.scala | 102 ++--
.../javadsl/UnixDomainSocket.scala | 22 +-
.../scaladsl/UnixDomainSocket.scala | 21 +-
.../scala/docs/scaladsl/UnixDomainSocketSpec.scala | 27 +-
.../akka/stream/alpakka/xml/impl/Coalesce.scala | 10 +-
.../alpakka/xml/impl/StreamingXmlParser.scala | 27 +-
.../alpakka/xml/impl/StreamingXmlWriter.scala | 6 +-
.../akka/stream/alpakka/xml/impl/Subslice.scala | 6 +-
.../akka/stream/alpakka/xml/impl/Subtree.scala | 6 +-
.../stream/alpakka/xml/javadsl/XmlParsing.scala | 6 +-
.../stream/alpakka/xml/javadsl/XmlWriting.scala | 4 +-
.../main/scala/akka/stream/alpakka/xml/model.scala | 46 +-
.../stream/alpakka/xml/scaladsl/XmlParsing.scala | 2 +-
.../stream/alpakka/xml/scaladsl/XmlWriting.scala | 2 +-
.../test/scala/docs/scaladsl/XmlCoalesceSpec.scala | 14 +-
.../scala/docs/scaladsl/XmlProcessingSpec.scala | 92 ++--
.../test/scala/docs/scaladsl/XmlSubsliceSpec.scala | 22 +-
.../test/scala/docs/scaladsl/XmlSubtreeSpec.scala | 22 +-
.../test/scala/docs/scaladsl/XmlWritingSpec.scala | 47 +-
756 files changed, 10194 insertions(+), 13296 deletions(-)
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectionProvider.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectionProvider.scala
index c2758232..33fa6cd1 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectionProvider.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectionProvider.scala
@@ -8,8 +8,8 @@ import java.util.ConcurrentModificationException
import java.util.concurrent.atomic.AtomicReference
import akka.annotation.DoNotInherit
-import com.rabbitmq.client.{Address, Connection, ConnectionFactory, ExceptionHandler}
-import javax.net.ssl.{SSLContext, TrustManager}
+import com.rabbitmq.client.{ Address, Connection, ConnectionFactory, ExceptionHandler }
+import javax.net.ssl.{ SSLContext, TrustManager }
import scala.annotation.tailrec
import scala.collection.immutable
@@ -72,8 +72,7 @@ final class AmqpDetailsConnectionProvider private (
val automaticRecoveryEnabled: Boolean = false,
val topologyRecoveryEnabled: Boolean = false,
val exceptionHandler: Option[ExceptionHandler] = None,
- val connectionName: Option[String] = None
-) extends AmqpConnectionProvider {
+ val connectionName: Option[String] = None) extends AmqpConnectionProvider {
def withHostAndPort(host: String, port: Int): AmqpDetailsConnectionProvider =
copy(hostAndPortList = immutable.Seq(host -> port))
@@ -152,18 +151,18 @@ final class AmqpDetailsConnectionProvider private (
}
private def copy(hostAndPortList: immutable.Seq[(String, Int)] = hostAndPortList,
- credentials: Option[AmqpCredentials] = credentials,
- virtualHost: Option[String] = virtualHost,
- sslConfiguration: Option[AmqpSSLConfiguration] = sslConfiguration,
- requestedHeartbeat: Option[Int] = requestedHeartbeat,
- connectionTimeout: Option[Int] = connectionTimeout,
- handshakeTimeout: Option[Int] = handshakeTimeout,
- shutdownTimeout: Option[Int] = shutdownTimeout,
- networkRecoveryInterval: Option[Int] = networkRecoveryInterval,
- automaticRecoveryEnabled: Boolean = automaticRecoveryEnabled,
- topologyRecoveryEnabled: Boolean = topologyRecoveryEnabled,
- exceptionHandler: Option[ExceptionHandler] = exceptionHandler,
- connectionName: Option[String] = connectionName): AmqpDetailsConnectionProvider =
+ credentials: Option[AmqpCredentials] = credentials,
+ virtualHost: Option[String] = virtualHost,
+ sslConfiguration: Option[AmqpSSLConfiguration] = sslConfiguration,
+ requestedHeartbeat: Option[Int] = requestedHeartbeat,
+ connectionTimeout: Option[Int] = connectionTimeout,
+ handshakeTimeout: Option[Int] = handshakeTimeout,
+ shutdownTimeout: Option[Int] = shutdownTimeout,
+ networkRecoveryInterval: Option[Int] = networkRecoveryInterval,
+ automaticRecoveryEnabled: Boolean = automaticRecoveryEnabled,
+ topologyRecoveryEnabled: Boolean = topologyRecoveryEnabled,
+ exceptionHandler: Option[ExceptionHandler] = exceptionHandler,
+ connectionName: Option[String] = connectionName): AmqpDetailsConnectionProvider =
new AmqpDetailsConnectionProvider(
hostAndPortList,
credentials,
@@ -177,8 +176,7 @@ final class AmqpDetailsConnectionProvider private (
automaticRecoveryEnabled,
topologyRecoveryEnabled,
exceptionHandler,
- connectionName
- )
+ connectionName)
override def toString: String =
"AmqpDetailsConnectionProvider(" +
@@ -233,8 +231,8 @@ object AmqpCredentials {
}
final class AmqpSSLConfiguration private (val protocol: Option[String] = None,
- val trustManager: Option[TrustManager] = None,
- val context: Option[SSLContext] = None) {
+ val trustManager: Option[TrustManager] = None,
+ val context: Option[SSLContext] = None) {
if (protocol.isDefined && context.isDefined) {
throw new IllegalArgumentException("Protocol and context can't be defined in the same AmqpSSLConfiguration.")
}
@@ -249,8 +247,8 @@ final class AmqpSSLConfiguration private (val protocol: Option[String] = None,
copy(context = context)
private def copy(protocol: Option[String] = protocol,
- trustManager: Option[TrustManager] = trustManager,
- context: Option[SSLContext] = context): AmqpSSLConfiguration =
+ trustManager: Option[TrustManager] = trustManager,
+ context: Option[SSLContext] = context): AmqpSSLConfiguration =
new AmqpSSLConfiguration(protocol, trustManager, context)
override def toString: String =
@@ -302,8 +300,8 @@ object AmqpSSLConfiguration {
* If empty, it defaults to the host and port in the underlying factory.
*/
final class AmqpConnectionFactoryConnectionProvider private (val factory: ConnectionFactory,
- private val hostAndPorts: immutable.Seq[(String, Int)] =
- Nil)
+ private val hostAndPorts: immutable.Seq[(String, Int)] =
+ Nil)
extends AmqpConnectionProvider {
/**
@@ -326,8 +324,7 @@ final class AmqpConnectionFactoryConnectionProvider private (val factory: Connec
* Java API
*/
def withHostsAndPorts(
- hostAndPorts: java.util.List[akka.japi.Pair[String, Int]]
- ): AmqpConnectionFactoryConnectionProvider =
+ hostAndPorts: java.util.List[akka.japi.Pair[String, Int]]): AmqpConnectionFactoryConnectionProvider =
copy(hostAndPorts = hostAndPorts.asScala.map(_.toScala).toIndexedSeq)
override def get: Connection = {
@@ -358,7 +355,7 @@ object AmqpConnectionFactoryConnectionProvider {
}
final class AmqpCachedConnectionProvider private (val provider: AmqpConnectionProvider,
- val automaticRelease: Boolean = true)
+ val automaticRelease: Boolean = true)
extends AmqpConnectionProvider {
import akka.stream.alpakka.amqp.AmqpCachedConnectionProvider._
@@ -375,8 +372,7 @@ final class AmqpCachedConnectionProvider private (val provider: AmqpConnectionPr
val connection = provider.get
if (!state.compareAndSet(Connecting, Connected(connection, 1)))
throw new ConcurrentModificationException(
- "Unexpected concurrent modification while creating the connection."
- )
+ "Unexpected concurrent modification while creating the connection.")
connection
} catch {
case e: ConcurrentModificationException => throw e
@@ -394,7 +390,7 @@ final class AmqpCachedConnectionProvider private (val provider: AmqpConnectionPr
@tailrec
override def release(connection: Connection): Unit = state.get match {
- case Empty => throw new IllegalStateException("There is no connection to release.")
+ case Empty => throw new IllegalStateException("There is no connection to release.")
case Connecting => release(connection)
case c @ Connected(cachedConnection, clients) =>
if (cachedConnection != connection)
@@ -405,8 +401,7 @@ final class AmqpCachedConnectionProvider private (val provider: AmqpConnectionPr
provider.release(connection)
if (!state.compareAndSet(Closing, Empty))
throw new ConcurrentModificationException(
- "Unexpected concurrent modification while closing the connection."
- )
+ "Unexpected concurrent modification while closing the connection.")
}
} else {
if (!state.compareAndSet(c, Connected(cachedConnection, clients - 1))) release(connection)
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectorSettings.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectorSettings.scala
index 95223fb9..f608d5a1 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectorSettings.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/AmqpConnectorSettings.scala
@@ -30,8 +30,7 @@ final class NamedQueueSourceSettings private (
val exclusive: Boolean = false,
val ackRequired: Boolean = true,
val consumerTag: String = "default",
- val arguments: Map[String, AnyRef] = Map.empty
-) extends AmqpSourceSettings {
+ val arguments: Map[String, AnyRef] = Map.empty) extends AmqpSourceSettings {
def withDeclaration(declaration: Declaration): NamedQueueSourceSettings =
copy(declarations = immutable.Seq(declaration))
@@ -71,11 +70,11 @@ final class NamedQueueSourceSettings private (
copy(arguments = arguments.asScala.toMap)
private def copy(declarations: immutable.Seq[Declaration] = declarations,
- noLocal: Boolean = noLocal,
- exclusive: Boolean = exclusive,
- ackRequired: Boolean = ackRequired,
- consumerTag: String = consumerTag,
- arguments: Map[String, AnyRef] = arguments) =
+ noLocal: Boolean = noLocal,
+ exclusive: Boolean = exclusive,
+ ackRequired: Boolean = ackRequired,
+ consumerTag: String = consumerTag,
+ arguments: Map[String, AnyRef] = arguments) =
new NamedQueueSourceSettings(
connectionProvider,
queue,
@@ -84,8 +83,7 @@ final class NamedQueueSourceSettings private (
exclusive = exclusive,
ackRequired = ackRequired,
consumerTag = consumerTag,
- arguments = arguments
- )
+ arguments = arguments)
override def toString: String =
"NamedQueueSourceSettings(" +
@@ -115,8 +113,7 @@ final class TemporaryQueueSourceSettings private (
val connectionProvider: AmqpConnectionProvider,
val exchange: String,
val declarations: immutable.Seq[Declaration] = Nil,
- val routingKey: Option[String] = None
-) extends AmqpSourceSettings {
+ val routingKey: Option[String] = None) extends AmqpSourceSettings {
def withDeclaration(declaration: Declaration): TemporaryQueueSourceSettings =
copy(declarations = immutable.Seq(declaration))
@@ -157,8 +154,7 @@ object TemporaryQueueSourceSettings {
final class AmqpReplyToSinkSettings private (
val connectionProvider: AmqpConnectionProvider,
- val failIfReplyToMissing: Boolean = true
-) extends AmqpConnectorSettings {
+ val failIfReplyToMissing: Boolean = true) extends AmqpConnectorSettings {
override final val declarations = Nil
def withFailIfReplyToMissing(failIfReplyToMissing: Boolean): AmqpReplyToSinkSettings =
@@ -191,8 +187,7 @@ final class AmqpWriteSettings private (
val routingKey: Option[String] = None,
val declarations: immutable.Seq[Declaration] = Nil,
val bufferSize: Int = 10,
- val confirmationTimeout: FiniteDuration = 100.millis
-) extends AmqpConnectorSettings {
+ val confirmationTimeout: FiniteDuration = 100.millis) extends AmqpConnectorSettings {
def withExchange(exchange: String): AmqpWriteSettings =
copy(exchange = Some(exchange))
@@ -225,11 +220,11 @@ final class AmqpWriteSettings private (
copy(confirmationTimeout = confirmationTimeout.asScala)
private def copy(connectionProvider: AmqpConnectionProvider = connectionProvider,
- exchange: Option[String] = exchange,
- routingKey: Option[String] = routingKey,
- declarations: immutable.Seq[Declaration] = declarations,
- bufferSize: Int = bufferSize,
- confirmationTimeout: FiniteDuration = confirmationTimeout) =
+ exchange: Option[String] = exchange,
+ routingKey: Option[String] = routingKey,
+ declarations: immutable.Seq[Declaration] = declarations,
+ bufferSize: Int = bufferSize,
+ confirmationTimeout: FiniteDuration = confirmationTimeout) =
new AmqpWriteSettings(connectionProvider, exchange, routingKey, declarations, bufferSize, confirmationTimeout)
override def toString: String =
@@ -261,8 +256,7 @@ final class QueueDeclaration private (
val durable: Boolean = false,
val exclusive: Boolean = false,
val autoDelete: Boolean = false,
- val arguments: Map[String, AnyRef] = Map.empty
-) extends Declaration {
+ val arguments: Map[String, AnyRef] = Map.empty) extends Declaration {
def withDurable(durable: Boolean): QueueDeclaration =
copy(durable = durable)
@@ -283,10 +277,10 @@ final class QueueDeclaration private (
copy(arguments = arguments.asScala.toMap)
private def copy(name: String = name,
- durable: Boolean = durable,
- exclusive: Boolean = exclusive,
- autoDelete: Boolean = autoDelete,
- arguments: Map[String, AnyRef] = arguments) =
+ durable: Boolean = durable,
+ exclusive: Boolean = exclusive,
+ autoDelete: Boolean = autoDelete,
+ arguments: Map[String, AnyRef] = arguments) =
new QueueDeclaration(name, durable, exclusive, autoDelete, arguments)
override def toString: String =
@@ -312,8 +306,7 @@ final class BindingDeclaration private (
val queue: String,
val exchange: String,
val routingKey: Option[String] = None,
- val arguments: Map[String, AnyRef] = Map.empty
-) extends Declaration {
+ val arguments: Map[String, AnyRef] = Map.empty) extends Declaration {
def withRoutingKey(routingKey: String): BindingDeclaration = copy(routingKey = Some(routingKey))
@@ -355,8 +348,7 @@ final class ExchangeDeclaration private (
val durable: Boolean = false,
val autoDelete: Boolean = false,
val internal: Boolean = false,
- val arguments: Map[String, AnyRef] = Map.empty
-) extends Declaration {
+ val arguments: Map[String, AnyRef] = Map.empty) extends Declaration {
def withDurable(durable: Boolean): ExchangeDeclaration = copy(durable = durable)
@@ -374,9 +366,9 @@ final class ExchangeDeclaration private (
copy(arguments = arguments.asScala.toMap)
private def copy(durable: Boolean = durable,
- autoDelete: Boolean = autoDelete,
- internal: Boolean = internal,
- arguments: Map[String, AnyRef] = arguments) =
+ autoDelete: Boolean = autoDelete,
+ internal: Boolean = internal,
+ arguments: Map[String, AnyRef] = arguments) =
new ExchangeDeclaration(name, exchangeType, durable, autoDelete, internal, arguments)
override def toString: String =
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpAsyncFlowStageLogic.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpAsyncFlowStageLogic.scala
index f703f853..87e2894b 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpAsyncFlowStageLogic.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpAsyncFlowStageLogic.scala
@@ -8,7 +8,7 @@ import akka.Done
import akka.annotation.InternalApi
import akka.stream._
import akka.stream.alpakka.amqp.impl.AbstractAmqpAsyncFlowStageLogic.DeliveryTag
-import akka.stream.alpakka.amqp.{AmqpWriteSettings, WriteMessage, WriteResult}
+import akka.stream.alpakka.amqp.{ AmqpWriteSettings, WriteMessage, WriteResult }
import akka.stream.stage._
import com.rabbitmq.client.ConfirmCallback
@@ -21,8 +21,7 @@ import scala.concurrent.Promise
@InternalApi private final case class AwaitingMessage[T](
tag: DeliveryTag,
passThrough: T,
- ready: Boolean = false
-)
+ ready: Boolean = false)
/**
* Internal API.
@@ -39,8 +38,7 @@ import scala.concurrent.Promise
@InternalApi private abstract class AbstractAmqpAsyncFlowStageLogic[T](
override val settings: AmqpWriteSettings,
streamCompletion: Promise[Done],
- shape: FlowShape[(WriteMessage, T), (WriteResult, T)]
-) extends TimerGraphStageLogic(shape)
+ shape: FlowShape[(WriteMessage, T), (WriteResult, T)]) extends TimerGraphStageLogic(shape)
with AmqpConnectorLogic
with StageLogging {
@@ -77,8 +75,7 @@ import scala.concurrent.Promise
dequeued.foreach(m => cancelTimer(m.tag))
pushOrEnqueueResults(
- dequeued.map(m => (WriteResult.confirmed, m.passThrough))
- )
+ dequeued.map(m => (WriteResult.confirmed, m.passThrough)))
}
private def onRejection(tag: DeliveryTag, multiple: Boolean): Unit = {
@@ -89,21 +86,18 @@ import scala.concurrent.Promise
dequeued.foreach(m => cancelTimer(m.tag))
pushOrEnqueueResults(
- dequeued.map(m => (WriteResult.rejected, m.passThrough))
- )
+ dequeued.map(m => (WriteResult.rejected, m.passThrough)))
}
private def pushOrEnqueueResults(results: Iterable[(WriteResult, T)]): Unit = {
- results.foreach(
- result =>
- if (isAvailable(out) && exitQueue.isEmpty) {
- log.debug("Pushing {} downstream.", result)
- push(out, result)
- } else {
- log.debug("Message {} queued for downstream push.", result)
- exitQueue.enqueue(result)
- }
- )
+ results.foreach(result =>
+ if (isAvailable(out) && exitQueue.isEmpty) {
+ log.debug("Pushing {} downstream.", result)
+ push(out, result)
+ } else {
+ log.debug("Message {} queued for downstream push.", result)
+ exitQueue.enqueue(result)
+ })
if (isFinished) closeStage()
}
@@ -166,13 +160,11 @@ import scala.concurrent.Promise
message.mandatory,
message.immediate,
message.properties.orNull,
- message.bytes.toArray
- )
+ message.bytes.toArray)
tag
}
- }
- )
+ })
setHandler(
out,
@@ -187,8 +179,7 @@ import scala.concurrent.Promise
if (isFinished) closeStage()
else if (!hasBeenPulled(in)) tryPull(in)
}
- }
- )
+ })
override protected def onTimer(timerKey: Any): Unit =
timerKey match {
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpFlowStageLogic.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpFlowStageLogic.scala
index 474451ad..cb15d008 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpFlowStageLogic.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AbstractAmqpFlowStageLogic.scala
@@ -7,8 +7,8 @@ package akka.stream.alpakka.amqp.impl
import akka.Done
import akka.annotation.InternalApi
import akka.stream._
-import akka.stream.alpakka.amqp.{AmqpWriteSettings, WriteMessage, WriteResult}
-import akka.stream.stage.{GraphStageLogic, InHandler, OutHandler, StageLogging}
+import akka.stream.alpakka.amqp.{ AmqpWriteSettings, WriteMessage, WriteResult }
+import akka.stream.stage.{ GraphStageLogic, InHandler, OutHandler, StageLogging }
import scala.concurrent.Promise
@@ -20,8 +20,7 @@ import scala.concurrent.Promise
@InternalApi private abstract class AbstractAmqpFlowStageLogic[T](
override val settings: AmqpWriteSettings,
streamCompletion: Promise[Done],
- shape: FlowShape[(WriteMessage, T), (WriteResult, T)]
-) extends GraphStageLogic(shape)
+ shape: FlowShape[(WriteMessage, T), (WriteResult, T)]) extends GraphStageLogic(shape)
with AmqpConnectorLogic
with StageLogging {
@@ -47,15 +46,15 @@ import scala.concurrent.Promise
val (message, passThrough) = grab(in)
publish(message, passThrough)
}
- }
- )
+ })
protected def publish(message: WriteMessage, passThrough: T): Unit
- setHandler(out, new OutHandler {
- override def onPull(): Unit =
- if (!hasBeenPulled(in)) tryPull(in)
- })
+ setHandler(out,
+ new OutHandler {
+ override def onPull(): Unit =
+ if (!hasBeenPulled(in)) tryPull(in)
+ })
override def postStop(): Unit = {
streamCompletion.tryFailure(new RuntimeException("Stage stopped unexpectedly."))
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncFlowStage.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncFlowStage.scala
index f54ae4ce..6f91a0a5 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncFlowStage.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncFlowStage.scala
@@ -8,12 +8,12 @@ import akka.Done
import akka.annotation.InternalApi
import akka.event.Logging
import akka.stream.alpakka.amqp.impl.AbstractAmqpAsyncFlowStageLogic.DeliveryTag
-import akka.stream.alpakka.amqp.{AmqpWriteSettings, WriteMessage, WriteResult}
-import akka.stream.stage.{GraphStageLogic, GraphStageWithMaterializedValue}
+import akka.stream.alpakka.amqp.{ AmqpWriteSettings, WriteMessage, WriteResult }
+import akka.stream.stage.{ GraphStageLogic, GraphStageWithMaterializedValue }
import akka.stream._
import scala.collection.immutable.TreeMap
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* Internal API.
@@ -27,8 +27,8 @@ import scala.concurrent.{Future, Promise}
* this delivery tag can be safely dequeued.
*/
@InternalApi private[amqp] final class AmqpAsyncFlowStage[T](
- settings: AmqpWriteSettings
-) extends GraphStageWithMaterializedValue[FlowShape[(WriteMessage, T), (WriteResult, T)], Future[Done]] {
+ settings: AmqpWriteSettings)
+ extends GraphStageWithMaterializedValue[FlowShape[(WriteMessage, T), (WriteResult, T)], Future[Done]] {
val in: Inlet[(WriteMessage, T)] = Inlet(Logging.simpleName(this) + ".in")
val out: Outlet[(WriteResult, T)] = Outlet(Logging.simpleName(this) + ".out")
@@ -42,41 +42,40 @@ import scala.concurrent.{Future, Promise}
val streamCompletion = Promise[Done]()
(new AbstractAmqpAsyncFlowStageLogic(settings, streamCompletion, shape) {
- private var buffer = TreeMap[DeliveryTag, AwaitingMessage[T]]()
+ private var buffer = TreeMap[DeliveryTag, AwaitingMessage[T]]()
- override def enqueueMessage(tag: DeliveryTag, passThrough: T): Unit =
- buffer += (tag -> AwaitingMessage(tag, passThrough))
+ override def enqueueMessage(tag: DeliveryTag, passThrough: T): Unit =
+ buffer += (tag -> AwaitingMessage(tag, passThrough))
- override def dequeueAwaitingMessages(tag: DeliveryTag, multiple: Boolean): Iterable[AwaitingMessage[T]] =
- if (multiple) {
- dequeueWhile((t, _) => t <= tag)
- } else {
- setReady(tag)
- if (isAtHead(tag)) {
- dequeueWhile((_, message) => message.ready)
+ override def dequeueAwaitingMessages(tag: DeliveryTag, multiple: Boolean): Iterable[AwaitingMessage[T]] =
+ if (multiple) {
+ dequeueWhile((t, _) => t <= tag)
} else {
- Seq.empty
+ setReady(tag)
+ if (isAtHead(tag)) {
+ dequeueWhile((_, message) => message.ready)
+ } else {
+ Seq.empty
+ }
}
- }
- private def dequeueWhile(
- predicate: (DeliveryTag, AwaitingMessage[T]) => Boolean
- ): Iterable[AwaitingMessage[T]] = {
- val dequeued = buffer.takeWhile { case (k, v) => predicate(k, v) }
- buffer --= dequeued.keys
- dequeued.values
- }
+ private def dequeueWhile(
+ predicate: (DeliveryTag, AwaitingMessage[T]) => Boolean): Iterable[AwaitingMessage[T]] = {
+ val dequeued = buffer.takeWhile { case (k, v) => predicate(k, v) }
+ buffer --= dequeued.keys
+ dequeued.values
+ }
- private def isAtHead(tag: DeliveryTag): Boolean =
- buffer.headOption.exists { case (tag, _) => tag == tag }
+ private def isAtHead(tag: DeliveryTag): Boolean =
+ buffer.headOption.exists { case (tag, _) => tag == tag }
- private def setReady(tag: DeliveryTag): Unit =
- buffer.get(tag).foreach(message => buffer += (tag -> message.copy(ready = true)))
+ private def setReady(tag: DeliveryTag): Unit =
+ buffer.get(tag).foreach(message => buffer += (tag -> message.copy(ready = true)))
- override def messagesAwaitingDelivery: Int = buffer.size
+ override def messagesAwaitingDelivery: Int = buffer.size
- override def noAwaitingMessages: Boolean = buffer.isEmpty
+ override def noAwaitingMessages: Boolean = buffer.isEmpty
- }, streamCompletion.future)
+ }, streamCompletion.future)
}
}
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncUnorderedFlowStage.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncUnorderedFlowStage.scala
index 34e7da2b..d718871e 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncUnorderedFlowStage.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpAsyncUnorderedFlowStage.scala
@@ -8,12 +8,12 @@ import akka.Done
import akka.annotation.InternalApi
import akka.event.Logging
import akka.stream.alpakka.amqp.impl.AbstractAmqpAsyncFlowStageLogic.DeliveryTag
-import akka.stream.alpakka.amqp.{AmqpWriteSettings, WriteMessage, WriteResult}
-import akka.stream.stage.{GraphStageLogic, GraphStageWithMaterializedValue}
+import akka.stream.alpakka.amqp.{ AmqpWriteSettings, WriteMessage, WriteResult }
+import akka.stream.stage.{ GraphStageLogic, GraphStageWithMaterializedValue }
import akka.stream._
import scala.collection.mutable
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* Internal API.
@@ -24,8 +24,8 @@ import scala.concurrent.{Future, Promise}
* given delivery tag, which means that so all messages up to (and including) this delivery tag can be safely dequeued.
*/
@InternalApi private[amqp] final class AmqpAsyncUnorderedFlowStage[T](
- settings: AmqpWriteSettings
-) extends GraphStageWithMaterializedValue[FlowShape[(WriteMessage, T), (WriteResult, T)], Future[Done]] {
+ settings: AmqpWriteSettings)
+ extends GraphStageWithMaterializedValue[FlowShape[(WriteMessage, T), (WriteResult, T)], Future[Done]] {
private val in: Inlet[(WriteMessage, T)] = Inlet(Logging.simpleName(this) + ".in")
private val out: Outlet[(WriteResult, T)] = Outlet(Logging.simpleName(this) + ".out")
@@ -39,23 +39,23 @@ import scala.concurrent.{Future, Promise}
val streamCompletion = Promise[Done]()
(new AbstractAmqpAsyncFlowStageLogic(settings, streamCompletion, shape) {
- private val buffer = mutable.Queue.empty[AwaitingMessage[T]]
+ private val buffer = mutable.Queue.empty[AwaitingMessage[T]]
- override def enqueueMessage(tag: DeliveryTag, passThrough: T): Unit =
- buffer += AwaitingMessage(tag, passThrough)
+ override def enqueueMessage(tag: DeliveryTag, passThrough: T): Unit =
+ buffer += AwaitingMessage(tag, passThrough)
- override def dequeueAwaitingMessages(tag: DeliveryTag, multiple: Boolean): Iterable[AwaitingMessage[T]] =
- if (multiple)
- buffer.dequeueAll(_.tag <= tag)
- else
- buffer
- .dequeueFirst(_.tag == tag)
- .fold(Seq.empty[AwaitingMessage[T]])(Seq(_))
+ override def dequeueAwaitingMessages(tag: DeliveryTag, multiple: Boolean): Iterable[AwaitingMessage[T]] =
+ if (multiple)
+ buffer.dequeueAll(_.tag <= tag)
+ else
+ buffer
+ .dequeueFirst(_.tag == tag)
+ .fold(Seq.empty[AwaitingMessage[T]])(Seq(_))
- override def messagesAwaitingDelivery: Int = buffer.length
+ override def messagesAwaitingDelivery: Int = buffer.length
- override def noAwaitingMessages: Boolean = buffer.isEmpty
+ override def noAwaitingMessages: Boolean = buffer.isEmpty
- }, streamCompletion.future)
+ }, streamCompletion.future)
}
}
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpConnectorLogic.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpConnectorLogic.scala
index bf93f56a..5729ead7 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpConnectorLogic.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpConnectorLogic.scala
@@ -4,8 +4,8 @@
package akka.stream.alpakka.amqp.impl
-import akka.stream.alpakka.amqp.{AmqpConnectorSettings, BindingDeclaration, ExchangeDeclaration, QueueDeclaration}
-import akka.stream.stage.{AsyncCallback, GraphStageLogic}
+import akka.stream.alpakka.amqp.{ AmqpConnectorSettings, BindingDeclaration, ExchangeDeclaration, QueueDeclaration }
+import akka.stream.stage.{ AsyncCallback, GraphStageLogic }
import com.rabbitmq.client._
import scala.util.control.NonFatal
@@ -41,16 +41,14 @@ private trait AmqpConnectorLogic { this: GraphStageLogic =>
d.durable,
d.exclusive,
d.autoDelete,
- d.arguments.asJava
- )
+ d.arguments.asJava)
case d: BindingDeclaration =>
channel.queueBind(
d.queue,
d.exchange,
d.routingKey.getOrElse(""),
- d.arguments.asJava
- )
+ d.arguments.asJava)
case d: ExchangeDeclaration =>
channel.exchangeDeclare(
@@ -59,8 +57,7 @@ private trait AmqpConnectorLogic { this: GraphStageLogic =>
d.durable,
d.autoDelete,
d.internal,
- d.arguments.asJava
- )
+ d.arguments.asJava)
}
whenConnected()
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpReplyToSinkStage.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpReplyToSinkStage.scala
index e230c75e..52f0fa3f 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpReplyToSinkStage.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpReplyToSinkStage.scala
@@ -6,11 +6,11 @@ package akka.stream.alpakka.amqp.impl
import akka.Done
import akka.annotation.InternalApi
-import akka.stream.alpakka.amqp.{AmqpReplyToSinkSettings, WriteMessage}
-import akka.stream.stage.{GraphStageLogic, GraphStageWithMaterializedValue, InHandler}
-import akka.stream.{ActorAttributes, Attributes, Inlet, SinkShape}
+import akka.stream.alpakka.amqp.{ AmqpReplyToSinkSettings, WriteMessage }
+import akka.stream.stage.{ GraphStageLogic, GraphStageWithMaterializedValue, InHandler }
+import akka.stream.{ ActorAttributes, Attributes, Inlet, SinkShape }
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* Connects to an AMQP server upon materialization and sends write messages to the server.
@@ -31,58 +31,56 @@ private[amqp] final class AmqpReplyToSinkStage(settings: AmqpReplyToSinkSettings
override def createLogicAndMaterializedValue(inheritedAttributes: Attributes): (GraphStageLogic, Future[Done]) = {
val streamCompletion = Promise[Done]()
(new GraphStageLogic(shape) with AmqpConnectorLogic {
- override val settings = stage.settings
-
- override def whenConnected(): Unit = pull(in)
-
- override def postStop(): Unit = {
- streamCompletion.tryFailure(new RuntimeException("stage stopped unexpectedly"))
- super.postStop()
- }
-
- override def onFailure(ex: Throwable): Unit = {
- streamCompletion.tryFailure(ex)
- super.onFailure(ex)
- }
-
- setHandler(
- in,
- new InHandler {
-
- override def onUpstreamFailure(ex: Throwable): Unit = {
- streamCompletion.failure(ex)
- super.onUpstreamFailure(ex)
- }
-
- override def onUpstreamFinish(): Unit = {
- streamCompletion.success(Done)
- super.onUpstreamFinish()
- }
-
- override def onPush(): Unit = {
- val elem = grab(in)
-
- val replyTo = elem.properties.flatMap(properties => Option(properties.getReplyTo))
-
- if (replyTo.isDefined) {
- channel.basicPublish(
- elem.routingKey.getOrElse(""),
- replyTo.get,
- elem.mandatory,
- elem.immediate,
- elem.properties.orNull,
- elem.bytes.toArray
- )
- } else if (settings.failIfReplyToMissing) {
- onFailure(new RuntimeException("Reply-to header was not set"))
- }
+ override val settings = stage.settings
+
+ override def whenConnected(): Unit = pull(in)
+
+ override def postStop(): Unit = {
+ streamCompletion.tryFailure(new RuntimeException("stage stopped unexpectedly"))
+ super.postStop()
+ }
- tryPull(in)
- }
+ override def onFailure(ex: Throwable): Unit = {
+ streamCompletion.tryFailure(ex)
+ super.onFailure(ex)
}
- )
- }, streamCompletion.future)
+ setHandler(
+ in,
+ new InHandler {
+
+ override def onUpstreamFailure(ex: Throwable): Unit = {
+ streamCompletion.failure(ex)
+ super.onUpstreamFailure(ex)
+ }
+
+ override def onUpstreamFinish(): Unit = {
+ streamCompletion.success(Done)
+ super.onUpstreamFinish()
+ }
+
+ override def onPush(): Unit = {
+ val elem = grab(in)
+
+ val replyTo = elem.properties.flatMap(properties => Option(properties.getReplyTo))
+
+ if (replyTo.isDefined) {
+ channel.basicPublish(
+ elem.routingKey.getOrElse(""),
+ replyTo.get,
+ elem.mandatory,
+ elem.immediate,
+ elem.properties.orNull,
+ elem.bytes.toArray)
+ } else if (settings.failIfReplyToMissing) {
+ onFailure(new RuntimeException("Reply-to header was not set"))
+ }
+
+ tryPull(in)
+ }
+ })
+
+ }, streamCompletion.future)
}
override def toString: String = "AmqpReplyToSink"
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpRpcFlowStage.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpRpcFlowStage.scala
index 01576960..c46e4a61 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpRpcFlowStage.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpRpcFlowStage.scala
@@ -17,7 +17,7 @@ import com.rabbitmq.client.AMQP.BasicProperties
import com.rabbitmq.client._
import scala.collection.mutable
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
import scala.util.Success
/**
@@ -43,191 +43,185 @@ private[amqp] final class AmqpRpcFlowStage(settings: AmqpWriteSettings, bufferSi
val streamCompletion = Promise[String]()
(new GraphStageLogic(shape) with AmqpConnectorLogic {
- override val settings = stage.settings
- private val exchange = settings.exchange.getOrElse("")
- private val routingKey = settings.routingKey.getOrElse("")
- private val queue = mutable.Queue[CommittableReadResult]()
- private var queueName: String = _
- private var unackedMessages = 0
- private var outstandingMessages = 0
-
- override def whenConnected(): Unit = {
-
- pull(in)
-
- channel.basicQos(bufferSize)
- val consumerCallback = getAsyncCallback(handleDelivery)
-
- val commitCallback = getAsyncCallback[AckArguments] {
- case AckArguments(deliveryTag, multiple, promise) => {
- try {
- channel.basicAck(deliveryTag, multiple)
- unackedMessages -= 1
- if (unackedMessages == 0 && (isClosed(out) || (isClosed(in) && queue.isEmpty && outstandingMessages == 0)))
- completeStage()
- promise.complete(Success(Done))
- } catch {
- case e: Throwable => promise.failure(e)
+ override val settings = stage.settings
+ private val exchange = settings.exchange.getOrElse("")
+ private val routingKey = settings.routingKey.getOrElse("")
+ private val queue = mutable.Queue[CommittableReadResult]()
+ private var queueName: String = _
+ private var unackedMessages = 0
+ private var outstandingMessages = 0
+
+ override def whenConnected(): Unit = {
+
+ pull(in)
+
+ channel.basicQos(bufferSize)
+ val consumerCallback = getAsyncCallback(handleDelivery)
+
+ val commitCallback = getAsyncCallback[AckArguments] {
+ case AckArguments(deliveryTag, multiple, promise) => {
+ try {
+ channel.basicAck(deliveryTag, multiple)
+ unackedMessages -= 1
+ if (unackedMessages == 0 && (isClosed(out) || (isClosed(
+ in) && queue.isEmpty && outstandingMessages == 0)))
+ completeStage()
+ promise.complete(Success(Done))
+ } catch {
+ case e: Throwable => promise.failure(e)
+ }
}
}
- }
- val nackCallback = getAsyncCallback[NackArguments] {
- case NackArguments(deliveryTag, multiple, requeue, promise) => {
- try {
- channel.basicNack(deliveryTag, multiple, requeue)
- unackedMessages -= 1
- if (unackedMessages == 0 && (isClosed(out) || (isClosed(in) && queue.isEmpty && outstandingMessages == 0)))
- completeStage()
- promise.complete(Success(Done))
- } catch {
- case e: Throwable => promise.failure(e)
+ val nackCallback = getAsyncCallback[NackArguments] {
+ case NackArguments(deliveryTag, multiple, requeue, promise) => {
+ try {
+ channel.basicNack(deliveryTag, multiple, requeue)
+ unackedMessages -= 1
+ if (unackedMessages == 0 && (isClosed(out) || (isClosed(
+ in) && queue.isEmpty && outstandingMessages == 0)))
+ completeStage()
+ promise.complete(Success(Done))
+ } catch {
+ case e: Throwable => promise.failure(e)
+ }
}
}
- }
- val amqpSourceConsumer = new DefaultConsumer(channel) {
- override def handleDelivery(consumerTag: String,
- envelope: Envelope,
- properties: BasicProperties,
- body: Array[Byte]): Unit =
- consumerCallback.invoke(
- new CommittableReadResult {
- override val message = ReadResult(ByteString(body), envelope, properties)
-
- override def ack(multiple: Boolean): Future[Done] = {
- val promise = Promise[Done]()
- commitCallback.invoke(AckArguments(message.envelope.getDeliveryTag, multiple, promise))
- promise.future
- }
+ val amqpSourceConsumer = new DefaultConsumer(channel) {
+ override def handleDelivery(consumerTag: String,
+ envelope: Envelope,
+ properties: BasicProperties,
+ body: Array[Byte]): Unit =
+ consumerCallback.invoke(
+ new CommittableReadResult {
+ override val message = ReadResult(ByteString(body), envelope, properties)
+
+ override def ack(multiple: Boolean): Future[Done] = {
+ val promise = Promise[Done]()
+ commitCallback.invoke(AckArguments(message.envelope.getDeliveryTag, multiple, promise))
+ promise.future
+ }
+
+ override def nack(multiple: Boolean, requeue: Boolean): Future[Done] = {
+ val promise = Promise[Done]()
+ nackCallback.invoke(NackArguments(message.envelope.getDeliveryTag, multiple, requeue, promise))
+ promise.future
+ }
+ })
+
+ override def handleCancel(consumerTag: String): Unit =
+ // non consumer initiated cancel, for example happens when the queue has been deleted.
+ shutdownCallback.invoke(
+ new RuntimeException(s"Consumer $queueName with consumerTag $consumerTag shut down unexpectedly"))
+
+ override def handleShutdownSignal(consumerTag: String, sig: ShutdownSignalException): Unit =
+ // "Called when either the channel or the underlying connection has been shut down."
+ shutdownCallback.invoke(
+ new RuntimeException(s"Consumer $queueName with consumerTag $consumerTag shut down unexpectedly", sig))
+ }
- override def nack(multiple: Boolean, requeue: Boolean): Future[Done] = {
- val promise = Promise[Done]()
- nackCallback.invoke(NackArguments(message.envelope.getDeliveryTag, multiple, requeue, promise))
- promise.future
- }
- }
- )
-
- override def handleCancel(consumerTag: String): Unit =
- // non consumer initiated cancel, for example happens when the queue has been deleted.
- shutdownCallback.invoke(
- new RuntimeException(s"Consumer $queueName with consumerTag $consumerTag shut down unexpectedly")
- )
-
- override def handleShutdownSignal(consumerTag: String, sig: ShutdownSignalException): Unit =
- // "Called when either the channel or the underlying connection has been shut down."
- shutdownCallback.invoke(
- new RuntimeException(s"Consumer $queueName with consumerTag $consumerTag shut down unexpectedly", sig)
- )
+ // Create an exclusive queue with a randomly generated name for use as the replyTo portion of RPC
+ queueName = channel
+ .queueDeclare(
+ "",
+ false,
+ true,
+ true,
+ Collections.emptyMap())
+ .getQueue
+
+ channel.basicConsume(
+ queueName,
+ amqpSourceConsumer)
+ streamCompletion.success(queueName)
}
- // Create an exclusive queue with a randomly generated name for use as the replyTo portion of RPC
- queueName = channel
- .queueDeclare(
- "",
- false,
- true,
- true,
- Collections.emptyMap()
- )
- .getQueue
-
- channel.basicConsume(
- queueName,
- amqpSourceConsumer
- )
- streamCompletion.success(queueName)
- }
-
- def handleDelivery(message: CommittableReadResult): Unit =
- if (isAvailable(out)) {
- pushMessage(message)
- } else if (queue.size + 1 > bufferSize) {
- onFailure(new RuntimeException(s"Reached maximum buffer size $bufferSize"))
- } else {
- queue.enqueue(message)
- }
+ def handleDelivery(message: CommittableReadResult): Unit =
+ if (isAvailable(out)) {
+ pushMessage(message)
+ } else if (queue.size + 1 > bufferSize) {
+ onFailure(new RuntimeException(s"Reached maximum buffer size $bufferSize"))
+ } else {
+ queue.enqueue(message)
+ }
+
+ setHandler(
+ out,
+ new OutHandler {
+ override def onPull(): Unit =
+ if (queue.nonEmpty) {
+ pushMessage(queue.dequeue())
+ }
- setHandler(
- out,
- new OutHandler {
- override def onPull(): Unit =
- if (queue.nonEmpty) {
- pushMessage(queue.dequeue())
+ override def onDownstreamFinish(cause: Throwable): Unit = {
+ setKeepGoing(true)
+ if (unackedMessages == 0) super.onDownstreamFinish(cause)
}
+ })
- override def onDownstreamFinish(cause: Throwable): Unit = {
- setKeepGoing(true)
- if (unackedMessages == 0) super.onDownstreamFinish(cause)
- }
+ def pushMessage(message: CommittableReadResult): Unit = {
+ push(out, message)
+ unackedMessages += 1
+ outstandingMessages -= 1
}
- )
-
- def pushMessage(message: CommittableReadResult): Unit = {
- push(out, message)
- unackedMessages += 1
- outstandingMessages -= 1
- }
-
- setHandler(
- in,
- new InHandler {
- // We don't want to finish since we're still waiting
- // on incoming messages from rabbit. However, if we
- // haven't processed a message yet, we do want to complete
- // so that we don't hang.
- override def onUpstreamFinish(): Unit = {
- setKeepGoing(true)
- if (queue.isEmpty && outstandingMessages == 0 && unackedMessages == 0) super.onUpstreamFinish()
- }
- override def onUpstreamFailure(ex: Throwable): Unit = {
- setKeepGoing(true)
- if (queue.isEmpty && outstandingMessages == 0 && unackedMessages == 0)
- super.onUpstreamFailure(ex)
- }
+ setHandler(
+ in,
+ new InHandler {
+ // We don't want to finish since we're still waiting
+ // on incoming messages from rabbit. However, if we
+ // haven't processed a message yet, we do want to complete
+ // so that we don't hang.
+ override def onUpstreamFinish(): Unit = {
+ setKeepGoing(true)
+ if (queue.isEmpty && outstandingMessages == 0 && unackedMessages == 0) super.onUpstreamFinish()
+ }
- override def onPush(): Unit = {
- val elem = grab(in)
- val props = elem.properties.getOrElse(new BasicProperties()).builder.replyTo(queueName).build()
- channel.basicPublish(
- exchange,
- elem.routingKey.getOrElse(routingKey),
- elem.mandatory,
- elem.immediate,
- props,
- elem.bytes.toArray
- )
-
- val expectedResponses: Int = {
- val headers = props.getHeaders
- if (headers == null) {
- responsesPerMessage
- } else {
- val r = headers.get("expectedReplies")
- if (r != null) {
- r.asInstanceOf[Int]
- } else {
+ override def onUpstreamFailure(ex: Throwable): Unit = {
+ setKeepGoing(true)
+ if (queue.isEmpty && outstandingMessages == 0 && unackedMessages == 0)
+ super.onUpstreamFailure(ex)
+ }
+
+ override def onPush(): Unit = {
+ val elem = grab(in)
+ val props = elem.properties.getOrElse(new BasicProperties()).builder.replyTo(queueName).build()
+ channel.basicPublish(
+ exchange,
+ elem.routingKey.getOrElse(routingKey),
+ elem.mandatory,
+ elem.immediate,
+ props,
+ elem.bytes.toArray)
+
+ val expectedResponses: Int = {
+ val headers = props.getHeaders
+ if (headers == null) {
responsesPerMessage
+ } else {
+ val r = headers.get("expectedReplies")
+ if (r != null) {
+ r.asInstanceOf[Int]
+ } else {
+ responsesPerMessage
+ }
}
}
+
+ outstandingMessages += expectedResponses
+ pull(in)
}
+ })
+ override def postStop(): Unit = {
+ streamCompletion.tryFailure(new RuntimeException("stage stopped unexpectedly"))
+ super.postStop()
+ }
- outstandingMessages += expectedResponses
- pull(in)
- }
+ override def onFailure(ex: Throwable): Unit = {
+ streamCompletion.tryFailure(ex)
+ super.onFailure(ex)
}
- )
- override def postStop(): Unit = {
- streamCompletion.tryFailure(new RuntimeException("stage stopped unexpectedly"))
- super.postStop()
- }
-
- override def onFailure(ex: Throwable): Unit = {
- streamCompletion.tryFailure(ex)
- super.onFailure(ex)
- }
- }, streamCompletion.future)
+ }, streamCompletion.future)
}
override def toString: String = "AmqpRpcFlow"
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSimpleFlowStage.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSimpleFlowStage.scala
index 0d3a855f..149e0808 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSimpleFlowStage.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSimpleFlowStage.scala
@@ -7,11 +7,11 @@ package akka.stream.alpakka.amqp.impl
import akka.Done
import akka.annotation.InternalApi
import akka.event.Logging
-import akka.stream.{ActorAttributes, Attributes, FlowShape, Inlet, Outlet}
-import akka.stream.alpakka.amqp.{AmqpWriteSettings, WriteMessage, WriteResult}
-import akka.stream.stage.{GraphStageLogic, GraphStageWithMaterializedValue}
+import akka.stream.{ ActorAttributes, Attributes, FlowShape, Inlet, Outlet }
+import akka.stream.alpakka.amqp.{ AmqpWriteSettings, WriteMessage, WriteResult }
+import akka.stream.stage.{ GraphStageLogic, GraphStageWithMaterializedValue }
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* Internal API.
@@ -36,19 +36,18 @@ import scala.concurrent.{Future, Promise}
override def createLogicAndMaterializedValue(inheritedAttributes: Attributes): (GraphStageLogic, Future[Done]) = {
val streamCompletion = Promise[Done]()
(new AbstractAmqpFlowStageLogic[T](settings, streamCompletion, shape) {
- override def publish(message: WriteMessage, passThrough: T): Unit = {
- log.debug("Publishing message {}.", message)
-
- channel.basicPublish(
- settings.exchange.getOrElse(""),
- message.routingKey.orElse(settings.routingKey).getOrElse(""),
- message.mandatory,
- message.immediate,
- message.properties.orNull,
- message.bytes.toArray
- )
- push(out, (WriteResult.confirmed, passThrough))
- }
- }, streamCompletion.future)
+ override def publish(message: WriteMessage, passThrough: T): Unit = {
+ log.debug("Publishing message {}.", message)
+
+ channel.basicPublish(
+ settings.exchange.getOrElse(""),
+ message.routingKey.orElse(settings.routingKey).getOrElse(""),
+ message.mandatory,
+ message.immediate,
+ message.properties.orNull,
+ message.bytes.toArray)
+ push(out, (WriteResult.confirmed, passThrough))
+ }
+ }, streamCompletion.future)
}
}
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSourceStage.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSourceStage.scala
index bb01248d..ab7c4754 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSourceStage.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/impl/AmqpSourceStage.scala
@@ -9,14 +9,14 @@ import akka.annotation.InternalApi
import akka.stream.alpakka.amqp._
import akka.stream.alpakka.amqp.impl.AmqpSourceStage.AutoAckedReadResult
import akka.stream.alpakka.amqp.scaladsl.CommittableReadResult
-import akka.stream.stage.{GraphStage, GraphStageLogic, OutHandler, StageLogging}
-import akka.stream.{Attributes, Outlet, SourceShape}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, OutHandler, StageLogging }
+import akka.stream.{ Attributes, Outlet, SourceShape }
import akka.util.ByteString
import com.rabbitmq.client.AMQP.BasicProperties
-import com.rabbitmq.client.{DefaultConsumer, Envelope, ShutdownSignalException}
+import com.rabbitmq.client.{ DefaultConsumer, Envelope, ShutdownSignalException }
import scala.collection.mutable
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
import scala.util.Success
private final case class AckArguments(deliveryTag: Long, multiple: Boolean, promise: Promise[Done])
@@ -80,9 +80,9 @@ private[amqp] final class AmqpSourceStage(settings: AmqpSourceSettings, bufferSi
val amqpSourceConsumer = new DefaultConsumer(channel) {
override def handleDelivery(consumerTag: String,
- envelope: Envelope,
- properties: BasicProperties,
- body: Array[Byte]): Unit = {
+ envelope: Envelope,
+ properties: BasicProperties,
+ body: Array[Byte]): Unit = {
val message = if (ackRequired) {
new CommittableReadResult {
@@ -107,14 +107,12 @@ private[amqp] final class AmqpSourceStage(settings: AmqpSourceSettings, bufferSi
override def handleCancel(consumerTag: String): Unit =
// non consumer initiated cancel, for example happens when the queue has been deleted.
shutdownCallback.invoke(
- new RuntimeException(s"Consumer with consumerTag $consumerTag shut down unexpectedly")
- )
+ new RuntimeException(s"Consumer with consumerTag $consumerTag shut down unexpectedly"))
override def handleShutdownSignal(consumerTag: String, sig: ShutdownSignalException): Unit =
// "Called when either the channel or the underlying connection has been shut down."
shutdownCallback.invoke(
- new RuntimeException(s"Consumer with consumerTag $consumerTag shut down unexpectedly", sig)
- )
+ new RuntimeException(s"Consumer with consumerTag $consumerTag shut down unexpectedly", sig))
}
def setupNamedQueue(settings: NamedQueueSourceSettings): Unit =
@@ -125,8 +123,7 @@ private[amqp] final class AmqpSourceStage(settings: AmqpSourceSettings, bufferSi
settings.noLocal,
settings.exclusive,
settings.arguments.asJava,
- amqpSourceConsumer
- )
+ amqpSourceConsumer)
def setupTemporaryQueue(settings: TemporaryQueueSourceSettings): Unit = {
// this is a weird case that required dynamic declaration, the queue name is not known
@@ -135,8 +132,7 @@ private[amqp] final class AmqpSourceStage(settings: AmqpSourceSettings, bufferSi
channel.queueBind(queueName, settings.exchange, settings.routingKey.getOrElse(""))
channel.basicConsume(
queueName,
- amqpSourceConsumer
- )
+ amqpSourceConsumer)
}
settings match {
@@ -171,8 +167,7 @@ private[amqp] final class AmqpSourceStage(settings: AmqpSourceSettings, bufferSi
setKeepGoing(true)
log.debug("Awaiting {} acks before finishing.", unackedMessages)
}
- }
- )
+ })
def pushMessage(message: CommittableReadResult): Unit = {
push(out, message)
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlow.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlow.scala
index 07e7d073..8c61ede6 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlow.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlow.scala
@@ -27,8 +27,7 @@ object AmqpFlow {
* @param settings `bufferSize` and `confirmationTimeout` properties are ignored by this connector
*/
def create(
- settings: AmqpWriteSettings
- ): akka.stream.javadsl.Flow[WriteMessage, WriteResult, CompletionStage[Done]] =
+ settings: AmqpWriteSettings): akka.stream.javadsl.Flow[WriteMessage, WriteResult, CompletionStage[Done]] =
akka.stream.alpakka.amqp.scaladsl.AmqpFlow(settings).mapMaterializedValue(f => f.toJava).asJava
/**
@@ -50,8 +49,7 @@ object AmqpFlow {
* supposed to be used with another AMQP brokers.
*/
def createWithConfirm(
- settings: AmqpWriteSettings
- ): akka.stream.javadsl.Flow[WriteMessage, WriteResult, CompletionStage[Done]] =
+ settings: AmqpWriteSettings): akka.stream.javadsl.Flow[WriteMessage, WriteResult, CompletionStage[Done]] =
akka.stream.alpakka.amqp.scaladsl.AmqpFlow
.withConfirm(settings = settings)
.mapMaterializedValue(_.toJava)
@@ -76,8 +74,7 @@ object AmqpFlow {
* supposed to be used with another AMQP brokers.
*/
def createWithConfirmUnordered(
- settings: AmqpWriteSettings
- ): akka.stream.javadsl.Flow[WriteMessage, WriteResult, CompletionStage[Done]] =
+ settings: AmqpWriteSettings): akka.stream.javadsl.Flow[WriteMessage, WriteResult, CompletionStage[Done]] =
akka.stream.alpakka.amqp.scaladsl.AmqpFlow
.withConfirmUnordered(settings)
.mapMaterializedValue(_.toJava)
@@ -93,15 +90,14 @@ object AmqpFlow {
* supposed to be used with another AMQP brokers.
*/
def createWithConfirmAndPassThroughUnordered[T](
- settings: AmqpWriteSettings
- ): akka.stream.javadsl.Flow[Pair[WriteMessage, T], Pair[WriteResult, T], CompletionStage[Done]] =
+ settings: AmqpWriteSettings)
+ : akka.stream.javadsl.Flow[Pair[WriteMessage, T], Pair[WriteResult, T], CompletionStage[Done]] =
akka.stream.scaladsl
.Flow[Pair[WriteMessage, T]]
.map((p: Pair[WriteMessage, T]) => p.toScala)
.viaMat(
akka.stream.alpakka.amqp.scaladsl.AmqpFlow
- .withConfirmAndPassThroughUnordered[T](settings = settings)
- )(Keep.right)
+ .withConfirmAndPassThroughUnordered[T](settings = settings))(Keep.right)
.map { case (writeResult, passThrough) => Pair(writeResult, passThrough) }
.mapMaterializedValue(_.toJava)
.asJava
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlowWithContext.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlowWithContext.scala
index 943e5f99..53535496 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlowWithContext.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpFlowWithContext.scala
@@ -19,8 +19,8 @@ object AmqpFlowWithContext {
* @see [[AmqpFlow.create]]
*/
def create[T](
- settings: AmqpWriteSettings
- ): akka.stream.javadsl.FlowWithContext[WriteMessage, T, WriteResult, T, CompletionStage[Done]] =
+ settings: AmqpWriteSettings)
+ : akka.stream.javadsl.FlowWithContext[WriteMessage, T, WriteResult, T, CompletionStage[Done]] =
akka.stream.alpakka.amqp.scaladsl.AmqpFlowWithContext
.apply(settings)
.mapMaterializedValue(_.toJava)
@@ -36,8 +36,8 @@ object AmqpFlowWithContext {
* supposed to be used with another AMQP brokers.
*/
def createWithConfirm[T](
- settings: AmqpWriteSettings
- ): akka.stream.javadsl.FlowWithContext[WriteMessage, T, WriteResult, T, CompletionStage[Done]] =
+ settings: AmqpWriteSettings)
+ : akka.stream.javadsl.FlowWithContext[WriteMessage, T, WriteResult, T, CompletionStage[Done]] =
akka.stream.alpakka.amqp.scaladsl.AmqpFlowWithContext
.withConfirm(settings)
.mapMaterializedValue(_.toJava)
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpRpcFlow.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpRpcFlow.scala
index 91f25271..9f28e57d 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpRpcFlow.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpRpcFlow.scala
@@ -24,7 +24,7 @@ object AmqpRpcFlow {
* @param repliesPerMessage The number of responses that should be expected for each message placed on the queue.
*/
def createSimple(settings: AmqpWriteSettings,
- repliesPerMessage: Int): Flow[ByteString, ByteString, CompletionStage[String]] =
+ repliesPerMessage: Int): Flow[ByteString, ByteString, CompletionStage[String]] =
akka.stream.alpakka.amqp.scaladsl.AmqpRpcFlow
.simple(settings, repliesPerMessage)
.mapMaterializedValue(f => f.toJava)
@@ -36,7 +36,7 @@ object AmqpRpcFlow {
* before its read result is emitted downstream.
*/
def atMostOnceFlow(settings: AmqpWriteSettings,
- bufferSize: Int): Flow[WriteMessage, ReadResult, CompletionStage[String]] =
+ bufferSize: Int): Flow[WriteMessage, ReadResult, CompletionStage[String]] =
akka.stream.alpakka.amqp.scaladsl.AmqpRpcFlow
.atMostOnceFlow(settings, bufferSize)
.mapMaterializedValue(f => f.toJava)
@@ -48,8 +48,8 @@ object AmqpRpcFlow {
* before its read result is emitted downstream.
*/
def atMostOnceFlow(settings: AmqpWriteSettings,
- bufferSize: Int,
- repliesPerMessage: Int): Flow[WriteMessage, ReadResult, CompletionStage[String]] =
+ bufferSize: Int,
+ repliesPerMessage: Int): Flow[WriteMessage, ReadResult, CompletionStage[String]] =
akka.stream.alpakka.amqp.scaladsl.AmqpRpcFlow
.atMostOnceFlow(settings, bufferSize, repliesPerMessage)
.mapMaterializedValue(f => f.toJava)
@@ -69,8 +69,7 @@ object AmqpRpcFlow {
def committableFlow(
settings: AmqpWriteSettings,
bufferSize: Int,
- repliesPerMessage: Int = 1
- ): Flow[WriteMessage, CommittableReadResult, CompletionStage[String]] =
+ repliesPerMessage: Int = 1): Flow[WriteMessage, CommittableReadResult, CompletionStage[String]] =
akka.stream.alpakka.amqp.scaladsl.AmqpRpcFlow
.committableFlow(settings, bufferSize, repliesPerMessage)
.mapMaterializedValue(f => f.toJava)
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSink.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSink.scala
index 0e75fefe..52db94f8 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSink.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSink.scala
@@ -41,8 +41,7 @@ object AmqpSink {
* either normally or because of an amqp failure.
*/
def createReplyTo(
- settings: AmqpReplyToSinkSettings
- ): akka.stream.javadsl.Sink[WriteMessage, CompletionStage[Done]] =
+ settings: AmqpReplyToSinkSettings): akka.stream.javadsl.Sink[WriteMessage, CompletionStage[Done]] =
akka.stream.alpakka.amqp.scaladsl.AmqpSink.replyTo(settings).mapMaterializedValue(f => f.toJava).asJava
}
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSource.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSource.scala
index 30accf60..ec7d15a1 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSource.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/javadsl/AmqpSource.scala
@@ -5,7 +5,7 @@
package akka.stream.alpakka.amqp.javadsl
import akka.NotUsed
-import akka.stream.alpakka.amqp.{AmqpSourceSettings, ReadResult}
+import akka.stream.alpakka.amqp.{ AmqpSourceSettings, ReadResult }
import akka.stream.javadsl.Source
object AmqpSource {
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/model.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/model.scala
index b86a2594..b66e6f36 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/model.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/model.scala
@@ -13,8 +13,7 @@ import com.rabbitmq.client.Envelope
final class ReadResult private (
val bytes: ByteString,
val envelope: Envelope,
- val properties: BasicProperties
-) {
+ val properties: BasicProperties) {
override def toString: String =
s"ReadResult(bytes=$bytes, envelope=$envelope, properties=$properties)"
}
@@ -35,8 +34,7 @@ final class WriteMessage private (
val immediate: Boolean,
val mandatory: Boolean,
val properties: Option[BasicProperties] = None,
- val routingKey: Option[String] = None
-) {
+ val routingKey: Option[String] = None) {
def withImmediate(value: Boolean): WriteMessage =
if (value == immediate) this
@@ -53,9 +51,9 @@ final class WriteMessage private (
copy(routingKey = Some(routingKey))
private def copy(immediate: Boolean = immediate,
- mandatory: Boolean = mandatory,
- properties: Option[BasicProperties] = properties,
- routingKey: Option[String] = routingKey) =
+ mandatory: Boolean = mandatory,
+ properties: Option[BasicProperties] = properties,
+ routingKey: Option[String] = routingKey) =
new WriteMessage(bytes, immediate, mandatory, properties, routingKey)
override def toString: String =
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlow.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlow.scala
index 035e0ad1..6fb7235a 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlow.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlow.scala
@@ -5,8 +5,8 @@
package akka.stream.alpakka.amqp.scaladsl
import akka.stream.alpakka.amqp._
-import akka.stream.scaladsl.{Flow, Keep}
-import akka.{Done, NotUsed}
+import akka.stream.scaladsl.{ Flow, Keep }
+import akka.{ Done, NotUsed }
import scala.concurrent.Future
@@ -24,11 +24,9 @@ object AmqpFlow {
* @param settings `bufferSize` and `confirmationTimeout` properties are ignored by this connector
*/
def apply(
- settings: AmqpWriteSettings
- ): Flow[WriteMessage, WriteResult, Future[Done]] =
+ settings: AmqpWriteSettings): Flow[WriteMessage, WriteResult, Future[Done]] =
asFlowWithoutContext(
- Flow.fromGraph(new impl.AmqpSimpleFlowStage(settings))
- )
+ Flow.fromGraph(new impl.AmqpSimpleFlowStage(settings)))
/**
* Creates an `AmqpFlow` that accepts `WriteMessage` elements and emits `WriteResult`.
@@ -45,11 +43,9 @@ object AmqpFlow {
* either normally or because of an amqp failure.
*/
def withConfirm(
- settings: AmqpWriteSettings
- ): Flow[WriteMessage, WriteResult, Future[Done]] =
+ settings: AmqpWriteSettings): Flow[WriteMessage, WriteResult, Future[Done]] =
asFlowWithoutContext(
- Flow.fromGraph(new impl.AmqpAsyncFlowStage(settings))
- )
+ Flow.fromGraph(new impl.AmqpAsyncFlowStage(settings)))
/**
* Creates an `AmqpFlow` that accepts `WriteMessage` elements and emits `WriteResult`.
@@ -70,11 +66,9 @@ object AmqpFlow {
* supposed to be used with another AMQP brokers.
*/
def withConfirmUnordered(
- settings: AmqpWriteSettings
- ): Flow[WriteMessage, WriteResult, Future[Done]] =
+ settings: AmqpWriteSettings): Flow[WriteMessage, WriteResult, Future[Done]] =
asFlowWithoutContext(
- Flow.fromGraph(new impl.AmqpAsyncUnorderedFlowStage(settings))
- )
+ Flow.fromGraph(new impl.AmqpAsyncUnorderedFlowStage(settings)))
/**
* Variant of `AmqpFlow.withConfirmUnordered` with additional support for pass-through elements.
@@ -86,8 +80,7 @@ object AmqpFlow {
* supposed to be used with another AMQP brokers.
*/
def withConfirmAndPassThroughUnordered[T](
- settings: AmqpWriteSettings
- ): Flow[(WriteMessage, T), (WriteResult, T), Future[Done]] =
+ settings: AmqpWriteSettings): Flow[(WriteMessage, T), (WriteResult, T), Future[Done]] =
Flow.fromGraph(new impl.AmqpAsyncUnorderedFlowStage(settings))
private def asFlowWithoutContext(flow: Flow[(WriteMessage, NotUsed), (WriteResult, NotUsed), Future[Done]]) =
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowWithContext.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowWithContext.scala
index e9ff77cc..358845c5 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowWithContext.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowWithContext.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.amqp.scaladsl
import akka.Done
import akka.stream.alpakka.amqp._
-import akka.stream.scaladsl.{Flow, FlowWithContext}
+import akka.stream.scaladsl.{ Flow, FlowWithContext }
import scala.concurrent.Future
@@ -18,11 +18,9 @@ object AmqpFlowWithContext {
* @see [[AmqpFlow.apply]]
*/
def apply[T](
- settings: AmqpWriteSettings
- ): FlowWithContext[WriteMessage, T, WriteResult, T, Future[Done]] =
+ settings: AmqpWriteSettings): FlowWithContext[WriteMessage, T, WriteResult, T, Future[Done]] =
FlowWithContext.fromTuples(
- Flow.fromGraph(new impl.AmqpSimpleFlowStage[T](settings))
- )
+ Flow.fromGraph(new impl.AmqpSimpleFlowStage[T](settings)))
/**
* Creates a contextual variant of corresponding [[AmqpFlow]].
@@ -34,9 +32,7 @@ object AmqpFlowWithContext {
* supposed to be used with another AMQP brokers.
*/
def withConfirm[T](
- settings: AmqpWriteSettings
- ): FlowWithContext[WriteMessage, T, WriteResult, T, Future[Done]] =
+ settings: AmqpWriteSettings): FlowWithContext[WriteMessage, T, WriteResult, T, Future[Done]] =
FlowWithContext.fromTuples(
- Flow.fromGraph(new impl.AmqpAsyncFlowStage(settings))
- )
+ Flow.fromGraph(new impl.AmqpAsyncFlowStage(settings)))
}
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpRpcFlow.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpRpcFlow.scala
index f6636037..cb711f12 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpRpcFlow.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpRpcFlow.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.amqp.scaladsl
import akka.dispatch.ExecutionContexts
import akka.stream.alpakka.amqp._
-import akka.stream.scaladsl.{Flow, Keep}
+import akka.stream.scaladsl.{ Flow, Keep }
import akka.util.ByteString
import scala.concurrent.Future
@@ -35,8 +35,8 @@ object AmqpRpcFlow {
* before it is emitted downstream.
*/
def atMostOnceFlow(settings: AmqpWriteSettings,
- bufferSize: Int,
- repliesPerMessage: Int = 1): Flow[WriteMessage, ReadResult, Future[String]] =
+ bufferSize: Int,
+ repliesPerMessage: Int = 1): Flow[WriteMessage, ReadResult, Future[String]] =
committableFlow(settings, bufferSize, repliesPerMessage)
.mapAsync(1) { cm =>
cm.ack().map(_ => cm.message)(ExecutionContexts.parasitic)
@@ -54,8 +54,8 @@ object AmqpRpcFlow {
* Compared to auto-commit, this gives exact control over when a message is considered consumed.
*/
def committableFlow(settings: AmqpWriteSettings,
- bufferSize: Int,
- repliesPerMessage: Int = 1): Flow[WriteMessage, CommittableReadResult, Future[String]] =
+ bufferSize: Int,
+ repliesPerMessage: Int = 1): Flow[WriteMessage, CommittableReadResult, Future[String]] =
Flow.fromGraph(new impl.AmqpRpcFlowStage(settings, bufferSize, repliesPerMessage))
}
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSink.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSink.scala
index 9a58df1f..5f874325 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSink.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSink.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.amqp.scaladsl
import akka.Done
import akka.stream.alpakka.amqp._
-import akka.stream.scaladsl.{Keep, Sink}
+import akka.stream.scaladsl.{ Keep, Sink }
import akka.util.ByteString
import scala.concurrent.Future
diff --git a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSource.scala b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSource.scala
index 302127f1..967ecc86 100644
--- a/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSource.scala
+++ b/amqp/src/main/scala/akka/stream/alpakka/amqp/scaladsl/AmqpSource.scala
@@ -7,7 +7,7 @@ package akka.stream.alpakka.amqp.scaladsl
import akka.NotUsed
import akka.dispatch.ExecutionContexts
import akka.stream.alpakka.amqp.impl
-import akka.stream.alpakka.amqp.{AmqpSourceSettings, ReadResult}
+import akka.stream.alpakka.amqp.{ AmqpSourceSettings, ReadResult }
import akka.stream.scaladsl.Source
object AmqpSource {
diff --git a/amqp/src/test/scala/akka/stream/alpakka/amqp/AmqpProxyConnection.scala b/amqp/src/test/scala/akka/stream/alpakka/amqp/AmqpProxyConnection.scala
index abedac8b..93e876f8 100644
--- a/amqp/src/test/scala/akka/stream/alpakka/amqp/AmqpProxyConnection.scala
+++ b/amqp/src/test/scala/akka/stream/alpakka/amqp/AmqpProxyConnection.scala
@@ -59,7 +59,7 @@ class AmqpProxyConnection(protected val delegate: Connection) extends Connection
override def addBlockedListener(blockedListener: BlockedListener): Unit = delegate.addBlockedListener(blockedListener)
override def addBlockedListener(blockedCallback: BlockedCallback,
- unblockedCallback: UnblockedCallback): BlockedListener =
+ unblockedCallback: UnblockedCallback): BlockedListener =
delegate.addBlockedListener(blockedCallback, unblockedCallback)
override def removeBlockedListener(blockedListener: BlockedListener): Boolean =
diff --git a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectionProvidersSpec.scala b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectionProvidersSpec.scala
index 00a042de..509ec1a6 100644
--- a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectionProvidersSpec.scala
+++ b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectionProvidersSpec.scala
@@ -15,7 +15,7 @@ class AmqpConnectionProvidersSpec extends AmqpSpec {
val connectionProvider = AmqpLocalConnectionProvider
val connection1 = connectionProvider.get
val connection2 = connectionProvider.get
- connection1 should not equal connection2
+ (connection1 should not).equal(connection2)
connectionProvider.release(connection1)
connectionProvider.release(connection2)
}
@@ -31,7 +31,7 @@ class AmqpConnectionProvidersSpec extends AmqpSpec {
val connectionProvider = AmqpUriConnectionProvider("amqp://localhost:5672")
val connection1 = connectionProvider.get
val connection2 = connectionProvider.get
- connection1 should not equal connection2
+ (connection1 should not).equal(connection2)
connectionProvider.release(connection1)
connectionProvider.release(connection2)
}
@@ -47,7 +47,7 @@ class AmqpConnectionProvidersSpec extends AmqpSpec {
val connectionProvider = AmqpDetailsConnectionProvider("localhost", 5672)
val connection1 = connectionProvider.get
val connection2 = connectionProvider.get
- connection1 should not equal connection2
+ (connection1 should not).equal(connection2)
connectionProvider.release(connection1)
connectionProvider.release(connection2)
}
@@ -65,7 +65,7 @@ class AmqpConnectionProvidersSpec extends AmqpSpec {
AmqpConnectionFactoryConnectionProvider(connectionFactory).withHostAndPort("localhost", 5672)
val connection1 = connectionProvider.get
val connection2 = connectionProvider.get
- connection1 should not equal connection2
+ (connection1 should not).equal(connection2)
connectionProvider.release(connection1)
connectionProvider.release(connection2)
}
diff --git a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectorsSpec.scala b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectorsSpec.scala
index a7b508a4..72933fe6 100644
--- a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectorsSpec.scala
+++ b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpConnectorsSpec.scala
@@ -9,10 +9,10 @@ import java.net.ConnectException
import akka.Done
import akka.stream._
import akka.stream.alpakka.amqp._
-import akka.stream.scaladsl.{GraphDSL, Keep, Merge, Sink, Source}
+import akka.stream.scaladsl.{ GraphDSL, Keep, Merge, Sink, Source }
import akka.stream.testkit.scaladsl.TestSink
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
-import akka.stream.testkit.{TestPublisher, TestSubscriber}
+import akka.stream.testkit.{ TestPublisher, TestSubscriber }
import akka.util.ByteString
import com.rabbitmq.client.AMQP.BasicProperties
import com.rabbitmq.client.AuthenticationFailureException
@@ -41,8 +41,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
val result = Source(input).map(s => ByteString(s)).runWith(amqpSink)
@@ -61,8 +60,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
val result = Source(input).map(s => ByteString(s)).runWith(amqpSink)
@@ -77,13 +75,11 @@ class AmqpConnectorsSpec extends AmqpSpec {
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
.withDeclaration(queueDeclaration),
- 2
- )
+ 2)
val amqpSource = AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName),
- bufferSize = 1
- )
+ bufferSize = 1)
val input = Vector("one", "two", "three", "four", "five")
val (rpcQueueF, probe) =
@@ -91,16 +87,14 @@ class AmqpConnectorsSpec extends AmqpSpec {
rpcQueueF.futureValue
val amqpSink = AmqpSink.replyTo(
- AmqpReplyToSinkSettings(connectionProvider)
- )
+ AmqpReplyToSinkSettings(connectionProvider))
val sourceToSink = amqpSource
.viaMat(KillSwitches.single)(Keep.right)
.mapConcat { b =>
List(
WriteMessage(b.bytes.concat(ByteString("a"))).withProperties(b.properties),
- WriteMessage(b.bytes.concat(ByteString("aa"))).withProperties(b.properties)
- )
+ WriteMessage(b.bytes.concat(ByteString("aa"))).withProperties(b.properties))
}
.to(amqpSink)
.run()
@@ -139,8 +133,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
Source
.single(outgoingMessage)
.toMat(AmqpSink.replyTo(AmqpReplyToSinkSettings(connectionProvider).withFailIfReplyToMissing(true)))(
- Keep.right
- )
+ Keep.right)
.run()
.futureValue
}
@@ -157,8 +150,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
Source
.single(outgoingMessageWithEmptyReplyTo)
.toMat(AmqpSink.replyTo(AmqpReplyToSinkSettings(connectionProvider).withFailIfReplyToMissing(false)))(
- Keep.right
- )
+ Keep.right)
.run()
.futureValue shouldBe Done
@@ -170,8 +162,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
Source(input).map(s => ByteString(s)).runWith(amqpSink)
@@ -185,9 +176,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName)
.withDeclaration(queueDeclaration),
- bufferSize = 1
- )
- )
+ bufferSize = 1))
source.out ~> merge.in(n)
}
@@ -204,14 +193,12 @@ class AmqpConnectorsSpec extends AmqpSpec {
val queueDeclaration = QueueDeclaration(queueName)
val amqpSource = AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName).withDeclaration(queueDeclaration),
- bufferSize = 2
- )
+ bufferSize = 2)
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val publisher = TestPublisher.probe[ByteString]()
val subscriber = TestSubscriber.probe[ReadResult]()
@@ -260,13 +247,11 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionSettings)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val amqpSource = AmqpSource.committableSource(
NamedQueueSourceSettings(connectionSettings, queueName).withDeclaration(queueDeclaration),
- bufferSize = 10
- )
+ bufferSize = 10)
val input = Vector("one", "two", "three", "four", "five")
Source(input).map(s => ByteString(s)).runWith(amqpSink).futureValue shouldEqual Done
@@ -287,15 +272,13 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
Source(input).map(s => ByteString(s)).runWith(amqpSink).futureValue shouldEqual Done
val amqpSource = AmqpSource.committableSource(
NamedQueueSourceSettings(connectionProvider, queueName).withDeclaration(queueDeclaration),
- bufferSize = 10
- )
+ bufferSize = 10)
val result1 = amqpSource
.mapAsync(1)(cm => cm.nack(requeue = false).map(_ => cm))
@@ -326,8 +309,7 @@ class AmqpConnectorsSpec extends AmqpSpec {
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
.withDeclaration(queueDeclaration),
- bufferSize = 10
- )
+ bufferSize = 10)
val (rpcQueueF, probe) =
Source(input)
.map(s => ByteString(s))
@@ -339,13 +321,11 @@ class AmqpConnectorsSpec extends AmqpSpec {
rpcQueueF.futureValue
val amqpSink = AmqpSink.replyTo(
- AmqpReplyToSinkSettings(connectionProvider)
- )
+ AmqpReplyToSinkSettings(connectionProvider))
val amqpSource = AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName),
- bufferSize = 1
- )
+ bufferSize = 1)
val sourceToSink = amqpSource
.viaMat(KillSwitches.single)(Keep.right)
.map(b => WriteMessage(b.bytes).withProperties(b.properties))
@@ -368,14 +348,12 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink(
AmqpWriteSettings(connectionProvider)
.withExchange(exchangeName)
- .withDeclarations(immutable.Seq(exchangeDeclaration, queueDeclaration, bindingDeclaration))
- )
+ .withDeclarations(immutable.Seq(exchangeDeclaration, queueDeclaration, bindingDeclaration)))
val amqpSource = AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName)
.withDeclarations(immutable.Seq(exchangeDeclaration, queueDeclaration, bindingDeclaration)),
- bufferSize = 10
- )
+ bufferSize = 10)
val input = Vector("one", "two", "three", "four", "five")
val routingKeys = input.map(s => getRoutingKey(s))
@@ -403,16 +381,14 @@ class AmqpConnectorsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val amqpSource = AmqpSource
.committableSource(
NamedQueueSourceSettings(connectionProvider, queueName)
.withAckRequired(false)
.withDeclaration(queueDeclaration),
- bufferSize = 10
- )
+ bufferSize = 10)
val input = Vector("one", "two", "three", "four", "five")
Source(input).map(s => ByteString(s)).runWith(amqpSink).futureValue shouldEqual Done
diff --git a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowSpec.scala b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowSpec.scala
index 157010c6..d84d9992 100644
--- a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowSpec.scala
+++ b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpFlowSpec.scala
@@ -6,15 +6,15 @@ package akka.stream.alpakka.amqp.scaladsl
import akka.Done
import akka.stream.alpakka.amqp._
-import akka.stream.scaladsl.{Flow, FlowWithContext, Keep, Sink, Source}
+import akka.stream.scaladsl.{ Flow, FlowWithContext, Keep, Sink, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
-import akka.stream.testkit.scaladsl.{TestSink, TestSource}
+import akka.stream.testkit.scaladsl.{ TestSink, TestSource }
import akka.util.ByteString
import com.rabbitmq.client.AMQP.BasicProperties
import com.rabbitmq.client.ConfirmCallback
import org.mockito.ArgumentMatchers._
import org.mockito.Mockito._
-import org.mockito.{ArgumentCaptor, Mockito}
+import org.mockito.{ ArgumentCaptor, Mockito }
import org.scalatest.BeforeAndAfterEach
import scala.concurrent.Future
@@ -91,8 +91,7 @@ class AmqpFlowSpec extends AmqpSpec with AmqpMocking with BeforeAndAfterEach {
val channelError = new RuntimeException("channel error")
when(
- connectionMock.createChannel()
- ).thenThrow(channelError)
+ connectionMock.createChannel()).thenThrow(channelError)
val completion =
Source
@@ -175,8 +174,7 @@ class AmqpFlowSpec extends AmqpSpec with AmqpMocking with BeforeAndAfterEach {
(WriteResult.rejected, input(3)),
(WriteResult.rejected, input(4)),
(WriteResult.rejected, input(5)),
- (WriteResult.rejected, input(6))
- )
+ (WriteResult.rejected, input(6)))
messages should contain theSameElementsInOrderAs expectedResult
completion.futureValue shouldBe an[Done]
@@ -262,8 +260,7 @@ class AmqpFlowSpec extends AmqpSpec with AmqpMocking with BeforeAndAfterEach {
(WriteResult.rejected, input(6)),
(WriteResult.rejected, input(3)),
(WriteResult.rejected, input(4)),
- (WriteResult.rejected, input(5))
- )
+ (WriteResult.rejected, input(5)))
messages should contain theSameElementsInOrderAs expectedResult
completion.futureValue shouldBe an[Done]
@@ -327,8 +324,8 @@ class AmqpFlowSpec extends AmqpSpec with AmqpMocking with BeforeAndAfterEach {
when(
channelMock
- .basicPublish(any[String], any[String], any[Boolean], any[Boolean], any[BasicProperties], any[Array[Byte]])
- ).thenThrow(publicationError)
+ .basicPublish(any[String], any[String], any[Boolean], any[Boolean], any[BasicProperties],
+ any[Array[Byte]])).thenThrow(publicationError)
val completion =
Source
@@ -419,8 +416,7 @@ class AmqpFlowSpec extends AmqpSpec with AmqpMocking with BeforeAndAfterEach {
WriteResult.confirmed,
WriteResult.confirmed,
WriteResult.rejected,
- WriteResult.rejected
- )
+ WriteResult.rejected)
messages should contain theSameElementsAs expectedResult
completion.futureValue shouldBe an[Done]
diff --git a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpGraphStageLogicConnectionShutdownSpec.scala b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpGraphStageLogicConnectionShutdownSpec.scala
index a57fe40c..65eb63dc 100644
--- a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpGraphStageLogicConnectionShutdownSpec.scala
+++ b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpGraphStageLogicConnectionShutdownSpec.scala
@@ -19,7 +19,7 @@ import akka.stream.alpakka.amqp.{
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
import akka.stream.scaladsl.Source
import akka.util.ByteString
-import com.rabbitmq.client.{AddressResolver, Connection, ConnectionFactory, ShutdownListener}
+import com.rabbitmq.client.{ AddressResolver, Connection, ConnectionFactory, ShutdownListener }
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.BeforeAndAfterEach
@@ -82,8 +82,7 @@ class AmqpGraphStageLogicConnectionShutdownSpec
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(reusableConnectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four")
diff --git a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpMocking.scala b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpMocking.scala
index 2e1d3a93..6138ff9f 100644
--- a/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpMocking.scala
+++ b/amqp/src/test/scala/akka/stream/alpakka/amqp/scaladsl/AmqpMocking.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.amqp.scaladsl
-import com.rabbitmq.client.{Address, Channel, ConfirmCallback, ConfirmListener, Connection, ConnectionFactory}
+import com.rabbitmq.client.{ Address, Channel, ConfirmCallback, ConfirmListener, Connection, ConnectionFactory }
import org.mockito.ArgumentMatchers._
import org.scalatestplus.mockito.MockitoSugar
import org.mockito.Mockito.when
diff --git a/amqp/src/test/scala/docs/scaladsl/AmqpDocsSpec.scala b/amqp/src/test/scala/docs/scaladsl/AmqpDocsSpec.scala
index 8447cd05..b0d7d644 100644
--- a/amqp/src/test/scala/docs/scaladsl/AmqpDocsSpec.scala
+++ b/amqp/src/test/scala/docs/scaladsl/AmqpDocsSpec.scala
@@ -4,18 +4,18 @@
package docs.scaladsl
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import akka.stream.KillSwitches
import akka.stream.alpakka.amqp._
-import akka.stream.alpakka.amqp.scaladsl.{AmqpFlow, AmqpRpcFlow, AmqpSink, AmqpSource, CommittableReadResult}
-import akka.stream.scaladsl.{Flow, Keep, Sink, Source}
+import akka.stream.alpakka.amqp.scaladsl.{ AmqpFlow, AmqpRpcFlow, AmqpSink, AmqpSource, CommittableReadResult }
+import akka.stream.scaladsl.{ Flow, Keep, Sink, Source }
import akka.stream.testkit.TestSubscriber
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.stream.testkit.scaladsl.TestSink
import akka.util.ByteString
import scala.concurrent.duration._
-import scala.concurrent.{Await, Future, Promise}
+import scala.concurrent.{ Await, Future, Promise }
import scala.collection.immutable
/**
@@ -38,41 +38,39 @@ class AmqpDocsSpec extends AmqpSpec {
val connectionProvider =
AmqpDetailsConnectionProvider("invalid", 5673).withHostsAndPorts(immutable.Seq("localhost" -> 5672))
- //#queue-declaration
+ // #queue-declaration
val queueName = "amqp-conn-it-spec-simple-queue-" + System.currentTimeMillis()
val queueDeclaration = QueueDeclaration(queueName)
- //#queue-declaration
+ // #queue-declaration
- //#create-sink
+ // #create-sink
val amqpSink: Sink[ByteString, Future[Done]] =
AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
val writing: Future[Done] =
Source(input)
.map(s => ByteString(s))
.runWith(amqpSink)
- //#create-sink
+ // #create-sink
writing.futureValue shouldEqual Done
- //#create-source
+ // #create-source
val amqpSource: Source[ReadResult, NotUsed] =
AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName)
.withDeclaration(queueDeclaration)
.withAckRequired(false),
- bufferSize = 10
- )
+ bufferSize = 10)
val result: Future[immutable.Seq[ReadResult]] =
amqpSource
.take(input.size)
.runWith(Sink.seq)
- //#create-source
+ // #create-source
result.futureValue.map(_.bytes.utf8String) shouldEqual input
}
@@ -84,27 +82,24 @@ class AmqpDocsSpec extends AmqpSpec {
val amqpSource = AmqpSource.atMostOnceSource(
NamedQueueSourceSettings(connectionProvider, queueName),
- bufferSize = 1
- )
+ bufferSize = 1)
val input = Vector("one", "two", "three", "four", "five")
- //#create-rpc-flow
+ // #create-rpc-flow
val amqpRpcFlow = AmqpRpcFlow.simple(
- AmqpWriteSettings(connectionProvider).withRoutingKey(queueName).withDeclaration(queueDeclaration)
- )
+ AmqpWriteSettings(connectionProvider).withRoutingKey(queueName).withDeclaration(queueDeclaration))
val (rpcQueueF: Future[String], probe: TestSubscriber.Probe[ByteString]) = Source(input)
.map(s => ByteString(s))
.viaMat(amqpRpcFlow)(Keep.right)
.toMat(TestSink.probe)(Keep.both)
.run()
- //#create-rpc-flow
+ // #create-rpc-flow
rpcQueueF.futureValue
val amqpSink = AmqpSink.replyTo(
- AmqpReplyToSinkSettings(connectionProvider)
- )
+ AmqpReplyToSinkSettings(connectionProvider))
val sourceToSink = amqpSource
.viaMat(KillSwitches.single)(Keep.right)
@@ -121,20 +116,19 @@ class AmqpDocsSpec extends AmqpSpec {
// and then one queue for each source which subscribes to the
// exchange - all this described by the declarations
- //#exchange-declaration
+ // #exchange-declaration
val exchangeName = "amqp-conn-it-spec-pub-sub-" + System.currentTimeMillis()
val exchangeDeclaration = ExchangeDeclaration(exchangeName, "fanout")
- //#exchange-declaration
+ // #exchange-declaration
- //#create-exchange-sink
+ // #create-exchange-sink
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withExchange(exchangeName)
- .withDeclaration(exchangeDeclaration)
- )
- //#create-exchange-sink
+ .withDeclaration(exchangeDeclaration))
+ // #create-exchange-sink
- //#create-exchange-source
+ // #create-exchange-source
val fanoutSize = 4
val mergedSources = (0 until fanoutSize).foldLeft(Source.empty[(Int, String)]) {
@@ -144,14 +138,11 @@ class AmqpDocsSpec extends AmqpSpec {
.atMostOnceSource(
TemporaryQueueSourceSettings(
connectionProvider,
- exchangeName
- ).withDeclaration(exchangeDeclaration),
- bufferSize = 1
- )
- .map(msg => (fanoutBranch, msg.bytes.utf8String))
- )
+ exchangeName).withDeclaration(exchangeDeclaration),
+ bufferSize = 1)
+ .map(msg => (fanoutBranch, msg.bytes.utf8String)))
}
- //#create-exchange-source
+ // #create-exchange-source
val completion = Promise[Done]()
val mergingFlow = mergedSources
@@ -164,8 +155,7 @@ class AmqpDocsSpec extends AmqpSpec {
.run()
system.scheduler.scheduleOnce(5.seconds)(
- completion.tryFailure(new Error("Did not get at least one element from every fanout branch"))
- )
+ completion.tryFailure(new Error("Did not get at least one element from every fanout branch")))
val dataSender = Source
.repeat("stuff")
@@ -186,25 +176,23 @@ class AmqpDocsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
Source(input).map(s => ByteString(s)).runWith(amqpSink).futureValue shouldEqual Done
- //#create-source-withoutautoack
+ // #create-source-withoutautoack
val amqpSource = AmqpSource.committableSource(
NamedQueueSourceSettings(connectionProvider, queueName)
.withDeclaration(queueDeclaration),
- bufferSize = 10
- )
+ bufferSize = 10)
val result: Future[immutable.Seq[ReadResult]] = amqpSource
.mapAsync(1)(businessLogic)
.mapAsync(1)(cm => cm.ack().map(_ => cm.message))
.take(input.size)
.runWith(Sink.seq)
- //#create-source-withoutautoack
+ // #create-source-withoutautoack
result.futureValue.map(_.bytes.utf8String) shouldEqual input
}
@@ -217,25 +205,23 @@ class AmqpDocsSpec extends AmqpSpec {
val amqpSink = AmqpSink.simple(
AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
- .withDeclaration(queueDeclaration)
- )
+ .withDeclaration(queueDeclaration))
val input = Vector("one", "two", "three", "four", "five")
Source(input).map(s => ByteString(s)).runWith(amqpSink).futureValue shouldEqual Done
val amqpSource = AmqpSource.committableSource(
NamedQueueSourceSettings(connectionProvider, queueName).withDeclaration(queueDeclaration),
- bufferSize = 10
- )
+ bufferSize = 10)
- //#create-source-withoutautoack
+ // #create-source-withoutautoack
val nackedResults: Future[immutable.Seq[ReadResult]] = amqpSource
.mapAsync(1)(businessLogic)
.take(input.size)
.mapAsync(1)(cm => cm.nack(multiple = false, requeue = true).map(_ => cm.message))
.runWith(Sink.seq)
- //#create-source-withoutautoack
+ // #create-source-withoutautoack
Await.ready(nackedResults, 3.seconds)
@@ -251,7 +237,7 @@ class AmqpDocsSpec extends AmqpSpec {
val queueName = "amqp-conn-it-spec-flow-" + System.currentTimeMillis()
val queueDeclaration = QueueDeclaration(queueName)
- //#create-flow
+ // #create-flow
val settings = AmqpWriteSettings(connectionProvider)
.withRoutingKey(queueName)
.withDeclaration(queueDeclaration)
@@ -267,7 +253,7 @@ class AmqpDocsSpec extends AmqpSpec {
.map(message => WriteMessage(ByteString(message)))
.via(amqpFlow)
.runWith(Sink.seq)
- //#create-flow
+ // #create-flow
result.futureValue should contain theSameElementsAs input.map(_ => WriteResult.confirmed)
}
diff --git a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetFlow.scala b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetFlow.scala
index 03e7cba9..901eaa2f 100644
--- a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetFlow.scala
+++ b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetFlow.scala
@@ -5,7 +5,7 @@
package akka.stream.alpakka.avroparquet.impl
import akka.annotation.InternalApi
import akka.stream._
-import akka.stream.stage.{GraphStage, GraphStageLogic, InHandler, OutHandler}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, InHandler, OutHandler }
import org.apache.avro.generic.GenericRecord
import org.apache.parquet.hadoop.ParquetWriter
@@ -30,7 +30,7 @@ private[avroparquet] class AvroParquetFlow[T <: GenericRecord](writer: ParquetWr
new InHandler {
override def onUpstreamFinish(): Unit =
- //super.onUpstreamFinish()
+ // super.onUpstreamFinish()
completeStage()
override def onUpstreamFailure(ex: Throwable): Unit = {
@@ -43,13 +43,13 @@ private[avroparquet] class AvroParquetFlow[T <: GenericRecord](writer: ParquetWr
writer.write(obtainedValue)
push(out, obtainedValue)
}
- }
- )
+ })
- setHandler(out, new OutHandler {
- override def onPull(): Unit =
- pull(in)
- })
+ setHandler(out,
+ new OutHandler {
+ override def onPull(): Unit =
+ pull(in)
+ })
override def postStop(): Unit = writer.close()
}
diff --git a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetSource.scala b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetSource.scala
index cc044412..331a13c0 100644
--- a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetSource.scala
+++ b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/impl/AvroParquetSource.scala
@@ -4,8 +4,8 @@
package akka.stream.alpakka.avroparquet.impl
import akka.annotation.InternalApi
-import akka.stream.{ActorAttributes, Attributes, Outlet, SourceShape}
-import akka.stream.stage.{GraphStage, GraphStageLogic, OutHandler}
+import akka.stream.{ ActorAttributes, Attributes, Outlet, SourceShape }
+import akka.stream.stage.{ GraphStage, GraphStageLogic, OutHandler }
import org.apache.avro.generic.GenericRecord
import org.apache.parquet.hadoop.ParquetReader
@@ -38,8 +38,7 @@ private[avroparquet] class AvroParquetSource[T <: GenericRecord](reader: Parquet
complete(out)
}(push(out, _))
}
- }
- )
+ })
override def postStop(): Unit = reader.close()
diff --git a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/javadsl/AvroParquetSink.scala b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/javadsl/AvroParquetSink.scala
index 46b4b5a5..8c594fd1 100644
--- a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/javadsl/AvroParquetSink.scala
+++ b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/javadsl/AvroParquetSink.scala
@@ -5,8 +5,8 @@
package akka.stream.alpakka.avroparquet.javadsl
import java.util.concurrent.CompletionStage
-import akka.stream.javadsl.{Flow, Keep, Sink}
-import akka.{Done, NotUsed}
+import akka.stream.javadsl.{ Flow, Keep, Sink }
+import akka.{ Done, NotUsed }
import org.apache.avro.generic.GenericRecord
import org.apache.parquet.hadoop.ParquetWriter
diff --git a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/scaladsl/AvroParquetSink.scala b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/scaladsl/AvroParquetSink.scala
index 1ab2cbb3..c7fde124 100644
--- a/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/scaladsl/AvroParquetSink.scala
+++ b/avroparquet/src/main/scala/akka/stream/alpakka/avroparquet/scaladsl/AvroParquetSink.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.avroparquet.scaladsl
import akka.Done
-import akka.stream.scaladsl.{Flow, Keep, Sink}
+import akka.stream.scaladsl.{ Flow, Keep, Sink }
import org.apache.avro.generic.GenericRecord
import org.apache.parquet.hadoop.ParquetWriter
diff --git a/avroparquet/src/test/scala/docs/scaladsl/AbstractAvroParquet.scala b/avroparquet/src/test/scala/docs/scaladsl/AbstractAvroParquet.scala
index b55cf80e..593abc19 100644
--- a/avroparquet/src/test/scala/docs/scaladsl/AbstractAvroParquet.scala
+++ b/avroparquet/src/test/scala/docs/scaladsl/AbstractAvroParquet.scala
@@ -9,14 +9,14 @@ import java.io.File
import akka.testkit.TestKit
import com.sksamuel.avro4s.RecordFormat
import org.apache.avro.Schema
-import org.apache.avro.generic.{GenericRecord, GenericRecordBuilder}
+import org.apache.avro.generic.{ GenericRecord, GenericRecordBuilder }
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.fs.Path
-import org.apache.parquet.avro.{AvroParquetReader, AvroParquetWriter, AvroReadSupport}
-import org.apache.parquet.hadoop.{ParquetReader, ParquetWriter}
+import org.apache.parquet.avro.{ AvroParquetReader, AvroParquetWriter, AvroReadSupport }
+import org.apache.parquet.hadoop.{ ParquetReader, ParquetWriter }
import org.apache.parquet.hadoop.util.HadoopInputFile
import org.scalacheck.Gen
-import org.scalatest.{BeforeAndAfterAll, Suite}
+import org.scalatest.{ BeforeAndAfterAll, Suite }
import scala.reflect.io.Directory
import scala.util.Random
@@ -27,8 +27,7 @@ trait AbstractAvroParquet extends BeforeAndAfterAll {
case class Document(id: String, body: String)
val schema: Schema = new Schema.Parser().parse(
- "{\"type\":\"record\",\"name\":\"Document\",\"fields\":[{\"name\":\"id\",\"type\":\"string\"},{\"name\":\"body\",\"type\":\"string\"}]}"
- )
+ "{\"type\":\"record\",\"name\":\"Document\",\"fields\":[{\"name\":\"id\",\"type\":\"string\"},{\"name\":\"body\",\"type\":\"string\"}]}")
val genDocument: Gen[Document] =
Gen.oneOf(Seq(Document(id = Gen.alphaStr.sample.get, body = Gen.alphaLowerStr.sample.get)))
diff --git a/avroparquet/src/test/scala/docs/scaladsl/AvroParquetFlowSpec.scala b/avroparquet/src/test/scala/docs/scaladsl/AvroParquetFlowSpec.scala
index 7023280f..20c39ed7 100644
--- a/avroparquet/src/test/scala/docs/scaladsl/AvroParquetFlowSpec.scala
+++ b/avroparquet/src/test/scala/docs/scaladsl/AvroParquetFlowSpec.scala
@@ -7,7 +7,7 @@ package docs.scaladsl
import akka.NotUsed
import akka.actor.ActorSystem
import akka.stream.alpakka.avroparquet.scaladsl.AvroParquetFlow
-import akka.stream.scaladsl.{Flow, Sink, Source}
+import akka.stream.scaladsl.{ Flow, Sink, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.testkit.TestKit
import com.sksamuel.avro4s.Record
@@ -29,7 +29,7 @@ class AvroParquetFlowSpec
"Parquet Flow" should {
"insert avro records in parquet from `GenericRecord`" in assertAllStagesStopped {
- //given
+ // given
val n: Int = 2
val file: String = genFinalFile.sample.get
// #init-flow
@@ -38,7 +38,7 @@ class AvroParquetFlowSpec
= genDocuments(n).sample.get.map(docToGenericRecord)
val writer: ParquetWriter[GenericRecord] = parquetWriter(file, conf, schema)
- //when
+ // when
// #init-flow
val source: Source[GenericRecord, NotUsed] = Source(records)
val avroParquet: Flow[GenericRecord, GenericRecord, NotUsed] = AvroParquetFlow(writer)
@@ -50,27 +50,27 @@ class AvroParquetFlowSpec
result.futureValue
- //then
+ // then
val parquetContent: List[GenericRecord] = fromParquet(file, conf)
parquetContent.length shouldEqual n
parquetContent should contain theSameElementsAs records
}
"insert avro records in parquet from a subtype of `GenericRecord`" in assertAllStagesStopped {
- //given
+ // given
val n: Int = 2
val file: String = genFinalFile.sample.get
val documents: List[Document] = genDocuments(n).sample.get
val avroDocuments: List[Record] = documents.map(format.to(_))
val writer: ParquetWriter[Record] = parquetWriter[Record](file, conf, schema)
- //when
+ // when
Source(avroDocuments)
.via(AvroParquetFlow[Record](writer))
.runWith(Sink.seq)
.futureValue
- //then
+ // then
val parquetContent: List[GenericRecord] = fromParquet(file, conf)
parquetContent.length shouldEqual n
parquetContent.map(format.from(_)) should contain theSameElementsAs documents
diff --git a/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSinkSpec.scala b/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSinkSpec.scala
index 4b9eae51..8a2085a9 100644
--- a/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSinkSpec.scala
+++ b/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSinkSpec.scala
@@ -4,13 +4,13 @@
package docs.scaladsl
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import akka.actor.ActorSystem
import akka.stream.alpakka.avroparquet.scaladsl.AvroParquetSink
import akka.stream.scaladsl.Source
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.testkit.TestKit
-import com.sksamuel.avro4s.{Record, RecordFormat}
+import com.sksamuel.avro4s.{ Record, RecordFormat }
import org.scalatest.concurrent.ScalaFutures
import org.apache.avro.generic.GenericRecord
import org.apache.parquet.hadoop.ParquetWriter
@@ -31,23 +31,23 @@ class AvroParquetSinkSpec
"Parquet Sink" should {
"create new parquet file from `GenericRecords`" in assertAllStagesStopped {
- //given
+ // given
val n: Int = 3
val file: String = genFinalFile.sample.get
val records: List[GenericRecord] = genDocuments(n).sample.get.map(docToGenericRecord)
Source(records).runWith(AvroParquetSink(parquetWriter(file, conf, schema))).futureValue
- //when
+ // when
val parquetContent: List[GenericRecord] = fromParquet(file, conf)
- //then
+ // then
parquetContent.length shouldEqual n
parquetContent should contain theSameElementsAs records
}
"create new parquet file from any subtype of `GenericRecord` " in assertAllStagesStopped {
- //given
+ // given
val n: Int = 3
val file: String = genFinalFile.sample.get
val documents: List[Document] = genDocuments(n).sample.get
@@ -60,10 +60,10 @@ class AvroParquetSinkSpec
// #init-sink
result.futureValue shouldBe Done
- //when
+ // when
val parquetContent: List[GenericRecord] = fromParquet(file, conf)
- //then
+ // then
parquetContent.length shouldEqual n
parquetContent.map(format.from(_)) should contain theSameElementsAs documents
}
diff --git a/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSourceSpec.scala b/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSourceSpec.scala
index 15fe5bbe..e372e659 100644
--- a/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSourceSpec.scala
+++ b/avroparquet/src/test/scala/docs/scaladsl/AvroParquetSourceSpec.scala
@@ -6,8 +6,8 @@ package docs.scaladsl
import akka.NotUsed
import akka.actor.ActorSystem
-import akka.stream.alpakka.avroparquet.scaladsl.{AvroParquetSink, AvroParquetSource}
-import akka.stream.scaladsl.{Keep, Source}
+import akka.stream.alpakka.avroparquet.scaladsl.{ AvroParquetSink, AvroParquetSource }
+import akka.stream.scaladsl.{ Keep, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.stream.testkit.scaladsl.TestSink
import akka.testkit.TestKit
@@ -32,7 +32,7 @@ class AvroParquetSourceSpec
"AvroParquetSource" should {
"read from parquet file as a `GenericRecord` type" in assertAllStagesStopped {
- //given
+ // given
val n: Int = 4
val file: String = genFinalFile.sample.get
val records: List[GenericRecord] = genDocuments(n).sample.get.map(docToGenericRecord)
@@ -41,21 +41,21 @@ class AvroParquetSourceSpec
.run()
.futureValue
- //when
+ // when
val reader: ParquetReader[GenericRecord] = parquetReader(file, conf)
// #init-source
val source: Source[GenericRecord, NotUsed] = AvroParquetSource(reader)
// #init-source
val sink = source.runWith(TestSink.probe)
- //then
+ // then
val result: Seq[GenericRecord] = sink.toStrict(3.seconds)
result.length shouldEqual n
result should contain theSameElementsAs records
}
"read from parquet file as any subtype of `GenericRecord` " in assertAllStagesStopped {
- //given
+ // given
val n: Int = 4
val file: String = genFinalFile.sample.get
val documents: List[Document] = genDocuments(n).sample.get
@@ -65,14 +65,14 @@ class AvroParquetSourceSpec
.run()
.futureValue
- //when
+ // when
val reader: ParquetReader[GenericRecord] = parquetReader(file, conf)
// #init-source
val source: Source[GenericRecord, NotUsed] = AvroParquetSource(reader)
// #init-source
val sink = source.runWith(TestSink.probe)
- //then
+ // then
val result: Seq[GenericRecord] = sink.toStrict(3.seconds)
result.length shouldEqual n
result.map(format.from(_)) should contain theSameElementsAs documents
diff --git a/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishSettings.scala b/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishSettings.scala
index f92d2725..eee63cb6 100644
--- a/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishSettings.scala
+++ b/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishSettings.scala
@@ -12,7 +12,7 @@ package akka.stream.alpakka.aws.eventbridge
* put into the stream. Use concurrency 1 for having control over failures.
*
* @param concurrency maps to parallelism in in async stream operations
- * */
+ */
final class EventBridgePublishSettings private (val concurrency: Int) {
require(concurrency > 0)
diff --git a/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/javadsl/EventBridgePublisher.scala b/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/javadsl/EventBridgePublisher.scala
index 51ebb98d..8e360536 100644
--- a/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/javadsl/EventBridgePublisher.scala
+++ b/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/javadsl/EventBridgePublisher.scala
@@ -7,8 +7,8 @@ package akka.stream.alpakka.aws.eventbridge.javadsl
import java.util.concurrent.CompletionStage
import akka.stream.alpakka.aws.eventbridge.EventBridgePublishSettings
-import akka.stream.javadsl.{Flow, Keep, Sink}
-import akka.{Done, NotUsed}
+import akka.stream.javadsl.{ Flow, Keep, Sink }
+import akka.{ Done, NotUsed }
import software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient
import software.amazon.awssdk.services.eventbridge.model._
@@ -25,7 +25,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def flow(settings: EventBridgePublishSettings,
- eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequestEntry, PutEventsResponse, NotUsed] =
+ eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequestEntry, PutEventsResponse, NotUsed] =
akka.stream.alpakka.aws.eventbridge.scaladsl.EventBridgePublisher.flow(settings)(eventBridgeClient).asJava
/**
@@ -46,8 +46,7 @@ object EventBridgePublisher {
*/
def flowSeq(
settings: EventBridgePublishSettings,
- eventBridgeClient: EventBridgeAsyncClient
- ): Flow[PutEventsRequestEntry, PutEventsResponse, NotUsed] =
+ eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequestEntry, PutEventsResponse, NotUsed] =
akka.stream.alpakka.aws.eventbridge.scaladsl.EventBridgePublisher.flow(settings)(eventBridgeClient).asJava
/**
@@ -56,8 +55,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def flowSeq(
- eventBridgeClient: EventBridgeAsyncClient
- ): Flow[Seq[PutEventsRequestEntry], PutEventsResponse, NotUsed] =
+ eventBridgeClient: EventBridgeAsyncClient): Flow[Seq[PutEventsRequestEntry], PutEventsResponse, NotUsed] =
akka.stream.alpakka.aws.eventbridge.scaladsl.EventBridgePublisher
.flowSeq(EventBridgePublishSettings())(eventBridgeClient)
.asJava
@@ -69,7 +67,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def publishFlow(settings: EventBridgePublishSettings,
- eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequest, PutEventsResponse, NotUsed] =
+ eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequest, PutEventsResponse, NotUsed] =
akka.stream.alpakka.aws.eventbridge.scaladsl.EventBridgePublisher.publishFlow(settings)(eventBridgeClient).asJava
/**
@@ -98,7 +96,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def sink(settings: EventBridgePublishSettings,
- eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequestEntry, CompletionStage[Done]] =
+ eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequestEntry, CompletionStage[Done]] =
flow(settings, eventBridgeClient)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
@@ -109,7 +107,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def publishSink(settings: EventBridgePublishSettings,
- eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequest, CompletionStage[Done]] =
+ eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequest, CompletionStage[Done]] =
publishFlow(settings, eventBridgeClient)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
diff --git a/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/scaladsl/EventBridgePublisher.scala b/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/scaladsl/EventBridgePublisher.scala
index c619171a..c31e27d3 100644
--- a/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/scaladsl/EventBridgePublisher.scala
+++ b/aws-event-bridge/src/main/scala/akka/stream/alpakka/aws/eventbridge/scaladsl/EventBridgePublisher.scala
@@ -5,8 +5,8 @@
package akka.stream.alpakka.aws.eventbridge.scaladsl
import akka.stream.alpakka.aws.eventbridge.EventBridgePublishSettings
-import akka.stream.scaladsl.{Flow, Keep, Sink}
-import akka.{Done, NotUsed}
+import akka.stream.scaladsl.{ Flow, Keep, Sink }
+import akka.{ Done, NotUsed }
import software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient
import software.amazon.awssdk.services.eventbridge.model._
@@ -26,8 +26,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def flow(settings: EventBridgePublishSettings = EventBridgePublishSettings())(
- implicit eventBridgeClient: EventBridgeAsyncClient
- ): Flow[PutEventsRequestEntry, PutEventsResponse, NotUsed] =
+ implicit eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequestEntry, PutEventsResponse, NotUsed] =
Flow
.fromFunction((message: PutEventsRequestEntry) => PutEventsRequest.builder().entries(message).build())
.via(publishFlow(settings))
@@ -39,8 +38,8 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def flowSeq(settings: EventBridgePublishSettings = EventBridgePublishSettings())(
- implicit eventBridgeClient: EventBridgeAsyncClient
- ): Flow[Seq[PutEventsRequestEntry], PutEventsResponse, NotUsed] =
+ implicit eventBridgeClient: EventBridgeAsyncClient)
+ : Flow[Seq[PutEventsRequestEntry], PutEventsResponse, NotUsed] =
Flow
.fromFunction((messages: Seq[PutEventsRequestEntry]) => PutEventsRequest.builder().entries(messages: _*).build())
.via(publishFlow(settings))
@@ -52,8 +51,8 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def publishFlow(
- settings: EventBridgePublishSettings
- )(implicit eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequest, PutEventsResponse, NotUsed] =
+ settings: EventBridgePublishSettings)(
+ implicit eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequest, PutEventsResponse, NotUsed] =
Flow[PutEventsRequest]
.mapAsync(settings.concurrency)(eventBridgeClient.putEvents(_).toScala)
@@ -63,8 +62,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def publishFlow()(
- implicit eventBridgeClient: EventBridgeAsyncClient
- ): Flow[PutEventsRequest, PutEventsResponse, NotUsed] =
+ implicit eventBridgeClient: EventBridgeAsyncClient): Flow[PutEventsRequest, PutEventsResponse, NotUsed] =
publishFlow(EventBridgePublishSettings())
/**
@@ -74,8 +72,7 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def sink(settings: EventBridgePublishSettings = EventBridgePublishSettings())(
- implicit eventBridgeClient: EventBridgeAsyncClient
- ): Sink[PutEventsRequestEntry, Future[Done]] =
+ implicit eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequestEntry, Future[Done]] =
flow(settings).toMat(Sink.ignore)(Keep.right)
/**
@@ -85,8 +82,8 @@ object EventBridgePublisher {
* @param eventBridgeClient [[software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient EventBridgeAsyncClient]] client for publishing
*/
def publishSink(
- settings: EventBridgePublishSettings = EventBridgePublishSettings()
- )(implicit eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequest, Future[Done]] =
+ settings: EventBridgePublishSettings = EventBridgePublishSettings())(
+ implicit eventBridgeClient: EventBridgeAsyncClient): Sink[PutEventsRequest, Future[Done]] =
publishFlow(settings).toMat(Sink.ignore)(Keep.right)
/**
diff --git a/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/DefaultTestContext.scala b/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/DefaultTestContext.scala
index 338e8e3d..f1034b4f 100644
--- a/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/DefaultTestContext.scala
+++ b/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/DefaultTestContext.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.aws.eventbridge
import akka.actor.ActorSystem
import org.mockito.Mockito.reset
-import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach, Suite}
+import org.scalatest.{ BeforeAndAfterAll, BeforeAndAfterEach, Suite }
import org.scalatestplus.mockito.MockitoSugar
import software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient
diff --git a/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishMockSpec.scala b/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishMockSpec.scala
index addfa896..dd03c815 100644
--- a/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishMockSpec.scala
+++ b/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/EventBridgePublishMockSpec.scala
@@ -7,9 +7,9 @@ package akka.stream.alpakka.aws.eventbridge
import java.util.concurrent.CompletableFuture
import akka.stream.alpakka.aws.eventbridge.scaladsl.EventBridgePublisher
-import akka.stream.scaladsl.{Keep, Sink}
+import akka.stream.scaladsl.{ Keep, Sink }
import akka.stream.testkit.scaladsl.TestSource
-import org.mockito.ArgumentMatchers.{any, eq => meq}
+import org.mockito.ArgumentMatchers.{ any, eq => meq }
import org.mockito.Mockito._
import org.scalatest.flatspec._
import org.scalatest.matchers.must.Matchers
@@ -30,8 +30,7 @@ class EventBridgePublishMockSpec extends AnyFlatSpec with DefaultTestContext wit
PutEventsRequest
.builder()
.entries(
- details.map(detail => entryDetail(detail, eventBusName)): _*
- )
+ details.map(detail => entryDetail(detail, eventBusName)): _*)
.build()
private def resultResponse(eventId: String): PutEventsResponse = {
diff --git a/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/IntegrationTestContext.scala b/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/IntegrationTestContext.scala
index f92c3d5e..8c80598c 100644
--- a/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/IntegrationTestContext.scala
+++ b/aws-event-bridge/src/test/scala/akka/stream/alpakka/aws/eventbridge/IntegrationTestContext.scala
@@ -9,7 +9,7 @@ import java.util.UUID
import akka.actor.ActorSystem
import akka.testkit.TestKit
import org.scalatest.concurrent.ScalaFutures
-import org.scalatest.{BeforeAndAfterAll, Suite}
+import org.scalatest.{ BeforeAndAfterAll, Suite }
import software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient
import software.amazon.awssdk.services.eventbridge.model.CreateEventBusRequest
@@ -18,9 +18,9 @@ import scala.concurrent.duration.FiniteDuration
trait IntegrationTestContext extends BeforeAndAfterAll with ScalaFutures {
this: Suite =>
- //#init-system
+ // #init-system
implicit val system: ActorSystem = ActorSystem()
- //#init-system
+ // #init-system
def eventBusEndpoint: String = s"http://localhost:4587"
@@ -30,8 +30,7 @@ trait IntegrationTestContext extends BeforeAndAfterAll with ScalaFutures {
def createEventBus(): String =
eventBridgeClient
.createEventBus(
- CreateEventBusRequest.builder().name(s"alpakka-topic-${UUID.randomUUID().toString}").build()
- )
+ CreateEventBusRequest.builder().name(s"alpakka-topic-${UUID.randomUUID().toString}").build())
.get()
.eventBusArn()
@@ -43,11 +42,11 @@ trait IntegrationTestContext extends BeforeAndAfterAll with ScalaFutures {
override protected def afterAll(): Unit = TestKit.shutdownActorSystem(system)
def createAsyncClient(endEndpoint: String): EventBridgeAsyncClient = {
- //#init-client
+ // #init-client
import java.net.URI
import com.github.matsluni.akkahttpspi.AkkaHttpClient
- import software.amazon.awssdk.auth.credentials.{AwsBasicCredentials, StaticCredentialsProvider}
+ import software.amazon.awssdk.auth.credentials.{ AwsBasicCredentials, StaticCredentialsProvider }
import software.amazon.awssdk.regions.Region
import software.amazon.awssdk.services.eventbridge.EventBridgeAsyncClient
@@ -61,7 +60,7 @@ trait IntegrationTestContext extends BeforeAndAfterAll with ScalaFutures {
.build()
system.registerOnTermination(awsEventBridgeClient.close())
- //#init-client
+ // #init-client
awsEventBridgeClient
}
diff --git a/aws-event-bridge/src/test/scala/docs/scaladsl/EventBridgePublisherSpec.scala b/aws-event-bridge/src/test/scala/docs/scaladsl/EventBridgePublisherSpec.scala
index d03f8075..ae06c56b 100644
--- a/aws-event-bridge/src/test/scala/docs/scaladsl/EventBridgePublisherSpec.scala
+++ b/aws-event-bridge/src/test/scala/docs/scaladsl/EventBridgePublisherSpec.scala
@@ -7,11 +7,11 @@ package docs.scaladsl
import akka.Done
import akka.stream.alpakka.aws.eventbridge.IntegrationTestContext
import akka.stream.alpakka.aws.eventbridge.scaladsl.EventBridgePublisher
-import akka.stream.scaladsl.{Sink, Source}
+import akka.stream.scaladsl.{ Sink, Source }
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.flatspec._
import org.scalatest.matchers.should.Matchers
-import software.amazon.awssdk.services.eventbridge.model.{PutEventsRequest, PutEventsRequestEntry}
+import software.amazon.awssdk.services.eventbridge.model.{ PutEventsRequest, PutEventsRequestEntry }
import scala.concurrent.Future
import scala.concurrent.duration._
diff --git a/awslambda/src/main/scala/akka/stream/alpakka/awslambda/javadsl/AwsLambdaFlow.scala b/awslambda/src/main/scala/akka/stream/alpakka/awslambda/javadsl/AwsLambdaFlow.scala
index 800dd131..10805e0c 100644
--- a/awslambda/src/main/scala/akka/stream/alpakka/awslambda/javadsl/AwsLambdaFlow.scala
+++ b/awslambda/src/main/scala/akka/stream/alpakka/awslambda/javadsl/AwsLambdaFlow.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.awslambda.javadsl
import akka.NotUsed
import akka.stream.javadsl.Flow
-import software.amazon.awssdk.services.lambda.model.{InvokeRequest, InvokeResponse}
+import software.amazon.awssdk.services.lambda.model.{ InvokeRequest, InvokeResponse }
import software.amazon.awssdk.services.lambda.LambdaAsyncClient
object AwsLambdaFlow {
diff --git a/awslambda/src/main/scala/akka/stream/alpakka/awslambda/scaladsl/AwsLambdaFlow.scala b/awslambda/src/main/scala/akka/stream/alpakka/awslambda/scaladsl/AwsLambdaFlow.scala
index 19a509f5..30d3c63d 100644
--- a/awslambda/src/main/scala/akka/stream/alpakka/awslambda/scaladsl/AwsLambdaFlow.scala
+++ b/awslambda/src/main/scala/akka/stream/alpakka/awslambda/scaladsl/AwsLambdaFlow.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.awslambda.scaladsl
import akka.NotUsed
import akka.stream.scaladsl.Flow
-import software.amazon.awssdk.services.lambda.model.{InvokeRequest, InvokeResponse}
+import software.amazon.awssdk.services.lambda.model.{ InvokeRequest, InvokeResponse }
import software.amazon.awssdk.services.lambda.LambdaAsyncClient
import scala.compat.java8.FutureConverters._
@@ -16,8 +16,7 @@ object AwsLambdaFlow {
* Scala API: creates a [[AwsLambdaFlowStage]] for a AWS Lambda function invocation using [[LambdaAsyncClient]]
*/
def apply(
- parallelism: Int
- )(implicit awsLambdaClient: LambdaAsyncClient): Flow[InvokeRequest, InvokeResponse, NotUsed] =
+ parallelism: Int)(implicit awsLambdaClient: LambdaAsyncClient): Flow[InvokeRequest, InvokeResponse, NotUsed] =
Flow[InvokeRequest].mapAsyncUnordered(parallelism)(awsLambdaClient.invoke(_).toScala)
}
diff --git a/awslambda/src/test/scala/docs/scaladsl/AwsLambdaFlowSpec.scala b/awslambda/src/test/scala/docs/scaladsl/AwsLambdaFlowSpec.scala
index b313bae8..68acd756 100644
--- a/awslambda/src/test/scala/docs/scaladsl/AwsLambdaFlowSpec.scala
+++ b/awslambda/src/test/scala/docs/scaladsl/AwsLambdaFlowSpec.scala
@@ -9,22 +9,22 @@ import java.util.concurrent.CompletableFuture
import akka.actor.ActorSystem
import akka.stream.alpakka.awslambda.scaladsl.AwsLambdaFlow
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
-import akka.stream.scaladsl.{Keep, Sink}
+import akka.stream.scaladsl.{ Keep, Sink }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.stream.testkit.scaladsl.TestSource
import akka.testkit.TestKit
-import org.mockito.ArgumentMatchers.{any => mockitoAny, eq => mockitoEq}
+import org.mockito.ArgumentMatchers.{ any => mockitoAny, eq => mockitoEq }
import org.mockito.Mockito._
import org.mockito.invocation.InvocationOnMock
import org.mockito.stubbing.Answer
-import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach}
+import org.scalatest.{ BeforeAndAfterAll, BeforeAndAfterEach }
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.matchers.should.Matchers
import org.scalatest.wordspec.AnyWordSpecLike
import org.scalatestplus.mockito.MockitoSugar
import software.amazon.awssdk.core.SdkBytes
import software.amazon.awssdk.services.lambda.LambdaAsyncClient
-import software.amazon.awssdk.services.lambda.model.{InvokeRequest, InvokeResponse}
+import software.amazon.awssdk.services.lambda.model.{ InvokeRequest, InvokeResponse }
import scala.concurrent.Await
import scala.concurrent.duration._
@@ -63,8 +63,7 @@ class AwsLambdaFlowSpec
"call a single invoke request" in assertAllStagesStopped {
when(
- awsLambdaClient.invoke(mockitoEq(invokeRequest))
- ).thenAnswer(new Answer[CompletableFuture[InvokeResponse]] {
+ awsLambdaClient.invoke(mockitoEq(invokeRequest))).thenAnswer(new Answer[CompletableFuture[InvokeResponse]] {
override def answer(invocation: InvocationOnMock): CompletableFuture[InvokeResponse] =
CompletableFuture.completedFuture(invokeResponse)
})
@@ -81,8 +80,7 @@ class AwsLambdaFlowSpec
"call with exception" in assertAllStagesStopped {
when(
- awsLambdaClient.invoke(mockitoAny[InvokeRequest]())
- ).thenAnswer(new Answer[CompletableFuture[InvokeResponse]] {
+ awsLambdaClient.invoke(mockitoAny[InvokeRequest]())).thenAnswer(new Answer[CompletableFuture[InvokeResponse]] {
override def answer(invocation: InvocationOnMock): CompletableFuture[InvokeResponse] = {
val exception = new RuntimeException("Error in lambda")
val future = new CompletableFuture[InvokeResponse]()
@@ -98,7 +96,7 @@ class AwsLambdaFlowSpec
val ex = Await.result(future.failed, 3.seconds)
ex shouldBe a[RuntimeException]
- ex.getMessage shouldBe ("Error in lambda")
+ ex.getMessage shouldBe "Error in lambda"
}
}
diff --git a/awslambda/src/test/scala/docs/scaladsl/Examples.scala b/awslambda/src/test/scala/docs/scaladsl/Examples.scala
index 6471ca0d..863cf664 100644
--- a/awslambda/src/test/scala/docs/scaladsl/Examples.scala
+++ b/awslambda/src/test/scala/docs/scaladsl/Examples.scala
@@ -5,21 +5,21 @@
package docs.scaladsl
import akka.stream.alpakka.awslambda.scaladsl.AwsLambdaFlow
-import akka.stream.scaladsl.{Sink, Source}
+import akka.stream.scaladsl.{ Sink, Source }
import software.amazon.awssdk.services.lambda.LambdaAsyncClient
object Examples {
- //#init-sys
+ // #init-sys
import akka.actor.ActorSystem
implicit val system: ActorSystem = ActorSystem()
- //#init-sys
+ // #init-sys
def initClient(): Unit = {
- //#init-client
+ // #init-client
import com.github.matsluni.akkahttpspi.AkkaHttpClient
- import software.amazon.awssdk.auth.credentials.{AwsBasicCredentials, StaticCredentialsProvider}
+ import software.amazon.awssdk.auth.credentials.{ AwsBasicCredentials, StaticCredentialsProvider }
import software.amazon.awssdk.services.lambda.LambdaAsyncClient
// Don't encode credentials in your source code!
@@ -35,11 +35,11 @@ object Examples {
.build()
system.registerOnTermination(lambdaClient.close())
- //#init-client
+ // #init-client
}
def run()(implicit lambdaClient: LambdaAsyncClient): Unit = {
- //#run
+ // #run
import software.amazon.awssdk.core.SdkBytes
import software.amazon.awssdk.services.lambda.model.InvokeRequest
@@ -49,6 +49,6 @@ object Examples {
.payload(SdkBytes.fromUtf8String("test-payload"))
.build()
Source.single(request).via(AwsLambdaFlow(1)).runWith(Sink.seq)
- //#run
+ // #run
}
}
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSinkFunctions.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSinkFunctions.scala
index eb01878f..3ccaca3d 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSinkFunctions.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSinkFunctions.scala
@@ -6,31 +6,29 @@ package akka.stream.alpakka.azure.storagequeue.impl
import akka.annotation.InternalApi
import akka.stream.alpakka.azure.storagequeue.DeleteOrUpdateMessage
-import akka.stream.alpakka.azure.storagequeue.DeleteOrUpdateMessage.{Delete, UpdateVisibility}
-import com.microsoft.azure.storage.queue.{CloudQueue, CloudQueueMessage}
+import akka.stream.alpakka.azure.storagequeue.DeleteOrUpdateMessage.{ Delete, UpdateVisibility }
+import com.microsoft.azure.storage.queue.{ CloudQueue, CloudQueueMessage }
/**
* INTERNAL API
*/
@InternalApi private[storagequeue] object AzureQueueSinkFunctions {
def addMessage(
- cloudQueue: () => CloudQueue
- )(msg: CloudQueueMessage, timeToLive: Int = 0, initialVisibilityTimeout: Int = 0): Unit =
+ cloudQueue: () => CloudQueue)(
+ msg: CloudQueueMessage, timeToLive: Int = 0, initialVisibilityTimeout: Int = 0): Unit =
cloudQueue().addMessage(msg, timeToLive, initialVisibilityTimeout, null, null)
def deleteMessage(
- cloudQueue: () => CloudQueue
- )(msg: CloudQueueMessage): Unit =
+ cloudQueue: () => CloudQueue)(msg: CloudQueueMessage): Unit =
cloudQueue().deleteMessage(msg)
def updateMessage(cloudQueue: () => CloudQueue)(msg: CloudQueueMessage, timeout: Int): Unit =
cloudQueue().updateMessage(msg, timeout)
def deleteOrUpdateMessage(
- cloudQueue: () => CloudQueue
- )(msg: CloudQueueMessage, op: DeleteOrUpdateMessage): Unit =
+ cloudQueue: () => CloudQueue)(msg: CloudQueueMessage, op: DeleteOrUpdateMessage): Unit =
op match {
- case _: Delete => deleteMessage(cloudQueue)(msg)
+ case _: Delete => deleteMessage(cloudQueue)(msg)
case m: UpdateVisibility => updateMessage(cloudQueue)(msg, m.timeout)
}
}
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSourceStage.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSourceStage.scala
index b727f084..c22c1495 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSourceStage.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/impl/AzureQueueSourceStage.scala
@@ -8,9 +8,9 @@ import akka.NotUsed
import akka.annotation.InternalApi
import akka.stream.alpakka.azure.storagequeue.AzureQueueSourceSettings
import akka.stream.impl.Stages.DefaultAttributes.IODispatcher
-import akka.stream.stage.{GraphStage, GraphStageLogic, OutHandler, TimerGraphStageLogic}
-import akka.stream.{Attributes, Outlet, SourceShape}
-import com.microsoft.azure.storage.queue.{CloudQueue, CloudQueueMessage}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, OutHandler, TimerGraphStageLogic }
+import akka.stream.{ Attributes, Outlet, SourceShape }
+import com.microsoft.azure.storage.queue.{ CloudQueue, CloudQueueMessage }
import scala.collection.mutable.Queue
@@ -18,7 +18,7 @@ import scala.collection.mutable.Queue
* INTERNAL API
*/
@InternalApi private[storagequeue] final class AzureQueueSourceStage(cloudQueue: () => CloudQueue,
- settings: AzureQueueSourceSettings)
+ settings: AzureQueueSourceSettings)
extends GraphStage[SourceShape[CloudQueueMessage]] {
val out: Outlet[CloudQueueMessage] = Outlet("AzureCloudQueue.out")
override val shape: SourceShape[CloudQueueMessage] = SourceShape(out)
@@ -64,7 +64,6 @@ import scala.collection.mutable.Queue
} else {
retrieveMessages()
}
- }
- )
+ })
}
}
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSink.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSink.scala
index d7265adf..63ff38f2 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSink.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSink.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.azure.storagequeue.javadsl
-import com.microsoft.azure.storage.queue.{CloudQueue, CloudQueueMessage}
+import com.microsoft.azure.storage.queue.{ CloudQueue, CloudQueueMessage }
import akka.stream.alpakka.azure.storagequeue.impl.AzureQueueSinkFunctions
import akka.stream.javadsl.Sink
import akka.Done
@@ -25,7 +25,7 @@ object AzureQueueSink {
* Internal API
*/
private[javadsl] def fromFunction[T](f: T => Unit): Sink[T, CompletionStage[Done]] = {
- import akka.stream.alpakka.azure.storagequeue.scaladsl.{AzureQueueSink => AzureQueueSinkScalaDSL}
+ import akka.stream.alpakka.azure.storagequeue.scaladsl.{ AzureQueueSink => AzureQueueSinkScalaDSL }
import scala.compat.java8.FutureConverters._
AzureQueueSinkScalaDSL.fromFunction(f).mapMaterializedValue(_.toJava).asJava
}
@@ -41,12 +41,10 @@ object AzureQueueWithTimeoutsSink {
* of a [[com.microsoft.azure.storage.queue.CouldQueueMessage]] a [[MessageWithTimeouts]].
*/
def create(cloudQueue: Supplier[CloudQueue]): Sink[MessageWithTimeouts, CompletionStage[Done]] =
- AzureQueueSink.fromFunction(
- { input: MessageWithTimeouts =>
- AzureQueueSinkFunctions
- .addMessage(() => cloudQueue.get)(input.message, input.timeToLive, input.initialVisibility)
- }
- )
+ AzureQueueSink.fromFunction { input: MessageWithTimeouts =>
+ AzureQueueSinkFunctions
+ .addMessage(() => cloudQueue.get)(input.message, input.timeToLive, input.initialVisibility)
+ }
}
object AzureQueueDeleteSink {
@@ -67,7 +65,6 @@ object AzureQueueDeleteOrUpdateSink {
* in an Azure Storage Queue.
*/
def create(cloudQueue: Supplier[CloudQueue]): Sink[MessageAndDeleteOrUpdate, CompletionStage[Done]] =
- AzureQueueSink.fromFunction[MessageAndDeleteOrUpdate](
- input => AzureQueueSinkFunctions.deleteOrUpdateMessage(() => cloudQueue.get)(input.message, input.op)
- )
+ AzureQueueSink.fromFunction[MessageAndDeleteOrUpdate](input =>
+ AzureQueueSinkFunctions.deleteOrUpdateMessage(() => cloudQueue.get)(input.message, input.op))
}
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSource.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSource.scala
index 4c3e3d22..56d530e4 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSource.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/javadsl/AzureQueueSource.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.azure.storagequeue.javadsl
-import com.microsoft.azure.storage.queue.{CloudQueue, CloudQueueMessage}
+import com.microsoft.azure.storage.queue.{ CloudQueue, CloudQueueMessage }
import akka.stream.alpakka.azure.storagequeue.AzureQueueSourceSettings
import akka.stream.javadsl.Source
import akka.NotUsed
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSink.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSink.scala
index ffb76bb6..fb73743b 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSink.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSink.scala
@@ -4,9 +4,9 @@
package akka.stream.alpakka.azure.storagequeue.scaladsl
-import com.microsoft.azure.storage.queue.{CloudQueue, CloudQueueMessage}
+import com.microsoft.azure.storage.queue.{ CloudQueue, CloudQueueMessage }
import akka.stream.alpakka.azure.storagequeue.impl.AzureQueueSinkFunctions
-import akka.stream.scaladsl.{Flow, Keep, Sink}
+import akka.stream.scaladsl.{ Flow, Keep, Sink }
import akka.Done
import scala.concurrent.Future
@@ -41,11 +41,8 @@ object AzureQueueWithTimeoutsSink {
* with (CouldQueueMessage, timeToLive, initialVisibilityTimeout).
*/
def apply(
- cloudQueue: () => CloudQueue
- ): Sink[(CloudQueueMessage, Int, Int), Future[Done]] =
- AzureQueueSink.fromFunction(
- tup => AzureQueueSinkFunctions.addMessage(cloudQueue)(tup._1, tup._2, tup._3)
- )
+ cloudQueue: () => CloudQueue): Sink[(CloudQueueMessage, Int, Int), Future[Done]] =
+ AzureQueueSink.fromFunction(tup => AzureQueueSinkFunctions.addMessage(cloudQueue)(tup._1, tup._2, tup._3))
}
object AzureQueueDeleteSink {
@@ -64,9 +61,6 @@ object AzureQueueDeleteOrUpdateSink {
* in an Azure Storage Queue.
*/
def apply(
- cloudQueue: () => CloudQueue
- ): Sink[(CloudQueueMessage, DeleteOrUpdateMessage), Future[Done]] =
- AzureQueueSink.fromFunction(
- input => AzureQueueSinkFunctions.deleteOrUpdateMessage(cloudQueue)(input._1, input._2)
- )
+ cloudQueue: () => CloudQueue): Sink[(CloudQueueMessage, DeleteOrUpdateMessage), Future[Done]] =
+ AzureQueueSink.fromFunction(input => AzureQueueSinkFunctions.deleteOrUpdateMessage(cloudQueue)(input._1, input._2))
}
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSource.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSource.scala
index 35a4d00f..54756788 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSource.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/scaladsl/AzureQueueSource.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.azure.storagequeue.scaladsl
-import com.microsoft.azure.storage.queue.{CloudQueue, CloudQueueMessage}
+import com.microsoft.azure.storage.queue.{ CloudQueue, CloudQueueMessage }
import akka.stream.alpakka.azure.storagequeue.AzureQueueSourceSettings
import akka.stream.scaladsl.Source
import akka.NotUsed
@@ -17,7 +17,6 @@ object AzureQueueSource {
*/
def apply(
cloudQueue: () => CloudQueue,
- settings: AzureQueueSourceSettings = AzureQueueSourceSettings()
- ): Source[CloudQueueMessage, NotUsed] =
+ settings: AzureQueueSourceSettings = AzureQueueSourceSettings()): Source[CloudQueueMessage, NotUsed] =
Source.fromGraph(new AzureQueueSourceStage(cloudQueue, settings))
}
diff --git a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/settings.scala b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/settings.scala
index de65edbc..495a98a8 100644
--- a/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/settings.scala
+++ b/azure-storage-queue/src/main/scala/akka/stream/alpakka/azure/storagequeue/settings.scala
@@ -4,13 +4,14 @@
package akka.stream.alpakka.azure.storagequeue
-import java.time.{Duration => JavaDuration}
+import java.time.{ Duration => JavaDuration }
import java.util.Optional
import scala.compat.java8.OptionConverters._
-import scala.concurrent.duration.{Duration, FiniteDuration}
+import scala.concurrent.duration.{ Duration, FiniteDuration }
-/** Settings for AzureQueueSource
+/**
+ * Settings for AzureQueueSource
*
* @param initalVisibilityTimeout Specifies how many seconds a message becomes invisible after it has been dequeued.
* See parameter of the same name in [[com.microsoft.azure.storage.queue.CloudQueue$.retrieveMessages]].
@@ -22,14 +23,12 @@ import scala.concurrent.duration.{Duration, FiniteDuration}
final class AzureQueueSourceSettings private (
val initialVisibilityTimeout: Int,
val batchSize: Int,
- val retrieveRetryTimeout: Option[FiniteDuration] = None
-) {
+ val retrieveRetryTimeout: Option[FiniteDuration] = None) {
def withBatchSize(batchSize: Int): AzureQueueSourceSettings =
copy(batchSize = batchSize)
/**
- *
* @param retrieveRetryTimeout in seconds. If <= 0 retrying of message retrieval is disabled.
* @return
*/
diff --git a/azure-storage-queue/src/test/scala/docs/scaladsl/AzureQueueSpec.scala b/azure-storage-queue/src/test/scala/docs/scaladsl/AzureQueueSpec.scala
index df2b8d94..44f13e62 100644
--- a/azure-storage-queue/src/test/scala/docs/scaladsl/AzureQueueSpec.scala
+++ b/azure-storage-queue/src/test/scala/docs/scaladsl/AzureQueueSpec.scala
@@ -5,7 +5,7 @@
package docs.scaladsl
import akka.actor.ActorSystem
-import akka.stream.alpakka.azure.storagequeue.DeleteOrUpdateMessage.{Delete, UpdateVisibility}
+import akka.stream.alpakka.azure.storagequeue.DeleteOrUpdateMessage.{ Delete, UpdateVisibility }
import akka.stream.alpakka.azure.storagequeue._
import akka.stream.alpakka.azure.storagequeue.scaladsl._
import akka.stream.scaladsl._
@@ -70,10 +70,8 @@ class AzureQueueSpec extends TestKit(ActorSystem()) with AsyncFlatSpecLike with
AzureQueueSource(queueFactory)
.runWith(Sink.seq)
- .map(
- dequeuedMsgs =>
- assert(msgs.map(_.getMessageContentAsString).toSet == dequeuedMsgs.map(_.getMessageContentAsString).toSet)
- )
+ .map(dequeuedMsgs =>
+ assert(msgs.map(_.getMessageContentAsString).toSet == dequeuedMsgs.map(_.getMessageContentAsString).toSet))
}
it should "observe retrieveRetryTimeout and retrieve messages queued later" in assertAllStagesStopped {
@@ -83,10 +81,8 @@ class AzureQueueSpec extends TestKit(ActorSystem()) with AsyncFlatSpecLike with
AzureQueueSource(queueFactory, AzureQueueSourceSettings().withRetrieveRetryTimeout(1.seconds))
.take(10)
.runWith(Sink.seq)
- .map(
- dequeuedMsgs =>
- assert(msgs.map(_.getMessageContentAsString).toSet == dequeuedMsgs.map(_.getMessageContentAsString).toSet)
- )
+ .map(dequeuedMsgs =>
+ assert(msgs.map(_.getMessageContentAsString).toSet == dequeuedMsgs.map(_.getMessageContentAsString).toSet))
Thread.sleep(3000)
msgs.foreach(m => queue.addMessage(m))
@@ -98,9 +94,9 @@ class AzureQueueSpec extends TestKit(ActorSystem()) with AsyncFlatSpecLike with
msgs.foreach(m => queue.addMessage(m))
Await.result(AzureQueueSource(queueFactory, AzureQueueSourceSettings().withBatchSize(2))
- .take(1)
- .runWith(Sink.seq),
- timeout)
+ .take(1)
+ .runWith(Sink.seq),
+ timeout)
assert(queue.retrieveMessage() != null, "There should be a 11th message on queue")
}
@@ -158,8 +154,7 @@ class AzureQueueSpec extends TestKit(ActorSystem()) with AsyncFlatSpecLike with
.take(10)
.map(msg => (msg, UpdateVisibility(120)))
.runWith(AzureQueueDeleteOrUpdateSink(queueFactory)),
- timeout
- )
+ timeout)
// Now we should not be able to get another one
assertCannotGetMessageFromQueue
@@ -176,8 +171,7 @@ class AzureQueueSpec extends TestKit(ActorSystem()) with AsyncFlatSpecLike with
.take(10)
.map(msg => (msg, Delete))
.runWith(AzureQueueDeleteOrUpdateSink(queueFactory)),
- timeout
- )
+ timeout)
// Now we should not be able to get another one
assertCannotGetMessageFromQueue
diff --git a/build.sbt b/build.sbt
index 0e0effef..2893f2c0 100644
--- a/build.sbt
+++ b/build.sbt
@@ -49,8 +49,7 @@ lazy val `pekko-connectors` = project
text,
udp,
unixdomainsocket,
- xml
- )
+ xml)
.aggregate(`doc-examples`)
.settings(
onLoadMessage :=
@@ -82,16 +81,15 @@ lazy val `pekko-connectors` = project
.filterNot(_.data.getAbsolutePath.contains("protobuf-java-2.6.1.jar"))
},
ScalaUnidoc / unidoc / unidocProjectFilter := inAnyProject
- -- inProjects(
- `doc-examples`,
- csvBench,
- mqttStreamingBench,
- // googleCloudPubSubGrpc and googleCloudBigQueryStorage contain the same gRPC generated classes
- // don't include ScalaDocs for googleCloudBigQueryStorage to make it work
- googleCloudBigQueryStorage,
- // springWeb triggers an esoteric ScalaDoc bug (from Java code)
- springWeb
- ),
+ -- inProjects(
+ `doc-examples`,
+ csvBench,
+ mqttStreamingBench,
+ // googleCloudPubSubGrpc and googleCloudBigQueryStorage contain the same gRPC generated classes
+ // don't include ScalaDocs for googleCloudBigQueryStorage to make it work
+ googleCloudBigQueryStorage,
+ // springWeb triggers an esoteric ScalaDoc bug (from Java code)
+ springWeb),
licenses := List(License.Apache2),
crossScalaVersions := List() // workaround for https://github.com/sbt/sbt/issues/3465
)
@@ -99,18 +97,15 @@ lazy val `pekko-connectors` = project
TaskKey[Unit]("verifyCodeFmt") := {
javafmtCheckAll.all(ScopeFilter(inAnyProject)).result.value.toEither.left.foreach { _ =>
throw new MessageOnlyException(
- "Unformatted Java code found. Please run 'javafmtAll' and commit the reformatted code"
- )
+ "Unformatted Java code found. Please run 'javafmtAll' and commit the reformatted code")
}
scalafmtCheckAll.all(ScopeFilter(inAnyProject)).result.value.toEither.left.foreach { _ =>
throw new MessageOnlyException(
- "Unformatted Scala code found. Please run 'scalafmtAll' and commit the reformatted code"
- )
+ "Unformatted Scala code found. Please run 'scalafmtAll' and commit the reformatted code")
}
(Compile / scalafmtSbtCheck).result.value.toEither.left.foreach { _ =>
throw new MessageOnlyException(
- "Unformatted sbt code found. Please run 'scalafmtSbt' and commit the reformatted code"
- )
+ "Unformatted sbt code found. Please run 'scalafmtSbt' and commit the reformatted code")
}
}
@@ -126,8 +121,7 @@ lazy val awslambda = pekkoConnectorProject("awslambda", "aws.lambda", Dependenci
lazy val azureStorageQueue = pekkoConnectorProject(
"azure-storage-queue",
"azure.storagequeue",
- Dependencies.AzureStorageQueue
-)
+ Dependencies.AzureStorageQueue)
lazy val cassandra =
pekkoConnectorProject("cassandra", "cassandra", Dependencies.Cassandra)
@@ -146,8 +140,7 @@ lazy val dynamodb = pekkoConnectorProject("dynamodb", "aws.dynamodb", Dependenci
lazy val elasticsearch = pekkoConnectorProject(
"elasticsearch",
"elasticsearch",
- Dependencies.Elasticsearch
-)
+ Dependencies.Elasticsearch)
// The name 'file' is taken by `sbt.file`, hence 'files'
lazy val files = pekkoConnectorProject("file", "file", Dependencies.File)
@@ -158,8 +151,7 @@ lazy val ftp = pekkoConnectorProject(
Dependencies.Ftp,
Test / fork := true,
// To avoid potential blocking in machines with low entropy (default is `/dev/random`)
- Test / javaOptions += "-Djava.security.egd=file:/dev/./urandom"
-)
+ Test / javaOptions += "-Djava.security.egd=file:/dev/./urandom")
lazy val geode =
pekkoConnectorProject(
@@ -168,23 +160,21 @@ lazy val geode =
Dependencies.Geode,
Test / fork := true,
// https://github.com/scala/bug/issues/12072
- Test / scalacOptions += "-Xlint:-byname-implicit"
- )
+ Test / scalacOptions += "-Xlint:-byname-implicit")
lazy val googleCommon = pekkoConnectorProject(
"google-common",
"google.common",
Dependencies.GoogleCommon,
- Test / fork := true
-)
+ Test / fork := true)
lazy val googleCloudBigQuery = pekkoConnectorProject(
"google-cloud-bigquery",
"google.cloud.bigquery",
Dependencies.GoogleBigQuery,
Test / fork := true,
- Compile / scalacOptions += "-Wconf:src=src_managed/.+:s"
-).dependsOn(googleCommon).enablePlugins(spray.boilerplate.BoilerplatePlugin)
+ Compile / scalacOptions += "-Wconf:src=src_managed/.+:s").dependsOn(googleCommon).enablePlugins(
+ spray.boilerplate.BoilerplatePlugin)
lazy val googleCloudBigQueryStorage = pekkoConnectorProject(
"google-cloud-bigquery-storage",
@@ -197,11 +187,10 @@ lazy val googleCloudBigQueryStorage = pekkoConnectorProject(
// Test / akkaGrpcGeneratedSources := Seq(AkkaGrpc.Server),
akkaGrpcGeneratedLanguages := Seq(AkkaGrpc.Scala, AkkaGrpc.Java),
Compile / scalacOptions ++= Seq(
- "-Wconf:src=.+/akka-grpc/main/.+:s",
- "-Wconf:src=.+/akka-grpc/test/.+:s"
- ),
- compile / javacOptions := (compile / javacOptions).value.filterNot(_ == "-Xlint:deprecation")
-).dependsOn(googleCommon).enablePlugins(AkkaGrpcPlugin)
+ "-Wconf:src=.+/akka-grpc/main/.+:s",
+ "-Wconf:src=.+/akka-grpc/test/.+:s"),
+ compile / javacOptions := (compile / javacOptions).value.filterNot(_ == "-Xlint:deprecation")).dependsOn(
+ googleCommon).enablePlugins(AkkaGrpcPlugin)
lazy val googleCloudPubSub = pekkoConnectorProject(
"google-cloud-pub-sub",
@@ -209,8 +198,8 @@ lazy val googleCloudPubSub = pekkoConnectorProject(
Dependencies.GooglePubSub,
Test / fork := true,
// See docker-compose.yml gcloud-pubsub-emulator_prep
- Test / envVars := Map("PUBSUB_EMULATOR_HOST" -> "localhost", "PUBSUB_EMULATOR_PORT" -> "8538")
-).dependsOn(googleCommon)
+ Test / envVars := Map("PUBSUB_EMULATOR_HOST" -> "localhost", "PUBSUB_EMULATOR_PORT" -> "8538")).dependsOn(
+ googleCommon)
lazy val googleCloudPubSubGrpc = pekkoConnectorProject(
"google-cloud-pub-sub-grpc",
@@ -222,18 +211,16 @@ lazy val googleCloudPubSubGrpc = pekkoConnectorProject(
// for the ExampleApp in the tests
run / connectInput := true,
Compile / scalacOptions ++= Seq(
- "-Wconf:src=.+/akka-grpc/main/.+:s",
- "-Wconf:src=.+/akka-grpc/test/.+:s"
- ),
- compile / javacOptions := (compile / javacOptions).value.filterNot(_ == "-Xlint:deprecation")
-).enablePlugins(AkkaGrpcPlugin).dependsOn(googleCommon)
+ "-Wconf:src=.+/akka-grpc/main/.+:s",
+ "-Wconf:src=.+/akka-grpc/test/.+:s"),
+ compile / javacOptions := (compile / javacOptions).value.filterNot(_ == "-Xlint:deprecation")).enablePlugins(
+ AkkaGrpcPlugin).dependsOn(googleCommon)
lazy val googleCloudStorage = pekkoConnectorProject(
"google-cloud-storage",
"google.cloud.storage",
Test / fork := true,
- Dependencies.GoogleStorage
-).dependsOn(googleCommon)
+ Dependencies.GoogleStorage).dependsOn(googleCommon)
lazy val googleFcm =
pekkoConnectorProject("google-fcm", "google.firebase.fcm", Dependencies.GoogleFcm, Test / fork := true)
@@ -251,17 +238,14 @@ lazy val influxdb = pekkoConnectorProject(
"influxdb",
Dependencies.InfluxDB,
Compile / scalacOptions ++= Seq(
- // JDK 11: method isAccessible in class AccessibleObject is deprecated
- "-Wconf:cat=deprecation:s"
- )
-)
+ // JDK 11: method isAccessible in class AccessibleObject is deprecated
+ "-Wconf:cat=deprecation:s"))
lazy val ironmq = pekkoConnectorProject(
"ironmq",
"ironmq",
Dependencies.IronMq,
- Test / fork := true
-)
+ Test / fork := true)
lazy val jms = pekkoConnectorProject("jms", "jms", Dependencies.Jms)
@@ -288,8 +272,7 @@ lazy val orientdb =
Dependencies.OrientDB,
Test / fork := true,
// note: orientdb client needs to be refactored to move off deprecated calls
- fatalWarnings := false
- )
+ fatalWarnings := false)
lazy val reference = internalProject("reference", Dependencies.Reference)
.dependsOn(testkit % Test)
@@ -300,14 +283,12 @@ lazy val pravega = pekkoConnectorProject(
"pravega",
"pravega",
Dependencies.Pravega,
- Test / fork := true
-)
+ Test / fork := true)
lazy val springWeb = pekkoConnectorProject(
"spring-web",
"spring.web",
- Dependencies.SpringWeb
-)
+ Dependencies.SpringWeb)
lazy val simpleCodecs = pekkoConnectorProject("simple-codecs", "simplecodecs")
@@ -344,77 +325,74 @@ lazy val docs = project
Preprocess / siteSubdirName := s"api/alpakka/${projectInfoVersion.value}",
Preprocess / sourceDirectory := (LocalRootProject / ScalaUnidoc / unidoc / target).value,
Preprocess / preprocessRules := Seq(
- ("http://www\\.eclipse\\.org/".r, _ => "https://www\\.eclipse\\.org/"),
- ("http://pravega\\.io/".r, _ => "https://pravega\\.io/"),
- ("http://www\\.scala-lang\\.org/".r, _ => "https://www\\.scala-lang\\.org/"),
- ("https://javadoc\\.io/page/".r, _ => "https://javadoc\\.io/static/")
- ),
+ ("http://www\\.eclipse\\.org/".r, _ => "https://www\\.eclipse\\.org/"),
+ ("http://pravega\\.io/".r, _ => "https://pravega\\.io/"),
+ ("http://www\\.scala-lang\\.org/".r, _ => "https://www\\.scala-lang\\.org/"),
+ ("https://javadoc\\.io/page/".r, _ => "https://javadoc\\.io/static/")),
Paradox / siteSubdirName := s"docs/alpakka/${projectInfoVersion.value}",
paradoxProperties ++= Map(
- "akka.version" -> Dependencies.AkkaVersion,
- "akka-http.version" -> Dependencies.AkkaHttpVersion,
- "hadoop.version" -> Dependencies.HadoopVersion,
- "extref.github.base_url" -> s"https://github.com/akka/alpakka/tree/${if (isSnapshot.value) "master"
+ "akka.version" -> Dependencies.AkkaVersion,
+ "akka-http.version" -> Dependencies.AkkaHttpVersion,
+ "hadoop.version" -> Dependencies.HadoopVersion,
+ "extref.github.base_url" -> s"https://github.com/akka/alpakka/tree/${if (isSnapshot.value) "master"
else "v" + version.value}/%s",
- "extref.akka.base_url" -> s"https://doc.akka.io/docs/akka/${Dependencies.AkkaBinaryVersion}/%s",
- "scaladoc.akka.base_url" -> s"https://doc.akka.io/api/akka/${Dependencies.AkkaBinaryVersion}",
- "javadoc.akka.base_url" -> s"https://doc.akka.io/japi/akka/${Dependencies.AkkaBinaryVersion}/",
- "javadoc.akka.link_style" -> "direct",
- "extref.akka-http.base_url" -> s"https://doc.akka.io/docs/akka-http/${Dependencies.AkkaHttpBinaryVersion}/%s",
- "scaladoc.akka.http.base_url" -> s"https://doc.akka.io/api/akka-http/${Dependencies.AkkaHttpBinaryVersion}/",
- "javadoc.akka.http.base_url" -> s"https://doc.akka.io/japi/akka-http/${Dependencies.AkkaHttpBinaryVersion}/",
- // Akka gRPC
- "akka-grpc.version" -> Dependencies.AkkaGrpcBinaryVersion,
- "extref.akka-grpc.base_url" -> s"https://doc.akka.io/docs/akka-grpc/${Dependencies.AkkaGrpcBinaryVersion}/%s",
- // Couchbase
- "couchbase.version" -> Dependencies.CouchbaseVersion,
- "extref.couchbase.base_url" -> s"https://docs.couchbase.com/java-sdk/${Dependencies.CouchbaseVersionForDocs}/%s",
- // Java
- "extref.java-api.base_url" -> "https://docs.oracle.com/javase/8/docs/api/index.html?%s.html",
- "extref.geode.base_url" -> s"https://geode.apache.org/docs/guide/${Dependencies.GeodeVersionForDocs}/%s",
- "extref.javaee-api.base_url" -> "https://docs.oracle.com/javaee/7/api/index.html?%s.html",
- "extref.paho-api.base_url" -> "https://www.eclipse.org/paho/files/javadoc/index.html?%s.html",
- "extref.pravega.base_url" -> s"https://cncf.pravega.io/docs/${Dependencies.PravegaVersionForDocs}/%s",
- "extref.slick.base_url" -> s"https://scala-slick.org/doc/${Dependencies.SlickVersion}/%s",
- // Cassandra
- "extref.cassandra.base_url" -> s"https://cassandra.apache.org/doc/${Dependencies.CassandraVersionInDocs}/%s",
- "extref.cassandra-driver.base_url" -> s"https://docs.datastax.com/en/developer/java-driver/${Dependencies.CassandraDriverVersionInDocs}/%s",
- "javadoc.com.datastax.oss.base_url" -> s"https://docs.datastax.com/en/drivers/java/${Dependencies.CassandraDriverVersionInDocs}/",
- // Solr
- "extref.solr.base_url" -> s"https://lucene.apache.org/solr/guide/${Dependencies.SolrVersionForDocs}/%s",
- "javadoc.org.apache.solr.base_url" -> s"https://lucene.apache.org/solr/${Dependencies.SolrVersionForDocs}_0/solr-solrj/",
- // Java
- "javadoc.base_url" -> "https://docs.oracle.com/javase/8/docs/api/",
- "javadoc.javax.jms.base_url" -> "https://docs.oracle.com/javaee/7/api/",
- "javadoc.com.couchbase.base_url" -> s"https://docs.couchbase.com/sdk-api/couchbase-java-client-${Dependencies.CouchbaseVersion}/",
- "javadoc.io.pravega.base_url" -> s"http://pravega.io/docs/${Dependencies.PravegaVersionForDocs}/javadoc/clients/",
- "javadoc.org.apache.kudu.base_url" -> s"https://kudu.apache.org/releases/${Dependencies.KuduVersion}/apidocs/",
- "javadoc.org.apache.hadoop.base_url" -> s"https://hadoop.apache.org/docs/r${Dependencies.HadoopVersion}/api/",
- "javadoc.software.amazon.awssdk.base_url" -> "https://sdk.amazonaws.com/java/api/latest/",
- "javadoc.com.google.auth.base_url" -> "https://www.javadoc.io/doc/com.google.auth/google-auth-library-credentials/latest/",
- "javadoc.com.google.auth.link_style" -> "direct",
- "javadoc.com.fasterxml.jackson.annotation.base_url" -> "https://javadoc.io/doc/com.fasterxml.jackson.core/jackson-annotations/latest/",
- "javadoc.com.fasterxml.jackson.annotation.link_style" -> "direct",
- // Scala
- "scaladoc.spray.json.base_url" -> s"https://javadoc.io/doc/io.spray/spray-json_${scalaBinaryVersion.value}/latest/",
- // Eclipse Paho client for MQTT
- "javadoc.org.eclipse.paho.client.mqttv3.base_url" -> "https://www.eclipse.org/paho/files/javadoc/",
- "javadoc.org.bson.codecs.configuration.base_url" -> "https://mongodb.github.io/mongo-java-driver/3.7/javadoc/",
- "scaladoc.scala.base_url" -> s"https://www.scala-lang.org/api/${scalaBinaryVersion.value}.x/",
- "scaladoc.akka.stream.alpakka.base_url" -> s"/${(Preprocess / siteSubdirName).value}/",
- "javadoc.akka.stream.alpakka.base_url" -> ""
- ),
+ "extref.akka.base_url" -> s"https://doc.akka.io/docs/akka/${Dependencies.AkkaBinaryVersion}/%s",
+ "scaladoc.akka.base_url" -> s"https://doc.akka.io/api/akka/${Dependencies.AkkaBinaryVersion}",
+ "javadoc.akka.base_url" -> s"https://doc.akka.io/japi/akka/${Dependencies.AkkaBinaryVersion}/",
+ "javadoc.akka.link_style" -> "direct",
+ "extref.akka-http.base_url" -> s"https://doc.akka.io/docs/akka-http/${Dependencies.AkkaHttpBinaryVersion}/%s",
+ "scaladoc.akka.http.base_url" -> s"https://doc.akka.io/api/akka-http/${Dependencies.AkkaHttpBinaryVersion}/",
+ "javadoc.akka.http.base_url" -> s"https://doc.akka.io/japi/akka-http/${Dependencies.AkkaHttpBinaryVersion}/",
+ // Akka gRPC
+ "akka-grpc.version" -> Dependencies.AkkaGrpcBinaryVersion,
+ "extref.akka-grpc.base_url" -> s"https://doc.akka.io/docs/akka-grpc/${Dependencies.AkkaGrpcBinaryVersion}/%s",
+ // Couchbase
+ "couchbase.version" -> Dependencies.CouchbaseVersion,
+ "extref.couchbase.base_url" -> s"https://docs.couchbase.com/java-sdk/${Dependencies.CouchbaseVersionForDocs}/%s",
+ // Java
+ "extref.java-api.base_url" -> "https://docs.oracle.com/javase/8/docs/api/index.html?%s.html",
+ "extref.geode.base_url" -> s"https://geode.apache.org/docs/guide/${Dependencies.GeodeVersionForDocs}/%s",
+ "extref.javaee-api.base_url" -> "https://docs.oracle.com/javaee/7/api/index.html?%s.html",
+ "extref.paho-api.base_url" -> "https://www.eclipse.org/paho/files/javadoc/index.html?%s.html",
+ "extref.pravega.base_url" -> s"https://cncf.pravega.io/docs/${Dependencies.PravegaVersionForDocs}/%s",
+ "extref.slick.base_url" -> s"https://scala-slick.org/doc/${Dependencies.SlickVersion}/%s",
+ // Cassandra
+ "extref.cassandra.base_url" -> s"https://cassandra.apache.org/doc/${Dependencies.CassandraVersionInDocs}/%s",
+ "extref.cassandra-driver.base_url" -> s"https://docs.datastax.com/en/developer/java-driver/${Dependencies.CassandraDriverVersionInDocs}/%s",
+ "javadoc.com.datastax.oss.base_url" -> s"https://docs.datastax.com/en/drivers/java/${Dependencies.CassandraDriverVersionInDocs}/",
+ // Solr
+ "extref.solr.base_url" -> s"https://lucene.apache.org/solr/guide/${Dependencies.SolrVersionForDocs}/%s",
+ "javadoc.org.apache.solr.base_url" -> s"https://lucene.apache.org/solr/${Dependencies.SolrVersionForDocs}_0/solr-solrj/",
+ // Java
+ "javadoc.base_url" -> "https://docs.oracle.com/javase/8/docs/api/",
+ "javadoc.javax.jms.base_url" -> "https://docs.oracle.com/javaee/7/api/",
+ "javadoc.com.couchbase.base_url" -> s"https://docs.couchbase.com/sdk-api/couchbase-java-client-${Dependencies.CouchbaseVersion}/",
+ "javadoc.io.pravega.base_url" -> s"http://pravega.io/docs/${Dependencies.PravegaVersionForDocs}/javadoc/clients/",
+ "javadoc.org.apache.kudu.base_url" -> s"https://kudu.apache.org/releases/${Dependencies.KuduVersion}/apidocs/",
+ "javadoc.org.apache.hadoop.base_url" -> s"https://hadoop.apache.org/docs/r${Dependencies.HadoopVersion}/api/",
+ "javadoc.software.amazon.awssdk.base_url" -> "https://sdk.amazonaws.com/java/api/latest/",
+ "javadoc.com.google.auth.base_url" -> "https://www.javadoc.io/doc/com.google.auth/google-auth-library-credentials/latest/",
+ "javadoc.com.google.auth.link_style" -> "direct",
+ "javadoc.com.fasterxml.jackson.annotation.base_url" -> "https://javadoc.io/doc/com.fasterxml.jackson.core/jackson-annotations/latest/",
+ "javadoc.com.fasterxml.jackson.annotation.link_style" -> "direct",
+ // Scala
+ "scaladoc.spray.json.base_url" -> s"https://javadoc.io/doc/io.spray/spray-json_${scalaBinaryVersion.value}/latest/",
+ // Eclipse Paho client for MQTT
+ "javadoc.org.eclipse.paho.client.mqttv3.base_url" -> "https://www.eclipse.org/paho/files/javadoc/",
+ "javadoc.org.bson.codecs.configuration.base_url" -> "https://mongodb.github.io/mongo-java-driver/3.7/javadoc/",
+ "scaladoc.scala.base_url" -> s"https://www.scala-lang.org/api/${scalaBinaryVersion.value}.x/",
+ "scaladoc.akka.stream.alpakka.base_url" -> s"/${(Preprocess / siteSubdirName).value}/",
+ "javadoc.akka.stream.alpakka.base_url" -> ""),
paradoxGroups := Map("Language" -> Seq("Java", "Scala")),
paradoxRoots := List("examples/elasticsearch-samples.html",
- "examples/ftp-samples.html",
- "examples/jms-samples.html",
- "examples/mqtt-samples.html",
- "index.html"),
+ "examples/ftp-samples.html",
+ "examples/jms-samples.html",
+ "examples/mqtt-samples.html",
+ "index.html"),
resolvers += Resolver.jcenterRepo,
publishRsyncArtifacts += makeSite.value -> "www/",
publishRsyncHost := "akkarepo@gustav.akka.io",
- apidocRootPackage := "akka"
- )
+ apidocRootPackage := "akka")
lazy val testkit = internalProject("testkit", Dependencies.testkit)
@@ -424,12 +402,11 @@ lazy val `doc-examples` = project
.settings(
name := s"pekko-connectors-doc-examples",
publish / skip := true,
- Dependencies.`Doc-examples`
- )
+ Dependencies.`Doc-examples`)
def pekkoConnectorProject(projectId: String,
- moduleName: String,
- additionalSettings: sbt.Def.SettingsDefinition*): Project = {
+ moduleName: String,
+ additionalSettings: sbt.Def.SettingsDefinition*): Project = {
import com.typesafe.tools.mima.core._
Project(id = projectId, base = file(projectId))
.enablePlugins(AutomateHeaderPlugin)
@@ -439,16 +416,13 @@ def pekkoConnectorProject(projectId: String,
licenses := List(License.Apache2),
AutomaticModuleName.settings(s"akka.stream.alpakka.$moduleName"),
mimaPreviousArtifacts := Set(
- organization.value %% name.value % previousStableVersion.value
- .getOrElse("0.0.0")
- ),
+ organization.value %% name.value % previousStableVersion.value
+ .getOrElse("0.0.0")),
mimaBinaryIssueFilters ++= Seq(
- ProblemFilters.exclude[Problem]("*.impl.*"),
- // generated code
- ProblemFilters.exclude[Problem]("com.google.*")
- ),
- Test / parallelExecution := false
- )
+ ProblemFilters.exclude[Problem]("*.impl.*"),
+ // generated code
+ ProblemFilters.exclude[Problem]("com.google.*")),
+ Test / parallelExecution := false)
.settings(additionalSettings: _*)
.dependsOn(testkit % Test)
}
@@ -465,7 +439,6 @@ Global / onLoad := (Global / onLoad).value.andThen { s =>
val log = sLog.value
if (dynverGitDescribeOutput.value.hasNoTags)
log.error(
- s"Failed to derive version from git tags. Maybe run `git fetch --unshallow` or `git fetch upstream` on a fresh git clone from a fork? Derived version: $v"
- )
+ s"Failed to derive version from git tags. Maybe run `git fetch --unshallow` or `git fetch upstream` on a fresh git clone from a fork? Derived version: $v")
s
}
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/AkkaDiscoverySessionProvider.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/AkkaDiscoverySessionProvider.scala
index 727be3a5..dac15aee 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/AkkaDiscoverySessionProvider.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/AkkaDiscoverySessionProvider.scala
@@ -5,16 +5,16 @@
package akka.stream.alpakka.cassandra
import akka.ConfigurationException
-import akka.actor.{ActorSystem, ClassicActorSystemProvider}
+import akka.actor.{ ActorSystem, ClassicActorSystemProvider }
import akka.discovery.Discovery
import akka.util.JavaDurationConverters._
import com.datastax.oss.driver.api.core.CqlSession
-import com.typesafe.config.{Config, ConfigFactory}
+import com.typesafe.config.{ Config, ConfigFactory }
import scala.collection.immutable
import scala.compat.java8.FutureConverters._
import scala.concurrent.duration.FiniteDuration
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
/**
* [[https://doc.akka.io/docs/akka/current/discovery/index.html Akka Discovery]]
@@ -73,7 +73,7 @@ private[cassandra] object AkkaDiscoverySessionProvider {
* Expect a `service` section in Config and use Akka Discovery to read the addresses for `name` within `lookup-timeout`.
*/
private def readNodes(config: Config)(implicit system: ActorSystem,
- ec: ExecutionContext): Future[immutable.Seq[String]] = {
+ ec: ExecutionContext): Future[immutable.Seq[String]] = {
val serviceConfig = config.getConfig("service-discovery")
val serviceName = serviceConfig.getString("name")
val lookupTimeout = serviceConfig.getDuration("lookup-timeout").asScala
@@ -85,14 +85,13 @@ private[cassandra] object AkkaDiscoverySessionProvider {
*/
private def readNodes(
serviceName: String,
- lookupTimeout: FiniteDuration
- )(implicit system: ActorSystem, ec: ExecutionContext): Future[immutable.Seq[String]] = {
+ lookupTimeout: FiniteDuration)(
+ implicit system: ActorSystem, ec: ExecutionContext): Future[immutable.Seq[String]] = {
Discovery(system).discovery.lookup(serviceName, lookupTimeout).map { resolved =>
resolved.addresses.map { target =>
target.host + ":" + target.port.getOrElse {
throw new ConfigurationException(
- s"Akka Discovery for Cassandra service [$serviceName] must provide a port for [${target.host}]"
- )
+ s"Akka Discovery for Cassandra service [$serviceName] must provide a port for [${target.host}]")
}
}
}
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraMetricsRegistry.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraMetricsRegistry.scala
index 2b32180d..5fa716a6 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraMetricsRegistry.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraMetricsRegistry.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.cassandra
-import akka.actor.{ClassicActorSystemProvider, ExtendedActorSystem, Extension, ExtensionId, ExtensionIdProvider}
+import akka.actor.{ ClassicActorSystemProvider, ExtendedActorSystem, Extension, ExtensionId, ExtensionIdProvider }
import akka.annotation.InternalApi
import com.codahale.metrics.MetricRegistry
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraSessionSettings.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraSessionSettings.scala
index 85bfb957..f71f1089 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraSessionSettings.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraSessionSettings.scala
@@ -14,8 +14,8 @@ import scala.compat.java8.FutureConverters._
import scala.concurrent.Future
class CassandraSessionSettings private (val configPath: String,
- _metricsCategory: Option[String] = None,
- val init: Option[CqlSession => Future[Done]] = None) {
+ _metricsCategory: Option[String] = None,
+ val init: Option[CqlSession => Future[Done]] = None) {
def metricsCategory: String = _metricsCategory.getOrElse(configPath)
@@ -40,8 +40,8 @@ class CassandraSessionSettings private (val configPath: String,
def withInit(value: CqlSession => Future[Done]): CassandraSessionSettings = copy(init = Some(value))
private def copy(configPath: String = configPath,
- metricsCategory: Option[String] = _metricsCategory,
- init: Option[CqlSession => Future[Done]] = init) =
+ metricsCategory: Option[String] = _metricsCategory,
+ init: Option[CqlSession => Future[Done]] = init) =
new CassandraSessionSettings(configPath, metricsCategory, init)
override def toString: String =
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraWriteSettings.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraWriteSettings.scala
index 8b5fbf3a..29e6d0f1 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraWriteSettings.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CassandraWriteSettings.scala
@@ -7,12 +7,12 @@ package akka.stream.alpakka.cassandra
import akka.util.JavaDurationConverters._
import com.datastax.oss.driver.api.core.cql.BatchType
-import scala.concurrent.duration.{FiniteDuration, _}
+import scala.concurrent.duration.{ FiniteDuration, _ }
class CassandraWriteSettings private (val parallelism: Int,
- val maxBatchSize: Int,
- val maxBatchWait: FiniteDuration,
- val batchType: BatchType) {
+ val maxBatchSize: Int,
+ val maxBatchWait: FiniteDuration,
+ val batchType: BatchType) {
require(parallelism > 0, s"Invalid value for parallelism: $parallelism. It should be > 0.")
require(maxBatchSize > 0, s"Invalid value for maxBatchSize: $maxBatchSize. It should be > 0.")
@@ -43,9 +43,9 @@ class CassandraWriteSettings private (val parallelism: Int,
copy(batchType = value)
private def copy(parallelism: Int = parallelism,
- maxBatchSize: Int = maxBatchSize,
- maxBatchWait: FiniteDuration = maxBatchWait,
- batchType: BatchType = batchType) =
+ maxBatchSize: Int = maxBatchSize,
+ maxBatchWait: FiniteDuration = maxBatchWait,
+ batchType: BatchType = batchType) =
new CassandraWriteSettings(parallelism, maxBatchSize, maxBatchWait, batchType)
override def toString: String =
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CqlSessionProvider.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CqlSessionProvider.scala
index 1bdaced4..b7ff21ae 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CqlSessionProvider.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/CqlSessionProvider.scala
@@ -4,13 +4,13 @@
package akka.stream.alpakka.cassandra
-import akka.actor.{ActorSystem, ClassicActorSystemProvider, ExtendedActorSystem}
+import akka.actor.{ ActorSystem, ClassicActorSystemProvider, ExtendedActorSystem }
import com.datastax.oss.driver.api.core.CqlSession
-import com.typesafe.config.{Config, ConfigFactory}
+import com.typesafe.config.{ Config, ConfigFactory }
import scala.collection.immutable
import scala.compat.java8.FutureConverters._
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
import scala.util.Failure
/**
@@ -82,9 +82,7 @@ object CqlSessionProvider {
new IllegalArgumentException(
s"Unable to create SessionProvider instance for class [$className], " +
"tried constructor with ActorSystem, Config, and only ActorSystem, and no parameters",
- ex
- )
- )
+ ex))
}
.get
}
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraFlow.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraFlow.scala
index 254c8421..f41e7c79 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraFlow.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraFlow.scala
@@ -7,8 +7,8 @@ package akka.stream.alpakka.cassandra.javadsl
import akka.NotUsed
import akka.stream.alpakka.cassandra.CassandraWriteSettings
import akka.stream.alpakka.cassandra.scaladsl
-import akka.stream.javadsl.{Flow, FlowWithContext}
-import com.datastax.oss.driver.api.core.cql.{BoundStatement, PreparedStatement}
+import akka.stream.javadsl.{ Flow, FlowWithContext }
+import com.datastax.oss.driver.api.core.cql.{ BoundStatement, PreparedStatement }
/**
* Java API to create Cassandra flows.
@@ -26,13 +26,12 @@ object CassandraFlow {
* @tparam T stream element type
*/
def create[T](session: CassandraSession,
- writeSettings: CassandraWriteSettings,
- cqlStatement: String,
- statementBinder: akka.japi.Function2[T, PreparedStatement, BoundStatement]): Flow[T, T, NotUsed] =
+ writeSettings: CassandraWriteSettings,
+ cqlStatement: String,
+ statementBinder: akka.japi.Function2[T, PreparedStatement, BoundStatement]): Flow[T, T, NotUsed] =
scaladsl.CassandraFlow
.create(writeSettings, cqlStatement, (t, preparedStatement) => statementBinder.apply(t, preparedStatement))(
- session.delegate
- )
+ session.delegate)
.asJava
/**
@@ -50,12 +49,11 @@ object CassandraFlow {
session: CassandraSession,
writeSettings: CassandraWriteSettings,
cqlStatement: String,
- statementBinder: akka.japi.Function2[T, PreparedStatement, BoundStatement]
- ): FlowWithContext[T, Ctx, T, Ctx, NotUsed] = {
+ statementBinder: akka.japi.Function2[T, PreparedStatement, BoundStatement])
+ : FlowWithContext[T, Ctx, T, Ctx, NotUsed] = {
scaladsl.CassandraFlow
.withContext(writeSettings, cqlStatement, (t, preparedStatement) => statementBinder.apply(t, preparedStatement))(
- session.delegate
- )
+ session.delegate)
.asJava
}
@@ -81,15 +79,15 @@ object CassandraFlow {
* @tparam K extracted key type for grouping into batches
*/
def createUnloggedBatch[T, K](session: CassandraSession,
- writeSettings: CassandraWriteSettings,
- cqlStatement: String,
- statementBinder: (T, PreparedStatement) => BoundStatement,
- groupingKey: akka.japi.Function[T, K]): Flow[T, T, NotUsed] = {
+ writeSettings: CassandraWriteSettings,
+ cqlStatement: String,
+ statementBinder: (T, PreparedStatement) => BoundStatement,
+ groupingKey: akka.japi.Function[T, K]): Flow[T, T, NotUsed] = {
scaladsl.CassandraFlow
.createBatch(writeSettings,
- cqlStatement,
- (t, preparedStatement) => statementBinder.apply(t, preparedStatement),
- t => groupingKey.apply(t))(session.delegate)
+ cqlStatement,
+ (t, preparedStatement) => statementBinder.apply(t, preparedStatement),
+ t => groupingKey.apply(t))(session.delegate)
.asJava
}
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSession.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSession.scala
index 991c47d2..241d760d 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSession.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSession.scala
@@ -4,10 +4,10 @@
package akka.stream.alpakka.cassandra.javadsl
-import java.util.{List => JList}
+import java.util.{ List => JList }
import java.util.Optional
-import java.util.concurrent.{CompletionStage, Executor}
-import java.util.function.{Function => JFunction}
+import java.util.concurrent.{ CompletionStage, Executor }
+import java.util.function.{ Function => JFunction }
import scala.annotation.varargs
import scala.jdk.CollectionConverters._
@@ -16,11 +16,11 @@ import scala.compat.java8.OptionConverters._
import scala.concurrent.ExecutionContext
import akka.Done
import akka.NotUsed
-import akka.actor.{ActorSystem, ClassicActorSystemProvider}
+import akka.actor.{ ActorSystem, ClassicActorSystemProvider }
import akka.annotation.InternalApi
import akka.event.LoggingAdapter
import akka.stream.alpakka.cassandra.CassandraServerMetaData
-import akka.stream.alpakka.cassandra.{scaladsl, CqlSessionProvider}
+import akka.stream.alpakka.cassandra.{ scaladsl, CqlSessionProvider }
import akka.stream.javadsl.Source
import com.datastax.oss.driver.api.core.CqlSession
import com.datastax.oss.driver.api.core.cql.BatchStatement
@@ -46,32 +46,31 @@ final class CassandraSession(@InternalApi private[akka] val delegate: scaladsl.C
* Use this constructor if you want to create a stand-alone `CassandraSession`.
*/
def this(system: ActorSystem,
- sessionProvider: CqlSessionProvider,
- executionContext: ExecutionContext,
- log: LoggingAdapter,
- metricsCategory: String,
- init: JFunction[CqlSession, CompletionStage[Done]],
- onClose: java.lang.Runnable) =
+ sessionProvider: CqlSessionProvider,
+ executionContext: ExecutionContext,
+ log: LoggingAdapter,
+ metricsCategory: String,
+ init: JFunction[CqlSession, CompletionStage[Done]],
+ onClose: java.lang.Runnable) =
this(
new scaladsl.CassandraSession(system,
- sessionProvider,
- executionContext,
- log,
- metricsCategory,
- session => init.apply(session).toScala,
- () => onClose.run())
- )
+ sessionProvider,
+ executionContext,
+ log,
+ metricsCategory,
+ session => init.apply(session).toScala,
+ () => onClose.run()))
/**
* Use this constructor if you want to create a stand-alone `CassandraSession`.
*/
def this(system: ClassicActorSystemProvider,
- sessionProvider: CqlSessionProvider,
- executionContext: ExecutionContext,
- log: LoggingAdapter,
- metricsCategory: String,
- init: JFunction[CqlSession, CompletionStage[Done]],
- onClose: java.lang.Runnable) =
+ sessionProvider: CqlSessionProvider,
+ executionContext: ExecutionContext,
+ log: LoggingAdapter,
+ metricsCategory: String,
+ init: JFunction[CqlSession, CompletionStage[Done]],
+ onClose: java.lang.Runnable) =
this(system.classicSystem, sessionProvider, executionContext, log, metricsCategory, init, onClose)
implicit private val ec = delegate.ec
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSessionRegistry.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSessionRegistry.scala
index fa0ab63a..34eba615 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSessionRegistry.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSessionRegistry.scala
@@ -8,7 +8,7 @@ import java.util.concurrent.CompletionStage
import akka.Done
import akka.actor.ClassicActorSystemProvider
-import akka.stream.alpakka.cassandra.{scaladsl, CassandraSessionSettings}
+import akka.stream.alpakka.cassandra.{ scaladsl, CassandraSessionSettings }
import com.datastax.oss.driver.api.core.CqlSession
import scala.compat.java8.FutureConverters._
@@ -56,7 +56,7 @@ final class CassandraSessionRegistry private (delegate: scaladsl.CassandraSessio
* Sessions in the session registry are closed after actor system termination.
*/
def sessionFor(configPath: String,
- init: java.util.function.Function[CqlSession, CompletionStage[Done]]): CassandraSession =
+ init: java.util.function.Function[CqlSession, CompletionStage[Done]]): CassandraSession =
new CassandraSession(delegate.sessionFor(configPath, ses => init(ses).toScala))
/**
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSource.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSource.scala
index 066aa7b0..33caceda 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSource.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/javadsl/CassandraSource.scala
@@ -8,7 +8,7 @@ import java.util.concurrent.CompletionStage
import akka.NotUsed
import akka.stream.javadsl.Source
-import com.datastax.oss.driver.api.core.cql.{Row, Statement}
+import com.datastax.oss.driver.api.core.cql.{ Row, Statement }
import scala.annotation.varargs
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraFlow.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraFlow.scala
index 695a6f5d..20747ab2 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraFlow.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraFlow.scala
@@ -7,8 +7,8 @@ package akka.stream.alpakka.cassandra.scaladsl
import akka.NotUsed
import akka.dispatch.ExecutionContexts
import akka.stream.alpakka.cassandra.CassandraWriteSettings
-import akka.stream.scaladsl.{Flow, FlowWithContext}
-import com.datastax.oss.driver.api.core.cql.{BatchStatement, BoundStatement, PreparedStatement}
+import akka.stream.scaladsl.{ Flow, FlowWithContext }
+import com.datastax.oss.driver.api.core.cql.{ BatchStatement, BoundStatement, PreparedStatement }
import scala.jdk.CollectionConverters._
import scala.concurrent.Future
@@ -31,8 +31,8 @@ object CassandraFlow {
def create[T](
writeSettings: CassandraWriteSettings,
cqlStatement: String,
- statementBinder: (T, PreparedStatement) => BoundStatement
- )(implicit session: CassandraSession): Flow[T, T, NotUsed] = {
+ statementBinder: (T, PreparedStatement) => BoundStatement)(
+ implicit session: CassandraSession): Flow[T, T, NotUsed] = {
Flow
.lazyFutureFlow { () =>
val prepare = session.prepare(cqlStatement)
@@ -61,8 +61,8 @@ object CassandraFlow {
def withContext[T, Ctx](
writeSettings: CassandraWriteSettings,
cqlStatement: String,
- statementBinder: (T, PreparedStatement) => BoundStatement
- )(implicit session: CassandraSession): FlowWithContext[T, Ctx, T, Ctx, NotUsed] = {
+ statementBinder: (T, PreparedStatement) => BoundStatement)(
+ implicit session: CassandraSession): FlowWithContext[T, Ctx, T, Ctx, NotUsed] = {
FlowWithContext.fromTuples {
Flow
.lazyFutureFlow { () =>
@@ -102,9 +102,9 @@ object CassandraFlow {
* @tparam K extracted key type for grouping into batches
*/
def createBatch[T, K](writeSettings: CassandraWriteSettings,
- cqlStatement: String,
- statementBinder: (T, PreparedStatement) => BoundStatement,
- groupingKey: T => K)(implicit session: CassandraSession): Flow[T, T, NotUsed] = {
+ cqlStatement: String,
+ statementBinder: (T, PreparedStatement) => BoundStatement,
+ groupingKey: T => K)(implicit session: CassandraSession): Flow[T, T, NotUsed] = {
Flow
.lazyFutureFlow { () =>
val prepareStatement: Future[PreparedStatement] = session.prepare(cqlStatement)
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSession.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSession.scala
index dc424155..d05219e8 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSession.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSession.scala
@@ -7,18 +7,18 @@ package akka.stream.alpakka.cassandra.scaladsl
import akka.actor.NoSerializationVerificationNeeded
import akka.annotation.InternalApi
import akka.event.LoggingAdapter
-import akka.stream.alpakka.cassandra.{CassandraMetricsRegistry, CassandraServerMetaData, CqlSessionProvider}
-import akka.stream.scaladsl.{Sink, Source}
-import akka.stream.{Materializer, SystemMaterializer}
+import akka.stream.alpakka.cassandra.{ CassandraMetricsRegistry, CassandraServerMetaData, CqlSessionProvider }
+import akka.stream.scaladsl.{ Sink, Source }
+import akka.stream.{ Materializer, SystemMaterializer }
import akka.util.OptionVal
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import com.datastax.oss.driver.api.core.CqlSession
import com.datastax.oss.driver.api.core.cql._
import com.datastax.oss.driver.api.core.servererrors.InvalidQueryException
import scala.collection.immutable
import scala.compat.java8.FutureConverters._
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
import scala.util.control.NonFatal
/**
@@ -34,12 +34,12 @@ import scala.util.control.NonFatal
* All methods are non-blocking.
*/
final class CassandraSession(system: akka.actor.ActorSystem,
- sessionProvider: CqlSessionProvider,
- executionContext: ExecutionContext,
- log: LoggingAdapter,
- metricsCategory: String,
- init: CqlSession => Future[Done],
- onClose: () => Unit)
+ sessionProvider: CqlSessionProvider,
+ executionContext: ExecutionContext,
+ log: LoggingAdapter,
+ metricsCategory: String,
+ init: CqlSession => Future[Done],
+ onClose: () => Unit)
extends NoSerializationVerificationNeeded {
implicit private[akka] val ec: ExecutionContext = executionContext
@@ -92,8 +92,8 @@ final class CassandraSession(system: akka.actor.ActorSystem,
val result = selectOne("select cluster_name, data_center, release_version from system.local").map {
case Some(row) =>
new CassandraServerMetaData(row.getString("cluster_name"),
- row.getString("data_center"),
- row.getString("release_version"))
+ row.getString("data_center"),
+ row.getString("release_version"))
case None =>
log.warning("Couldn't retrieve serverMetaData from system.local table. No rows found.")
new CassandraServerMetaData("", "", "")
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionRegistry.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionRegistry.scala
index 3c5c71a5..820cd81a 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionRegistry.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionRegistry.scala
@@ -10,10 +10,10 @@ import scala.jdk.CollectionConverters._
import scala.concurrent.ExecutionContext
import scala.concurrent.Future
import akka.Done
-import akka.actor.{ClassicActorSystemProvider, ExtendedActorSystem, Extension, ExtensionId, ExtensionIdProvider}
+import akka.actor.{ ClassicActorSystemProvider, ExtendedActorSystem, Extension, ExtensionId, ExtensionIdProvider }
import akka.annotation.InternalStableApi
import akka.event.Logging
-import akka.stream.alpakka.cassandra.{CassandraSessionSettings, CqlSessionProvider}
+import akka.stream.alpakka.cassandra.{ CassandraSessionSettings, CqlSessionProvider }
import com.datastax.oss.driver.api.core.CqlSession
import com.typesafe.config.Config
@@ -82,24 +82,24 @@ final class CassandraSessionRegistry(system: ExtendedActorSystem) extends Extens
* that is different from the ActorSystem's config section for the `configPath`.
*/
@InternalStableApi private[akka] def sessionFor(settings: CassandraSessionSettings,
- sessionProviderConfig: Config): CassandraSession = {
+ sessionProviderConfig: Config): CassandraSession = {
val key = sessionKey(settings)
sessions.computeIfAbsent(key, _ => startSession(settings, key, sessionProviderConfig))
}
private def startSession(settings: CassandraSessionSettings,
- key: SessionKey,
- sessionProviderConfig: Config): CassandraSession = {
+ key: SessionKey,
+ sessionProviderConfig: Config): CassandraSession = {
val sessionProvider = CqlSessionProvider(system, sessionProviderConfig)
val log = Logging(system, classOf[CassandraSession])
val executionContext = system.dispatchers.lookup(sessionProviderConfig.getString("session-dispatcher"))
new CassandraSession(system,
- sessionProvider,
- executionContext,
- log,
- metricsCategory = settings.metricsCategory,
- init = settings.init.getOrElse(_ => Future.successful(Done)),
- onClose = () => sessions.remove(key))
+ sessionProvider,
+ executionContext,
+ log,
+ metricsCategory = settings.metricsCategory,
+ init = settings.init.getOrElse(_ => Future.successful(Done)),
+ onClose = () => sessions.remove(key))
}
/**
diff --git a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSource.scala b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSource.scala
index ca41113e..a0d3148b 100644
--- a/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSource.scala
+++ b/cassandra/src/main/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSource.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.cassandra.scaladsl
import akka.NotUsed
import akka.stream.scaladsl.Source
-import com.datastax.oss.driver.api.core.cql.{Row, Statement}
+import com.datastax.oss.driver.api.core.cql.{ Row, Statement }
import scala.concurrent.Future
diff --git a/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraLifecycle.scala b/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraLifecycle.scala
index 75010305..87e09462 100644
--- a/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraLifecycle.scala
+++ b/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraLifecycle.scala
@@ -11,12 +11,12 @@ import akka.Done
import akka.testkit.TestKitBase
import com.datastax.oss.driver.api.core.cql._
import org.scalatest._
-import org.scalatest.concurrent.{PatienceConfiguration, ScalaFutures}
+import org.scalatest.concurrent.{ PatienceConfiguration, ScalaFutures }
import scala.jdk.CollectionConverters._
import scala.collection.immutable
import scala.concurrent.duration._
-import scala.concurrent.{Await, Future}
+import scala.concurrent.{ Await, Future }
import scala.util.control.NonFatal
import scala.compat.java8.FutureConverters._
@@ -40,16 +40,14 @@ trait CassandraLifecycleBase {
def createKeyspace(session: CassandraSession, name: String): Future[Done] = {
session.executeWrite(
new SimpleStatementBuilder(
- s"""CREATE KEYSPACE $name WITH replication = { 'class': 'SimpleStrategy', 'replication_factor': '1'};"""
- ).setTimeout(keyspaceTimeout)
- .build()
- )
+ s"""CREATE KEYSPACE $name WITH replication = { 'class': 'SimpleStrategy', 'replication_factor': '1'};""").setTimeout(
+ keyspaceTimeout)
+ .build())
}
def dropKeyspace(session: CassandraSession, name: String): Future[Done] =
session.executeWrite(
- new SimpleStatementBuilder(s"""DROP KEYSPACE IF EXISTS $name;""").setTimeout(keyspaceTimeout).build()
- )
+ new SimpleStatementBuilder(s"""DROP KEYSPACE IF EXISTS $name;""").setTimeout(keyspaceTimeout).build())
def createKeyspace(name: String): Future[Done] = withSchemaMetadataDisabled(createKeyspace(lifecycleSession, name))
diff --git a/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionPerformanceSpec.scala b/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionPerformanceSpec.scala
index 15374219..f7899220 100644
--- a/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionPerformanceSpec.scala
+++ b/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSessionPerformanceSpec.scala
@@ -8,8 +8,8 @@ import scala.concurrent.Await
import akka.actor.ActorSystem
import akka.event.Logging
-import akka.stream.alpakka.cassandra.{CassandraSessionSettings, CassandraWriteSettings}
-import akka.stream.scaladsl.{Sink, Source}
+import akka.stream.alpakka.cassandra.{ CassandraSessionSettings, CassandraWriteSettings }
+import akka.stream.scaladsl.{ Sink, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import scala.concurrent.duration._
@@ -51,8 +51,7 @@ final class CassandraSessionPerformanceSpec extends CassandraSpecBase(ActorSyste
CassandraWriteSettings.create().withMaxBatchSize(10000),
s"INSERT INTO $dataTable(partition_id, id, value, seq_nr) VALUES (?, ?, ?, ?)",
(d: Int, ps) => ps.bind(Long.box(partitionId), id, Long.box(d), Long.box(d)),
- (_: Int) => partitionId
- )(lifecycleSession)
+ (_: Int) => partitionId)(lifecycleSession)
}
.runWith(Sink.ignore)
}
diff --git a/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSpecBase.scala b/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSpecBase.scala
index feb8e1bf..d016464a 100644
--- a/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSpecBase.scala
+++ b/cassandra/src/test/scala/akka/stream/alpakka/cassandra/scaladsl/CassandraSpecBase.scala
@@ -9,10 +9,10 @@ import akka.stream.alpakka.testkit.scaladsl.LogCapturing
import akka.testkit.TestKit
import org.scalatest.matchers.must.Matchers
import org.scalatest.wordspec.AnyWordSpecLike
-import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach}
+import org.scalatest.{ BeforeAndAfterAll, BeforeAndAfterEach }
import scala.concurrent.ExecutionContext
-import akka.stream.{Materializer, SystemMaterializer}
+import akka.stream.{ Materializer, SystemMaterializer }
/**
* All the tests must be run with a local Cassandra running on default port 9042.
diff --git a/cassandra/src/test/scala/docs/javadsl/CassandraSessionSpec.scala b/cassandra/src/test/scala/docs/javadsl/CassandraSessionSpec.scala
index b968de45..d928add0 100644
--- a/cassandra/src/test/scala/docs/javadsl/CassandraSessionSpec.scala
+++ b/cassandra/src/test/scala/docs/javadsl/CassandraSessionSpec.scala
@@ -53,9 +53,7 @@ final class CassandraSessionSpec extends CassandraSpecBase(ActorSystem("Cassandr
s"INSERT INTO $dataTable (partition, key, count) VALUES ('A', 'c', 3);",
s"INSERT INTO $dataTable (partition, key, count) VALUES ('A', 'd', 4);",
s"INSERT INTO $dataTable (partition, key, count) VALUES ('B', 'e', 5);",
- s"INSERT INTO $dataTable (partition, key, count) VALUES ('B', 'f', 6);"
- )
- )
+ s"INSERT INTO $dataTable (partition, key, count) VALUES ('B', 'f', 6);"))
} yield Done
}.futureValue mustBe Done
}
@@ -138,15 +136,13 @@ final class CassandraSessionSpec extends CassandraSpecBase(ActorSystem("Cassandr
"create indexes" in {
withSchemaMetadataDisabled(
- lifecycleSession.executeDDL(s"CREATE INDEX IF NOT EXISTS count_idx ON $dataTable(count)")
- ).futureValue mustBe Done
+ lifecycleSession.executeDDL(
+ s"CREATE INDEX IF NOT EXISTS count_idx ON $dataTable(count)")).futureValue mustBe Done
val row =
await(
- session.selectOne("SELECT * FROM system_schema.indexes WHERE table_name = ? ALLOW FILTERING", dataTableName)
- )
+ session.selectOne("SELECT * FROM system_schema.indexes WHERE table_name = ? ALLOW FILTERING", dataTableName))
row.asScala.map(index => index.getString("table_name") -> index.getString("index_name")) mustBe Some(
- dataTableName -> "count_idx"
- )
+ dataTableName -> "count_idx")
}
}
diff --git a/cassandra/src/test/scala/docs/scaladsl/AkkaDiscoverySpec.scala b/cassandra/src/test/scala/docs/scaladsl/AkkaDiscoverySpec.scala
index cc2cac3f..b59e90a2 100644
--- a/cassandra/src/test/scala/docs/scaladsl/AkkaDiscoverySpec.scala
+++ b/cassandra/src/test/scala/docs/scaladsl/AkkaDiscoverySpec.scala
@@ -7,7 +7,7 @@ package docs.scaladsl
import akka.Done
import akka.actor.ActorSystem
import akka.stream.alpakka.cassandra.CassandraSessionSettings
-import akka.stream.alpakka.cassandra.scaladsl.{CassandraSession, CassandraSessionRegistry, CassandraSpecBase}
+import akka.stream.alpakka.cassandra.scaladsl.{ CassandraSession, CassandraSessionRegistry, CassandraSpecBase }
import akka.stream.scaladsl.Sink
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
diff --git a/cassandra/src/test/scala/docs/scaladsl/CassandraFlowSpec.scala b/cassandra/src/test/scala/docs/scaladsl/CassandraFlowSpec.scala
index 91ab3521..70c4bd3b 100644
--- a/cassandra/src/test/scala/docs/scaladsl/CassandraFlowSpec.scala
+++ b/cassandra/src/test/scala/docs/scaladsl/CassandraFlowSpec.scala
@@ -4,11 +4,11 @@
package docs.scaladsl
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import akka.actor.ActorSystem
-import akka.stream.alpakka.cassandra.{CassandraSessionSettings, CassandraWriteSettings}
-import akka.stream.alpakka.cassandra.scaladsl.{CassandraFlow, CassandraSession, CassandraSource, CassandraSpecBase}
-import akka.stream.scaladsl.{Sink, Source, SourceWithContext}
+import akka.stream.alpakka.cassandra.{ CassandraSessionSettings, CassandraWriteSettings }
+import akka.stream.alpakka.cassandra.scaladsl.{ CassandraFlow, CassandraSession, CassandraSource, CassandraSpecBase }
+import akka.stream.scaladsl.{ Sink, Source, SourceWithContext }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import scala.collection.immutable
@@ -16,9 +16,9 @@ import scala.concurrent.Future
class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec")) {
- //#element-to-insert
+ // #element-to-insert
case class ToInsert(id: Integer, cc: Integer)
- //#element-to-insert
+ // #element-to-insert
val sessionSettings = CassandraSessionSettings("alpakka.cassandra")
val data = 1 until 103
@@ -40,9 +40,8 @@ class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec
val written: Future[Done] = Source(data)
.via(
CassandraFlow.create(CassandraWriteSettings.defaults,
- s"INSERT INTO $table(id) VALUES (?)",
- (element, preparedStatement) => preparedStatement.bind(Int.box(element)))
- )
+ s"INSERT INTO $table(id) VALUES (?)",
+ (element, preparedStatement) => preparedStatement.bind(Int.box(element))))
.runWith(Sink.ignore)
written.futureValue mustBe Done
@@ -65,7 +64,7 @@ class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec
// #prepared
import akka.stream.alpakka.cassandra.CassandraWriteSettings
import akka.stream.alpakka.cassandra.scaladsl.CassandraFlow
- import com.datastax.oss.driver.api.core.cql.{BoundStatement, PreparedStatement}
+ import com.datastax.oss.driver.api.core.cql.{ BoundStatement, PreparedStatement }
case class Person(id: Int, name: String, city: String)
@@ -78,9 +77,8 @@ class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec
val written: Future[immutable.Seq[Person]] = Source(persons)
.via(
CassandraFlow.create(CassandraWriteSettings.defaults,
- s"INSERT INTO $table(id, name, city) VALUES (?, ?, ?)",
- statementBinder)
- )
+ s"INSERT INTO $table(id, name, city) VALUES (?, ?, ?)",
+ statementBinder))
.runWith(Sink.seq)
// #prepared
@@ -112,8 +110,8 @@ class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec
}
val persons =
immutable.Seq(Person(12, "John", "London") -> AckHandle(12),
- Person(43, "Umberto", "Roma") -> AckHandle(43),
- Person(56, "James", "Chicago") -> AckHandle(56))
+ Person(43, "Umberto", "Roma") -> AckHandle(43),
+ Person(56, "James", "Chicago") -> AckHandle(56))
// #withContext
val personsAndHandles: SourceWithContext[Person, AckHandle, NotUsed] = // ???
@@ -126,9 +124,7 @@ class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec
CassandraFlow.withContext(
CassandraWriteSettings.defaults,
s"INSERT INTO $table(id, name, city) VALUES (?, ?, ?)",
- (person, preparedStatement) => preparedStatement.bind(Int.box(person.id), person.name, person.city)
- )
- )
+ (person, preparedStatement) => preparedStatement.bind(Int.box(person.id), person.name, person.city)))
.asSource
.mapAsync(1) {
case (_, handle) => handle.ack()
@@ -169,9 +165,7 @@ class CassandraFlowSpec extends CassandraSpecBase(ActorSystem("CassandraFlowSpec
s"INSERT INTO $table(id, name, city) VALUES (?, ?, ?)",
statementBinder =
(person, preparedStatement) => preparedStatement.bind(Int.box(person.id), person.name, person.city),
- groupingKey = person => person.id
- )
- )
+ groupingKey = person => person.id))
.runWith(Sink.ignore)
written.futureValue mustBe Done
diff --git a/cassandra/src/test/scala/docs/scaladsl/CassandraSourceSpec.scala b/cassandra/src/test/scala/docs/scaladsl/CassandraSourceSpec.scala
index df808eac..0c2a97bd 100644
--- a/cassandra/src/test/scala/docs/scaladsl/CassandraSourceSpec.scala
+++ b/cassandra/src/test/scala/docs/scaladsl/CassandraSourceSpec.scala
@@ -6,7 +6,7 @@ package docs.scaladsl
import akka.Done
import akka.actor.ActorSystem
-import akka.stream.alpakka.cassandra.scaladsl.{CassandraSession, CassandraSource, CassandraSpecBase}
+import akka.stream.alpakka.cassandra.scaladsl.{ CassandraSession, CassandraSource, CassandraSpecBase }
import akka.stream.scaladsl.Sink
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
@@ -15,9 +15,9 @@ import scala.concurrent.Future
class CassandraSourceSpec extends CassandraSpecBase(ActorSystem("CassandraSourceSpec")) {
- //#element-to-insert
+ // #element-to-insert
case class ToInsert(id: Integer, cc: Integer)
- //#element-to-insert
+ // #element-to-insert
val sessionSettings = akka.stream.alpakka.cassandra.CassandraSessionSettings()
val data = 1 until 103
@@ -80,7 +80,7 @@ class CassandraSourceSpec extends CassandraSpecBase(ActorSystem("CassandraSource
"stream the result of a Cassandra statement with several pages" in assertAllStagesStopped {
// #statement
- import com.datastax.oss.driver.api.core.cql.{Row, SimpleStatement}
+ import com.datastax.oss.driver.api.core.cql.{ Row, SimpleStatement }
val stmt = SimpleStatement.newInstance(s"SELECT * FROM $intTable").setPageSize(20)
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseResponseException.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseResponseException.scala
index 9cd876bf..1ff29f69 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseResponseException.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseResponseException.scala
@@ -20,6 +20,5 @@ private[akka] object CouchbaseResponseException {
def apply(json: JsonObject): CouchbaseResponseException =
new CouchbaseResponseException(
msg = if (json.containsKey("msg")) json.getString("msg") else "",
- code = if (json.containsKey("code")) Some(json.getInt("code")) else None
- )
+ code = if (json.containsKey("code")) Some(json.getInt("code")) else None)
}
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseSessionRegistry.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseSessionRegistry.scala
index 342d025f..b806f5ea 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseSessionRegistry.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/CouchbaseSessionRegistry.scala
@@ -7,15 +7,15 @@ package akka.stream.alpakka.couchbase
import java.util.concurrent.CompletionStage
import java.util.concurrent.atomic.AtomicReference
-import akka.actor.{ClassicActorSystemProvider, ExtendedActorSystem, Extension, ExtensionId, ExtensionIdProvider}
+import akka.actor.{ ClassicActorSystemProvider, ExtendedActorSystem, Extension, ExtensionId, ExtensionIdProvider }
import akka.dispatch.ExecutionContexts
import akka.stream.alpakka.couchbase.impl.CouchbaseClusterRegistry
-import akka.stream.alpakka.couchbase.javadsl.{CouchbaseSession => JCouchbaseSession}
+import akka.stream.alpakka.couchbase.javadsl.{ CouchbaseSession => JCouchbaseSession }
import akka.stream.alpakka.couchbase.scaladsl.CouchbaseSession
import scala.annotation.tailrec
import scala.compat.java8.FutureConverters._
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* This Couchbase session registry makes it possible to share Couchbase sessions between multiple use sites
@@ -65,7 +65,7 @@ final class CouchbaseSessionRegistry(system: ExtendedActorSystem) extends Extens
val key = SessionKey(enrichedSettings, bucketName)
sessions.get.get(key) match {
case Some(futureSession) => futureSession
- case _ => startSession(key)
+ case _ => startSession(key)
}
}(system.dispatcher)
@@ -91,8 +91,7 @@ final class CouchbaseSessionRegistry(system: ExtendedActorSystem) extends Extens
val session = clusterRegistry
.clusterFor(key.settings)
.flatMap(cluster => CouchbaseSession(cluster, key.bucketName)(blockingDispatcher))(
- ExecutionContexts.parasitic
- )
+ ExecutionContexts.parasitic)
promise.completeWith(session)
promise.future
} else {
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseClusterRegistry.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseClusterRegistry.scala
index 529f914c..7adbc43a 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseClusterRegistry.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseClusterRegistry.scala
@@ -14,7 +14,7 @@ import akka.stream.alpakka.couchbase.scaladsl.CouchbaseSession
import com.couchbase.client.java.AsyncCluster
import scala.annotation.tailrec
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* Internal API
@@ -31,7 +31,7 @@ final private[couchbase] class CouchbaseClusterRegistry(system: ActorSystem) {
def clusterFor(settings: CouchbaseSessionSettings): Future[AsyncCluster] =
clusters.get.get(settings) match {
case Some(futureSession) => futureSession
- case _ => createClusterClient(settings)
+ case _ => createClusterClient(settings)
}
@tailrec
@@ -45,8 +45,7 @@ final private[couchbase] class CouchbaseClusterRegistry(system: ActorSystem) {
log.info("Starting Couchbase client for nodes [{}]", nodesAsString)
promise.completeWith(
CouchbaseSession
- .createClusterClient(settings)(blockingDispatcher)
- )
+ .createClusterClient(settings)(blockingDispatcher))
val future = promise.future
system.registerOnTermination {
future.foreach { cluster =>
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionImpl.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionImpl.scala
index 88bef879..cd3667d7 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionImpl.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionImpl.scala
@@ -9,15 +9,15 @@ import java.util.concurrent.TimeUnit
import akka.annotation.InternalApi
import akka.dispatch.ExecutionContexts
import akka.stream.alpakka.couchbase.scaladsl.CouchbaseSession
-import akka.stream.alpakka.couchbase.{javadsl, CouchbaseWriteSettings}
+import akka.stream.alpakka.couchbase.{ javadsl, CouchbaseWriteSettings }
import akka.stream.scaladsl.Source
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import com.couchbase.client.java.bucket.AsyncBucketManager
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import com.couchbase.client.java.document.{ Document, JsonDocument }
import com.couchbase.client.java.query.util.IndexInfo
-import com.couchbase.client.java.query.{N1qlQuery, Statement}
-import com.couchbase.client.java.{AsyncBucket, AsyncCluster}
+import com.couchbase.client.java.query.{ N1qlQuery, Statement }
+import com.couchbase.client.java.{ AsyncBucket, AsyncCluster }
import rx.RxReactiveStreams
import scala.concurrent.Future
@@ -47,11 +47,11 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
def insertDoc[T <: Document[_]](document: T, writeSettings: CouchbaseWriteSettings): Future[T] =
singleObservableToFuture(asyncBucket.insert(document,
- writeSettings.persistTo,
- writeSettings.replicateTo,
- writeSettings.timeout.toMillis,
- TimeUnit.MILLISECONDS),
- document)
+ writeSettings.persistTo,
+ writeSettings.replicateTo,
+ writeSettings.timeout.toMillis,
+ TimeUnit.MILLISECONDS),
+ document)
def get(id: String): Future[Option[JsonDocument]] =
zeroOrOneObservableToFuture(asyncBucket.get(id))
@@ -63,8 +63,8 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
zeroOrOneObservableToFuture(asyncBucket.get(id, timeout.toMillis, TimeUnit.MILLISECONDS))
def get[T <: Document[_]](id: String,
- timeout: FiniteDuration,
- documentClass: Class[T]): scala.concurrent.Future[Option[T]] =
+ timeout: FiniteDuration,
+ documentClass: Class[T]): scala.concurrent.Future[Option[T]] =
zeroOrOneObservableToFuture(asyncBucket.get(id, documentClass, timeout.toMillis, TimeUnit.MILLISECONDS))
def upsert(document: JsonDocument): Future[JsonDocument] = upsertDoc(document)
@@ -77,11 +77,11 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
def upsertDoc[T <: Document[_]](document: T, writeSettings: CouchbaseWriteSettings): Future[T] =
singleObservableToFuture(asyncBucket.upsert(document,
- writeSettings.persistTo,
- writeSettings.replicateTo,
- writeSettings.timeout.toMillis,
- TimeUnit.MILLISECONDS),
- document.id)
+ writeSettings.persistTo,
+ writeSettings.replicateTo,
+ writeSettings.timeout.toMillis,
+ TimeUnit.MILLISECONDS),
+ document.id)
def replace(document: JsonDocument): Future[JsonDocument] = replaceDoc(document)
@@ -93,11 +93,11 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
def replaceDoc[T <: Document[_]](document: T, writeSettings: CouchbaseWriteSettings): Future[T] =
singleObservableToFuture(asyncBucket.replace(document,
- writeSettings.persistTo,
- writeSettings.replicateTo,
- writeSettings.timeout.toMillis,
- TimeUnit.MILLISECONDS),
- document.id)
+ writeSettings.persistTo,
+ writeSettings.replicateTo,
+ writeSettings.timeout.toMillis,
+ TimeUnit.MILLISECONDS),
+ document.id)
def remove(id: String): Future[Done] =
singleObservableToFuture(asyncBucket.remove(id), id)
@@ -105,11 +105,11 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
def remove(id: String, writeSettings: CouchbaseWriteSettings): Future[Done] =
singleObservableToFuture(asyncBucket.remove(id,
- writeSettings.persistTo,
- writeSettings.replicateTo,
- writeSettings.timeout.toMillis,
- TimeUnit.MILLISECONDS),
- id)
+ writeSettings.persistTo,
+ writeSettings.replicateTo,
+ writeSettings.timeout.toMillis,
+ TimeUnit.MILLISECONDS),
+ id)
.map(_ => Done)(ExecutionContexts.parasitic)
def streamedQuery(query: N1qlQuery): Source[JsonObject, NotUsed] =
@@ -130,13 +130,13 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
def counter(id: String, delta: Long, initial: Long, writeSettings: CouchbaseWriteSettings): Future[Long] =
singleObservableToFuture(asyncBucket.counter(id,
- delta,
- initial,
- writeSettings.persistTo,
- writeSettings.replicateTo,
- writeSettings.timeout.toMillis,
- TimeUnit.MILLISECONDS),
- id)
+ delta,
+ initial,
+ writeSettings.persistTo,
+ writeSettings.replicateTo,
+ writeSettings.timeout.toMillis,
+ TimeUnit.MILLISECONDS),
+ id)
.map(_.content(): Long)(ExecutionContexts.parasitic)
def close(): Future[Done] =
@@ -162,11 +162,8 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
.flatMap(
func1Observable[AsyncBucketManager, Boolean](
_.createN1qlIndex(indexName, ignoreIfExist, false, fields: _*)
- .map(func1(Boolean.unbox))
- )
- ),
- s"Create index: $indexName"
- )
+ .map(func1(Boolean.unbox)))),
+ s"Create index: $indexName")
override def listIndexes(): Source[IndexInfo, NotUsed] =
Source.fromPublisher(
@@ -174,9 +171,6 @@ final private[couchbase] class CouchbaseSessionImpl(asyncBucket: AsyncBucket, cl
asyncBucket
.bucketManager()
.flatMap(
- func1Observable((abm: AsyncBucketManager) => abm.listN1qlIndexes())
- )
- )
- )
+ func1Observable((abm: AsyncBucketManager) => abm.listN1qlIndexes()))))
}
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionJavaAdapter.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionJavaAdapter.scala
index bf87ed35..43eb25a6 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionJavaAdapter.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/CouchbaseSessionJavaAdapter.scala
@@ -14,17 +14,17 @@ import akka.stream.alpakka.couchbase.CouchbaseWriteSettings
import akka.stream.alpakka.couchbase.javadsl
import akka.stream.alpakka.couchbase.scaladsl
import akka.stream.javadsl.Source
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import com.couchbase.client.java.AsyncBucket
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import com.couchbase.client.java.document.{ Document, JsonDocument }
import com.couchbase.client.java.query.util.IndexInfo
-import com.couchbase.client.java.query.{N1qlQuery, Statement}
+import com.couchbase.client.java.query.{ N1qlQuery, Statement }
import scala.compat.java8.FutureConverters._
import scala.compat.java8.OptionConverters._
import scala.concurrent.duration.FiniteDuration
-import scala.concurrent.{duration, Future}
+import scala.concurrent.{ duration, Future }
/**
* INTERNAL API
@@ -43,13 +43,12 @@ private[couchbase] final class CouchbaseSessionJavaAdapter(delegate: scaladsl.Co
override def insert(
document: JsonDocument,
- writeSettings: CouchbaseWriteSettings
- ): CompletionStage[JsonDocument] = delegate.insert(document, writeSettings).toJava
+ writeSettings: CouchbaseWriteSettings): CompletionStage[JsonDocument] =
+ delegate.insert(document, writeSettings).toJava
override def insertDoc[T <: Document[_]](
document: T,
- writeSettings: CouchbaseWriteSettings
- ): CompletionStage[T] = delegate.insertDoc(document, writeSettings).toJava
+ writeSettings: CouchbaseWriteSettings): CompletionStage[T] = delegate.insertDoc(document, writeSettings).toJava
override def get(id: String): CompletionStage[Optional[JsonDocument]] =
futureOptToJava(delegate.get(id))
@@ -107,8 +106,8 @@ private[couchbase] final class CouchbaseSessionJavaAdapter(delegate: scaladsl.Co
id: String,
delta: Long,
initial: Long,
- writeSettings: CouchbaseWriteSettings
- ): CompletionStage[Long] = delegate.counter(id, delta, initial, writeSettings).toJava
+ writeSettings: CouchbaseWriteSettings): CompletionStage[Long] =
+ delegate.counter(id, delta, initial, writeSettings).toJava
override def close(): CompletionStage[Done] = delegate.close().toJava
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/RxUtilities.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/RxUtilities.scala
index dc3411a8..b2858bf2 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/RxUtilities.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/impl/RxUtilities.scala
@@ -7,11 +7,11 @@ package akka.stream.alpakka.couchbase.impl
import akka.annotation.InternalApi
import akka.stream.alpakka.couchbase.CouchbaseResponseException
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.query.{AsyncN1qlQueryResult, AsyncN1qlQueryRow}
+import com.couchbase.client.java.query.{ AsyncN1qlQueryResult, AsyncN1qlQueryRow }
import rx.functions.Func1
-import rx.{Observable, Subscriber}
+import rx.{ Observable, Subscriber }
-import scala.concurrent.{Future, Promise}
+import scala.concurrent.{ Future, Promise }
/**
* INTERNAL API
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseFlow.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseFlow.scala
index c9cac312..efbfbe81 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseFlow.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseFlow.scala
@@ -7,7 +7,7 @@ package akka.stream.alpakka.couchbase.javadsl
import akka.NotUsed
import akka.stream.alpakka.couchbase._
import akka.stream.javadsl.Flow
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import com.couchbase.client.java.document.{ Document, JsonDocument }
/**
* Java API: Factory methods for Couchbase flows.
@@ -24,24 +24,24 @@ object CouchbaseFlow {
* Create a flow to query Couchbase for by `id` and emit documents of the given class.
*/
def fromId[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- bucketName: String,
- target: Class[T]): Flow[String, T, NotUsed] =
+ bucketName: String,
+ target: Class[T]): Flow[String, T, NotUsed] =
scaladsl.CouchbaseFlow.fromId(sessionSettings, bucketName, target).asJava
/**
* Create a flow to update or insert a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def upsert(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
scaladsl.CouchbaseFlow.upsert(sessionSettings, writeSettings, bucketName).asJava
/**
* Create a flow to update or insert a Couchbase document of the given class.
*/
def upsertDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, T, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, T, NotUsed] =
scaladsl.CouchbaseFlow.upsertDoc(sessionSettings, writeSettings, bucketName).asJava
/**
@@ -49,24 +49,24 @@ object CouchbaseFlow {
* can be handled in-stream.
*/
def upsertDocWithResult[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
scaladsl.CouchbaseFlow.upsertDocWithResult(sessionSettings, writeSettings, bucketName).asJava
/**
* Create a flow to replace a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def replace(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
scaladsl.CouchbaseFlow.replace(sessionSettings, writeSettings, bucketName).asJava
/**
* Create a flow to replace a Couchbase document of the given class.
*/
def replaceDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, T, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, T, NotUsed] =
scaladsl.CouchbaseFlow.replaceDoc(sessionSettings, writeSettings, bucketName).asJava
/**
@@ -74,24 +74,24 @@ object CouchbaseFlow {
* can be handled in-stream.
*/
def replaceDocWithResult[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
scaladsl.CouchbaseFlow.replaceDocWithResult(sessionSettings, writeSettings, bucketName).asJava
/**
* Create a flow to delete documents from Couchbase by `id`. Emits the same `id`.
*/
def delete(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[String, String, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[String, String, NotUsed] =
scaladsl.CouchbaseFlow.delete(sessionSettings, writeSettings, bucketName).asJava
/**
* Create a flow to delete documents from Couchbase by `id` and emit operation outcome containing the same `id`.
*/
def deleteWithResult(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[String, CouchbaseDeleteResult, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[String, CouchbaseDeleteResult, NotUsed] =
scaladsl.CouchbaseFlow.deleteWithResult(sessionSettings, writeSettings, bucketName).asJava
}
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSession.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSession.scala
index 560e8c0f..9a1e2f23 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSession.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSession.scala
@@ -6,20 +6,20 @@ package akka.stream.alpakka.couchbase.javadsl
import java.time.Duration
import java.util.Optional
-import java.util.concurrent.{CompletionStage, Executor}
+import java.util.concurrent.{ CompletionStage, Executor }
import akka.annotation.DoNotInherit
import akka.dispatch.ExecutionContexts
import akka.stream.alpakka.couchbase.impl.CouchbaseSessionJavaAdapter
-import akka.stream.alpakka.couchbase.scaladsl.{CouchbaseSession => ScalaDslCouchbaseSession}
-import akka.stream.alpakka.couchbase.{CouchbaseSessionSettings, CouchbaseWriteSettings}
+import akka.stream.alpakka.couchbase.scaladsl.{ CouchbaseSession => ScalaDslCouchbaseSession }
+import akka.stream.alpakka.couchbase.{ CouchbaseSessionSettings, CouchbaseWriteSettings }
import akka.stream.javadsl.Source
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import com.couchbase.client.java.document.{ Document, JsonDocument }
import com.couchbase.client.java.query.util.IndexInfo
-import com.couchbase.client.java.query.{N1qlQuery, Statement}
-import com.couchbase.client.java.{AsyncBucket, AsyncCluster, Bucket}
+import com.couchbase.client.java.query.{ N1qlQuery, Statement }
+import com.couchbase.client.java.{ AsyncBucket, AsyncCluster, Bucket }
import scala.compat.java8.FutureConverters._
import scala.concurrent.ExecutionContext
@@ -36,13 +36,12 @@ object CouchbaseSession {
* the session is closed.
*/
def create(settings: CouchbaseSessionSettings,
- bucketName: String,
- executor: Executor): CompletionStage[CouchbaseSession] =
+ bucketName: String,
+ executor: Executor): CompletionStage[CouchbaseSession] =
ScalaDslCouchbaseSession
.apply(settings, bucketName)(executionContext(executor))
.map(new CouchbaseSessionJavaAdapter(_): CouchbaseSession)(
- ExecutionContexts.parasitic
- )
+ ExecutionContexts.parasitic)
.toJava
/**
@@ -52,8 +51,7 @@ object CouchbaseSession {
def create(client: AsyncCluster, bucketName: String, executor: Executor): CompletionStage[CouchbaseSession] =
ScalaDslCouchbaseSession(client, bucketName)(executionContext(executor))
.map(new CouchbaseSessionJavaAdapter(_): CouchbaseSession)(
- ExecutionContexts.parasitic
- )
+ ExecutionContexts.parasitic)
.toJava
/**
@@ -68,7 +66,7 @@ object CouchbaseSession {
private def executionContext(executor: Executor): ExecutionContext =
executor match {
case ec: ExecutionContext => ec
- case _ => ExecutionContext.fromExecutor(executor)
+ case _ => ExecutionContext.fromExecutor(executor)
}
/**
@@ -269,7 +267,7 @@ abstract class CouchbaseSession {
* @return a [[java.util.concurrent.CompletionStage]] of `true` if the index was/will be effectively created, `false`
* if the index existed and ignoreIfExist` is true. Completion of the `CompletionStage` does not guarantee the index
* is online and ready to be used.
- **/
+ */
def createIndex(indexName: String, ignoreIfExist: Boolean, fields: AnyRef*): CompletionStage[Boolean]
/**
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSink.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSink.scala
index 144fc420..593f6a01 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSink.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSink.scala
@@ -7,9 +7,9 @@ package akka.stream.alpakka.couchbase.javadsl
import java.util.concurrent.CompletionStage
import akka.stream.alpakka.couchbase._
-import akka.stream.javadsl.{Keep, Sink}
-import akka.{Done, NotUsed}
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import akka.stream.javadsl.{ Keep, Sink }
+import akka.{ Done, NotUsed }
+import com.couchbase.client.java.document.{ Document, JsonDocument }
/**
* Java API: Factory methods for Couchbase sinks.
@@ -20,8 +20,8 @@ object CouchbaseSink {
* Create a sink to update or insert a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def upsert(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[JsonDocument, CompletionStage[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[JsonDocument, CompletionStage[Done]] =
CouchbaseFlow
.upsert(sessionSettings, writeSettings, bucketName)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
@@ -30,8 +30,8 @@ object CouchbaseSink {
* Create a sink to update or insert a Couchbase document of the given class.
*/
def upsertDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[T, CompletionStage[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[T, CompletionStage[Done]] =
CouchbaseFlow
.upsertDoc[T](sessionSettings, writeSettings, bucketName)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
@@ -40,8 +40,8 @@ object CouchbaseSink {
* Create a sink to replace a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def replace(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[JsonDocument, CompletionStage[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[JsonDocument, CompletionStage[Done]] =
CouchbaseFlow
.replace(sessionSettings, writeSettings, bucketName)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
@@ -50,8 +50,8 @@ object CouchbaseSink {
* Create a sink to replace a Couchbase document of the given class.
*/
def replaceDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[T, CompletionStage[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[T, CompletionStage[Done]] =
CouchbaseFlow
.replaceDoc[T](sessionSettings, writeSettings, bucketName)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
@@ -60,8 +60,8 @@ object CouchbaseSink {
* Create a sink to delete documents from Couchbase by `id`.
*/
def delete(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[String, CompletionStage[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[String, CompletionStage[Done]] =
CouchbaseFlow
.delete(sessionSettings, writeSettings, bucketName)
.toMat(Sink.ignore(), Keep.right[NotUsed, CompletionStage[Done]])
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSource.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSource.scala
index 8cac9af1..0750bdbf 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSource.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/javadsl/CouchbaseSource.scala
@@ -5,10 +5,10 @@
package akka.stream.alpakka.couchbase.javadsl
import akka.NotUsed
-import akka.stream.alpakka.couchbase.{scaladsl, CouchbaseSessionSettings}
+import akka.stream.alpakka.couchbase.{ scaladsl, CouchbaseSessionSettings }
import akka.stream.javadsl.Source
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.query.{N1qlQuery, Statement}
+import com.couchbase.client.java.query.{ N1qlQuery, Statement }
/**
* Java API: Factory methods for Couchbase sources.
@@ -19,16 +19,16 @@ object CouchbaseSource {
* Create a source query Couchbase by statement, emitted as [[com.couchbase.client.java.document.JsonDocument JsonDocument]]s.
*/
def fromStatement(sessionSettings: CouchbaseSessionSettings,
- statement: Statement,
- bucketName: String): Source[JsonObject, NotUsed] =
+ statement: Statement,
+ bucketName: String): Source[JsonObject, NotUsed] =
scaladsl.CouchbaseSource.fromStatement(sessionSettings, statement, bucketName).asJava
/**
* Create a source query Couchbase by statement, emitted as [[com.couchbase.client.java.document.JsonDocument JsonDocument]]s.
*/
def fromN1qlQuery(sessionSettings: CouchbaseSessionSettings,
- query: N1qlQuery,
- bucketName: String): Source[JsonObject, NotUsed] =
+ query: N1qlQuery,
+ bucketName: String): Source[JsonObject, NotUsed] =
scaladsl.CouchbaseSource.fromN1qlQuery(sessionSettings, query, bucketName).asJava
}
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/model.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/model.scala
index 71696d05..6eb1d283 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/model.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/model.scala
@@ -4,13 +4,13 @@
package akka.stream.alpakka.couchbase
-import java.util.concurrent.{CompletionStage, TimeUnit}
+import java.util.concurrent.{ CompletionStage, TimeUnit }
-import akka.actor.{ActorSystem, ClassicActorSystemProvider}
+import akka.actor.{ ActorSystem, ClassicActorSystemProvider }
import akka.annotation.InternalApi
import com.couchbase.client.java.document.Document
import com.couchbase.client.java.env.CouchbaseEnvironment
-import com.couchbase.client.java.{PersistTo, ReplicateTo}
+import com.couchbase.client.java.{ PersistTo, ReplicateTo }
import com.typesafe.config.Config
import scala.jdk.CollectionConverters._
@@ -32,21 +32,21 @@ object CouchbaseWriteSettings {
def apply(): CouchbaseWriteSettings = inMemory
def apply(parallelism: Int,
- replicateTo: ReplicateTo,
- persistTo: PersistTo,
- timeout: FiniteDuration): CouchbaseWriteSettings =
+ replicateTo: ReplicateTo,
+ persistTo: PersistTo,
+ timeout: FiniteDuration): CouchbaseWriteSettings =
new CouchbaseWriteSettings(parallelism, replicateTo, persistTo, timeout)
def create(): CouchbaseWriteSettings = inMemory
def create(parallelism: Int,
- replicateTo: ReplicateTo,
- persistTo: PersistTo,
- timeout: java.time.Duration): CouchbaseWriteSettings =
+ replicateTo: ReplicateTo,
+ persistTo: PersistTo,
+ timeout: java.time.Duration): CouchbaseWriteSettings =
new CouchbaseWriteSettings(parallelism,
- replicateTo,
- persistTo,
- FiniteDuration(timeout.toMillis, TimeUnit.MILLISECONDS))
+ replicateTo,
+ persistTo,
+ FiniteDuration(timeout.toMillis, TimeUnit.MILLISECONDS))
}
@@ -54,9 +54,9 @@ object CouchbaseWriteSettings {
* Configure Couchbase writes.
*/
final class CouchbaseWriteSettings private (val parallelism: Int,
- val replicateTo: ReplicateTo,
- val persistTo: PersistTo,
- val timeout: FiniteDuration) {
+ val replicateTo: ReplicateTo,
+ val persistTo: PersistTo,
+ val timeout: FiniteDuration) {
def withParallelism(parallelism: Int): CouchbaseWriteSettings = copy(parallelism = parallelism)
@@ -76,9 +76,9 @@ final class CouchbaseWriteSettings private (val parallelism: Int,
def withTimeout(timeout: FiniteDuration): CouchbaseWriteSettings = copy(timeout = timeout)
private[this] def copy(parallelism: Int = parallelism,
- replicateTo: ReplicateTo = replicateTo,
- persistTo: PersistTo = persistTo,
- timeout: FiniteDuration = timeout) =
+ replicateTo: ReplicateTo = replicateTo,
+ persistTo: PersistTo = persistTo,
+ timeout: FiniteDuration = timeout) =
new CouchbaseWriteSettings(parallelism, replicateTo, persistTo, timeout)
override def equals(other: Any): Boolean = other match {
@@ -174,8 +174,7 @@ final class CouchbaseSessionSettings private (
val password: String,
val nodes: immutable.Seq[String],
val environment: Option[CouchbaseEnvironment],
- val enrichAsync: CouchbaseSessionSettings => Future[CouchbaseSessionSettings]
-) {
+ val enrichAsync: CouchbaseSessionSettings => Future[CouchbaseSessionSettings]) {
def withUsername(username: String): CouchbaseSessionSettings =
copy(username = username)
@@ -193,18 +192,20 @@ final class CouchbaseSessionSettings private (
def withNodes(nodes: java.util.List[String]): CouchbaseSessionSettings =
copy(nodes = nodes.asScala.toList)
- /** Scala API:
+ /**
+ * Scala API:
* Allows to provide an asynchronous method to update the settings.
*/
def withEnrichAsync(value: CouchbaseSessionSettings => Future[CouchbaseSessionSettings]): CouchbaseSessionSettings =
copy(enrichAsync = value)
- /** Java API:
+ /**
+ * Java API:
* Allows to provide an asynchronous method to update the settings.
*/
def withEnrichAsyncCs(
- value: java.util.function.Function[CouchbaseSessionSettings, CompletionStage[CouchbaseSessionSettings]]
- ): CouchbaseSessionSettings =
+ value: java.util.function.Function[CouchbaseSessionSettings, CompletionStage[CouchbaseSessionSettings]])
+ : CouchbaseSessionSettings =
copy(enrichAsync = (s: CouchbaseSessionSettings) => value.apply(s).toScala)
def withEnvironment(environment: CouchbaseEnvironment): CouchbaseSessionSettings =
@@ -222,8 +223,8 @@ final class CouchbaseSessionSettings private (
password: String = password,
nodes: immutable.Seq[String] = nodes,
environment: Option[CouchbaseEnvironment] = environment,
- enrichAsync: CouchbaseSessionSettings => Future[CouchbaseSessionSettings] = enrichAsync
- ): CouchbaseSessionSettings =
+ enrichAsync: CouchbaseSessionSettings => Future[CouchbaseSessionSettings] = enrichAsync)
+ : CouchbaseSessionSettings =
new CouchbaseSessionSettings(username, password, nodes, environment, enrichAsync)
override def equals(other: Any): Boolean = other match {
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseFlow.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseFlow.scala
index 0e570ce5..e872713a 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseFlow.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseFlow.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.couchbase.scaladsl
import akka.NotUsed
import akka.stream.alpakka.couchbase._
import akka.stream.scaladsl.Flow
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import com.couchbase.client.java.document.{ Document, JsonDocument }
/**
* Scala API: Factory methods for Couchbase flows.
@@ -30,8 +30,8 @@ object CouchbaseFlow {
* Create a flow to query Couchbase for by `id` and emit documents of the given class.
*/
def fromId[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- bucketName: String,
- target: Class[T]): Flow[String, T, NotUsed] =
+ bucketName: String,
+ target: Class[T]): Flow[String, T, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
@@ -45,15 +45,14 @@ object CouchbaseFlow {
* Create a flow to update or insert a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def upsert(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[JsonDocument]
- .mapAsync(writeSettings.parallelism)(
- doc => session.flatMap(_.upsert(doc, writeSettings))(materializer.system.dispatcher)
- )
+ .mapAsync(writeSettings.parallelism)(doc =>
+ session.flatMap(_.upsert(doc, writeSettings))(materializer.system.dispatcher))
}
.mapMaterializedValue(_ => NotUsed)
@@ -61,15 +60,14 @@ object CouchbaseFlow {
* Create a flow to update or insert a Couchbase document of the given class.
*/
def upsertDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, T, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, T, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[T]
- .mapAsync(writeSettings.parallelism)(
- doc => session.flatMap(_.upsertDoc(doc, writeSettings))(materializer.system.dispatcher)
- )
+ .mapAsync(writeSettings.parallelism)(doc =>
+ session.flatMap(_.upsertDoc(doc, writeSettings))(materializer.system.dispatcher))
}
.mapMaterializedValue(_ => NotUsed)
@@ -78,23 +76,21 @@ object CouchbaseFlow {
* can be handled in-stream.
*/
def upsertDocWithResult[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[T]
- .mapAsync(writeSettings.parallelism)(
- doc => {
- implicit val executor = materializer.system.dispatcher
- session
- .flatMap(_.upsertDoc(doc, writeSettings))
- .map(_ => CouchbaseWriteSuccess(doc))
- .recover {
- case exception => CouchbaseWriteFailure(doc, exception)
- }
- }
- )
+ .mapAsync(writeSettings.parallelism)(doc => {
+ implicit val executor = materializer.system.dispatcher
+ session
+ .flatMap(_.upsertDoc(doc, writeSettings))
+ .map(_ => CouchbaseWriteSuccess(doc))
+ .recover {
+ case exception => CouchbaseWriteFailure(doc, exception)
+ }
+ })
}
.mapMaterializedValue(_ => NotUsed)
@@ -102,15 +98,14 @@ object CouchbaseFlow {
* Create a flow to replace a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def replace(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[JsonDocument, JsonDocument, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[JsonDocument]
- .mapAsync(writeSettings.parallelism)(
- doc => session.flatMap(_.replace(doc, writeSettings))(materializer.system.dispatcher)
- )
+ .mapAsync(writeSettings.parallelism)(doc =>
+ session.flatMap(_.replace(doc, writeSettings))(materializer.system.dispatcher))
}
.mapMaterializedValue(_ => NotUsed)
@@ -118,15 +113,14 @@ object CouchbaseFlow {
* Create a flow to replace a Couchbase document of the given class.
*/
def replaceDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, T, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, T, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[T]
- .mapAsync(writeSettings.parallelism)(
- doc => session.flatMap(_.replaceDoc(doc, writeSettings))(materializer.system.dispatcher)
- )
+ .mapAsync(writeSettings.parallelism)(doc =>
+ session.flatMap(_.replaceDoc(doc, writeSettings))(materializer.system.dispatcher))
}
.mapMaterializedValue(_ => NotUsed)
@@ -135,23 +129,21 @@ object CouchbaseFlow {
* can be handled in-stream.
*/
def replaceDocWithResult[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[T, CouchbaseWriteResult[T], NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[T]
- .mapAsync(writeSettings.parallelism)(
- doc => {
- implicit val executor = materializer.system.dispatcher
- session
- .flatMap(_.replaceDoc(doc, writeSettings))
- .map(_ => CouchbaseWriteSuccess(doc))
- .recover {
- case exception => CouchbaseWriteFailure(doc, exception)
- }
- }
- )
+ .mapAsync(writeSettings.parallelism)(doc => {
+ implicit val executor = materializer.system.dispatcher
+ session
+ .flatMap(_.replaceDoc(doc, writeSettings))
+ .map(_ => CouchbaseWriteSuccess(doc))
+ .recover {
+ case exception => CouchbaseWriteFailure(doc, exception)
+ }
+ })
}
.mapMaterializedValue(_ => NotUsed)
@@ -159,20 +151,18 @@ object CouchbaseFlow {
* Create a flow to delete documents from Couchbase by `id`. Emits the same `id`.
*/
def delete(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[String, String, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[String, String, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[String]
- .mapAsync(writeSettings.parallelism)(
- id => {
- implicit val executor = materializer.system.dispatcher
- session
- .flatMap(_.remove(id, writeSettings))
- .map(_ => id)
- }
- )
+ .mapAsync(writeSettings.parallelism)(id => {
+ implicit val executor = materializer.system.dispatcher
+ session
+ .flatMap(_.remove(id, writeSettings))
+ .map(_ => id)
+ })
}
.mapMaterializedValue(_ => NotUsed)
@@ -180,23 +170,21 @@ object CouchbaseFlow {
* Create a flow to delete documents from Couchbase by `id` and emit operation outcome containing the same `id`.
*/
def deleteWithResult(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Flow[String, CouchbaseDeleteResult, NotUsed] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Flow[String, CouchbaseDeleteResult, NotUsed] =
Flow
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
Flow[String]
- .mapAsync(writeSettings.parallelism)(
- id => {
- implicit val executor = materializer.system.dispatcher
- session
- .flatMap(_.remove(id, writeSettings))
- .map(_ => CouchbaseDeleteSuccess(id))
- .recover {
- case exception => CouchbaseDeleteFailure(id, exception)
- }
- }
- )
+ .mapAsync(writeSettings.parallelism)(id => {
+ implicit val executor = materializer.system.dispatcher
+ session
+ .flatMap(_.remove(id, writeSettings))
+ .map(_ => CouchbaseDeleteSuccess(id))
+ .recover {
+ case exception => CouchbaseDeleteFailure(id, exception)
+ }
+ })
}
.mapMaterializedValue(_ => NotUsed)
}
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSession.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSession.scala
index b3679a27..8700bd2a 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSession.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSession.scala
@@ -4,19 +4,19 @@
package akka.stream.alpakka.couchbase.scaladsl
-import akka.annotation.{DoNotInherit, InternalApi}
-import akka.stream.alpakka.couchbase.impl.{CouchbaseSessionImpl, RxUtilities}
-import akka.stream.alpakka.couchbase.javadsl.{CouchbaseSession => JavaDslCouchbaseSession}
-import akka.stream.alpakka.couchbase.{CouchbaseSessionSettings, CouchbaseWriteSettings}
+import akka.annotation.{ DoNotInherit, InternalApi }
+import akka.stream.alpakka.couchbase.impl.{ CouchbaseSessionImpl, RxUtilities }
+import akka.stream.alpakka.couchbase.javadsl.{ CouchbaseSession => JavaDslCouchbaseSession }
+import akka.stream.alpakka.couchbase.{ CouchbaseSessionSettings, CouchbaseWriteSettings }
import akka.stream.scaladsl.Source
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import com.couchbase.client.java._
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import com.couchbase.client.java.document.{ Document, JsonDocument }
import com.couchbase.client.java.query._
import com.couchbase.client.java.query.util.IndexInfo
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
import scala.concurrent.duration.FiniteDuration
/**
@@ -31,7 +31,7 @@ object CouchbaseSession {
* the session is closed.
*/
def apply(settings: CouchbaseSessionSettings,
- bucketName: String)(implicit ec: ExecutionContext): Future[CouchbaseSession] =
+ bucketName: String)(implicit ec: ExecutionContext): Future[CouchbaseSession] =
createClusterClient(settings).flatMap(c => openBucket(c, disconnectClusterOnClose = true, bucketName))
/**
@@ -56,8 +56,7 @@ object CouchbaseSession {
*/
@InternalApi
private[couchbase] def createClusterClient(
- settings: CouchbaseSessionSettings
- )(implicit ec: ExecutionContext): Future[AsyncCluster] =
+ settings: CouchbaseSessionSettings)(implicit ec: ExecutionContext): Future[AsyncCluster] =
settings.enriched
.flatMap { enrichedSettings =>
Future(enrichedSettings.environment match {
@@ -69,8 +68,7 @@ object CouchbaseSession {
}
private def openBucket(cluster: AsyncCluster, disconnectClusterOnClose: Boolean, bucketName: String)(
- implicit ec: ExecutionContext
- ): Future[CouchbaseSession] =
+ implicit ec: ExecutionContext): Future[CouchbaseSession] =
RxUtilities
.singleObservableToFuture(cluster.openBucket(bucketName), "openBucket")
.map { bucket =>
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSink.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSink.scala
index 8d8542e0..7d058951 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSink.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSink.scala
@@ -6,8 +6,8 @@ package akka.stream.alpakka.couchbase.scaladsl
import akka.Done
import akka.stream.alpakka.couchbase._
-import akka.stream.scaladsl.{Keep, Sink}
-import com.couchbase.client.java.document.{Document, JsonDocument}
+import akka.stream.scaladsl.{ Keep, Sink }
+import com.couchbase.client.java.document.{ Document, JsonDocument }
import scala.concurrent.Future
@@ -20,16 +20,16 @@ object CouchbaseSink {
* Create a sink to update or insert a Couchbase [[com.couchbase.client.java.document.JsonDocument JsonDocument]].
*/
def upsert(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[JsonDocument, Future[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[JsonDocument, Future[Done]] =
CouchbaseFlow.upsert(sessionSettings, writeSettings, bucketName).toMat(Sink.ignore)(Keep.right)
/**
* Create a sink to update or insert a Couchbase document of the given class.
*/
def upsertDoc[T <: Document[_]](sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[T, Future[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[T, Future[Done]] =
CouchbaseFlow
.upsertDoc(sessionSettings, writeSettings, bucketName)
.toMat(Sink.ignore)(Keep.right)
@@ -38,8 +38,8 @@ object CouchbaseSink {
* Create a sink to delete documents from Couchbase by `id`.
*/
def delete(sessionSettings: CouchbaseSessionSettings,
- writeSettings: CouchbaseWriteSettings,
- bucketName: String): Sink[String, Future[Done]] =
+ writeSettings: CouchbaseWriteSettings,
+ bucketName: String): Sink[String, Future[Done]] =
CouchbaseFlow.delete(sessionSettings, writeSettings, bucketName).toMat(Sink.ignore)(Keep.right)
}
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSource.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSource.scala
index 61f801e8..62aadca2 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSource.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/CouchbaseSource.scala
@@ -5,10 +5,10 @@
package akka.stream.alpakka.couchbase.scaladsl
import akka.NotUsed
-import akka.stream.alpakka.couchbase.{CouchbaseSessionRegistry, CouchbaseSessionSettings}
+import akka.stream.alpakka.couchbase.{ CouchbaseSessionRegistry, CouchbaseSessionSettings }
import akka.stream.scaladsl.Source
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.query.{N1qlQuery, Statement}
+import com.couchbase.client.java.query.{ N1qlQuery, Statement }
/**
* Scala API: Factory methods for Couchbase sources.
@@ -19,8 +19,8 @@ object CouchbaseSource {
* Create a source query Couchbase by statement, emitted as [[com.couchbase.client.java.document.JsonDocument JsonDocument]]s.
*/
def fromStatement(sessionSettings: CouchbaseSessionSettings,
- statement: Statement,
- bucketName: String): Source[JsonObject, NotUsed] =
+ statement: Statement,
+ bucketName: String): Source[JsonObject, NotUsed] =
Source
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
@@ -34,8 +34,8 @@ object CouchbaseSource {
* Create a source query Couchbase by statement, emitted as [[com.couchbase.client.java.document.JsonDocument JsonDocument]]s.
*/
def fromN1qlQuery(sessionSettings: CouchbaseSessionSettings,
- query: N1qlQuery,
- bucketName: String): Source[JsonObject, NotUsed] =
+ query: N1qlQuery,
+ bucketName: String): Source[JsonObject, NotUsed] =
Source
.fromMaterializer { (materializer, _) =>
val session = CouchbaseSessionRegistry(materializer.system).sessionFor(sessionSettings, bucketName)
diff --git a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/DiscoverySupport.scala b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/DiscoverySupport.scala
index a77a0030..b026c8b8 100644
--- a/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/DiscoverySupport.scala
+++ b/couchbase/src/main/scala/akka/stream/alpakka/couchbase/scaladsl/DiscoverySupport.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.couchbase.scaladsl
import java.util.concurrent.CompletionStage
-import akka.actor.{ActorSystem, ClassicActorSystemProvider}
+import akka.actor.{ ActorSystem, ClassicActorSystemProvider }
import akka.annotation.InternalApi
import akka.discovery.Discovery
import akka.stream.alpakka.couchbase.CouchbaseSessionSettings
@@ -29,8 +29,7 @@ sealed class DiscoverySupport private {
*/
private def readNodes(
serviceName: String,
- lookupTimeout: FiniteDuration
- )(implicit system: ClassicActorSystemProvider): Future[immutable.Seq[String]] = {
+ lookupTimeout: FiniteDuration)(implicit system: ClassicActorSystemProvider): Future[immutable.Seq[String]] = {
implicit val ec = system.classicSystem.dispatcher
val discovery = Discovery(system).discovery
discovery.lookup(serviceName, lookupTimeout).map { resolved =>
@@ -53,8 +52,8 @@ sealed class DiscoverySupport private {
* to be used as Couchbase `nodes`.
*/
def nodes(
- config: Config
- )(implicit system: ClassicActorSystemProvider): CouchbaseSessionSettings => Future[CouchbaseSessionSettings] = {
+ config: Config)(
+ implicit system: ClassicActorSystemProvider): CouchbaseSessionSettings => Future[CouchbaseSessionSettings] = {
implicit val ec = system.classicSystem.dispatcher
settings =>
readNodes(config)
@@ -64,7 +63,7 @@ sealed class DiscoverySupport private {
}
private[couchbase] def nodes(config: Config,
- system: ActorSystem): CouchbaseSessionSettings => Future[CouchbaseSessionSettings] =
+ system: ActorSystem): CouchbaseSessionSettings => Future[CouchbaseSessionSettings] =
nodes(config)(system)
/**
@@ -73,8 +72,8 @@ sealed class DiscoverySupport private {
@InternalApi
private[couchbase] def getNodes(
config: Config,
- system: ClassicActorSystemProvider
- ): java.util.function.Function[CouchbaseSessionSettings, CompletionStage[CouchbaseSessionSettings]] =
+ system: ClassicActorSystemProvider)
+ : java.util.function.Function[CouchbaseSessionSettings, CompletionStage[CouchbaseSessionSettings]] =
nodes(config)(system).andThen(_.toJava).asJava
/**
@@ -82,8 +81,7 @@ sealed class DiscoverySupport private {
* to be used as Couchbase `nodes`.
*/
def nodes()(
- implicit system: ClassicActorSystemProvider
- ): CouchbaseSessionSettings => Future[CouchbaseSessionSettings] =
+ implicit system: ClassicActorSystemProvider): CouchbaseSessionSettings => Future[CouchbaseSessionSettings] =
nodes(system.classicSystem)
/**
diff --git a/couchbase/src/test/scala/akka/stream/alpakka/couchbase/testing/CouchbaseSupport.scala b/couchbase/src/test/scala/akka/stream/alpakka/couchbase/testing/CouchbaseSupport.scala
index 587e3485..89cc28a4 100644
--- a/couchbase/src/test/scala/akka/stream/alpakka/couchbase/testing/CouchbaseSupport.scala
+++ b/couchbase/src/test/scala/akka/stream/alpakka/couchbase/testing/CouchbaseSupport.scala
@@ -7,13 +7,13 @@ package akka.stream.alpakka.couchbase.testing
import akka.Done
import akka.actor.ActorSystem
import akka.stream.alpakka.couchbase.scaladsl._
-import akka.stream.alpakka.couchbase.{CouchbaseSessionSettings, CouchbaseWriteSettings}
-import akka.stream.scaladsl.{Sink, Source}
+import akka.stream.alpakka.couchbase.{ CouchbaseSessionSettings, CouchbaseWriteSettings }
+import akka.stream.scaladsl.{ Sink, Source }
import com.couchbase.client.deps.io.netty.buffer.Unpooled
import com.couchbase.client.deps.io.netty.util.CharsetUtil
import com.couchbase.client.java.ReplicateTo
import com.couchbase.client.java.document.json.JsonObject
-import com.couchbase.client.java.document.{BinaryDocument, JsonDocument, RawJsonDocument, StringDocument}
+import com.couchbase.client.java.document.{ BinaryDocument, JsonDocument, RawJsonDocument, StringDocument }
import org.slf4j.LoggerFactory
import play.api.libs.json.Json
@@ -21,7 +21,7 @@ import scala.jdk.CollectionConverters._
import scala.collection.immutable.Seq
import scala.concurrent.ExecutionContext.Implicits.global
import scala.concurrent.duration._
-import scala.concurrent.{Await, Future}
+import scala.concurrent.{ Await, Future }
case class TestObject(id: String, value: String)
@@ -29,15 +29,15 @@ trait CouchbaseSupport {
private val log = LoggerFactory.getLogger(classOf[CouchbaseSupport])
- //#init-actor-system
+ // #init-actor-system
implicit val actorSystem: ActorSystem = ActorSystem()
- //#init-actor-system
+ // #init-actor-system
val sampleData = TestObject("First", "First")
val sampleSequence: Seq[TestObject] = sampleData +: Seq[TestObject](TestObject("Second", "Second"),
- TestObject("Third", "Third"),
- TestObject("Fourth", "Fourth"))
+ TestObject("Third", "Third"),
+ TestObject("Fourth", "Fourth"))
val sampleJavaList: java.util.List[TestObject] = sampleSequence.asJava
@@ -78,7 +78,7 @@ trait CouchbaseSupport {
.via(CouchbaseFlow.upsert(sessionSettings, CouchbaseWriteSettings.inMemory, bucketName))
.runWith(Sink.ignore)
Await.result(bulkUpsertResult, 5.seconds)
- //all queries are Eventual Consistent, se we need to wait for index refresh!!
+ // all queries are Eventual Consistent, se we need to wait for index refresh!!
Thread.sleep(2000)
}
diff --git a/couchbase/src/test/scala/docs/scaladsl/CouchbaseFlowSpec.scala b/couchbase/src/test/scala/docs/scaladsl/CouchbaseFlowSpec.scala
index f91a55d9..2bf8f701 100644
--- a/couchbase/src/test/scala/docs/scaladsl/CouchbaseFlowSpec.scala
+++ b/couchbase/src/test/scala/docs/scaladsl/CouchbaseFlowSpec.scala
@@ -5,22 +5,22 @@
package docs.scaladsl
import akka.Done
-import akka.stream.alpakka.couchbase.{CouchbaseDeleteFailure, CouchbaseDeleteResult}
+import akka.stream.alpakka.couchbase.{ CouchbaseDeleteFailure, CouchbaseDeleteResult }
import akka.stream.alpakka.couchbase.scaladsl.CouchbaseFlow
-import akka.stream.alpakka.couchbase.testing.{CouchbaseSupport, TestObject}
+import akka.stream.alpakka.couchbase.testing.{ CouchbaseSupport, TestObject }
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
-import akka.stream.scaladsl.{Sink, Source}
+import akka.stream.scaladsl.{ Sink, Source }
import com.couchbase.client.java.error.DocumentDoesNotExistException
import org.scalatest.concurrent.ScalaFutures
import org.scalatest._
//#write-settings
import akka.stream.alpakka.couchbase.CouchbaseWriteSettings
-import com.couchbase.client.java.{PersistTo, ReplicateTo}
+import com.couchbase.client.java.{ PersistTo, ReplicateTo }
//#write-settings
import akka.stream.testkit.scaladsl.StreamTestKit._
-import com.couchbase.client.java.document.{BinaryDocument, RawJsonDocument, StringDocument}
+import com.couchbase.client.java.document.{ BinaryDocument, RawJsonDocument, StringDocument }
import scala.collection.immutable
import scala.collection.immutable.Seq
@@ -60,13 +60,13 @@ class CouchbaseFlowSpec
"create custom writeSettings object" in {
- //#write-settings
+ // #write-settings
val writeSettings = CouchbaseWriteSettings()
.withParallelism(3)
.withPersistTo(PersistTo.FOUR)
.withReplicateTo(ReplicateTo.THREE)
.withTimeout(5.seconds)
- //#write-settings
+ // #write-settings
val expectedwriteSettings = CouchbaseWriteSettings(3, ReplicateTo.THREE, PersistTo.FOUR, 5.seconds)
writeSettings shouldEqual expectedwriteSettings
@@ -84,9 +84,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDoc(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
result.futureValue
@@ -110,9 +108,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsert(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
// #upsert
jsonDocumentUpsert.futureValue
@@ -132,9 +128,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDoc(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
// #upsert
stringDocumentUpsert.futureValue
@@ -163,9 +157,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDoc(
sessionSettings,
writeSettings.withParallelism(2),
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
bulkUpsertResult.futureValue
@@ -182,8 +174,7 @@ class CouchbaseFlowSpec
val bulkUpsertResult: Future[Done] = Source(sampleSequence)
.map(toJsonDocument)
.via(
- CouchbaseFlow.upsert(sessionSettings, writeSettings.withParallelism(2), bucketName)
- )
+ CouchbaseFlow.upsert(sessionSettings, writeSettings.withParallelism(2), bucketName))
.runWith(Sink.ignore)
bulkUpsertResult.futureValue
@@ -196,9 +187,7 @@ class CouchbaseFlowSpec
.via(
CouchbaseFlow.fromId(
sessionSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.seq)
// #fromId
@@ -212,9 +201,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDoc(
sessionSettings,
writeSettings.withParallelism(2),
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
bulkUpsertResult.futureValue
@@ -224,9 +211,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.fromId(
sessionSettings,
bucketName,
- classOf[StringDocument]
- )
- )
+ classOf[StringDocument]))
.runWith(Sink.seq)
resultsAsFuture.futureValue.map(_.id()) should contain.inOrder("First", "Second", "Third", "Fourth")
@@ -239,9 +224,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDoc(
sessionSettings,
writeSettings.withParallelism(2),
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
bulkUpsertResult.futureValue
@@ -251,9 +234,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.fromId(
sessionSettings,
bucketName,
- classOf[BinaryDocument]
- )
- )
+ classOf[BinaryDocument]))
.runWith(Sink.seq)
resultsAsFuture.futureValue.map(_.id()) shouldBe Seq("First", "Second", "Third", "Fourth")
}
@@ -263,12 +244,11 @@ class CouchbaseFlowSpec
.map(toJsonDocument)
.via(
CouchbaseFlow.upsert(sessionSettings,
- writeSettings
- .withParallelism(2)
- .withPersistTo(PersistTo.THREE)
- .withTimeout(1.seconds),
- bucketName)
- )
+ writeSettings
+ .withParallelism(2)
+ .withPersistTo(PersistTo.THREE)
+ .withTimeout(1.seconds),
+ bucketName))
.runWith(Sink.seq)
bulkUpsertResult.failed.futureValue shouldBe a[com.couchbase.client.java.error.DurabilityException]
@@ -286,11 +266,9 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDoc(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
- //wait til operation completed
+ // wait til operation completed
upsertFuture.futureValue
// #delete
@@ -301,9 +279,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.delete(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
// #delete
deleteFuture.futureValue
@@ -321,9 +297,7 @@ class CouchbaseFlowSpec
.fromId(
sessionSettings,
bucketName,
- classOf[RawJsonDocument]
- )
- )
+ classOf[RawJsonDocument]))
.runWith(Sink.head)
getFuture.failed.futureValue shouldBe a[NoSuchElementException]
}
@@ -332,23 +306,20 @@ class CouchbaseFlowSpec
val bulkUpsertResult: Future[Done] = Source(sampleSequence)
.map(toRawJsonDocument)
.via(
- CouchbaseFlow.upsertDoc(sessionSettings, writeSettings.withParallelism(2), bucketName)
- )
+ CouchbaseFlow.upsertDoc(sessionSettings, writeSettings.withParallelism(2), bucketName))
.runWith(Sink.ignore)
bulkUpsertResult.futureValue
val deleteFuture: Future[Done] = Source(sampleSequence.map(_.id) :+ "NoneExisting")
.via(
- CouchbaseFlow.delete(sessionSettings, writeSettings.withParallelism(2), bucketName)
- )
+ CouchbaseFlow.delete(sessionSettings, writeSettings.withParallelism(2), bucketName))
.runWith(Sink.ignore)
deleteFuture.failed.futureValue shouldBe a[DocumentDoesNotExistException]
val getFuture: Future[Seq[RawJsonDocument]] =
Source(sampleSequence.map(_.id))
.via(
- CouchbaseFlow.fromId(sessionSettings, bucketName, classOf[RawJsonDocument])
- )
+ CouchbaseFlow.fromId(sessionSettings, bucketName, classOf[RawJsonDocument]))
.runWith(Sink.seq)
getFuture.futureValue shouldBe empty
}
@@ -364,8 +335,7 @@ class CouchbaseFlowSpec
val result: Future[JsonDocument] = Source
.single(id)
.via(
- CouchbaseFlow.fromId(sessionSettings, queryBucketName)
- )
+ CouchbaseFlow.fromId(sessionSettings, queryBucketName))
.runWith(Sink.head)
result.futureValue.id() shouldEqual id
}
@@ -416,9 +386,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.replace(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
// #replace
replaceFuture.futureValue
@@ -432,7 +400,7 @@ class CouchbaseFlowSpec
"replace multiple RawJsonDocuments" in assertAllStagesStopped {
val replaceSequence: Seq[TestObject] = sampleData +: Seq[TestObject](TestObject("Second", "SecondReplace"),
- TestObject("Third", "ThirdReplace"))
+ TestObject("Third", "ThirdReplace"))
upsertSampleData(bucketName)
@@ -442,9 +410,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.replaceDoc(
sessionSettings,
writeSettings.withParallelism(2),
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
bulkReplaceResult.futureValue
@@ -455,9 +421,9 @@ class CouchbaseFlowSpec
.runWith(Sink.seq)
resultsAsFuture.futureValue.map(doc => doc.content().get("value")) should contain.inOrderOnly("First",
- "SecondReplace",
- "ThirdReplace",
- "Fourth")
+ "SecondReplace",
+ "ThirdReplace",
+ "Fourth")
}
"replace RawJsonDocument" in assertAllStagesStopped {
@@ -474,9 +440,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.replaceDoc(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.ignore)
// #replaceDocreplace
@@ -496,12 +460,11 @@ class CouchbaseFlowSpec
.map(toJsonDocument)
.via(
CouchbaseFlow.replace(sessionSettings,
- writeSettings
- .withParallelism(2)
- .withPersistTo(PersistTo.THREE)
- .withTimeout(1.seconds),
- bucketName)
- )
+ writeSettings
+ .withParallelism(2)
+ .withPersistTo(PersistTo.THREE)
+ .withTimeout(1.seconds),
+ bucketName))
.runWith(Sink.seq)
bulkReplaceResult.failed.futureValue shouldBe a[com.couchbase.client.java.error.DurabilityException]
@@ -511,7 +474,7 @@ class CouchbaseFlowSpec
"Couchbase upsert with result" should {
"write documents" in assertAllStagesStopped {
// #upsertDocWithResult
- import akka.stream.alpakka.couchbase.{CouchbaseWriteFailure, CouchbaseWriteResult}
+ import akka.stream.alpakka.couchbase.{ CouchbaseWriteFailure, CouchbaseWriteResult }
val result: Future[immutable.Seq[CouchbaseWriteResult[RawJsonDocument]]] =
Source(sampleSequence)
@@ -520,9 +483,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.upsertDocWithResult(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.seq)
val failedDocs: immutable.Seq[CouchbaseWriteFailure[RawJsonDocument]] = result.futureValue.collect {
@@ -536,18 +497,17 @@ class CouchbaseFlowSpec
}
"expose failures in-stream" in assertAllStagesStopped {
- import akka.stream.alpakka.couchbase.{CouchbaseWriteFailure, CouchbaseWriteResult}
+ import akka.stream.alpakka.couchbase.{ CouchbaseWriteFailure, CouchbaseWriteResult }
val result: Future[immutable.Seq[CouchbaseWriteResult[JsonDocument]]] = Source(sampleSequence)
.map(toJsonDocument)
.via(
CouchbaseFlow.upsertDocWithResult(sessionSettings,
- writeSettings
- .withParallelism(2)
- .withPersistTo(PersistTo.THREE)
- .withTimeout(1.seconds),
- bucketName)
- )
+ writeSettings
+ .withParallelism(2)
+ .withPersistTo(PersistTo.THREE)
+ .withTimeout(1.seconds),
+ bucketName))
.runWith(Sink.seq)
result.futureValue should have size sampleSequence.size
@@ -571,9 +531,7 @@ class CouchbaseFlowSpec
.withParallelism(2)
.withReplicateTo(ReplicateTo.THREE)
.withTimeout(1.seconds),
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.head)
// #deleteWithResult
@@ -584,9 +542,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.deleteWithResult(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.head)
// #deleteWithResult
deleteFuture.failed.futureValue shouldBe a[DocumentDoesNotExistException]
@@ -603,7 +559,7 @@ class CouchbaseFlowSpec
upsertSampleData(bucketName)
// #replaceDocWithResult
- import akka.stream.alpakka.couchbase.{CouchbaseWriteFailure, CouchbaseWriteResult}
+ import akka.stream.alpakka.couchbase.{ CouchbaseWriteFailure, CouchbaseWriteResult }
val result: Future[immutable.Seq[CouchbaseWriteResult[RawJsonDocument]]] =
Source(sampleSequence)
@@ -612,9 +568,7 @@ class CouchbaseFlowSpec
CouchbaseFlow.replaceDocWithResult(
sessionSettings,
writeSettings,
- bucketName
- )
- )
+ bucketName))
.runWith(Sink.seq)
val failedDocs: immutable.Seq[CouchbaseWriteFailure[RawJsonDocument]] = result.futureValue.collect {
@@ -631,18 +585,17 @@ class CouchbaseFlowSpec
cleanAllInBucket(bucketName)
- import akka.stream.alpakka.couchbase.{CouchbaseWriteFailure, CouchbaseWriteResult}
+ import akka.stream.alpakka.couchbase.{ CouchbaseWriteFailure, CouchbaseWriteResult }
val result: Future[immutable.Seq[CouchbaseWriteResult[JsonDocument]]] = Source(sampleSequence)
.map(toJsonDocument)
.via(
CouchbaseFlow.replaceDocWithResult(sessionSettings,
- writeSettings
- .withParallelism(2)
- .withPersistTo(PersistTo.THREE)
- .withTimeout(1.seconds),
- bucketName)
- )
+ writeSettings
+ .withParallelism(2)
+ .withPersistTo(PersistTo.THREE)
+ .withTimeout(1.seconds),
+ bucketName))
.runWith(Sink.seq)
result.futureValue should have size sampleSequence.size
diff --git a/couchbase/src/test/scala/docs/scaladsl/CouchbaseSessionExamplesSpec.scala b/couchbase/src/test/scala/docs/scaladsl/CouchbaseSessionExamplesSpec.scala
index 9267777d..b36dd921 100644
--- a/couchbase/src/test/scala/docs/scaladsl/CouchbaseSessionExamplesSpec.scala
+++ b/couchbase/src/test/scala/docs/scaladsl/CouchbaseSessionExamplesSpec.scala
@@ -12,7 +12,7 @@ import org.scalatest.concurrent.ScalaFutures
import org.scalatest.BeforeAndAfterAll
import scala.concurrent.duration._
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
import org.scalatest.matchers.should.Matchers
import org.scalatest.wordspec.AnyWordSpec
@@ -35,7 +35,7 @@ class CouchbaseSessionExamplesSpec
import akka.stream.alpakka.couchbase.CouchbaseSessionRegistry
import akka.stream.alpakka.couchbase.CouchbaseSessionSettings
import akka.stream.alpakka.couchbase.scaladsl.CouchbaseSession
- import com.couchbase.client.java.env.{CouchbaseEnvironment, DefaultCouchbaseEnvironment}
+ import com.couchbase.client.java.env.{ CouchbaseEnvironment, DefaultCouchbaseEnvironment }
// Akka extension (singleton per actor system)
val registry = CouchbaseSessionRegistry(actorSystem)
@@ -81,7 +81,7 @@ class CouchbaseSessionExamplesSpec
implicit val ec: ExecutionContext = actorSystem.dispatcher
// #fromBucket
import com.couchbase.client.java.auth.PasswordAuthenticator
- import com.couchbase.client.java.{Bucket, CouchbaseCluster}
+ import com.couchbase.client.java.{ Bucket, CouchbaseCluster }
val cluster: CouchbaseCluster = CouchbaseCluster.create("localhost")
cluster.authenticate(new PasswordAuthenticator("Administrator", "password"))
diff --git a/couchbase/src/test/scala/docs/scaladsl/CouchbaseSourceSpec.scala b/couchbase/src/test/scala/docs/scaladsl/CouchbaseSourceSpec.scala
index 70adf9e3..9ace5986 100644
--- a/couchbase/src/test/scala/docs/scaladsl/CouchbaseSourceSpec.scala
+++ b/couchbase/src/test/scala/docs/scaladsl/CouchbaseSourceSpec.scala
@@ -4,13 +4,13 @@
package docs.scaladsl
-import akka.stream.alpakka.couchbase.scaladsl.{CouchbaseSession, CouchbaseSource}
+import akka.stream.alpakka.couchbase.scaladsl.{ CouchbaseSession, CouchbaseSource }
import akka.stream.alpakka.couchbase.testing.CouchbaseSupport
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
import akka.stream.scaladsl.Sink
import akka.stream.testkit.scaladsl.StreamTestKit._
import com.couchbase.client.java.auth.PasswordAuthenticator
-import com.couchbase.client.java.{Bucket, CouchbaseCluster}
+import com.couchbase.client.java.{ Bucket, CouchbaseCluster }
import com.couchbase.client.java.document.json.JsonObject
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.BeforeAndAfterAll
@@ -70,8 +70,8 @@ class CouchbaseSourceSpec
"run simple N1QL query" in assertAllStagesStopped {
- //#n1ql
- import com.couchbase.client.java.query.{N1qlParams, N1qlQuery}
+ // #n1ql
+ import com.couchbase.client.java.query.{ N1qlParams, N1qlQuery }
val params = N1qlParams.build.adhoc(false)
val query = N1qlQuery.simple(s"select count(*) from $queryBucketName", params)
@@ -80,7 +80,7 @@ class CouchbaseSourceSpec
CouchbaseSource
.fromN1qlQuery(sessionSettings, query, bucketName)
.runWith(Sink.seq)
- //#n1ql
+ // #n1ql
resultAsFuture.futureValue.head.get("$1") shouldEqual 4
}
diff --git a/couchbase/src/test/scala/docs/scaladsl/DiscoverySpec.scala b/couchbase/src/test/scala/docs/scaladsl/DiscoverySpec.scala
index 0ca616e3..22ac2c18 100644
--- a/couchbase/src/test/scala/docs/scaladsl/DiscoverySpec.scala
+++ b/couchbase/src/test/scala/docs/scaladsl/DiscoverySpec.scala
@@ -7,14 +7,14 @@ package docs.scaladsl
import akka.actor.ActorSystem
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
import com.couchbase.client.java.document.JsonDocument
-import com.typesafe.config.{Config, ConfigFactory}
+import com.typesafe.config.{ Config, ConfigFactory }
import org.scalatest.BeforeAndAfterAll
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.matchers.should.Matchers
import org.scalatest.wordspec.AnyWordSpec
import scala.concurrent.duration._
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
class DiscoverySpec extends AnyWordSpec with Matchers with BeforeAndAfterAll with ScalaFutures with LogCapturing {
@@ -32,8 +32,8 @@ class DiscoverySpec extends AnyWordSpec with Matchers with BeforeAndAfterAll wit
"a Couchbasesession" should {
"be managed by the registry" in {
// #registry
- import akka.stream.alpakka.couchbase.scaladsl.{CouchbaseSession, DiscoverySupport}
- import akka.stream.alpakka.couchbase.{CouchbaseSessionRegistry, CouchbaseSessionSettings}
+ import akka.stream.alpakka.couchbase.scaladsl.{ CouchbaseSession, DiscoverySupport }
+ import akka.stream.alpakka.couchbase.{ CouchbaseSessionRegistry, CouchbaseSessionSettings }
val registry = CouchbaseSessionRegistry(actorSystem)
@@ -47,7 +47,7 @@ class DiscoverySpec extends AnyWordSpec with Matchers with BeforeAndAfterAll wit
"be created from settings" in {
// #create
import akka.stream.alpakka.couchbase.CouchbaseSessionSettings
- import akka.stream.alpakka.couchbase.scaladsl.{CouchbaseSession, DiscoverySupport}
+ import akka.stream.alpakka.couchbase.scaladsl.{ CouchbaseSession, DiscoverySupport }
implicit val ec: ExecutionContext = actorSystem.dispatcher
val sessionSettings = CouchbaseSessionSettings(actorSystem)
diff --git a/csv-bench/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvBench.scala b/csv-bench/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvBench.scala
index 3efd95d9..fe3ddb2c 100644
--- a/csv-bench/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvBench.scala
+++ b/csv-bench/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvBench.scala
@@ -67,8 +67,7 @@ class CsvBench {
"1024", // ~8x smaller than row
"8192", // ~same size as row
"65536" // ~8k larger than row
- )
- )
+ ))
var bsSize: Int = _
var source: Source[ByteString, NotUsed] = _
@@ -132,8 +131,7 @@ object CsvBench {
def main(args: Array[String]): Unit = {
val bench = new CsvBench
bench.parse(
- new Blackhole("Today's password is swordfish. I understand instantiating Blackholes directly is dangerous.")
- )
+ new Blackhole("Today's password is swordfish. I understand instantiating Blackholes directly is dangerous."))
bench.tearDown()
}
}
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvFormatter.scala b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvFormatter.scala
index 0c707e51..d8efe7a2 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvFormatter.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvFormatter.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.csv.impl
-import java.nio.charset.{Charset, StandardCharsets}
+import java.nio.charset.{ Charset, StandardCharsets }
import akka.annotation.InternalApi
import akka.stream.alpakka.csv.scaladsl.CsvQuotingStyle
@@ -16,11 +16,11 @@ import scala.collection.immutable
* Internal API
*/
@InternalApi private[csv] class CsvFormatter(delimiter: Char,
- quoteChar: Char,
- escapeChar: Char,
- endOfLine: String,
- quotingStyle: CsvQuotingStyle,
- charset: Charset = StandardCharsets.UTF_8) {
+ quoteChar: Char,
+ escapeChar: Char,
+ endOfLine: String,
+ quotingStyle: CsvQuotingStyle,
+ charset: Charset = StandardCharsets.UTF_8) {
private[this] val charsetName = charset.name()
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParser.scala b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParser.scala
index 69962dfc..13ca679b 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParser.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParser.scala
@@ -9,7 +9,7 @@ import java.nio.charset.UnsupportedCharsetException
import akka.annotation.InternalApi
import akka.stream.alpakka.csv.MalformedCsvException
import akka.stream.alpakka.csv.scaladsl.ByteOrderMark
-import akka.util.{ByteIterator, ByteString, ByteStringBuilder}
+import akka.util.{ ByteIterator, ByteString, ByteStringBuilder }
import scala.collection.mutable
@@ -38,9 +38,9 @@ import scala.collection.mutable
* INTERNAL API: Use [[akka.stream.alpakka.csv.scaladsl.CsvParsing]] instead.
*/
@InternalApi private[csv] final class CsvParser(delimiter: Byte,
- quoteChar: Byte,
- escapeChar: Byte,
- maximumLineLength: Int) {
+ quoteChar: Byte,
+ escapeChar: Byte,
+ maximumLineLength: Int) {
import CsvParser._
@@ -144,7 +144,8 @@ import scala.collection.mutable
fieldStart = 0
}
- /** FieldBuilder will just cut the required part out of the incoming ByteBuffer
+ /**
+ * FieldBuilder will just cut the required part out of the incoming ByteBuffer
* as long as non escaping is used.
*/
private final class FieldBuilder {
@@ -155,7 +156,8 @@ import scala.collection.mutable
private[this] var useBuilder = false
private[this] var builder: ByteStringBuilder = _
- /** Set up the ByteString builder instead of relying on `ByteString.slice`.
+ /**
+ * Set up the ByteString builder instead of relying on `ByteString.slice`.
*/
@inline def init(): Unit =
if (!useBuilder) {
@@ -176,8 +178,8 @@ import scala.collection.mutable
private[this] def noCharEscaped() =
throw new MalformedCsvException(currentLineNo,
- lineLength,
- s"wrong escaping at $currentLineNo:$lineLength, no character after escape")
+ lineLength,
+ s"wrong escaping at $currentLineNo:$lineLength, no character after escape")
private[this] def checkForByteOrderMark(): Unit =
if (buffer.length >= 2) {
@@ -211,8 +213,7 @@ import scala.collection.mutable
throw new MalformedCsvException(
currentLineNo,
lineLength,
- s"no line end encountered within $maximumLineLength bytes on line $currentLineNo"
- )
+ s"no line end encountered within $maximumLineLength bytes on line $currentLineNo")
val byte = current.head
state match {
case LineStart =>
@@ -317,8 +318,7 @@ import scala.collection.mutable
throw new MalformedCsvException(
currentLineNo,
lineLength,
- s"wrong escaping at $currentLineNo:$lineLength, quote is escaped as ${quoteChar.toChar}${quoteChar.toChar}"
- )
+ s"wrong escaping at $currentLineNo:$lineLength, quote is escaped as ${quoteChar.toChar}${quoteChar.toChar}")
case b =>
fieldBuilder.add(escapeChar)
@@ -412,8 +412,7 @@ import scala.collection.mutable
throw new MalformedCsvException(
currentLineNo,
lineLength,
- s"unclosed quote at end of input $currentLineNo:$lineLength, no matching quote found"
- )
+ s"unclosed quote at end of input $currentLineNo:$lineLength, no matching quote found")
case WithinField =>
columns += fieldBuilder.result(pos)
Some(columns.toList)
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParsingStage.scala b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParsingStage.scala
index 57ee808f..f1c9f0a5 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParsingStage.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvParsingStage.scala
@@ -6,8 +6,8 @@ package akka.stream.alpakka.csv.impl
import akka.annotation.InternalApi
import akka.event.Logging
-import akka.stream.stage.{GraphStage, GraphStageLogic, InHandler, OutHandler}
-import akka.stream.{Attributes, FlowShape, Inlet, Outlet}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, InHandler, OutHandler }
+import akka.stream.{ Attributes, FlowShape, Inlet, Outlet }
import akka.util.ByteString
import scala.annotation.tailrec
@@ -17,9 +17,9 @@ import scala.util.control.NonFatal
* Internal API: Use [[akka.stream.alpakka.csv.scaladsl.CsvParsing]] instead.
*/
@InternalApi private[csv] class CsvParsingStage(delimiter: Byte,
- quoteChar: Byte,
- escapeChar: Byte,
- maximumLineLength: Int)
+ quoteChar: Byte,
+ escapeChar: Byte,
+ maximumLineLength: Int)
extends GraphStage[FlowShape[ByteString, List[ByteString]]] {
private val in = Inlet[ByteString](Logging.simpleName(this) + ".in")
@@ -49,13 +49,14 @@ import scala.util.control.NonFatal
private def tryPollBuffer() =
try buffer.poll(requireLineEnd = true) match {
- case Some(csvLine) => push(out, csvLine)
- case _ =>
- if (isClosed(in)) {
- emitRemaining()
- completeStage()
- } else pull(in)
- } catch {
+ case Some(csvLine) => push(out, csvLine)
+ case _ =>
+ if (isClosed(in)) {
+ emitRemaining()
+ completeStage()
+ } else pull(in)
+ }
+ catch {
case NonFatal(ex) => failStage(ex)
}
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapJavaStage.scala b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapJavaStage.scala
index 7c14c36f..3240d234 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapJavaStage.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapJavaStage.scala
@@ -6,10 +6,10 @@ package akka.stream.alpakka.csv.impl
import java.nio.charset.Charset
import java.util.stream.Collectors
-import java.{util => ju}
+import java.{ util => ju }
import akka.annotation.InternalApi
import akka.stream._
-import akka.stream.stage.{GraphStage, GraphStageLogic, InHandler, OutHandler}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, InHandler, OutHandler }
import akka.util.ByteString
/**
@@ -22,10 +22,10 @@ import akka.util.ByteString
* @param headerPlaceholder placeholder used when there are more headers than data.
*/
@InternalApi private[csv] abstract class CsvToMapJavaStageBase[V](columnNames: ju.Optional[ju.Collection[String]],
- charset: Charset,
- combineAll: Boolean,
- customFieldValuePlaceholder: ju.Optional[V],
- headerPlaceholder: ju.Optional[String])
+ charset: Charset,
+ combineAll: Boolean,
+ customFieldValuePlaceholder: ju.Optional[V],
+ headerPlaceholder: ju.Optional[String])
extends GraphStage[FlowShape[ju.Collection[ByteString], ju.Map[String, V]]] {
override protected def initialAttributes: Attributes = Attributes.name("CsvToMap")
@@ -60,8 +60,7 @@ import akka.util.ByteString
process(elem, zipWithHeaders)
}
}
- }
- )
+ })
private def process(elem: ju.Collection[ByteString], combine: ju.Collection[V] => ju.Map[String, V]) = {
if (headers.isPresent) {
@@ -73,9 +72,10 @@ import akka.util.ByteString
}
}
- setHandler(out, new OutHandler {
- override def onPull(): Unit = pull(in)
- })
+ setHandler(out,
+ new OutHandler {
+ override def onPull(): Unit = pull(in)
+ })
private def zipWithHeaders(elem: ju.Collection[V]): ju.Map[String, V] = {
val map = new ju.HashMap[String, V]()
@@ -126,15 +126,15 @@ import akka.util.ByteString
* Internal API
*/
@InternalApi private[csv] class CsvToMapJavaStage(columnNames: ju.Optional[ju.Collection[String]],
- charset: Charset,
- combineAll: Boolean,
- customFieldValuePlaceholder: ju.Optional[ByteString],
- headerPlaceholder: ju.Optional[String])
+ charset: Charset,
+ combineAll: Boolean,
+ customFieldValuePlaceholder: ju.Optional[ByteString],
+ headerPlaceholder: ju.Optional[String])
extends CsvToMapJavaStageBase[ByteString](columnNames,
- charset,
- combineAll,
- customFieldValuePlaceholder,
- headerPlaceholder) {
+ charset,
+ combineAll,
+ customFieldValuePlaceholder,
+ headerPlaceholder) {
override val fieldValuePlaceholder: ByteString = ByteString("")
@@ -146,15 +146,15 @@ import akka.util.ByteString
* Internal API
*/
@InternalApi private[csv] class CsvToMapAsStringsJavaStage(columnNames: ju.Optional[ju.Collection[String]],
- charset: Charset,
- combineAll: Boolean,
- customFieldValuePlaceholder: ju.Optional[String],
- headerPlaceholder: ju.Optional[String])
+ charset: Charset,
+ combineAll: Boolean,
+ customFieldValuePlaceholder: ju.Optional[String],
+ headerPlaceholder: ju.Optional[String])
extends CsvToMapJavaStageBase[String](columnNames,
- charset,
- combineAll,
- customFieldValuePlaceholder,
- headerPlaceholder) {
+ charset,
+ combineAll,
+ customFieldValuePlaceholder,
+ headerPlaceholder) {
override val fieldValuePlaceholder: String = ""
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapStage.scala b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapStage.scala
index cc7f0f2f..2366a1f0 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapStage.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/impl/CsvToMapStage.scala
@@ -6,8 +6,8 @@ package akka.stream.alpakka.csv.impl
import java.nio.charset.Charset
import akka.annotation.InternalApi
-import akka.stream.stage.{GraphStage, GraphStageLogic, InHandler, OutHandler}
-import akka.stream.{Attributes, FlowShape, Inlet, Outlet}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, InHandler, OutHandler }
+import akka.stream.{ Attributes, FlowShape, Inlet, Outlet }
import akka.util.ByteString
import scala.collection.immutable
@@ -23,10 +23,10 @@ import scala.collection.immutable
* @param headerPlaceholder placeholder used when there are more headers than data.
*/
@InternalApi private[csv] abstract class CsvToMapStageBase[V](columnNames: Option[immutable.Seq[String]],
- charset: Charset,
- combineAll: Boolean,
- customFieldValuePlaceholder: Option[V],
- headerPlaceholder: Option[String])
+ charset: Charset,
+ combineAll: Boolean,
+ customFieldValuePlaceholder: Option[V],
+ headerPlaceholder: Option[String])
extends GraphStage[FlowShape[immutable.Seq[ByteString], Map[String, V]]] {
override protected def initialAttributes: Attributes = Attributes.name("CsvToMap")
@@ -71,8 +71,8 @@ import scala.collection.immutable
private def combineUsingPlaceholder(elem: immutable.Seq[ByteString]): Headers => Map[String, V] = headers => {
val combined = headers.get
.zipAll(transformElements(elem),
- headerPlaceholder.getOrElse("MissingHeader"),
- customFieldValuePlaceholder.getOrElse(fieldValuePlaceholder))
+ headerPlaceholder.getOrElse("MissingHeader"),
+ customFieldValuePlaceholder.getOrElse(fieldValuePlaceholder))
val filtering: String => Boolean = key =>
headerPlaceholder.map(_.equalsIgnoreCase(key)).fold(key.equalsIgnoreCase("MissingHeader"))(identity)
val missingHeadersContent =
@@ -101,15 +101,15 @@ import scala.collection.immutable
* Internal API
*/
@InternalApi private[csv] class CsvToMapStage(columnNames: Option[immutable.Seq[String]],
- charset: Charset,
- combineAll: Boolean,
- customFieldValuePlaceholder: Option[ByteString],
- headerPlaceholder: Option[String])
+ charset: Charset,
+ combineAll: Boolean,
+ customFieldValuePlaceholder: Option[ByteString],
+ headerPlaceholder: Option[String])
extends CsvToMapStageBase[ByteString](columnNames,
- charset,
- combineAll,
- customFieldValuePlaceholder,
- headerPlaceholder) {
+ charset,
+ combineAll,
+ customFieldValuePlaceholder,
+ headerPlaceholder) {
override val fieldValuePlaceholder: ByteString = ByteString("")
@@ -121,11 +121,12 @@ import scala.collection.immutable
* Internal API
*/
@InternalApi private[csv] class CsvToMapAsStringsStage(columnNames: Option[immutable.Seq[String]],
- charset: Charset,
- combineAll: Boolean,
- customFieldValuePlaceholder: Option[String],
- headerPlaceholder: Option[String])
- extends CsvToMapStageBase[String](columnNames, charset, combineAll, customFieldValuePlaceholder, headerPlaceholder) {
+ charset: Charset,
+ combineAll: Boolean,
+ customFieldValuePlaceholder: Option[String],
+ headerPlaceholder: Option[String])
+ extends CsvToMapStageBase[String](columnNames, charset, combineAll, customFieldValuePlaceholder,
+ headerPlaceholder) {
override val fieldValuePlaceholder: String = ""
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvFormatting.scala b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvFormatting.scala
index d3df181e..6ab6b0c5 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvFormatting.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvFormatting.scala
@@ -4,16 +4,17 @@
package akka.stream.alpakka.csv.scaladsl
-import java.nio.charset.{Charset, StandardCharsets}
+import java.nio.charset.{ Charset, StandardCharsets }
import akka.NotUsed
import akka.stream.alpakka.csv.impl.CsvFormatter
-import akka.stream.scaladsl.{Flow, Source}
+import akka.stream.scaladsl.{ Flow, Source }
import akka.util.ByteString
import scala.collection.immutable
-/** Provides CSV formatting flows that convert a sequence of String into their CSV representation
+/**
+ * Provides CSV formatting flows that convert a sequence of String into their CSV representation
* in [[akka.util.ByteString]].
*/
object CsvFormatting {
@@ -39,8 +40,7 @@ object CsvFormatting {
endOfLine: String = "\r\n",
quotingStyle: CsvQuotingStyle = CsvQuotingStyle.Required,
charset: Charset = StandardCharsets.UTF_8,
- byteOrderMark: Option[ByteString] = None
- ): Flow[T, ByteString, NotUsed] = {
+ byteOrderMark: Option[ByteString] = None): Flow[T, ByteString, NotUsed] = {
val formatter =
new CsvFormatter(delimiter, quoteChar, escapeChar, endOfLine, quotingStyle, charset)
byteOrderMark.fold {
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvParsing.scala b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvParsing.scala
index 2a96adaa..27f5063a 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvParsing.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvParsing.scala
@@ -19,12 +19,13 @@ object CsvParsing {
val DoubleQuote: Byte = '"'
val maximumLineLengthDefault: Int = 10 * 1024
- /** Creates CSV parsing flow that reads CSV lines from incoming
+ /**
+ * Creates CSV parsing flow that reads CSV lines from incoming
* [[akka.util.ByteString]] objects.
*/
def lineScanner(delimiter: Byte = Comma,
- quoteChar: Byte = DoubleQuote,
- escapeChar: Byte = Backslash,
- maximumLineLength: Int = maximumLineLengthDefault): Flow[ByteString, List[ByteString], NotUsed] =
+ quoteChar: Byte = DoubleQuote,
+ escapeChar: Byte = Backslash,
+ maximumLineLength: Int = maximumLineLengthDefault): Flow[ByteString, List[ByteString], NotUsed] =
Flow.fromGraph(new CsvParsingStage(delimiter, quoteChar, escapeChar, maximumLineLength))
}
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvQuotingStyle.scala b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvQuotingStyle.scala
index 1b1c0188..f9d13ff5 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvQuotingStyle.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvQuotingStyle.scala
@@ -21,7 +21,7 @@ object CsvQuotingStyle {
/** Java to Scala conversion helper */
def asScala(qs: javadsl.CsvQuotingStyle): CsvQuotingStyle = qs match {
- case javadsl.CsvQuotingStyle.ALWAYS => CsvQuotingStyle.Always
+ case javadsl.CsvQuotingStyle.ALWAYS => CsvQuotingStyle.Always
case javadsl.CsvQuotingStyle.REQUIRED => CsvQuotingStyle.Required
}
diff --git a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvToMap.scala b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvToMap.scala
index 98a24827..b74d8aef 100644
--- a/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvToMap.scala
+++ b/csv/src/main/scala/akka/stream/alpakka/csv/scaladsl/CsvToMap.scala
@@ -4,10 +4,10 @@
package akka.stream.alpakka.csv.scaladsl
-import java.nio.charset.{Charset, StandardCharsets}
+import java.nio.charset.{ Charset, StandardCharsets }
import akka.NotUsed
-import akka.stream.alpakka.csv.impl.{CsvToMapAsStringsStage, CsvToMapStage}
+import akka.stream.alpakka.csv.impl.{ CsvToMapAsStringsStage, CsvToMapStage }
import akka.stream.scaladsl.Flow
import akka.util.ByteString
@@ -21,11 +21,10 @@ object CsvToMap {
def toMap(charset: Charset = StandardCharsets.UTF_8): Flow[List[ByteString], Map[String, ByteString], NotUsed] =
Flow.fromGraph(
new CsvToMapStage(columnNames = None,
- charset,
- combineAll = false,
- customFieldValuePlaceholder = Option.empty,
- headerPlaceholder = Option.empty)
- )
+ charset,
+ combineAll = false,
+ customFieldValuePlaceholder = Option.empty,
+ headerPlaceholder = Option.empty))
/**
* A flow translating incoming [[scala.List]] of [[akka.util.ByteString]] to a map of String keys and values using the stream's first
@@ -35,11 +34,10 @@ object CsvToMap {
def toMapAsStrings(charset: Charset = StandardCharsets.UTF_8): Flow[List[ByteString], Map[String, String], NotUsed] =
Flow.fromGraph(
new CsvToMapAsStringsStage(columnNames = None,
- charset,
- combineAll = false,
- customFieldValuePlaceholder = Option.empty,
- headerPlaceholder = Option.empty)
- )
+ charset,
+ combineAll = false,
+ customFieldValuePlaceholder = Option.empty,
+ headerPlaceholder = Option.empty))
/**
* A flow translating incoming [[scala.List]] of [[akka.util.ByteString]] to a map of String and ByteString
@@ -52,15 +50,13 @@ object CsvToMap {
def toMapCombineAll(
charset: Charset = StandardCharsets.UTF_8,
customFieldValuePlaceholder: Option[ByteString] = None,
- headerPlaceholder: Option[String] = None
- ): Flow[List[ByteString], Map[String, ByteString], NotUsed] =
+ headerPlaceholder: Option[String] = None): Flow[List[ByteString], Map[String, ByteString], NotUsed] =
Flow.fromGraph(
new CsvToMapStage(columnNames = None,
- charset,
- combineAll = true,
- customFieldValuePlaceholder = customFieldValuePlaceholder,
- headerPlaceholder = headerPlaceholder)
- )
+ charset,
+ combineAll = true,
+ customFieldValuePlaceholder = customFieldValuePlaceholder,
+ headerPlaceholder = headerPlaceholder))
/**
* A flow translating incoming [[scala.List]] of [[akka.util.ByteString]] to a map of String keys and values
@@ -73,15 +69,13 @@ object CsvToMap {
def toMapAsStringsCombineAll(
charset: Charset = StandardCharsets.UTF_8,
customFieldValuePlaceholder: Option[String] = None,
- headerPlaceholder: Option[String] = None
- ): Flow[List[ByteString], Map[String, String], NotUsed] =
+ headerPlaceholder: Option[String] = None): Flow[List[ByteString], Map[String, String], NotUsed] =
Flow.fromGraph(
new CsvToMapAsStringsStage(columnNames = None,
- charset,
- combineAll = true,
- customFieldValuePlaceholder = customFieldValuePlaceholder,
- headerPlaceholder = headerPlaceholder)
- )
+ charset,
+ combineAll = true,
+ customFieldValuePlaceholder = customFieldValuePlaceholder,
+ headerPlaceholder = headerPlaceholder))
/**
* A flow translating incoming [[scala.List]] of [[akka.util.ByteString]] to a map of String and ByteString using the given headers
@@ -91,11 +85,10 @@ object CsvToMap {
def withHeaders(headers: String*): Flow[List[ByteString], Map[String, ByteString], NotUsed] =
Flow.fromGraph(
new CsvToMapStage(Some(headers.toList),
- StandardCharsets.UTF_8,
- combineAll = false,
- customFieldValuePlaceholder = Option.empty,
- headerPlaceholder = Option.empty)
- )
+ StandardCharsets.UTF_8,
+ combineAll = false,
+ customFieldValuePlaceholder = Option.empty,
+ headerPlaceholder = Option.empty))
/**
* A flow translating incoming [[scala.List]] of [[akka.util.ByteString]] to a map of String keys and values using the given headers
@@ -105,13 +98,11 @@ object CsvToMap {
*/
def withHeadersAsStrings(
charset: Charset,
- headers: String*
- ): Flow[List[ByteString], Map[String, String], NotUsed] =
+ headers: String*): Flow[List[ByteString], Map[String, String], NotUsed] =
Flow.fromGraph(
new CsvToMapAsStringsStage(Some(headers.toList),
- charset,
- combineAll = false,
- customFieldValuePlaceholder = Option.empty,
- headerPlaceholder = Option.empty)
- )
+ charset,
+ combineAll = false,
+ customFieldValuePlaceholder = Option.empty,
+ headerPlaceholder = Option.empty))
}
diff --git a/csv/src/test/scala/akka/stream/alpakka/csv/CsvParserSpec.scala b/csv/src/test/scala/akka/stream/alpakka/csv/CsvParserSpec.scala
index 3eff5afc..852bab1f 100644
--- a/csv/src/test/scala/akka/stream/alpakka/csv/CsvParserSpec.scala
+++ b/csv/src/test/scala/akka/stream/alpakka/csv/CsvParserSpec.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.csv
-import java.nio.charset.{StandardCharsets, UnsupportedCharsetException}
+import java.nio.charset.{ StandardCharsets, UnsupportedCharsetException }
import akka.stream.alpakka.csv.impl.CsvParser
import akka.stream.alpakka.csv.scaladsl.ByteOrderMark
@@ -93,7 +93,7 @@ class CsvParserSpec extends AnyWordSpec with Matchers with OptionValues with Log
"parse double quote chars within quotes inte one quoute at end of value" in {
expectInOut("\"Venture \"\"Extended Edition\"\"\",\"\",4900.00\n",
- List("Venture \"Extended Edition\"", "", "4900.00"))
+ List("Venture \"Extended Edition\"", "", "4900.00"))
}
"parse double escape chars into one escape char" in {
@@ -298,8 +298,8 @@ class CsvParserSpec extends AnyWordSpec with Matchers with OptionValues with Log
"read values with different separator" in {
expectInOut("$Foo $#$Bar $#$Baz $\n", List("Foo ", "Bar ", "Baz "))(delimiter = '#',
- quoteChar = '$',
- escapeChar = '\\')
+ quoteChar = '$',
+ escapeChar = '\\')
}
"fail on a very 'long' line" in {
@@ -348,17 +348,17 @@ class CsvParserSpec extends AnyWordSpec with Matchers with OptionValues with Log
}
def expectInOut(in: String, expected: List[String]*)(implicit delimiter: Byte = ',',
- quoteChar: Byte = '"',
- escapeChar: Byte = '\\',
- requireLineEnd: Boolean = true): Unit = {
+ quoteChar: Byte = '"',
+ escapeChar: Byte = '\\',
+ requireLineEnd: Boolean = true): Unit = {
val bsIn = ByteString(in)
expectBsInOut(bsIn, expected: _*)(delimiter, quoteChar, escapeChar, requireLineEnd)
}
def expectBsInOut(bsIn: ByteString, expected: List[String]*)(implicit delimiter: Byte = ',',
- quoteChar: Byte = '"',
- escapeChar: Byte = '\\',
- requireLineEnd: Boolean = true): Unit = {
+ quoteChar: Byte = '"',
+ escapeChar: Byte = '\\',
+ requireLineEnd: Boolean = true): Unit = {
val parser = new CsvParser(delimiter, quoteChar, escapeChar, maximumLineLength)
parser.offer(bsIn)
expected.foreach { out =>
diff --git a/csv/src/test/scala/docs/scaladsl/CsvFormattingSpec.scala b/csv/src/test/scala/docs/scaladsl/CsvFormattingSpec.scala
index fa1540c3..a8fed803 100644
--- a/csv/src/test/scala/docs/scaladsl/CsvFormattingSpec.scala
+++ b/csv/src/test/scala/docs/scaladsl/CsvFormattingSpec.scala
@@ -6,7 +6,7 @@ package docs.scaladsl
import java.nio.charset.StandardCharsets
-import akka.stream.scaladsl.{Flow, Sink, Source}
+import akka.stream.scaladsl.{ Flow, Sink, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.util.ByteString
@@ -16,7 +16,7 @@ class CsvFormattingSpec extends CsvSpec {
def documentation(): Unit = {
// #flow-type
- import akka.stream.alpakka.csv.scaladsl.{CsvFormatting, CsvQuotingStyle}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvFormatting, CsvQuotingStyle }
// #flow-type
import CsvFormatting._
@@ -59,7 +59,7 @@ class CsvFormattingSpec extends CsvSpec {
"include Byte Order Mark" in assertAllStagesStopped {
// #formatting-bom
- import akka.stream.alpakka.csv.scaladsl.{ByteOrderMark, CsvFormatting}
+ import akka.stream.alpakka.csv.scaladsl.{ ByteOrderMark, CsvFormatting }
// #formatting-bom
val fut =
@@ -72,8 +72,7 @@ class CsvFormattingSpec extends CsvSpec {
// #formatting-bom
// format: on
fut.futureValue should be(
- List(ByteOrderMark.UTF_8, ByteString("eins,zwei,drei\r\n"), ByteString("uno,dos,tres\r\n"))
- )
+ List(ByteOrderMark.UTF_8, ByteString("eins,zwei,drei\r\n"), ByteString("uno,dos,tres\r\n")))
}
}
diff --git a/csv/src/test/scala/docs/scaladsl/CsvParsingSpec.scala b/csv/src/test/scala/docs/scaladsl/CsvParsingSpec.scala
index 23046034..605ef953 100644
--- a/csv/src/test/scala/docs/scaladsl/CsvParsingSpec.scala
+++ b/csv/src/test/scala/docs/scaladsl/CsvParsingSpec.scala
@@ -7,10 +7,10 @@ package docs.scaladsl
import java.nio.file.Paths
import akka.NotUsed
-import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
-import akka.stream.scaladsl.{FileIO, Flow, Keep, Sink, Source}
+import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
+import akka.stream.scaladsl.{ FileIO, Flow, Keep, Sink, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
-import akka.stream.testkit.scaladsl.{TestSink, TestSource}
+import akka.stream.testkit.scaladsl.{ TestSink, TestSource }
import akka.util.ByteString
import scala.collection.immutable.Seq
@@ -111,8 +111,7 @@ class CsvParsingSpec extends CsvSpec {
val input = Seq(
"eins,zw",
"ei,drei\nuno",
- ",dos,tres\n"
- ).map(ByteString(_))
+ ",dos,tres\n").map(ByteString(_))
val fut = Source.apply(input).via(CsvParsing.lineScanner()).map(_.map(_.utf8String)).runWith(Sink.seq)
val res = fut.futureValue
res.head should be(List("eins", "zwei", "drei"))
@@ -189,18 +188,14 @@ class CsvParsingSpec extends CsvSpec {
"Make" -> "Ford",
"Model" -> "E350",
"Description" -> "ac, abs, moon",
- "Price" -> "3000.00"
- )
- )
+ "Price" -> "3000.00"))
res(1) should contain allElementsOf (
Map(
"Year" -> "1999",
"Make" -> "Chevy",
"Model" -> "Venture \"Extended Edition\"",
"Description" -> "",
- "Price" -> "4900.00"
- )
- )
+ "Price" -> "4900.00"))
res(2) should contain allElementsOf (
Map(
"Year" -> "1996",
@@ -208,45 +203,35 @@ class CsvParsingSpec extends CsvSpec {
"Model" -> "Grand Cherokee",
"Description" -> """MUST SELL!
|air, moon roof, loaded""".stripMargin,
- "Price" -> "4799.00"
- )
- )
+ "Price" -> "4799.00"))
res(3) should contain allElementsOf (
Map(
"Year" -> "1999",
"Make" -> "Chevy",
"Model" -> "Venture \"Extended Edition, Very Large\"",
"Description" -> "",
- "Price" -> "5000.00"
- )
- )
+ "Price" -> "5000.00"))
res(4) should contain allElementsOf (
Map(
"Year" -> "",
"Make" -> "",
"Model" -> "Venture \"Extended Edition\"",
"Description" -> "",
- "Price" -> "4900.00"
- )
- )
+ "Price" -> "4900.00"))
res(5) should contain allElementsOf (
Map(
"Year" -> "1995",
"Make" -> "VW",
"Model" -> "Golf \"GTE\"",
"Description" -> "",
- "Price" -> "5000.00"
- )
- )
+ "Price" -> "5000.00"))
res(6) should contain allElementsOf (
Map(
"Year" -> "1996",
"Make" -> "VW",
"Model" -> "Golf GTE",
"Description" -> "",
- "Price" -> "5000.00"
- )
- )
+ "Price" -> "5000.00"))
}
}
}
diff --git a/csv/src/test/scala/docs/scaladsl/CsvSpec.scala b/csv/src/test/scala/docs/scaladsl/CsvSpec.scala
index d5eec85c..0e7d7469 100644
--- a/csv/src/test/scala/docs/scaladsl/CsvSpec.scala
+++ b/csv/src/test/scala/docs/scaladsl/CsvSpec.scala
@@ -8,7 +8,7 @@ import akka.actor.ActorSystem
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
import akka.testkit.TestKit
import org.scalatest.concurrent.ScalaFutures
-import org.scalatest.{BeforeAndAfterAll, BeforeAndAfterEach}
+import org.scalatest.{ BeforeAndAfterAll, BeforeAndAfterEach }
import org.scalatest.matchers.should.Matchers
import org.scalatest.wordspec.AnyWordSpec
diff --git a/csv/src/test/scala/docs/scaladsl/CsvToMapSpec.scala b/csv/src/test/scala/docs/scaladsl/CsvToMapSpec.scala
index 0838114a..ef693dec 100644
--- a/csv/src/test/scala/docs/scaladsl/CsvToMapSpec.scala
+++ b/csv/src/test/scala/docs/scaladsl/CsvToMapSpec.scala
@@ -7,8 +7,8 @@ package docs.scaladsl
import java.nio.charset.StandardCharsets
import akka.NotUsed
-import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
-import akka.stream.scaladsl.{Flow, Sink, Source}
+import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
+import akka.stream.scaladsl.{ Flow, Sink, Source }
import akka.stream.testkit.scaladsl.StreamTestKit.assertAllStagesStopped
import akka.util.ByteString
@@ -54,7 +54,7 @@ class CsvToMapSpec extends CsvSpec {
"CSV to Map" should {
"parse header line and data line into map" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -77,9 +77,7 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> ByteString("11"), "zwei" -> ByteString("12"), "drei" -> ByteString("13")),
- Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "drei" -> ByteString("23"))
- )
- )
+ Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "drei" -> ByteString("23"))))
// #header-line
}
@@ -105,8 +103,7 @@ class CsvToMapSpec extends CsvSpec {
.via(CsvToMap.toMap())
.runWith(Sink.head)
future.futureValue should be(
- Map("eins" -> ByteString("1"), "zwei" -> ByteString("2"), "drei" -> ByteString("3"))
- )
+ Map("eins" -> ByteString("1"), "zwei" -> ByteString("2"), "drei" -> ByteString("3")))
}
"parse header line and decode data line" in assertAllStagesStopped {
@@ -131,15 +128,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> "11", "zwei" -> "12", "drei" -> "13"),
- Map("eins" -> "21", "zwei" -> "22", "drei" -> "23")
- )
- )
+ Map("eins" -> "21", "zwei" -> "22", "drei" -> "23")))
// #header-line
}
"use column names and data line into map" in assertAllStagesStopped {
// #column-names
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #column-names
val future =
@@ -162,9 +157,7 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> ByteString("11"), "zwei" -> ByteString("12"), "drei" -> ByteString("13")),
- Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "drei" -> ByteString("23"))
- )
- )
+ Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "drei" -> ByteString("23"))))
// #column-names
}
@@ -189,15 +182,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> "11", "zwei" -> "12", "drei" -> "13"),
- Map("eins" -> "21", "zwei" -> "22", "drei" -> "23")
- )
- )
+ Map("eins" -> "21", "zwei" -> "22", "drei" -> "23")))
// #column-names
}
"parse header and decode data line. Be OK with more headers column than data (including the header in the result)" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -220,15 +211,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> "11", "zwei" -> "12", "drei" -> "13", "vier" -> "", "fünt" -> ""),
- Map("eins" -> "21", "zwei" -> "22", "drei" -> "23", "vier" -> "", "fünt" -> "")
- )
- )
+ Map("eins" -> "21", "zwei" -> "22", "drei" -> "23", "vier" -> "", "fünt" -> "")))
// #header-line
}
"parse header and decode data line. Be OK when there are more data than header column, set a default header in the result" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -251,15 +240,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> "11", "zwei" -> "12", "drei" -> "13", "MissingHeader0" -> "14"),
- Map("eins" -> "21", "zwei" -> "22", "drei" -> "23")
- )
- )
+ Map("eins" -> "21", "zwei" -> "22", "drei" -> "23")))
// #header-line
}
"parse header and decode data line. Be OK when there are more data than header column, set the user configured header in the result" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -282,15 +269,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> "11", "zwei" -> "12", "MyCustomHeader0" -> "13"),
- Map("eins" -> "21", "zwei" -> "22", "MyCustomHeader0" -> "")
- )
- )
+ Map("eins" -> "21", "zwei" -> "22", "MyCustomHeader0" -> "")))
// #header-line
}
"parse header and decode data line. Be OK when there are more headers than data column, set the user configured field value in the result" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -313,16 +298,14 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> "11", "zwei" -> "12", "drei" -> "13", "fünt" -> "missing"),
- Map("eins" -> "21", "zwei" -> "22", "drei" -> "23", "fünt" -> "missing")
- )
- )
+ Map("eins" -> "21", "zwei" -> "22", "drei" -> "23", "fünt" -> "missing")))
// #header-line
}
}
"be OK with more headers column than data (including the header in the result)" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -345,23 +328,21 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> ByteString("11"),
- "zwei" -> ByteString("12"),
- "drei" -> ByteString("13"),
- "vier" -> ByteString(""),
- "fünt" -> ByteString("")),
+ "zwei" -> ByteString("12"),
+ "drei" -> ByteString("13"),
+ "vier" -> ByteString(""),
+ "fünt" -> ByteString("")),
Map("eins" -> ByteString("21"),
- "zwei" -> ByteString("22"),
- "drei" -> ByteString("23"),
- "vier" -> ByteString(""),
- "fünt" -> ByteString(""))
- )
- )
+ "zwei" -> ByteString("22"),
+ "drei" -> ByteString("23"),
+ "vier" -> ByteString(""),
+ "fünt" -> ByteString(""))))
// #header-line
}
"be OK when there are more data than header column, set a default header in the result" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -384,19 +365,17 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> ByteString("11"),
- "zwei" -> ByteString("12"),
- "drei" -> ByteString("13"),
- "MissingHeader0" -> ByteString("14"),
- "MissingHeader1" -> ByteString("15")),
- Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "drei" -> ByteString("23"))
- )
- )
+ "zwei" -> ByteString("12"),
+ "drei" -> ByteString("13"),
+ "MissingHeader0" -> ByteString("14"),
+ "MissingHeader1" -> ByteString("15")),
+ Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "drei" -> ByteString("23"))))
// #header-line
}
"be OK when there are more data than header column, set the user configured header in the result" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -419,15 +398,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> ByteString("11"), "zwei" -> ByteString("12"), "MyCustomHeader0" -> ByteString("13")),
- Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "MyCustomHeader0" -> ByteString(""))
- )
- )
+ Map("eins" -> ByteString("21"), "zwei" -> ByteString("22"), "MyCustomHeader0" -> ByteString(""))))
// #header-line
}
"be OK when there are more headers than data column, set the user configured field value in the result" in assertAllStagesStopped {
// #header-line
- import akka.stream.alpakka.csv.scaladsl.{CsvParsing, CsvToMap}
+ import akka.stream.alpakka.csv.scaladsl.{ CsvParsing, CsvToMap }
// #header-line
val future =
@@ -450,15 +427,13 @@ class CsvToMapSpec extends CsvSpec {
result should be(
Seq(
Map("eins" -> ByteString("11"),
- "zwei" -> ByteString("12"),
- "drei" -> ByteString("13"),
- "fünt" -> ByteString("missing")),
+ "zwei" -> ByteString("12"),
+ "drei" -> ByteString("13"),
+ "fünt" -> ByteString("missing")),
Map("eins" -> ByteString("21"),
- "zwei" -> ByteString("22"),
- "drei" -> ByteString(""),
- "fünt" -> ByteString("missing"))
- )
- )
+ "zwei" -> ByteString("22"),
+ "drei" -> ByteString(""),
+ "fünt" -> ByteString("missing"))))
// #header-line
}
}
diff --git a/doc-examples/src/test/scala/akka/stream/alpakka/eip/scaladsl/PassThroughExamples.scala b/doc-examples/src/test/scala/akka/stream/alpakka/eip/scaladsl/PassThroughExamples.scala
index 78dcd393..8a6ef83c 100644
--- a/doc-examples/src/test/scala/akka/stream/alpakka/eip/scaladsl/PassThroughExamples.scala
+++ b/doc-examples/src/test/scala/akka/stream/alpakka/eip/scaladsl/PassThroughExamples.scala
@@ -7,11 +7,11 @@ package akka.stream.alpakka.eip.scaladsl
import akka.NotUsed
import akka.actor.ActorSystem
import akka.kafka.scaladsl.Consumer.DrainingControl
-import akka.kafka.scaladsl.{Committer, Consumer}
-import akka.kafka.{CommitterSettings, ConsumerMessage, ConsumerSettings, Subscriptions}
+import akka.kafka.scaladsl.{ Committer, Consumer }
+import akka.kafka.{ CommitterSettings, ConsumerMessage, ConsumerSettings, Subscriptions }
import akka.stream.scaladsl._
-import akka.stream.{FlowShape, Graph}
-import org.apache.kafka.common.serialization.{ByteArrayDeserializer, StringDeserializer}
+import akka.stream.{ FlowShape, Graph }
+import org.apache.kafka.common.serialization.{ ByteArrayDeserializer, StringDeserializer }
import org.scalatest.BeforeAndAfterAll
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.matchers.should.Matchers
@@ -24,8 +24,8 @@ class PassThroughExamples extends AnyWordSpec with BeforeAndAfterAll with Matche
"PassThroughFlow" should {
" original message is maintained " in {
- //#PassThroughWithKeep
- //Sample Source
+ // #PassThroughWithKeep
+ // Sample Source
val source = Source(List(1, 2, 3))
// Pass through this flow maintaining the original message
@@ -37,16 +37,16 @@ class PassThroughExamples extends AnyWordSpec with BeforeAndAfterAll with Matche
.via(PassThroughFlow(passThroughMe, Keep.right))
.runWith(Sink.seq)
- //Verify results
+ // Verify results
ret.futureValue should be(Vector(1, 2, 3))
- //#PassThroughWithKeep
+ // #PassThroughWithKeep
}
" original message and pass through flow output are returned " in {
- //#PassThroughTuple
- //Sample Source
+ // #PassThroughTuple
+ // Sample Source
val source = Source(List(1, 2, 3))
// Pass through this flow maintaining the original message
@@ -58,9 +58,9 @@ class PassThroughExamples extends AnyWordSpec with BeforeAndAfterAll with Matche
.via(PassThroughFlow(passThroughMe))
.runWith(Sink.seq)
- //Verify results
+ // Verify results
ret.futureValue should be(Vector((10, 1), (20, 2), (30, 3)))
- //#PassThroughTuple
+ // #PassThroughTuple
}
}
diff --git a/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/DynamoDbOp.scala b/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/DynamoDbOp.scala
index c0860656..81a5d1f5 100644
--- a/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/DynamoDbOp.scala
+++ b/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/DynamoDbOp.scala
@@ -29,8 +29,7 @@ import scala.concurrent.Future
* @tparam Out dynamodb response type
*/
sealed class DynamoDbOp[In <: DynamoDbRequest, Out <: DynamoDbResponse](
- sdkExecute: DynamoDbAsyncClient => In => CompletableFuture[Out]
-) {
+ sdkExecute: DynamoDbAsyncClient => In => CompletableFuture[Out]) {
def execute(request: In)(implicit client: DynamoDbAsyncClient): Future[Out] = sdkExecute(client)(request).toScala
}
@@ -45,8 +44,7 @@ sealed class DynamoDbOp[In <: DynamoDbRequest, Out <: DynamoDbResponse](
*/
sealed class DynamoDbPaginatedOp[In <: DynamoDbRequest, Out <: DynamoDbResponse, Pub <: SdkPublisher[Out]](
sdkExecute: DynamoDbAsyncClient => In => CompletableFuture[Out],
- sdkPublisher: DynamoDbAsyncClient => In => Pub
-) extends DynamoDbOp[In, Out](sdkExecute) {
+ sdkPublisher: DynamoDbAsyncClient => In => Pub) extends DynamoDbOp[In, Out](sdkExecute) {
def publisher(request: In)(implicit client: DynamoDbAsyncClient): Publisher[Out] = sdkPublisher(client)(request)
}
diff --git a/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/javadsl/DynamoDb.scala b/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/javadsl/DynamoDb.scala
index fd3c0306..e5f0886a 100644
--- a/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/javadsl/DynamoDb.scala
+++ b/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/javadsl/DynamoDb.scala
@@ -10,11 +10,11 @@ import akka.NotUsed
import akka.actor.ClassicActorSystemProvider
import akka.annotation.ApiMayChange
import akka.stream.Materializer
-import akka.stream.alpakka.dynamodb.{scaladsl, DynamoDbOp, DynamoDbPaginatedOp}
-import akka.stream.javadsl.{Flow, FlowWithContext, Sink, Source}
+import akka.stream.alpakka.dynamodb.{ scaladsl, DynamoDbOp, DynamoDbPaginatedOp }
+import akka.stream.javadsl.{ Flow, FlowWithContext, Sink, Source }
import software.amazon.awssdk.core.async.SdkPublisher
import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient
-import software.amazon.awssdk.services.dynamodb.model.{DynamoDbRequest, DynamoDbResponse}
+import software.amazon.awssdk.services.dynamodb.model.{ DynamoDbRequest, DynamoDbResponse }
import scala.util.Try
@@ -29,8 +29,8 @@ object DynamoDb {
* @param parallelism maximum number of in-flight requests at any given time
*/
def flow[In <: DynamoDbRequest, Out <: DynamoDbResponse](client: DynamoDbAsyncClient,
- operation: DynamoDbOp[In, Out],
- parallelism: Int): Flow[In, Out, NotUsed] =
+ operation: DynamoDbOp[In, Out],
+ parallelism: Int): Flow[In, Out, NotUsed] =
scaladsl.DynamoDb.flow(parallelism)(client, operation).asJava
/**
@@ -47,19 +47,16 @@ object DynamoDb {
def flowWithContext[In <: DynamoDbRequest, Out <: DynamoDbResponse, Ctx](
client: DynamoDbAsyncClient,
operation: DynamoDbOp[In, Out],
- parallelism: Int
- ): FlowWithContext[In, Ctx, Try[Out], Ctx, NotUsed] =
+ parallelism: Int): FlowWithContext[In, Ctx, Try[Out], Ctx, NotUsed] =
scaladsl.DynamoDb.flowWithContext[In, Out, Ctx](parallelism)(client, operation).asJava
/**
* Create a Source that will emit potentially multiple responses for a given request.
- *
*/
def source[In <: DynamoDbRequest, Out <: DynamoDbResponse, Pub <: SdkPublisher[Out]](
client: DynamoDbAsyncClient,
operation: DynamoDbPaginatedOp[In, Out, Pub],
- request: In
- ): Source[Out, NotUsed] =
+ request: In): Source[Out, NotUsed] =
scaladsl.DynamoDb.source(request)(client, operation).asJava
/**
@@ -69,8 +66,7 @@ object DynamoDb {
*/
def flowPaginated[In <: DynamoDbRequest, Out <: DynamoDbResponse](
client: DynamoDbAsyncClient,
- operation: DynamoDbPaginatedOp[In, Out, _]
- ): Flow[In, Out, NotUsed] =
+ operation: DynamoDbPaginatedOp[In, Out, _]): Flow[In, Out, NotUsed] =
scaladsl.DynamoDb.flowPaginated()(client, operation).asJava
/**
@@ -79,9 +75,9 @@ object DynamoDb {
*/
@deprecated("pass in the actor system instead of the materializer", "3.0.0")
def single[In <: DynamoDbRequest, Out <: DynamoDbResponse](client: DynamoDbAsyncClient,
- operation: DynamoDbOp[In, Out],
- request: In,
- mat: Materializer): CompletionStage[Out] =
+ operation: DynamoDbOp[In, Out],
+ request: In,
+ mat: Materializer): CompletionStage[Out] =
single(client, operation, request, mat.system)
/**
@@ -91,8 +87,7 @@ object DynamoDb {
client: DynamoDbAsyncClient,
operation: DynamoDbOp[In, Out],
request: In,
- system: ClassicActorSystemProvider
- ): CompletionStage[Out] = {
+ system: ClassicActorSystemProvider): CompletionStage[Out] = {
val sink: Sink[Out, CompletionStage[Out]] = Sink.head()
Source.single(request).via(flow(client, operation, 1)).runWith(sink, system)
}
diff --git a/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/scaladsl/DynamoDb.scala b/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/scaladsl/DynamoDb.scala
index 366d0e0c..e7fe2036 100644
--- a/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/scaladsl/DynamoDb.scala
+++ b/dynamodb/src/main/scala/akka/stream/alpakka/dynamodb/scaladsl/DynamoDb.scala
@@ -9,13 +9,13 @@ import akka.actor.ClassicActorSystemProvider
import akka.dispatch.ExecutionContexts
import scala.annotation.implicitNotFound
-import akka.stream.alpakka.dynamodb.{DynamoDbOp, DynamoDbPaginatedOp}
-import akka.stream.scaladsl.{Flow, FlowWithContext, Sink, Source}
+import akka.stream.alpakka.dynamodb.{ DynamoDbOp, DynamoDbPaginatedOp }
+import akka.stream.scaladsl.{ Flow, FlowWithContext, Sink, Source }
import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient
import software.amazon.awssdk.services.dynamodb.model._
import scala.concurrent.Future
-import scala.util.{Failure, Success, Try}
+import scala.util.{ Failure, Success, Try }
/**
* Factory of DynamoDb Akka Stream operators.
@@ -28,8 +28,7 @@ object DynamoDb {
* @param parallelism maximum number of in-flight requests at any given time
*/
def flow[In <: DynamoDbRequest, Out <: DynamoDbResponse](
- parallelism: Int
- )(implicit client: DynamoDbAsyncClient, operation: DynamoDbOp[In, Out]): Flow[In, Out, NotUsed] =
+ parallelism: Int)(implicit client: DynamoDbAsyncClient, operation: DynamoDbOp[In, Out]): Flow[In, Out, NotUsed] =
Flow[In].mapAsync(parallelism)(operation.execute(_))
/**
@@ -43,9 +42,8 @@ object DynamoDb {
* @tparam Ctx context (or pass-through)
*/
def flowWithContext[In <: DynamoDbRequest, Out <: DynamoDbResponse, Ctx](
- parallelism: Int
- )(implicit client: DynamoDbAsyncClient,
- operation: DynamoDbOp[In, Out]): FlowWithContext[In, Ctx, Try[Out], Ctx, NotUsed] =
+ parallelism: Int)(implicit client: DynamoDbAsyncClient,
+ operation: DynamoDbOp[In, Out]): FlowWithContext[In, Ctx, Try[Out], Ctx, NotUsed] =
FlowWithContext.fromTuples(
Flow[(In, Ctx)]
.mapAsync(parallelism) {
@@ -54,15 +52,14 @@ object DynamoDb {
.execute(in)
.map[(Try[Out], Ctx)](res => (Success(res), ctx))(ExecutionContexts.parasitic)
.recover { case t => (Failure(t), ctx) }(ExecutionContexts.parasitic)
- }
- )
+ })
/**
* Create a Source that will emit potentially multiple responses for a given request.
*/
def source[In <: DynamoDbRequest, Out <: DynamoDbResponse](
- request: In
- )(implicit client: DynamoDbAsyncClient, operation: DynamoDbPaginatedOp[In, Out, _]): Source[Out, NotUsed] =
+ request: In)(
+ implicit client: DynamoDbAsyncClient, operation: DynamoDbPaginatedOp[In, Out, _]): Source[Out, NotUsed] =
Source.fromPublisher(operation.publisher(request))
/**
@@ -72,19 +69,16 @@ object DynamoDb {
*/
def flowPaginated[In <: DynamoDbRequest, Out <: DynamoDbResponse]()(
implicit client: DynamoDbAsyncClient,
- operation: DynamoDbPaginatedOp[In, Out, _]
- ): Flow[In, Out, NotUsed] = Flow[In].flatMapConcat(source(_))
+ operation: DynamoDbPaginatedOp[In, Out, _]): Flow[In, Out, NotUsed] = Flow[In].flatMapConcat(source(_))
/**
* Create a Future that will be completed with a response to a given request.
*/
@implicitNotFound(
- "a `ClassicActorSystemProvider` is a classic or new API actor system, provide this instead of a `Materializer`"
- )
+ "a `ClassicActorSystemProvider` is a classic or new API actor system, provide this instead of a `Materializer`")
def single[In <: DynamoDbRequest, Out <: DynamoDbResponse](
- request: In
- )(implicit client: DynamoDbAsyncClient,
- operation: DynamoDbOp[In, Out],
- system: ClassicActorSystemProvider): Future[Out] =
+ request: In)(implicit client: DynamoDbAsyncClient,
+ operation: DynamoDbOp[In, Out],
+ system: ClassicActorSystemProvider): Future[Out] =
Source.single(request).via(flow(1)).runWith(Sink.head)
}
diff --git a/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/ItemSpec.scala b/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/ItemSpec.scala
index 86e57e1a..e71424a0 100644
--- a/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/ItemSpec.scala
+++ b/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/ItemSpec.scala
@@ -13,7 +13,7 @@ import com.github.matsluni.akkahttpspi.AkkaHttpClient
import org.scalatest._
import org.scalatest.matchers.should.Matchers
import org.scalatest.wordspec.AsyncWordSpecLike
-import software.amazon.awssdk.auth.credentials.{AwsBasicCredentials, StaticCredentialsProvider}
+import software.amazon.awssdk.auth.credentials.{ AwsBasicCredentials, StaticCredentialsProvider }
import software.amazon.awssdk.regions.Region
import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient
import software.amazon.awssdk.services.dynamodb.model.TableStatus
@@ -114,7 +114,7 @@ class ItemSpec extends TestKit(ActorSystem("ItemSpec")) with AsyncWordSpecLike w
for {
_ <- DynamoDb.single(deleteTableRequest)
list <- DynamoDb.single(listTablesRequest)
- } yield list.tableNames.asScala should not contain (tableName)
+ } yield list.tableNames.asScala should not contain tableName
}
}
diff --git a/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TableSpec.scala b/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TableSpec.scala
index fafc2c2d..2da2a374 100644
--- a/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TableSpec.scala
+++ b/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TableSpec.scala
@@ -11,7 +11,7 @@ import akka.stream.alpakka.dynamodb.scaladsl.DynamoDb
import akka.testkit.TestKit
import com.github.matsluni.akkahttpspi.AkkaHttpClient
import org.scalatest.BeforeAndAfterAll
-import software.amazon.awssdk.auth.credentials.{AwsBasicCredentials, StaticCredentialsProvider}
+import software.amazon.awssdk.auth.credentials.{ AwsBasicCredentials, StaticCredentialsProvider }
import software.amazon.awssdk.regions.Region
import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient
@@ -80,7 +80,7 @@ class TableSpec extends TestKit(ActorSystem("TableSpec")) with AsyncWordSpecLike
for {
_ <- DynamoDb.single(deleteTableRequest)
list <- DynamoDb.single(listTablesRequest)
- } yield list.tableNames.asScala should not contain (tableName)
+ } yield list.tableNames.asScala should not contain tableName
}
}
diff --git a/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TestOps.scala b/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TestOps.scala
index 91093b3a..328e5b30 100644
--- a/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TestOps.scala
+++ b/dynamodb/src/test/scala/akka/stream/alpakka/dynamodb/TestOps.scala
@@ -19,16 +19,14 @@ trait TestOps {
def N(n: Int) = AttributeValue.builder().n(n.toString).build()
def keyMap(hash: String, sort: Int): Map[String, AttributeValue] = Map(
keyCol -> S(hash),
- sortCol -> N(sort)
- )
+ sortCol -> N(sort))
def keyEQ(hash: String): Map[String, Condition] = Map(
keyCol -> Condition
.builder()
.comparisonOperator(ComparisonOperator.EQ)
.attributeValueList(S(hash))
- .build()
- )
+ .build())
object common {
val listTablesRequest = ListTablesRequest.builder().build()
@@ -38,15 +36,12 @@ trait TestOps {
.tableName(tableName)
.keySchema(
KeySchemaElement.builder().attributeName(keyCol).keyType(KeyType.HASH).build(),
- KeySchemaElement.builder().attributeName(sortCol).keyType(KeyType.RANGE).build()
- )
+ KeySchemaElement.builder().attributeName(sortCol).keyType(KeyType.RANGE).build())
.attributeDefinitions(
AttributeDefinition.builder().attributeName(keyCol).attributeType(ScalarAttributeType.S).build(),
- AttributeDefinition.builder().attributeName(sortCol).attributeType(ScalarAttributeType.N).build()
- )
+ AttributeDefinition.builder().attributeName(sortCol).attributeType(ScalarAttributeType.N).build())
.provisionedThroughput(
- ProvisionedThroughput.builder().readCapacityUnits(10L).writeCapacityUnits(10L).build()
- )
+ ProvisionedThroughput.builder().readCapacityUnits(10L).writeCapacityUnits(10L).build())
.build()
val describeTableRequest = DescribeTableRequest.builder().tableName(tableName).build()
@@ -93,10 +88,7 @@ abstract class ItemSpecOps extends TestOps {
WriteRequest
.builder()
.putRequest(PutRequest.builder().item((keyMap("B", 1) + ("data" -> S(test5Data))).asJava).build())
- .build()
- ).asJava
- ).asJava
- )
+ .build()).asJava).asJava)
.build()
def batchWriteLargeItemRequest(from: Int, to: Int) =
@@ -110,12 +102,10 @@ abstract class ItemSpecOps extends TestOps {
WriteRequest
.builder()
.putRequest(
- PutRequest.builder().item((keyMap(i.toString, i) + ("data1" -> S("0123456789" * 39000))).asJava).build()
- )
+ PutRequest.builder().item((keyMap(i.toString, i) + ("data1" -> S(
+ "0123456789" * 39000))).asJava).build())
.build()
- }.asJava
- ).asJava
- )
+ }.asJava).asJava)
.build()
def batchGetLargeItemRequest(from: Int, to: Int) =
@@ -132,9 +122,7 @@ abstract class ItemSpecOps extends TestOps {
}.asJava
}
.attributesToGet("data1")
- .build()
- ).asJava
- )
+ .build()).asJava)
.build()
def batchGetItemRequest(items: java.util.Map[String, KeysAndAttributes]) =
@@ -177,9 +165,7 @@ abstract class ItemSpecOps extends TestOps {
TransactWriteItem
.builder()
.put(Put.builder().tableName(tableName).item((keyMap("C", 1) + ("data" -> S(test8Data))).asJava).build())
- .build()
- ).asJava
- )
+ .build()).asJava)
.build()
val transactGetItemsRequest = TransactGetItemsRequest
@@ -187,9 +173,8 @@ abstract class ItemSpecOps extends TestOps {
.transactItems(
List(
TransactGetItem.builder().get(Get.builder().tableName(tableName).key(keyMap("C", 0).asJava).build()).build(),
- TransactGetItem.builder().get(Get.builder().tableName(tableName).key(keyMap("C", 1).asJava).build()).build()
- ).asJava
- )
+ TransactGetItem.builder().get(Get.builder().tableName(tableName).key(keyMap("C",
+ 1).asJava).build()).build()).asJava)
.build()
val transactDeleteItemsRequest = TransactWriteItemsRequest
@@ -203,9 +188,7 @@ abstract class ItemSpecOps extends TestOps {
TransactWriteItem
.builder()
.delete(Delete.builder().tableName(tableName).key(keyMap("C", 1).asJava).build())
- .build()
- ).asJava
- )
+ .build()).asJava)
.build()
val deleteTableRequest = common.deleteTableRequest
@@ -229,8 +212,7 @@ object TableSpecOps extends TestOps {
.builder()
.tableName(tableName)
.provisionedThroughput(
- ProvisionedThroughput.builder().writeCapacityUnits(newMaxLimit).readCapacityUnits(newMaxLimit).build()
- )
+ ProvisionedThroughput.builder().writeCapacityUnits(newMaxLimit).readCapacityUnits(newMaxLimit).build())
.build()
val describeTimeToLiveRequest = DescribeTimeToLiveRequest.builder().build()
diff --git a/dynamodb/src/test/scala/docs/scaladsl/ExampleSpec.scala b/dynamodb/src/test/scala/docs/scaladsl/ExampleSpec.scala
index 10ac506f..98f863c8 100644
--- a/dynamodb/src/test/scala/docs/scaladsl/ExampleSpec.scala
+++ b/dynamodb/src/test/scala/docs/scaladsl/ExampleSpec.scala
@@ -8,22 +8,22 @@ import java.net.URI
import akka.NotUsed
import akka.stream.alpakka.testkit.scaladsl.LogCapturing
-import akka.stream.scaladsl.{FlowWithContext, SourceWithContext}
+import akka.stream.scaladsl.{ FlowWithContext, SourceWithContext }
-import scala.util.{Failure, Success, Try}
+import scala.util.{ Failure, Success, Try }
//#init-client
import akka.actor.ActorSystem
//#init-client
import akka.stream.alpakka.dynamodb.DynamoDbOp._
import akka.stream.alpakka.dynamodb.scaladsl._
-import akka.stream.scaladsl.{Sink, Source}
+import akka.stream.scaladsl.{ Sink, Source }
import akka.testkit.TestKit
import org.scalatest.concurrent.ScalaFutures
import org.scalatest.BeforeAndAfterAll
//#init-client
import com.github.matsluni.akkahttpspi.AkkaHttpClient
-import software.amazon.awssdk.auth.credentials.{AwsBasicCredentials, StaticCredentialsProvider}
+import software.amazon.awssdk.auth.credentials.{ AwsBasicCredentials, StaticCredentialsProvider }
import software.amazon.awssdk.regions.Region
import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient
@@ -46,7 +46,7 @@ class ExampleSpec
override implicit val patienceConfig: PatienceConfig = PatienceConfig(5.seconds, 100.millis)
- //#init-client
+ // #init-client
// Don't encode credentials in your source code!
// see https://doc.akka.io/docs/alpakka/current/aws-shared-configuration.html
@@ -59,14 +59,14 @@ class ExampleSpec
// Possibility to configure the retry policy
// see https://doc.akka.io/docs/alpakka/current/aws-shared-configuration.html
// .overrideConfiguration(...)
- //#init-client
+ // #init-client
.endpointOverride(new URI("http://localhost:8001/"))
- //#init-client
+ // #init-client
.build()
system.registerOnTermination(client.close())
- //#init-client
+ // #init-client
override def afterAll(): Unit = {
client.close();
@@ -105,8 +105,7 @@ class ExampleSpec
val source: SourceWithContext[PutItemRequest, SomeContext, NotUsed] = // ???
// #withContext
SourceWithContext.fromTuples(
- Source.single(PutItemRequest.builder().build() -> SomeContext())
- )
+ Source.single(PutItemRequest.builder().build() -> SomeContext()))
// #withContext
@@ -116,7 +115,7 @@ class ExampleSpec
val writtenSource: SourceWithContext[PutItemResponse, SomeContext, NotUsed] = source
.via(flow)
.map {
- case Success(response) => response
+ case Success(response) => response
case Failure(exception) => throw exception
}
// #withContext
@@ -128,8 +127,7 @@ class ExampleSpec
(for {
create <- DynamoDb.single(CreateTableRequest.builder().tableName("testTable").build())
describe <- DynamoDb.single(
- DescribeTableRequest.builder().tableName(create.tableDescription.tableName).build()
- )
+ DescribeTableRequest.builder().tableName(create.tableDescription.tableName).build())
} yield describe.table.itemCount).failed.futureValue
}
diff --git a/dynamodb/src/test/scala/docs/scaladsl/RetrySpec.scala b/dynamodb/src/test/scala/docs/scaladsl/RetrySpec.scala
index 70af732c..f1203a61 100644
--- a/dynamodb/src/test/scala/docs/scaladsl/RetrySpec.scala
+++ b/dynamodb/src/test/scala/docs/scaladsl/RetrySpec.scala
@@ -9,7 +9,7 @@ import akka.stream.alpakka.testkit.scaladsl.LogCapturing
import akka.testkit.TestKit
import com.github.matsluni.akkahttpspi.AkkaHttpClient
import org.scalatest.BeforeAndAfterAll
-import software.amazon.awssdk.auth.credentials.{AwsBasicCredentials, StaticCredentialsProvider}
+import software.amazon.awssdk.auth.credentials.{ AwsBasicCredentials, StaticCredentialsProvider }
// #awsRetryConfiguration
import software.amazon.awssdk.core.client.config.ClientOverrideConfiguration
import software.amazon.awssdk.core.internal.retry.SdkDefaultRetrySetting
@@ -46,10 +46,8 @@ class RetrySpec
.throttlingBackoffStrategy(BackoffStrategy.defaultThrottlingStrategy)
.numRetries(SdkDefaultRetrySetting.defaultMaxAttempts)
.retryCondition(RetryCondition.defaultRetryCondition)
- .build
- )
- .build()
- )
+ .build)
+ .build())
// #awsRetryConfiguration
.build()
// #clientRetryConfig
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchConnectionSettings.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchConnectionSettings.scala
index d0bf6cd2..f1078e3e 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchConnectionSettings.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchConnectionSettings.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.elasticsearch
-import akka.http.scaladsl.{ConnectionContext, HttpsConnectionContext}
+import akka.http.scaladsl.{ ConnectionContext, HttpsConnectionContext }
import akka.http.scaladsl.model.HttpHeader
import akka.http.scaladsl.model.HttpHeader.ParsingResult
import akka.japi.Util
@@ -18,8 +18,7 @@ final class ElasticsearchConnectionSettings private (
val username: Option[String],
val password: Option[String],
val headers: List[HttpHeader],
- val connectionContext: Option[HttpsConnectionContext]
-) {
+ val connectionContext: Option[HttpsConnectionContext]) {
def withBaseUrl(value: String): ElasticsearchConnectionSettings = copy(baseUrl = value)
@@ -57,23 +56,20 @@ final class ElasticsearchConnectionSettings private (
@deprecated("prefer ElasticsearchConnectionSettings.withSSLContext", "3.1.0")
@Deprecated
def withConnectionContext(
- connectionContext: akka.http.javadsl.HttpsConnectionContext
- ): ElasticsearchConnectionSettings = {
+ connectionContext: akka.http.javadsl.HttpsConnectionContext): ElasticsearchConnectionSettings = {
val scalaContext = new HttpsConnectionContext(
connectionContext.getSslContext,
None,
OptionConverters.toScala(connectionContext.getEnabledCipherSuites).map(Util.immutableSeq(_)),
OptionConverters.toScala(connectionContext.getEnabledProtocols).map(Util.immutableSeq(_)),
OptionConverters.toScala(connectionContext.getClientAuth),
- OptionConverters.toScala(connectionContext.getSslParameters)
- )
+ OptionConverters.toScala(connectionContext.getSslParameters))
copy(connectionContext = Option(scalaContext))
}
def withSSLContext(
- sslContext: SSLContext
- ): ElasticsearchConnectionSettings = {
+ sslContext: SSLContext): ElasticsearchConnectionSettings = {
copy(connectionContext = Option(ConnectionContext.httpsClient(sslContext)))
}
@@ -84,18 +80,16 @@ final class ElasticsearchConnectionSettings private (
username: Option[String] = username,
password: Option[String] = password,
headers: List[HttpHeader] = headers,
- connectionContext: Option[HttpsConnectionContext] = connectionContext
- ): ElasticsearchConnectionSettings =
+ connectionContext: Option[HttpsConnectionContext] = connectionContext): ElasticsearchConnectionSettings =
new ElasticsearchConnectionSettings(baseUrl = baseUrl,
- username = username,
- password = password,
- headers = headers,
- connectionContext = connectionContext)
+ username = username,
+ password = password,
+ headers = headers,
+ connectionContext = connectionContext)
override def toString =
- s"""ElasticsearchConnectionSettings(baseUrl=$baseUrl,username=$username,password=${password.fold("")(
- _ => "***"
- )},headers=${headers.mkString(";")},connectionContext=$connectionContext)"""
+ s"""ElasticsearchConnectionSettings(baseUrl=$baseUrl,username=$username,password=${password.fold("")(_ =>
+ "***")},headers=${headers.mkString(";")},connectionContext=$connectionContext)"""
}
object ElasticsearchConnectionSettings {
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchSourceSettings.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchSourceSettings.scala
index 0215fa2f..e1394d81 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchSourceSettings.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchSourceSettings.scala
@@ -10,29 +10,28 @@ import scala.concurrent.duration.FiniteDuration
/**
* Configure Elastiscsearch sources.
- *
*/
final class ElasticsearchSourceSettings private (connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- includeDocumentVersion: Boolean,
- scrollDuration: FiniteDuration,
- apiVersion: ApiVersion)
+ bufferSize: Int,
+ includeDocumentVersion: Boolean,
+ scrollDuration: FiniteDuration,
+ apiVersion: ApiVersion)
extends SourceSettingsBase[ApiVersion, ElasticsearchSourceSettings](connection,
- bufferSize,
- includeDocumentVersion,
- scrollDuration,
- apiVersion) {
+ bufferSize,
+ includeDocumentVersion,
+ scrollDuration,
+ apiVersion) {
protected override def copy(connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- includeDocumentVersion: Boolean,
- scrollDuration: FiniteDuration,
- apiVersion: ApiVersion): ElasticsearchSourceSettings =
+ bufferSize: Int,
+ includeDocumentVersion: Boolean,
+ scrollDuration: FiniteDuration,
+ apiVersion: ApiVersion): ElasticsearchSourceSettings =
new ElasticsearchSourceSettings(connection = connection,
- bufferSize = bufferSize,
- includeDocumentVersion = includeDocumentVersion,
- scrollDuration = scrollDuration,
- apiVersion = apiVersion)
+ bufferSize = bufferSize,
+ includeDocumentVersion = includeDocumentVersion,
+ scrollDuration = scrollDuration,
+ apiVersion = apiVersion)
override def toString =
s"""ElasticsearchSourceSettings(connection=$connection,bufferSize=$bufferSize,includeDocumentVersion=$includeDocumentVersion,scrollDuration=$scrollDuration,apiVersion=$apiVersion)"""
@@ -44,16 +43,16 @@ object ElasticsearchSourceSettings {
/** Scala API */
def apply(connection: ElasticsearchConnectionSettings): ElasticsearchSourceSettings =
new ElasticsearchSourceSettings(connection,
- 10,
- includeDocumentVersion = false,
- FiniteDuration(5, TimeUnit.MINUTES),
- ApiVersion.V7)
+ 10,
+ includeDocumentVersion = false,
+ FiniteDuration(5, TimeUnit.MINUTES),
+ ApiVersion.V7)
/** Java API */
def create(connection: ElasticsearchConnectionSettings): ElasticsearchSourceSettings =
new ElasticsearchSourceSettings(connection,
- 10,
- includeDocumentVersion = false,
- FiniteDuration(5, TimeUnit.MINUTES),
- ApiVersion.V7)
+ 10,
+ includeDocumentVersion = false,
+ FiniteDuration(5, TimeUnit.MINUTES),
+ ApiVersion.V7)
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchWriteSettings.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchWriteSettings.scala
index 07979653..b51da675 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchWriteSettings.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ElasticsearchWriteSettings.scala
@@ -37,8 +37,8 @@ object RetryAtFixedRate {
}
final class RetryWithBackoff(_maxRetries: Int,
- _minBackoff: scala.concurrent.duration.FiniteDuration,
- _maxBackoff: scala.concurrent.duration.FiniteDuration)
+ _minBackoff: scala.concurrent.duration.FiniteDuration,
+ _maxBackoff: scala.concurrent.duration.FiniteDuration)
extends RetryLogic {
override val maxRetries: Int = _maxRetries
override val minBackoff: scala.concurrent.duration.FiniteDuration = _minBackoff
@@ -48,8 +48,8 @@ final class RetryWithBackoff(_maxRetries: Int,
object RetryWithBackoff {
def apply(maxRetries: Int,
- minBackoff: scala.concurrent.duration.FiniteDuration,
- maxBackoff: scala.concurrent.duration.FiniteDuration): RetryWithBackoff =
+ minBackoff: scala.concurrent.duration.FiniteDuration,
+ maxBackoff: scala.concurrent.duration.FiniteDuration): RetryWithBackoff =
new RetryWithBackoff(maxRetries, minBackoff, maxBackoff)
def create(maxRetries: Int, minBackoff: java.time.Duration, maxBackoff: java.time.Duration): RetryWithBackoff =
@@ -60,24 +60,24 @@ object RetryWithBackoff {
* Configure Elasticsearch sinks and flows.
*/
final class ElasticsearchWriteSettings private (connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- retryLogic: RetryLogic,
- versionType: Option[String],
- apiVersion: ApiVersion,
- allowExplicitIndex: Boolean)
+ bufferSize: Int,
+ retryLogic: RetryLogic,
+ versionType: Option[String],
+ apiVersion: ApiVersion,
+ allowExplicitIndex: Boolean)
extends WriteSettingsBase[ApiVersion, ElasticsearchWriteSettings](connection,
- bufferSize,
- retryLogic,
- versionType,
- apiVersion,
- allowExplicitIndex) {
+ bufferSize,
+ retryLogic,
+ versionType,
+ apiVersion,
+ allowExplicitIndex) {
protected override def copy(connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- retryLogic: RetryLogic,
- versionType: Option[String],
- apiVersion: ApiVersion,
- allowExplicitIndex: Boolean): ElasticsearchWriteSettings =
+ bufferSize: Int,
+ retryLogic: RetryLogic,
+ versionType: Option[String],
+ apiVersion: ApiVersion,
+ allowExplicitIndex: Boolean): ElasticsearchWriteSettings =
new ElasticsearchWriteSettings(connection, bufferSize, retryLogic, versionType, apiVersion, allowExplicitIndex)
override def toString: String =
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchSourceSettings.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchSourceSettings.scala
index 72353e28..846c573f 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchSourceSettings.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchSourceSettings.scala
@@ -9,28 +9,27 @@ import scala.concurrent.duration.FiniteDuration
/**
* Configure Opensearch sources.
- *
*/
final class OpensearchSourceSettings private (connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- includeDocumentVersion: Boolean,
- scrollDuration: FiniteDuration,
- apiVersion: OpensearchApiVersion)
+ bufferSize: Int,
+ includeDocumentVersion: Boolean,
+ scrollDuration: FiniteDuration,
+ apiVersion: OpensearchApiVersion)
extends SourceSettingsBase[OpensearchApiVersion, OpensearchSourceSettings](connection,
- bufferSize,
- includeDocumentVersion,
- scrollDuration,
- apiVersion) {
+ bufferSize,
+ includeDocumentVersion,
+ scrollDuration,
+ apiVersion) {
protected override def copy(connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- includeDocumentVersion: Boolean,
- scrollDuration: FiniteDuration,
- apiVersion: OpensearchApiVersion): OpensearchSourceSettings =
+ bufferSize: Int,
+ includeDocumentVersion: Boolean,
+ scrollDuration: FiniteDuration,
+ apiVersion: OpensearchApiVersion): OpensearchSourceSettings =
new OpensearchSourceSettings(connection = connection,
- bufferSize = bufferSize,
- includeDocumentVersion = includeDocumentVersion,
- scrollDuration = scrollDuration,
- apiVersion = apiVersion)
+ bufferSize = bufferSize,
+ includeDocumentVersion = includeDocumentVersion,
+ scrollDuration = scrollDuration,
+ apiVersion = apiVersion)
override def toString =
s"""OpensearchSourceSettings(connection=$connection,bufferSize=$bufferSize,includeDocumentVersion=$includeDocumentVersion,scrollDuration=$scrollDuration,apiVersion=$apiVersion)"""
@@ -42,16 +41,16 @@ object OpensearchSourceSettings {
/** Scala API */
def apply(connection: ElasticsearchConnectionSettings): OpensearchSourceSettings =
new OpensearchSourceSettings(connection,
- 10,
- includeDocumentVersion = false,
- FiniteDuration(5, TimeUnit.MINUTES),
- OpensearchApiVersion.V1)
+ 10,
+ includeDocumentVersion = false,
+ FiniteDuration(5, TimeUnit.MINUTES),
+ OpensearchApiVersion.V1)
/** Java API */
def create(connection: ElasticsearchConnectionSettings): OpensearchSourceSettings =
new OpensearchSourceSettings(connection,
- 10,
- includeDocumentVersion = false,
- FiniteDuration(5, TimeUnit.MINUTES),
- OpensearchApiVersion.V1)
+ 10,
+ includeDocumentVersion = false,
+ FiniteDuration(5, TimeUnit.MINUTES),
+ OpensearchApiVersion.V1)
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchWriteSettings.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchWriteSettings.scala
index 54b6d4fe..cb17923a 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchWriteSettings.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/OpensearchWriteSettings.scala
@@ -8,24 +8,24 @@ package akka.stream.alpakka.elasticsearch
* Configure Opensearch sinks and flows.
*/
final class OpensearchWriteSettings private (connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- retryLogic: RetryLogic,
- versionType: Option[String],
- apiVersion: OpensearchApiVersion,
- allowExplicitIndex: Boolean)
+ bufferSize: Int,
+ retryLogic: RetryLogic,
+ versionType: Option[String],
+ apiVersion: OpensearchApiVersion,
+ allowExplicitIndex: Boolean)
extends WriteSettingsBase[OpensearchApiVersion, OpensearchWriteSettings](connection,
- bufferSize,
- retryLogic,
- versionType,
- apiVersion,
- allowExplicitIndex) {
+ bufferSize,
+ retryLogic,
+ versionType,
+ apiVersion,
+ allowExplicitIndex) {
protected override def copy(connection: ElasticsearchConnectionSettings,
- bufferSize: Int,
- retryLogic: RetryLogic,
- versionType: Option[String],
- apiVersion: OpensearchApiVersion,
- allowExplicitIndex: Boolean): OpensearchWriteSettings =
+ bufferSize: Int,
+ retryLogic: RetryLogic,
+ versionType: Option[String],
+ apiVersion: OpensearchApiVersion,
+ allowExplicitIndex: Boolean): OpensearchWriteSettings =
new OpensearchWriteSettings(connection, bufferSize, retryLogic, versionType, apiVersion, allowExplicitIndex)
override def toString: String =
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ReadResult.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ReadResult.scala
index 45204630..6ed80da6 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ReadResult.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/ReadResult.scala
@@ -15,8 +15,8 @@ import scala.compat.java8.OptionConverters._
* [[akka.stream.alpakka.elasticsearch.testkit.MessageFactory]].
*/
final class ReadResult[T] @InternalApi private[elasticsearch] (val id: String,
- val source: T,
- val version: Option[Long]) {
+ val source: T,
+ val version: Option[Long]) {
/** Java API */
def getVersion: java.util.Optional[Long] = version.asJava
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/SourceSettingsBase.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/SourceSettingsBase.scala
index 123eafdf..1d67faa9 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/SourceSettingsBase.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/SourceSettingsBase.scala
@@ -13,15 +13,13 @@ import scala.concurrent.duration.FiniteDuration
/**
* Configure Elastiscsearch/OpenSearch sources.
- *
*/
abstract class SourceSettingsBase[Version <: ApiVersionBase, S <: SourceSettingsBase[Version, S]] private[alpakka] (
val connection: ElasticsearchConnectionSettings,
val bufferSize: Int,
val includeDocumentVersion: Boolean,
val scrollDuration: FiniteDuration,
- val apiVersion: Version
-) { this: S =>
+ val apiVersion: Version) { this: S =>
def withConnection(value: ElasticsearchConnectionSettings): S = copy(connection = value)
def withBufferSize(value: Int): S = copy(bufferSize = value)
@@ -43,22 +41,22 @@ abstract class SourceSettingsBase[Version <: ApiVersionBase, S <: SourceSettings
def scroll: String = {
val scrollString = scrollDuration.unit match {
- case TimeUnit.DAYS => "d"
- case TimeUnit.HOURS => "h"
- case TimeUnit.MINUTES => "m"
- case TimeUnit.SECONDS => "s"
+ case TimeUnit.DAYS => "d"
+ case TimeUnit.HOURS => "h"
+ case TimeUnit.MINUTES => "m"
+ case TimeUnit.SECONDS => "s"
case TimeUnit.MILLISECONDS => "ms"
case TimeUnit.MICROSECONDS => "micros"
- case TimeUnit.NANOSECONDS => "nanos"
+ case TimeUnit.NANOSECONDS => "nanos"
}
s"${scrollDuration.length}$scrollString"
}
protected def copy(connection: ElasticsearchConnectionSettings = connection,
- bufferSize: Int = bufferSize,
- includeDocumentVersion: Boolean = includeDocumentVersion,
- scrollDuration: FiniteDuration = scrollDuration,
- apiVersion: Version = apiVersion): S;
+ bufferSize: Int = bufferSize,
+ includeDocumentVersion: Boolean = includeDocumentVersion,
+ scrollDuration: FiniteDuration = scrollDuration,
+ apiVersion: Version = apiVersion): S;
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteMessage.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteMessage.scala
index 2b17533d..65ec975f 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteMessage.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteMessage.scala
@@ -32,23 +32,23 @@ private[elasticsearch] object Operation {
}
final class WriteMessage[T, PT] private (val operation: Operation,
- val id: Option[String],
- val source: Option[T],
- val passThrough: PT = NotUsed,
- val version: Option[Long] = None,
- val indexName: Option[String] = None,
- val customMetadata: Map[String, java.lang.String] = Map.empty) {
+ val id: Option[String],
+ val source: Option[T],
+ val passThrough: PT = NotUsed,
+ val version: Option[Long] = None,
+ val indexName: Option[String] = None,
+ val customMetadata: Map[String, java.lang.String] = Map.empty) {
def withSource(value: T): WriteMessage[T, PT] = copy(source = Option(value))
def withPassThrough[PT2](value: PT2): WriteMessage[T, PT2] =
new WriteMessage[T, PT2](operation = operation,
- id = id,
- source = source,
- value,
- version = version,
- indexName = indexName,
- customMetadata = customMetadata)
+ id = id,
+ source = source,
+ value,
+ version = version,
+ indexName = indexName,
+ customMetadata = customMetadata)
def withVersion(value: Long): WriteMessage[T, PT] = copy(version = Option(value))
def withIndexName(value: String): WriteMessage[T, PT] = copy(indexName = Option(value))
@@ -67,19 +67,19 @@ final class WriteMessage[T, PT] private (val operation: Operation,
this.copy(customMetadata = metadata.asScala.toMap)
private def copy(operation: Operation = operation,
- id: Option[String] = id,
- source: Option[T] = source,
- passThrough: PT = passThrough,
- version: Option[Long] = version,
- indexName: Option[String] = indexName,
- customMetadata: Map[String, String] = customMetadata): WriteMessage[T, PT] =
+ id: Option[String] = id,
+ source: Option[T] = source,
+ passThrough: PT = passThrough,
+ version: Option[Long] = version,
+ indexName: Option[String] = indexName,
+ customMetadata: Map[String, String] = customMetadata): WriteMessage[T, PT] =
new WriteMessage[T, PT](operation = operation,
- id = id,
- source = source,
- passThrough = passThrough,
- version = version,
- indexName = indexName,
- customMetadata = customMetadata)
+ id = id,
+ source = source,
+ passThrough = passThrough,
+ version = version,
+ indexName = indexName,
+ customMetadata = customMetadata)
override def toString =
s"""WriteMessage(operation=$operation,id=$id,source=$source,passThrough=$passThrough,version=$version,indexName=$indexName,customMetadata=$customMetadata)"""
@@ -138,8 +138,8 @@ object WriteMessage {
* [[akka.stream.alpakka.elasticsearch.testkit.MessageFactory]].
*/
final class WriteResult[T2, C2] @InternalApi private[elasticsearch] (val message: WriteMessage[T2, C2],
- /** JSON structure of the Elasticsearch error. */
- val error: Option[String]) {
+ /** JSON structure of the Elasticsearch error. */
+ val error: Option[String]) {
val success: Boolean = error.isEmpty
/** Java API: JSON structure of the Elasticsearch error. */
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteSettingsBase.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteSettingsBase.scala
index ac0d37a4..091d9c9a 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteSettingsBase.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/WriteSettingsBase.scala
@@ -16,8 +16,7 @@ abstract class WriteSettingsBase[Version <: ApiVersionBase, W <: WriteSettingsBa
val retryLogic: RetryLogic,
val versionType: Option[String],
val apiVersion: Version,
- val allowExplicitIndex: Boolean
-) { this: W =>
+ val allowExplicitIndex: Boolean) { this: W =>
def withConnection(value: ElasticsearchConnectionSettings): W = copy(connection = value)
@@ -34,9 +33,9 @@ abstract class WriteSettingsBase[Version <: ApiVersionBase, W <: WriteSettingsBa
def withAllowExplicitIndex(value: Boolean): W = copy(allowExplicitIndex = value)
protected def copy(connection: ElasticsearchConnectionSettings = connection,
- bufferSize: Int = bufferSize,
- retryLogic: RetryLogic = retryLogic,
- versionType: Option[String] = versionType,
- apiVersion: Version = apiVersion,
- allowExplicitIndex: Boolean = allowExplicitIndex): W;
+ bufferSize: Int = bufferSize,
+ retryLogic: RetryLogic = retryLogic,
+ versionType: Option[String] = versionType,
+ apiVersion: Version = apiVersion,
+ allowExplicitIndex: Boolean = allowExplicitIndex): W;
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchApi.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchApi.scala
index 9985c7b7..9719fe57 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchApi.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchApi.scala
@@ -15,16 +15,14 @@ import scala.concurrent.Future
@InternalApi private[impl] object ElasticsearchApi {
def executeRequest(
request: HttpRequest,
- connectionSettings: ElasticsearchConnectionSettings
- )(implicit http: HttpExt): Future[HttpResponse] = {
+ connectionSettings: ElasticsearchConnectionSettings)(implicit http: HttpExt): Future[HttpResponse] = {
if (connectionSettings.hasCredentialsDefined) {
http.singleRequest(
- request.addCredentials(BasicHttpCredentials(connectionSettings.username.get, connectionSettings.password.get))
- )
+ request.addCredentials(BasicHttpCredentials(connectionSettings.username.get, connectionSettings.password.get)))
} else {
http.singleRequest(request,
- connectionContext =
- connectionSettings.connectionContext.getOrElse(http.defaultClientHttpsContext))
+ connectionContext =
+ connectionSettings.connectionContext.getOrElse(http.defaultClientHttpsContext))
}
}
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSimpleFlowStage.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSimpleFlowStage.scala
index d2d722a7..241d9319 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSimpleFlowStage.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSimpleFlowStage.scala
@@ -15,7 +15,7 @@ import akka.stream._
import akka.stream.alpakka.elasticsearch
import scala.collection.immutable
-import scala.concurrent.{ExecutionContext, Future}
+import scala.concurrent.{ ExecutionContext, Future }
/**
* INTERNAL API.
@@ -26,11 +26,10 @@ import scala.concurrent.{ExecutionContext, Future}
private[elasticsearch] final class ElasticsearchSimpleFlowStage[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- writer: MessageWriter[T]
-)(implicit http: HttpExt, mat: Materializer, ec: ExecutionContext)
+ writer: MessageWriter[T])(implicit http: HttpExt, mat: Materializer, ec: ExecutionContext)
extends GraphStage[
- FlowShape[(immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]]), immutable.Seq[WriteResult[T, C]]]
- ] {
+ FlowShape[(immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]]), immutable.Seq[WriteResult[T,
+ C]]]] {
private val in =
Inlet[(immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]])]("messagesAndResultPassthrough")
@@ -40,10 +39,10 @@ private[elasticsearch] final class ElasticsearchSimpleFlowStage[T, C](
private val restApi: RestBulkApi[T, C] = settings.apiVersion match {
case ApiVersion.V5 =>
new RestBulkApiV5[T, C](elasticsearchParams.indexName,
- elasticsearchParams.typeName.get,
- settings.versionType,
- settings.allowExplicitIndex,
- writer)
+ elasticsearchParams.typeName.get,
+ settings.versionType,
+ settings.allowExplicitIndex,
+ writer)
case ApiVersion.V7 =>
new RestBulkApiV7[T, C](elasticsearchParams.indexName, settings.versionType, settings.allowExplicitIndex, writer)
@@ -88,8 +87,7 @@ private[elasticsearch] final class ElasticsearchSimpleFlowStage[T, C](
ElasticsearchApi
.executeRequest(
request,
- connectionSettings = settings.connection
- )
+ connectionSettings = settings.connection)
.map {
case HttpResponse(StatusCodes.OK, _, responseEntity, _) =>
Unmarshal(responseEntity)
@@ -99,8 +97,7 @@ private[elasticsearch] final class ElasticsearchSimpleFlowStage[T, C](
Unmarshal(response.entity).to[String].map { body =>
failureHandler.invoke(
(resultsPassthrough,
- new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body"))
- )
+ new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body")))
}
}
.recoverWith {
@@ -115,20 +112,18 @@ private[elasticsearch] final class ElasticsearchSimpleFlowStage[T, C](
}
private def handleFailure(
- args: (immutable.Seq[WriteResult[T, C]], Throwable)
- ): Unit = {
+ args: (immutable.Seq[WriteResult[T, C]], Throwable)): Unit = {
inflight = false
val (resultsPassthrough, exception) = args
log.error(s"Received error from elastic after having already processed {} documents. Error: {}",
- resultsPassthrough.size,
- exception)
+ resultsPassthrough.size,
+ exception)
failStage(exception)
}
private def handleResponse(
- args: (immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]], String)
- ): Unit = {
+ args: (immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]], String)): Unit = {
inflight = false
val (messages, resultsPassthrough, response) = args
@@ -142,7 +137,7 @@ private[elasticsearch] final class ElasticsearchSimpleFlowStage[T, C](
messageResults.filterNot(_.success).foreach { failure =>
if (failure.getError.isPresent) {
log.error(s"Received error from elastic when attempting to index documents. Error: {}",
- failure.getError.get)
+ failure.getError.get)
}
}
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSourceStage.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSourceStage.scala
index eb4075b9..d132b499 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSourceStage.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/ElasticsearchSourceStage.scala
@@ -16,13 +16,13 @@ import akka.stream.alpakka.elasticsearch.{
ReadResult,
SourceSettingsBase
}
-import akka.stream.stage.{GraphStage, GraphStageLogic, OutHandler, StageLogging}
-import akka.stream.{Attributes, Materializer, Outlet, SourceShape}
+import akka.stream.stage.{ GraphStage, GraphStageLogic, OutHandler, StageLogging }
+import akka.stream.{ Attributes, Materializer, Outlet, SourceShape }
import spray.json.DefaultJsonProtocol._
import spray.json._
import scala.concurrent.ExecutionContext
-import scala.util.{Failure, Success, Try}
+import scala.util.{ Failure, Success, Try }
/**
* INTERNAL API
@@ -52,8 +52,7 @@ private[elasticsearch] final class ElasticsearchSourceStage[T](
elasticsearchParams: ElasticsearchParams,
searchParams: Map[String, String],
settings: SourceSettingsBase[_, _],
- reader: MessageReader[T]
-)(implicit http: HttpExt, mat: Materializer, ec: ExecutionContext)
+ reader: MessageReader[T])(implicit http: HttpExt, mat: Materializer, ec: ExecutionContext)
extends GraphStage[SourceShape[ReadResult[T]]] {
val out: Outlet[ReadResult[T]] = Outlet("ElasticsearchSource.out")
@@ -80,8 +79,7 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
settings: SourceSettingsBase[_, _],
out: Outlet[ReadResult[T]],
shape: SourceShape[ReadResult[T]],
- reader: MessageReader[T]
-)(implicit http: HttpExt, mat: Materializer, ec: ExecutionContext)
+ reader: MessageReader[T])(implicit http: HttpExt, mat: Materializer, ec: ExecutionContext)
extends GraphStageLogic(shape)
with OutHandler
with StageLogging {
@@ -110,7 +108,7 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
// Add extra params to search
val extraParams = Seq(
if (!searchParams.contains("size")) {
- Some(("size" -> settings.bufferSize.toString))
+ Some("size" -> settings.bufferSize.toString)
} else {
None
},
@@ -118,11 +116,10 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
// http://nocf-www.elastic.co/guide/en/elasticsearch/reference/current/search-request-version.html
// https://www.elastic.co/guide/en/elasticsearch/guide/current/optimistic-concurrency-control.html
if (!searchParams.contains("version") && settings.includeDocumentVersion) {
- Some(("version" -> "true"))
+ Some("version" -> "true")
} else {
None
- }
- )
+ })
val baseMap = Map("scroll" -> settings.scroll)
@@ -139,17 +136,17 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
val completeParams = searchParams ++ extraParams.flatten - "routing"
val searchBody = "{" + completeParams
- .map {
- case (name, json) =>
- "\"" + name + "\":" + json
- }
- .mkString(",") + "}"
+ .map {
+ case (name, json) =>
+ "\"" + name + "\":" + json
+ }
+ .mkString(",") + "}"
val endpoint: String = settings.apiVersion match {
- case ApiVersion.V5 => s"/${elasticsearchParams.indexName}/${elasticsearchParams.typeName.get}/_search"
- case ApiVersion.V7 => s"/${elasticsearchParams.indexName}/_search"
+ case ApiVersion.V5 => s"/${elasticsearchParams.indexName}/${elasticsearchParams.typeName.get}/_search"
+ case ApiVersion.V7 => s"/${elasticsearchParams.indexName}/_search"
case OpensearchApiVersion.V1 => s"/${elasticsearchParams.indexName}/_search"
- case other => throw new IllegalArgumentException(s"API version $other is not supported")
+ case other => throw new IllegalArgumentException(s"API version $other is not supported")
}
val uri = prepareUri(Path(endpoint))
@@ -158,15 +155,13 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
val request = HttpRequest(HttpMethods.POST)
.withUri(uri)
.withEntity(
- HttpEntity(ContentTypes.`application/json`, searchBody)
- )
+ HttpEntity(ContentTypes.`application/json`, searchBody))
.withHeaders(settings.connection.headers)
ElasticsearchApi
.executeRequest(
request,
- settings.connection
- )
+ settings.connection)
.flatMap {
case HttpResponse(StatusCodes.OK, _, responseEntity, _) =>
Unmarshal(responseEntity)
@@ -176,8 +171,7 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
Unmarshal(response.entity).to[String].map { body =>
failureHandler
.invoke(
- new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body")
- )
+ new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body"))
}
}
.recover {
@@ -194,15 +188,13 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
.withUri(uri)
.withEntity(
HttpEntity(ContentTypes.`application/json`,
- Map("scroll" -> settings.scroll, "scroll_id" -> actualScrollId).toJson.compactPrint)
- )
+ Map("scroll" -> settings.scroll, "scroll_id" -> actualScrollId).toJson.compactPrint))
.withHeaders(settings.connection.headers)
ElasticsearchApi
.executeRequest(
request,
- settings.connection
- )
+ settings.connection)
.flatMap {
case HttpResponse(StatusCodes.OK, _, responseEntity, _) =>
Unmarshal(responseEntity)
@@ -213,8 +205,7 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
.to[String]
.map { body =>
failureHandler.invoke(
- new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body")
- )
+ new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body"))
}
}
.recover {
@@ -343,9 +334,7 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
clearScrollAsyncHandler
.invoke(
Failure(
- new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body")
- )
- )
+ new RuntimeException(s"Request failed for POST $uri, got ${response.status} with body: $body")))
}
}
.recover {
@@ -355,12 +344,12 @@ private[elasticsearch] final class ElasticsearchSourceLogic[T](
}
}
- private val clearScrollAsyncHandler = getAsyncCallback[Try[String]]({ result =>
+ private val clearScrollAsyncHandler = getAsyncCallback[Try[String]] { result =>
{
// Note: the scroll will expire, so there is no reason to consider a failed
// clear as a reason to fail the stream.
log.debug("Result of clearing the scroll: {}", result)
completeStage()
}
- })
+ }
}
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/NDJsonProtocol.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/NDJsonProtocol.scala
index 475b34fc..6c5b4a32 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/NDJsonProtocol.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/NDJsonProtocol.scala
@@ -4,7 +4,7 @@
package akka.stream.alpakka.elasticsearch.impl
-import akka.http.scaladsl.model.{ContentType, HttpCharsets, MediaType}
+import akka.http.scaladsl.model.{ ContentType, HttpCharsets, MediaType }
object NDJsonProtocol {
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApi.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApi.scala
index 1a800655..b8262a85 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApi.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApi.scala
@@ -5,8 +5,8 @@
package akka.stream.alpakka.elasticsearch.impl
import akka.annotation.InternalApi
-import akka.stream.alpakka.elasticsearch.Operation.{Create, Delete, Index, Nop, Update, Upsert}
-import akka.stream.alpakka.elasticsearch.{WriteMessage, WriteResult}
+import akka.stream.alpakka.elasticsearch.Operation.{ Create, Delete, Index, Nop, Update, Upsert }
+import akka.stream.alpakka.elasticsearch.{ WriteMessage, WriteResult }
import spray.json._
import scala.collection.immutable
@@ -20,7 +20,7 @@ private[impl] abstract class RestBulkApi[T, C] {
def toJson(messages: immutable.Seq[WriteMessage[T, C]]): String
def toWriteResults(messages: immutable.Seq[WriteMessage[T, C]],
- jsonString: String): immutable.Seq[WriteResult[T, C]] = {
+ jsonString: String): immutable.Seq[WriteResult[T, C]] = {
val responseJson = jsonString.parseJson
// If some commands in bulk request failed, pass failed messages to follows.
@@ -36,17 +36,17 @@ private[impl] abstract class RestBulkApi[T, C] {
def messageToJson(message: WriteMessage[T, C], messageSource: String): String = message.operation match {
case Index | Create => "\n" + messageSource
- case Upsert => "\n" + JsObject("doc" -> messageSource.parseJson, "doc_as_upsert" -> JsTrue).toString
- case Update => "\n" + JsObject("doc" -> messageSource.parseJson).toString
- case Delete => ""
- case Nop => ""
+ case Upsert => "\n" + JsObject("doc" -> messageSource.parseJson, "doc_as_upsert" -> JsTrue).toString
+ case Update => "\n" + JsObject("doc" -> messageSource.parseJson).toString
+ case Delete => ""
+ case Nop => ""
}
def constructSharedFields(message: WriteMessage[T, C]): Seq[(String, JsString)]
/** NOPs don't come back so slip them into the results like this: */
private def buildMessageResults(items: JsArray,
- messages: immutable.Seq[WriteMessage[T, C]]): immutable.Seq[WriteResult[T, C]] = {
+ messages: immutable.Seq[WriteMessage[T, C]]): immutable.Seq[WriteResult[T, C]] = {
val ret = new immutable.VectorBuilder[WriteResult[T, C]]
ret.sizeHint(messages)
val itemsIter = items.elements.iterator
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV5.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV5.scala
index 48af4ab4..f9c1c1fb 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV5.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV5.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.elasticsearch.impl
import akka.annotation.InternalApi
import akka.stream.alpakka.elasticsearch.Operation._
-import akka.stream.alpakka.elasticsearch.{MessageWriter, WriteMessage}
+import akka.stream.alpakka.elasticsearch.{ MessageWriter, WriteMessage }
import spray.json._
import scala.collection.immutable
@@ -19,10 +19,10 @@ import scala.collection.immutable
*/
@InternalApi
private[impl] final class RestBulkApiV5[T, C](indexName: String,
- typeName: String,
- versionType: Option[String],
- allowExplicitIndex: Boolean,
- messageWriter: MessageWriter[T])
+ typeName: String,
+ versionType: Option[String],
+ allowExplicitIndex: Boolean,
+ messageWriter: MessageWriter[T])
extends RestBulkApi[T, C] {
private lazy val typeNameTuple = "_type" -> JsString(typeName)
@@ -36,23 +36,20 @@ private[impl] final class RestBulkApiV5[T, C](indexName: String,
val fields = Seq(
optionalNumber("_version", message.version),
optionalString("version_type", versionType),
- optionalString("_id", message.id)
- ).flatten
+ optionalString("_id", message.id)).flatten
"index" -> JsObject(sharedFields ++ fields: _*)
case Create => "create" -> JsObject(sharedFields ++ optionalString("_id", message.id): _*)
case Update | Upsert =>
val fields =
("_id" -> JsString(message.id.get)) +: Seq(
optionalNumber("_version", message.version),
- optionalString("version_type", versionType)
- ).flatten
+ optionalString("version_type", versionType)).flatten
"update" -> JsObject(sharedFields ++ fields: _*)
case Delete =>
val fields =
("_id" -> JsString(message.id.get)) +: Seq(
optionalNumber("_version", message.version),
- optionalString("version_type", versionType)
- ).flatten
+ optionalString("version_type", versionType)).flatten
"delete" -> JsObject(sharedFields ++ fields: _*)
case Nop => "" -> JsObject()
}
@@ -63,7 +60,7 @@ private[impl] final class RestBulkApiV5[T, C](indexName: String,
}
.filter(_.nonEmpty) match {
case Nil => "" // if all NOPs
- case x => x.mkString("", "\n", "\n")
+ case x => x.mkString("", "\n", "\n")
}
override def constructSharedFields(message: WriteMessage[T, C]): Seq[(String, JsString)] = {
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV7.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV7.scala
index 1fdf8fd8..45f5c45d 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV7.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/impl/RestBulkApiV7.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.elasticsearch.impl
import akka.annotation.InternalApi
import akka.stream.alpakka.elasticsearch.Operation._
-import akka.stream.alpakka.elasticsearch.{MessageWriter, WriteMessage}
+import akka.stream.alpakka.elasticsearch.{ MessageWriter, WriteMessage }
import spray.json._
import scala.collection.immutable
@@ -19,9 +19,9 @@ import scala.collection.immutable
*/
@InternalApi
private[impl] final class RestBulkApiV7[T, C](indexName: String,
- versionType: Option[String],
- allowExplicitIndex: Boolean,
- messageWriter: MessageWriter[T])
+ versionType: Option[String],
+ allowExplicitIndex: Boolean,
+ messageWriter: MessageWriter[T])
extends RestBulkApi[T, C] {
def toJson(messages: immutable.Seq[WriteMessage[T, C]]): String =
@@ -33,17 +33,15 @@ private[impl] final class RestBulkApiV7[T, C](indexName: String,
val fields = Seq(
optionalNumber("version", message.version),
optionalString("version_type", versionType),
- optionalString("_id", message.id)
- ).flatten
+ optionalString("_id", message.id)).flatten
"index" -> JsObject(sharedFields ++ fields: _*)
- case Create => "create" -> JsObject(sharedFields ++ optionalString("_id", message.id): _*)
+ case Create => "create" -> JsObject(sharedFields ++ optionalString("_id", message.id): _*)
case Update | Upsert => "update" -> JsObject(sharedFields :+ ("_id" -> JsString(message.id.get)): _*)
case Delete =>
val fields =
("_id" -> JsString(message.id.get)) +: Seq(
optionalNumber("version", message.version),
- optionalString("version_type", versionType)
- ).flatten
+ optionalString("version_type", versionType)).flatten
"delete" -> JsObject(sharedFields ++ fields: _*)
case Nop => "" -> JsObject()
}
@@ -54,7 +52,7 @@ private[impl] final class RestBulkApiV7[T, C](indexName: String,
}
.filter(_.nonEmpty) match {
case Nil => "" // if all NOPs
- case x => x.mkString("", "\n", "\n")
+ case x => x.mkString("", "\n", "\n")
}
override def constructSharedFields(message: WriteMessage[T, C]): Seq[(String, JsString)] = {
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchFlow.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchFlow.scala
index 8f19c4a9..fa9759ce 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchFlow.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchFlow.scala
@@ -6,7 +6,7 @@ package akka.stream.alpakka.elasticsearch.javadsl
import akka.NotUsed
import akka.annotation.ApiMayChange
-import akka.stream.alpakka.elasticsearch.{scaladsl, _}
+import akka.stream.alpakka.elasticsearch.{ scaladsl, _ }
import com.fasterxml.jackson.databind.ObjectMapper
import scala.jdk.CollectionConverters._
@@ -28,8 +28,8 @@ object ElasticsearchFlow {
def create[T](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- objectMapper: ObjectMapper
- ): akka.stream.javadsl.Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] =
+ objectMapper: ObjectMapper)
+ : akka.stream.javadsl.Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] =
create(elasticsearchParams, settings, new JacksonWriter[T](objectMapper))
/**
@@ -44,8 +44,8 @@ object ElasticsearchFlow {
def create[T](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- messageWriter: MessageWriter[T]
- ): akka.stream.javadsl.Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] =
+ messageWriter: MessageWriter[T])
+ : akka.stream.javadsl.Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] =
scaladsl.ElasticsearchFlow
.create(elasticsearchParams, settings, messageWriter)
.asJava
@@ -63,8 +63,7 @@ object ElasticsearchFlow {
def createWithPassThrough[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- objectMapper: ObjectMapper
- ): akka.stream.javadsl.Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] =
+ objectMapper: ObjectMapper): akka.stream.javadsl.Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] =
createWithPassThrough(elasticsearchParams, settings, new JacksonWriter[T](objectMapper))
/**
@@ -80,8 +79,7 @@ object ElasticsearchFlow {
def createWithPassThrough[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- messageWriter: MessageWriter[T]
- ): akka.stream.javadsl.Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] =
+ messageWriter: MessageWriter[T]): akka.stream.javadsl.Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] =
scaladsl.ElasticsearchFlow
.createWithPassThrough(elasticsearchParams, settings, messageWriter)
.asJava
@@ -100,8 +98,8 @@ object ElasticsearchFlow {
def createBulk[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- objectMapper: ObjectMapper
- ): akka.stream.javadsl.Flow[java.util.List[WriteMessage[T, C]], java.util.List[WriteResult[T, C]], NotUsed] =
+ objectMapper: ObjectMapper)
+ : akka.stream.javadsl.Flow[java.util.List[WriteMessage[T, C]], java.util.List[WriteResult[T, C]], NotUsed] =
createBulk(elasticsearchParams, settings, new JacksonWriter[T](objectMapper))
/**
@@ -118,15 +116,14 @@ object ElasticsearchFlow {
def createBulk[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- messageWriter: MessageWriter[T]
- ): akka.stream.javadsl.Flow[java.util.List[WriteMessage[T, C]], java.util.List[WriteResult[T, C]], NotUsed] =
+ messageWriter: MessageWriter[T])
+ : akka.stream.javadsl.Flow[java.util.List[WriteMessage[T, C]], java.util.List[WriteResult[T, C]], NotUsed] =
akka.stream.scaladsl
.Flow[java.util.List[WriteMessage[T, C]]]
.map(_.asScala.toIndexedSeq)
.via(
scaladsl.ElasticsearchFlow
- .createBulk(elasticsearchParams, settings, messageWriter)
- )
+ .createBulk(elasticsearchParams, settings, messageWriter))
.map(_.asJava)
.asJava
@@ -143,8 +140,8 @@ object ElasticsearchFlow {
def createWithContext[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- objectMapper: ObjectMapper
- ): akka.stream.javadsl.FlowWithContext[WriteMessage[T, NotUsed], C, WriteResult[T, C], C, NotUsed] =
+ objectMapper: ObjectMapper)
+ : akka.stream.javadsl.FlowWithContext[WriteMessage[T, NotUsed], C, WriteResult[T, C], C, NotUsed] =
createWithContext(elasticsearchParams, settings, new JacksonWriter[T](objectMapper))
/**
@@ -160,8 +157,8 @@ object ElasticsearchFlow {
def createWithContext[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- messageWriter: MessageWriter[T]
- ): akka.stream.javadsl.FlowWithContext[WriteMessage[T, NotUsed], C, WriteResult[T, C], C, NotUsed] =
+ messageWriter: MessageWriter[T])
+ : akka.stream.javadsl.FlowWithContext[WriteMessage[T, NotUsed], C, WriteResult[T, C], C, NotUsed] =
scaladsl.ElasticsearchFlow
.createWithContext(elasticsearchParams, settings, messageWriter)
.asJava
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSink.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSink.scala
index e80459cd..a0cfc461 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSink.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSink.scala
@@ -8,7 +8,7 @@ import java.util.concurrent.CompletionStage
import akka.stream.alpakka.elasticsearch._
import akka.stream.javadsl._
-import akka.{Done, NotUsed}
+import akka.{ Done, NotUsed }
import com.fasterxml.jackson.databind.ObjectMapper
/**
@@ -22,8 +22,7 @@ object ElasticsearchSink {
def create[T](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- objectMapper: ObjectMapper
- ): akka.stream.javadsl.Sink[WriteMessage[T, NotUsed], CompletionStage[Done]] =
+ objectMapper: ObjectMapper): akka.stream.javadsl.Sink[WriteMessage[T, NotUsed], CompletionStage[Done]] =
ElasticsearchFlow
.create(elasticsearchParams, settings, objectMapper)
.toMat(Sink.ignore[WriteResult[T, NotUsed]](), Keep.right[NotUsed, CompletionStage[Done]])
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSource.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSource.scala
index dd1c4842..abb1a76e 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSource.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/javadsl/ElasticsearchSource.scala
@@ -6,12 +6,12 @@ package akka.stream.alpakka.elasticsearch.javadsl
import akka.NotUsed
import akka.actor.ActorSystem
-import akka.http.scaladsl.{Http, HttpExt}
-import akka.stream.alpakka.elasticsearch.{impl, _}
+import akka.http.scaladsl.{ Http, HttpExt }
+import akka.stream.alpakka.elasticsearch.{ impl, _ }
import akka.stream.javadsl.Source
-import akka.stream.{Attributes, Materializer}
+import akka.stream.{ Attributes, Materializer }
import com.fasterxml.jackson.databind.ObjectMapper
-import com.fasterxml.jackson.databind.node.{ArrayNode, NumericNode}
+import com.fasterxml.jackson.databind.node.{ ArrayNode, NumericNode }
import scala.jdk.CollectionConverters._
import scala.concurrent.ExecutionContext
@@ -26,8 +26,8 @@ object ElasticsearchSource {
* Using default objectMapper
*/
def create(elasticsearchParams: ElasticsearchParams,
- query: String,
- settings: SourceSettingsBase[_, _]): Source[ReadResult[java.util.Map[String, Object]], NotUsed] =
+ query: String,
+ settings: SourceSettingsBase[_, _]): Source[ReadResult[java.util.Map[String, Object]], NotUsed] =
create(elasticsearchParams, query, settings, new ObjectMapper())
/**
@@ -35,9 +35,9 @@ object ElasticsearchSource {
* Using custom objectMapper
*/
def create(elasticsearchParams: ElasticsearchParams,
- query: String,
- settings: SourceSettingsBase[_, _],
- objectMapper: ObjectMapper): Source[ReadResult[java.util.Map[String, Object]], NotUsed] =
+ query: String,
+ settings: SourceSettingsBase[_, _],
+ objectMapper: ObjectMapper): Source[ReadResult[java.util.Map[String, Object]], NotUsed] =
Source
.fromMaterializer { (mat: Materializer, _: Attributes) =>
{
@@ -51,9 +51,7 @@ object ElasticsearchSource {
elasticsearchParams,
Map("query" -> query),
settings,
- new JacksonReader[java.util.Map[String, Object]](objectMapper, classOf[java.util.Map[String, Object]])
- )
- )
+ new JacksonReader[java.util.Map[String, Object]](objectMapper, classOf[java.util.Map[String, Object]])))
}
}
.mapMaterializedValue(_ => NotUsed)
@@ -69,9 +67,9 @@ object ElasticsearchSource {
* searchParams.put("_source", "[\"fieldToInclude\", \"anotherFieldToInclude\"]");
*/
def create(elasticsearchParams: ElasticsearchParams,
- searchParams: java.util.Map[String, String],
- settings: SourceSettingsBase[_, _],
- objectMapper: ObjectMapper): Source[ReadResult[java.util.Map[String, Object]], NotUsed] =
+ searchParams: java.util.Map[String, String],
+ settings: SourceSettingsBase[_, _],
+ objectMapper: ObjectMapper): Source[ReadResult[java.util.Map[String, Object]], NotUsed] =
Source
.fromMaterializer { (mat: Materializer, _: Attributes) =>
{
@@ -84,9 +82,7 @@ object ElasticsearchSource {
elasticsearchParams,
searchParams.asScala.toMap,
settings,
- new JacksonReader[java.util.Map[String, Object]](objectMapper, classOf[java.util.Map[String, Object]])
- )
- )
+ new JacksonReader[java.util.Map[String, Object]](objectMapper, classOf[java.util.Map[String, Object]])))
}
}
.mapMaterializedValue(_ => NotUsed)
@@ -96,9 +92,9 @@ object ElasticsearchSource {
* Using default objectMapper
*/
def typed[T](elasticsearchParams: ElasticsearchParams,
- query: String,
- settings: SourceSettingsBase[_, _],
- clazz: Class[T]): Source[ReadResult[T], NotUsed] =
+ query: String,
+ settings: SourceSettingsBase[_, _],
+ clazz: Class[T]): Source[ReadResult[T], NotUsed] =
typed[T](elasticsearchParams, query, settings, clazz, new ObjectMapper())
/**
@@ -106,10 +102,10 @@ object ElasticsearchSource {
* Using custom objectMapper
*/
def typed[T](elasticsearchParams: ElasticsearchParams,
- query: String,
- settings: SourceSettingsBase[_, _],
- clazz: Class[T],
- objectMapper: ObjectMapper): Source[ReadResult[T], NotUsed] =
+ query: String,
+ settings: SourceSettingsBase[_, _],
+ clazz: Class[T],
+ objectMapper: ObjectMapper): Source[ReadResult[T], NotUsed] =
Source
.fromMaterializer { (mat: Materializer, _: Attributes) =>
{
@@ -122,9 +118,7 @@ object ElasticsearchSource {
elasticsearchParams,
Map("query" -> query),
settings,
- new JacksonReader[T](objectMapper, clazz)
- )
- )
+ new JacksonReader[T](objectMapper, clazz)))
}
}
.mapMaterializedValue(_ => NotUsed)
@@ -140,10 +134,10 @@ object ElasticsearchSource {
* searchParams.put("_source", "[\"fieldToInclude\", \"anotherFieldToInclude\"]");
*/
def typed[T](elasticsearchParams: ElasticsearchParams,
- searchParams: java.util.Map[String, String],
- settings: SourceSettingsBase[_, _],
- clazz: Class[T],
- objectMapper: ObjectMapper): Source[ReadResult[T], NotUsed] =
+ searchParams: java.util.Map[String, String],
+ settings: SourceSettingsBase[_, _],
+ clazz: Class[T],
+ objectMapper: ObjectMapper): Source[ReadResult[T], NotUsed] =
Source
.fromMaterializer { (mat: Materializer, _: Attributes) =>
{
@@ -156,9 +150,7 @@ object ElasticsearchSource {
elasticsearchParams,
searchParams.asScala.toMap,
settings,
- new JacksonReader[T](objectMapper, clazz)
- )
- )
+ new JacksonReader[T](objectMapper, clazz)))
}
}
.mapMaterializedValue(_ => NotUsed)
@@ -179,7 +171,7 @@ object ElasticsearchSource {
val source = element.get("_source")
val version: Option[Long] = element.get("_version") match {
case n: NumericNode => Some(n.asLong())
- case _ => None
+ case _ => None
}
new ReadResult[T](id, mapper.treeToValue(source, clazz), version)
diff --git a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/scaladsl/ElasticsearchFlow.scala b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/scaladsl/ElasticsearchFlow.scala
index a2163398..fb7c9e42 100644
--- a/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/scaladsl/ElasticsearchFlow.scala
+++ b/elasticsearch/src/main/scala/akka/stream/alpakka/elasticsearch/scaladsl/ElasticsearchFlow.scala
@@ -6,10 +6,10 @@ package akka.stream.alpakka.elasticsearch.scaladsl
import akka.NotUsed
import akka.actor.ActorSystem
-import akka.annotation.{ApiMayChange, InternalApi}
-import akka.http.scaladsl.{Http, HttpExt}
-import akka.stream.alpakka.elasticsearch.{impl, _}
-import akka.stream.scaladsl.{Flow, FlowWithContext, RetryFlow}
+import akka.annotation.{ ApiMayChange, InternalApi }
+import akka.http.scaladsl.{ Http, HttpExt }
+import akka.stream.alpakka.elasticsearch.{ impl, _ }
+import akka.stream.scaladsl.{ Flow, FlowWithContext, RetryFlow }
import spray.json._
import scala.collection.immutable
@@ -28,8 +28,7 @@ object ElasticsearchFlow {
* This factory method requires an implicit Spray JSON writer for `T`.
*/
def create[T](elasticsearchParams: ElasticsearchParams, settings: WriteSettingsBase[_, _])(
- implicit sprayJsonWriter: JsonWriter[T]
- ): Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] =
+ implicit sprayJsonWriter: JsonWriter[T]): Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] =
create[T](elasticsearchParams, settings, new SprayJsonWriter[T]()(sprayJsonWriter))
/**
@@ -38,8 +37,8 @@ object ElasticsearchFlow {
* successful execution.
*/
def create[T](elasticsearchParams: ElasticsearchParams,
- settings: WriteSettingsBase[_, _],
- writer: MessageWriter[T]): Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] = {
+ settings: WriteSettingsBase[_, _],
+ writer: MessageWriter[T]): Flow[WriteMessage[T, NotUsed], WriteResult[T, NotUsed], NotUsed] = {
Flow[WriteMessage[T, NotUsed]]
.batch(settings.bufferSize, immutable.Seq(_)) { case (seq, wm) => seq :+ wm }
.via(stageFlow(elasticsearchParams, settings, writer))
@@ -55,8 +54,7 @@ object ElasticsearchFlow {
* This factory method requires an implicit Spray JSON writer for `T`.
*/
def createWithPassThrough[T, C](elasticsearchParams: ElasticsearchParams, settings: WriteSettingsBase[_, _])(
- implicit sprayJsonWriter: JsonWriter[T]
- ): Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] =
+ implicit sprayJsonWriter: JsonWriter[T]): Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] =
createWithPassThrough[T, C](elasticsearchParams, settings, new SprayJsonWriter[T]()(sprayJsonWriter))
/**
@@ -66,8 +64,8 @@ object ElasticsearchFlow {
* successful execution.
*/
def createWithPassThrough[T, C](elasticsearchParams: ElasticsearchParams,
- settings: WriteSettingsBase[_, _],
- writer: MessageWriter[T]): Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] = {
+ settings: WriteSettingsBase[_, _],
+ writer: MessageWriter[T]): Flow[WriteMessage[T, C], WriteResult[T, C], NotUsed] = {
Flow[WriteMessage[T, C]]
.batch(settings.bufferSize, immutable.Seq(_)) { case (seq, wm) => seq :+ wm }
.via(stageFlow(elasticsearchParams, settings, writer))
@@ -84,8 +82,8 @@ object ElasticsearchFlow {
* This factory method requires an implicit Spray JSON writer for `T`.
*/
def createBulk[T, C](elasticsearchParams: ElasticsearchParams, settings: WriteSettingsBase[_, _])(
- implicit sprayJsonWriter: JsonWriter[T]
- ): Flow[immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]], NotUsed] =
+ implicit sprayJsonWriter: JsonWriter[T])
+ : Flow[immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]], NotUsed] =
createBulk[T, C](elasticsearchParams, settings, new SprayJsonWriter[T]()(sprayJsonWriter))
/**
@@ -98,8 +96,7 @@ object ElasticsearchFlow {
def createBulk[T, C](
elasticsearchParams: ElasticsearchParams,
settings: WriteSettingsBase[_, _],
- writer: MessageWriter[T]
- ): Flow[immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]], NotUsed] = {
+ writer: MessageWriter[T]): Flow[immutable.Seq[WriteMessage[T, C]], immutable.Seq[WriteResult[T, C]], NotUsed] = {
stageFlow(elasticsearchParams, settings, writer)
}
@@ -113,8 +110,8 @@ object ElasticsearchFlow {
*/
@ApiMayChange
def createWithContext[T, C](elasticsearchParams: ElasticsearchParams, settings: WriteSettingsBase[_, _])(
- implicit sprayJsonWriter: JsonWriter[T]
- ): FlowWithContext[WriteMessage[T, NotUsed], C, WriteResult[T, C], C, NotUsed] =
+ implicit sprayJsonWriter: JsonWriter[T])
... 50816 lines suppressed ...
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@pekko.apache.org
For additional commands, e-mail: commits-help@pekko.apache.org